xref: /openbmc/qemu/tcg/tcg.c (revision e4370165)
1 /*
2  * Tiny Code Generator for QEMU
3  *
4  * Copyright (c) 2008 Fabrice Bellard
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a copy
7  * of this software and associated documentation files (the "Software"), to deal
8  * in the Software without restriction, including without limitation the rights
9  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10  * copies of the Software, and to permit persons to whom the Software is
11  * furnished to do so, subject to the following conditions:
12  *
13  * The above copyright notice and this permission notice shall be included in
14  * all copies or substantial portions of the Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22  * THE SOFTWARE.
23  */
24 
25 /* define it to use liveness analysis (better code) */
26 #define USE_TCG_OPTIMIZATIONS
27 
28 #include "qemu/osdep.h"
29 
30 /* Define to jump the ELF file used to communicate with GDB.  */
31 #undef DEBUG_JIT
32 
33 #include "qemu/cutils.h"
34 #include "qemu/host-utils.h"
35 #include "qemu/timer.h"
36 
37 /* Note: the long term plan is to reduce the dependencies on the QEMU
38    CPU definitions. Currently they are used for qemu_ld/st
39    instructions */
40 #define NO_CPU_IO_DEFS
41 #include "cpu.h"
42 
43 #include "exec/cpu-common.h"
44 #include "exec/exec-all.h"
45 
46 #include "tcg-op.h"
47 
48 #if UINTPTR_MAX == UINT32_MAX
49 # define ELF_CLASS  ELFCLASS32
50 #else
51 # define ELF_CLASS  ELFCLASS64
52 #endif
53 #ifdef HOST_WORDS_BIGENDIAN
54 # define ELF_DATA   ELFDATA2MSB
55 #else
56 # define ELF_DATA   ELFDATA2LSB
57 #endif
58 
59 #include "elf.h"
60 #include "exec/log.h"
61 
62 /* Forward declarations for functions declared in tcg-target.inc.c and
63    used here. */
64 static void tcg_target_init(TCGContext *s);
65 static void tcg_target_qemu_prologue(TCGContext *s);
66 static void patch_reloc(tcg_insn_unit *code_ptr, int type,
67                         intptr_t value, intptr_t addend);
68 
69 /* The CIE and FDE header definitions will be common to all hosts.  */
70 typedef struct {
71     uint32_t len __attribute__((aligned((sizeof(void *)))));
72     uint32_t id;
73     uint8_t version;
74     char augmentation[1];
75     uint8_t code_align;
76     uint8_t data_align;
77     uint8_t return_column;
78 } DebugFrameCIE;
79 
80 typedef struct QEMU_PACKED {
81     uint32_t len __attribute__((aligned((sizeof(void *)))));
82     uint32_t cie_offset;
83     uintptr_t func_start;
84     uintptr_t func_len;
85 } DebugFrameFDEHeader;
86 
87 typedef struct QEMU_PACKED {
88     DebugFrameCIE cie;
89     DebugFrameFDEHeader fde;
90 } DebugFrameHeader;
91 
92 static void tcg_register_jit_int(void *buf, size_t size,
93                                  const void *debug_frame,
94                                  size_t debug_frame_size)
95     __attribute__((unused));
96 
97 /* Forward declarations for functions declared and used in tcg-target.inc.c. */
98 static int target_parse_constraint(TCGArgConstraint *ct, const char **pct_str);
99 static void tcg_out_ld(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg1,
100                        intptr_t arg2);
101 static void tcg_out_mov(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg);
102 static void tcg_out_movi(TCGContext *s, TCGType type,
103                          TCGReg ret, tcg_target_long arg);
104 static void tcg_out_op(TCGContext *s, TCGOpcode opc, const TCGArg *args,
105                        const int *const_args);
106 static void tcg_out_st(TCGContext *s, TCGType type, TCGReg arg, TCGReg arg1,
107                        intptr_t arg2);
108 static bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val,
109                         TCGReg base, intptr_t ofs);
110 static void tcg_out_call(TCGContext *s, tcg_insn_unit *target);
111 static int tcg_target_const_match(tcg_target_long val, TCGType type,
112                                   const TCGArgConstraint *arg_ct);
113 static void tcg_out_tb_init(TCGContext *s);
114 static bool tcg_out_tb_finalize(TCGContext *s);
115 
116 
117 
118 static TCGRegSet tcg_target_available_regs[2];
119 static TCGRegSet tcg_target_call_clobber_regs;
120 
121 #if TCG_TARGET_INSN_UNIT_SIZE == 1
122 static __attribute__((unused)) inline void tcg_out8(TCGContext *s, uint8_t v)
123 {
124     *s->code_ptr++ = v;
125 }
126 
127 static __attribute__((unused)) inline void tcg_patch8(tcg_insn_unit *p,
128                                                       uint8_t v)
129 {
130     *p = v;
131 }
132 #endif
133 
134 #if TCG_TARGET_INSN_UNIT_SIZE <= 2
135 static __attribute__((unused)) inline void tcg_out16(TCGContext *s, uint16_t v)
136 {
137     if (TCG_TARGET_INSN_UNIT_SIZE == 2) {
138         *s->code_ptr++ = v;
139     } else {
140         tcg_insn_unit *p = s->code_ptr;
141         memcpy(p, &v, sizeof(v));
142         s->code_ptr = p + (2 / TCG_TARGET_INSN_UNIT_SIZE);
143     }
144 }
145 
146 static __attribute__((unused)) inline void tcg_patch16(tcg_insn_unit *p,
147                                                        uint16_t v)
148 {
149     if (TCG_TARGET_INSN_UNIT_SIZE == 2) {
150         *p = v;
151     } else {
152         memcpy(p, &v, sizeof(v));
153     }
154 }
155 #endif
156 
157 #if TCG_TARGET_INSN_UNIT_SIZE <= 4
158 static __attribute__((unused)) inline void tcg_out32(TCGContext *s, uint32_t v)
159 {
160     if (TCG_TARGET_INSN_UNIT_SIZE == 4) {
161         *s->code_ptr++ = v;
162     } else {
163         tcg_insn_unit *p = s->code_ptr;
164         memcpy(p, &v, sizeof(v));
165         s->code_ptr = p + (4 / TCG_TARGET_INSN_UNIT_SIZE);
166     }
167 }
168 
169 static __attribute__((unused)) inline void tcg_patch32(tcg_insn_unit *p,
170                                                        uint32_t v)
171 {
172     if (TCG_TARGET_INSN_UNIT_SIZE == 4) {
173         *p = v;
174     } else {
175         memcpy(p, &v, sizeof(v));
176     }
177 }
178 #endif
179 
180 #if TCG_TARGET_INSN_UNIT_SIZE <= 8
181 static __attribute__((unused)) inline void tcg_out64(TCGContext *s, uint64_t v)
182 {
183     if (TCG_TARGET_INSN_UNIT_SIZE == 8) {
184         *s->code_ptr++ = v;
185     } else {
186         tcg_insn_unit *p = s->code_ptr;
187         memcpy(p, &v, sizeof(v));
188         s->code_ptr = p + (8 / TCG_TARGET_INSN_UNIT_SIZE);
189     }
190 }
191 
192 static __attribute__((unused)) inline void tcg_patch64(tcg_insn_unit *p,
193                                                        uint64_t v)
194 {
195     if (TCG_TARGET_INSN_UNIT_SIZE == 8) {
196         *p = v;
197     } else {
198         memcpy(p, &v, sizeof(v));
199     }
200 }
201 #endif
202 
203 /* label relocation processing */
204 
205 static void tcg_out_reloc(TCGContext *s, tcg_insn_unit *code_ptr, int type,
206                           TCGLabel *l, intptr_t addend)
207 {
208     TCGRelocation *r;
209 
210     if (l->has_value) {
211         /* FIXME: This may break relocations on RISC targets that
212            modify instruction fields in place.  The caller may not have
213            written the initial value.  */
214         patch_reloc(code_ptr, type, l->u.value, addend);
215     } else {
216         /* add a new relocation entry */
217         r = tcg_malloc(sizeof(TCGRelocation));
218         r->type = type;
219         r->ptr = code_ptr;
220         r->addend = addend;
221         r->next = l->u.first_reloc;
222         l->u.first_reloc = r;
223     }
224 }
225 
226 static void tcg_out_label(TCGContext *s, TCGLabel *l, tcg_insn_unit *ptr)
227 {
228     intptr_t value = (intptr_t)ptr;
229     TCGRelocation *r;
230 
231     tcg_debug_assert(!l->has_value);
232 
233     for (r = l->u.first_reloc; r != NULL; r = r->next) {
234         patch_reloc(r->ptr, r->type, value, r->addend);
235     }
236 
237     l->has_value = 1;
238     l->u.value_ptr = ptr;
239 }
240 
241 TCGLabel *gen_new_label(void)
242 {
243     TCGContext *s = &tcg_ctx;
244     TCGLabel *l = tcg_malloc(sizeof(TCGLabel));
245 
246     *l = (TCGLabel){
247         .id = s->nb_labels++
248     };
249 
250     return l;
251 }
252 
253 #include "tcg-target.inc.c"
254 
255 /* pool based memory allocation */
256 void *tcg_malloc_internal(TCGContext *s, int size)
257 {
258     TCGPool *p;
259     int pool_size;
260 
261     if (size > TCG_POOL_CHUNK_SIZE) {
262         /* big malloc: insert a new pool (XXX: could optimize) */
263         p = g_malloc(sizeof(TCGPool) + size);
264         p->size = size;
265         p->next = s->pool_first_large;
266         s->pool_first_large = p;
267         return p->data;
268     } else {
269         p = s->pool_current;
270         if (!p) {
271             p = s->pool_first;
272             if (!p)
273                 goto new_pool;
274         } else {
275             if (!p->next) {
276             new_pool:
277                 pool_size = TCG_POOL_CHUNK_SIZE;
278                 p = g_malloc(sizeof(TCGPool) + pool_size);
279                 p->size = pool_size;
280                 p->next = NULL;
281                 if (s->pool_current)
282                     s->pool_current->next = p;
283                 else
284                     s->pool_first = p;
285             } else {
286                 p = p->next;
287             }
288         }
289     }
290     s->pool_current = p;
291     s->pool_cur = p->data + size;
292     s->pool_end = p->data + p->size;
293     return p->data;
294 }
295 
296 void tcg_pool_reset(TCGContext *s)
297 {
298     TCGPool *p, *t;
299     for (p = s->pool_first_large; p; p = t) {
300         t = p->next;
301         g_free(p);
302     }
303     s->pool_first_large = NULL;
304     s->pool_cur = s->pool_end = NULL;
305     s->pool_current = NULL;
306 }
307 
308 typedef struct TCGHelperInfo {
309     void *func;
310     const char *name;
311     unsigned flags;
312     unsigned sizemask;
313 } TCGHelperInfo;
314 
315 #include "exec/helper-proto.h"
316 
317 static const TCGHelperInfo all_helpers[] = {
318 #include "exec/helper-tcg.h"
319 };
320 
321 static int indirect_reg_alloc_order[ARRAY_SIZE(tcg_target_reg_alloc_order)];
322 
323 void tcg_context_init(TCGContext *s)
324 {
325     int op, total_args, n, i;
326     TCGOpDef *def;
327     TCGArgConstraint *args_ct;
328     int *sorted_args;
329     GHashTable *helper_table;
330 
331     memset(s, 0, sizeof(*s));
332     s->nb_globals = 0;
333 
334     /* Count total number of arguments and allocate the corresponding
335        space */
336     total_args = 0;
337     for(op = 0; op < NB_OPS; op++) {
338         def = &tcg_op_defs[op];
339         n = def->nb_iargs + def->nb_oargs;
340         total_args += n;
341     }
342 
343     args_ct = g_malloc(sizeof(TCGArgConstraint) * total_args);
344     sorted_args = g_malloc(sizeof(int) * total_args);
345 
346     for(op = 0; op < NB_OPS; op++) {
347         def = &tcg_op_defs[op];
348         def->args_ct = args_ct;
349         def->sorted_args = sorted_args;
350         n = def->nb_iargs + def->nb_oargs;
351         sorted_args += n;
352         args_ct += n;
353     }
354 
355     /* Register helpers.  */
356     /* Use g_direct_hash/equal for direct pointer comparisons on func.  */
357     s->helpers = helper_table = g_hash_table_new(NULL, NULL);
358 
359     for (i = 0; i < ARRAY_SIZE(all_helpers); ++i) {
360         g_hash_table_insert(helper_table, (gpointer)all_helpers[i].func,
361                             (gpointer)&all_helpers[i]);
362     }
363 
364     tcg_target_init(s);
365 
366     /* Reverse the order of the saved registers, assuming they're all at
367        the start of tcg_target_reg_alloc_order.  */
368     for (n = 0; n < ARRAY_SIZE(tcg_target_reg_alloc_order); ++n) {
369         int r = tcg_target_reg_alloc_order[n];
370         if (tcg_regset_test_reg(tcg_target_call_clobber_regs, r)) {
371             break;
372         }
373     }
374     for (i = 0; i < n; ++i) {
375         indirect_reg_alloc_order[i] = tcg_target_reg_alloc_order[n - 1 - i];
376     }
377     for (; i < ARRAY_SIZE(tcg_target_reg_alloc_order); ++i) {
378         indirect_reg_alloc_order[i] = tcg_target_reg_alloc_order[i];
379     }
380 }
381 
382 void tcg_prologue_init(TCGContext *s)
383 {
384     size_t prologue_size, total_size;
385     void *buf0, *buf1;
386 
387     /* Put the prologue at the beginning of code_gen_buffer.  */
388     buf0 = s->code_gen_buffer;
389     s->code_ptr = buf0;
390     s->code_buf = buf0;
391     s->code_gen_prologue = buf0;
392 
393     /* Generate the prologue.  */
394     tcg_target_qemu_prologue(s);
395     buf1 = s->code_ptr;
396     flush_icache_range((uintptr_t)buf0, (uintptr_t)buf1);
397 
398     /* Deduct the prologue from the buffer.  */
399     prologue_size = tcg_current_code_size(s);
400     s->code_gen_ptr = buf1;
401     s->code_gen_buffer = buf1;
402     s->code_buf = buf1;
403     total_size = s->code_gen_buffer_size - prologue_size;
404     s->code_gen_buffer_size = total_size;
405 
406     /* Compute a high-water mark, at which we voluntarily flush the buffer
407        and start over.  The size here is arbitrary, significantly larger
408        than we expect the code generation for any one opcode to require.  */
409     s->code_gen_highwater = s->code_gen_buffer + (total_size - 1024);
410 
411     tcg_register_jit(s->code_gen_buffer, total_size);
412 
413 #ifdef DEBUG_DISAS
414     if (qemu_loglevel_mask(CPU_LOG_TB_OUT_ASM)) {
415         qemu_log("PROLOGUE: [size=%zu]\n", prologue_size);
416         log_disas(buf0, prologue_size);
417         qemu_log("\n");
418         qemu_log_flush();
419     }
420 #endif
421 }
422 
423 void tcg_func_start(TCGContext *s)
424 {
425     tcg_pool_reset(s);
426     s->nb_temps = s->nb_globals;
427 
428     /* No temps have been previously allocated for size or locality.  */
429     memset(s->free_temps, 0, sizeof(s->free_temps));
430 
431     s->nb_labels = 0;
432     s->current_frame_offset = s->frame_start;
433 
434 #ifdef CONFIG_DEBUG_TCG
435     s->goto_tb_issue_mask = 0;
436 #endif
437 
438     s->gen_op_buf[0].next = 1;
439     s->gen_op_buf[0].prev = 0;
440     s->gen_next_op_idx = 1;
441     s->gen_next_parm_idx = 0;
442 
443     s->be = tcg_malloc(sizeof(TCGBackendData));
444 }
445 
446 static inline int temp_idx(TCGContext *s, TCGTemp *ts)
447 {
448     ptrdiff_t n = ts - s->temps;
449     tcg_debug_assert(n >= 0 && n < s->nb_temps);
450     return n;
451 }
452 
453 static inline TCGTemp *tcg_temp_alloc(TCGContext *s)
454 {
455     int n = s->nb_temps++;
456     tcg_debug_assert(n < TCG_MAX_TEMPS);
457     return memset(&s->temps[n], 0, sizeof(TCGTemp));
458 }
459 
460 static inline TCGTemp *tcg_global_alloc(TCGContext *s)
461 {
462     tcg_debug_assert(s->nb_globals == s->nb_temps);
463     s->nb_globals++;
464     return tcg_temp_alloc(s);
465 }
466 
467 static int tcg_global_reg_new_internal(TCGContext *s, TCGType type,
468                                        TCGReg reg, const char *name)
469 {
470     TCGTemp *ts;
471 
472     if (TCG_TARGET_REG_BITS == 32 && type != TCG_TYPE_I32) {
473         tcg_abort();
474     }
475 
476     ts = tcg_global_alloc(s);
477     ts->base_type = type;
478     ts->type = type;
479     ts->fixed_reg = 1;
480     ts->reg = reg;
481     ts->name = name;
482     tcg_regset_set_reg(s->reserved_regs, reg);
483 
484     return temp_idx(s, ts);
485 }
486 
487 void tcg_set_frame(TCGContext *s, TCGReg reg, intptr_t start, intptr_t size)
488 {
489     int idx;
490     s->frame_start = start;
491     s->frame_end = start + size;
492     idx = tcg_global_reg_new_internal(s, TCG_TYPE_PTR, reg, "_frame");
493     s->frame_temp = &s->temps[idx];
494 }
495 
496 TCGv_i32 tcg_global_reg_new_i32(TCGReg reg, const char *name)
497 {
498     TCGContext *s = &tcg_ctx;
499     int idx;
500 
501     if (tcg_regset_test_reg(s->reserved_regs, reg)) {
502         tcg_abort();
503     }
504     idx = tcg_global_reg_new_internal(s, TCG_TYPE_I32, reg, name);
505     return MAKE_TCGV_I32(idx);
506 }
507 
508 TCGv_i64 tcg_global_reg_new_i64(TCGReg reg, const char *name)
509 {
510     TCGContext *s = &tcg_ctx;
511     int idx;
512 
513     if (tcg_regset_test_reg(s->reserved_regs, reg)) {
514         tcg_abort();
515     }
516     idx = tcg_global_reg_new_internal(s, TCG_TYPE_I64, reg, name);
517     return MAKE_TCGV_I64(idx);
518 }
519 
520 int tcg_global_mem_new_internal(TCGType type, TCGv_ptr base,
521                                 intptr_t offset, const char *name)
522 {
523     TCGContext *s = &tcg_ctx;
524     TCGTemp *base_ts = &s->temps[GET_TCGV_PTR(base)];
525     TCGTemp *ts = tcg_global_alloc(s);
526     int indirect_reg = 0, bigendian = 0;
527 #ifdef HOST_WORDS_BIGENDIAN
528     bigendian = 1;
529 #endif
530 
531     if (!base_ts->fixed_reg) {
532         /* We do not support double-indirect registers.  */
533         tcg_debug_assert(!base_ts->indirect_reg);
534         base_ts->indirect_base = 1;
535         s->nb_indirects += (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64
536                             ? 2 : 1);
537         indirect_reg = 1;
538     }
539 
540     if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
541         TCGTemp *ts2 = tcg_global_alloc(s);
542         char buf[64];
543 
544         ts->base_type = TCG_TYPE_I64;
545         ts->type = TCG_TYPE_I32;
546         ts->indirect_reg = indirect_reg;
547         ts->mem_allocated = 1;
548         ts->mem_base = base_ts;
549         ts->mem_offset = offset + bigendian * 4;
550         pstrcpy(buf, sizeof(buf), name);
551         pstrcat(buf, sizeof(buf), "_0");
552         ts->name = strdup(buf);
553 
554         tcg_debug_assert(ts2 == ts + 1);
555         ts2->base_type = TCG_TYPE_I64;
556         ts2->type = TCG_TYPE_I32;
557         ts2->indirect_reg = indirect_reg;
558         ts2->mem_allocated = 1;
559         ts2->mem_base = base_ts;
560         ts2->mem_offset = offset + (1 - bigendian) * 4;
561         pstrcpy(buf, sizeof(buf), name);
562         pstrcat(buf, sizeof(buf), "_1");
563         ts2->name = strdup(buf);
564     } else {
565         ts->base_type = type;
566         ts->type = type;
567         ts->indirect_reg = indirect_reg;
568         ts->mem_allocated = 1;
569         ts->mem_base = base_ts;
570         ts->mem_offset = offset;
571         ts->name = name;
572     }
573     return temp_idx(s, ts);
574 }
575 
576 static int tcg_temp_new_internal(TCGType type, int temp_local)
577 {
578     TCGContext *s = &tcg_ctx;
579     TCGTemp *ts;
580     int idx, k;
581 
582     k = type + (temp_local ? TCG_TYPE_COUNT : 0);
583     idx = find_first_bit(s->free_temps[k].l, TCG_MAX_TEMPS);
584     if (idx < TCG_MAX_TEMPS) {
585         /* There is already an available temp with the right type.  */
586         clear_bit(idx, s->free_temps[k].l);
587 
588         ts = &s->temps[idx];
589         ts->temp_allocated = 1;
590         tcg_debug_assert(ts->base_type == type);
591         tcg_debug_assert(ts->temp_local == temp_local);
592     } else {
593         ts = tcg_temp_alloc(s);
594         if (TCG_TARGET_REG_BITS == 32 && type == TCG_TYPE_I64) {
595             TCGTemp *ts2 = tcg_temp_alloc(s);
596 
597             ts->base_type = type;
598             ts->type = TCG_TYPE_I32;
599             ts->temp_allocated = 1;
600             ts->temp_local = temp_local;
601 
602             tcg_debug_assert(ts2 == ts + 1);
603             ts2->base_type = TCG_TYPE_I64;
604             ts2->type = TCG_TYPE_I32;
605             ts2->temp_allocated = 1;
606             ts2->temp_local = temp_local;
607         } else {
608             ts->base_type = type;
609             ts->type = type;
610             ts->temp_allocated = 1;
611             ts->temp_local = temp_local;
612         }
613         idx = temp_idx(s, ts);
614     }
615 
616 #if defined(CONFIG_DEBUG_TCG)
617     s->temps_in_use++;
618 #endif
619     return idx;
620 }
621 
622 TCGv_i32 tcg_temp_new_internal_i32(int temp_local)
623 {
624     int idx;
625 
626     idx = tcg_temp_new_internal(TCG_TYPE_I32, temp_local);
627     return MAKE_TCGV_I32(idx);
628 }
629 
630 TCGv_i64 tcg_temp_new_internal_i64(int temp_local)
631 {
632     int idx;
633 
634     idx = tcg_temp_new_internal(TCG_TYPE_I64, temp_local);
635     return MAKE_TCGV_I64(idx);
636 }
637 
638 static void tcg_temp_free_internal(int idx)
639 {
640     TCGContext *s = &tcg_ctx;
641     TCGTemp *ts;
642     int k;
643 
644 #if defined(CONFIG_DEBUG_TCG)
645     s->temps_in_use--;
646     if (s->temps_in_use < 0) {
647         fprintf(stderr, "More temporaries freed than allocated!\n");
648     }
649 #endif
650 
651     tcg_debug_assert(idx >= s->nb_globals && idx < s->nb_temps);
652     ts = &s->temps[idx];
653     tcg_debug_assert(ts->temp_allocated != 0);
654     ts->temp_allocated = 0;
655 
656     k = ts->base_type + (ts->temp_local ? TCG_TYPE_COUNT : 0);
657     set_bit(idx, s->free_temps[k].l);
658 }
659 
660 void tcg_temp_free_i32(TCGv_i32 arg)
661 {
662     tcg_temp_free_internal(GET_TCGV_I32(arg));
663 }
664 
665 void tcg_temp_free_i64(TCGv_i64 arg)
666 {
667     tcg_temp_free_internal(GET_TCGV_I64(arg));
668 }
669 
670 TCGv_i32 tcg_const_i32(int32_t val)
671 {
672     TCGv_i32 t0;
673     t0 = tcg_temp_new_i32();
674     tcg_gen_movi_i32(t0, val);
675     return t0;
676 }
677 
678 TCGv_i64 tcg_const_i64(int64_t val)
679 {
680     TCGv_i64 t0;
681     t0 = tcg_temp_new_i64();
682     tcg_gen_movi_i64(t0, val);
683     return t0;
684 }
685 
686 TCGv_i32 tcg_const_local_i32(int32_t val)
687 {
688     TCGv_i32 t0;
689     t0 = tcg_temp_local_new_i32();
690     tcg_gen_movi_i32(t0, val);
691     return t0;
692 }
693 
694 TCGv_i64 tcg_const_local_i64(int64_t val)
695 {
696     TCGv_i64 t0;
697     t0 = tcg_temp_local_new_i64();
698     tcg_gen_movi_i64(t0, val);
699     return t0;
700 }
701 
702 #if defined(CONFIG_DEBUG_TCG)
703 void tcg_clear_temp_count(void)
704 {
705     TCGContext *s = &tcg_ctx;
706     s->temps_in_use = 0;
707 }
708 
709 int tcg_check_temp_count(void)
710 {
711     TCGContext *s = &tcg_ctx;
712     if (s->temps_in_use) {
713         /* Clear the count so that we don't give another
714          * warning immediately next time around.
715          */
716         s->temps_in_use = 0;
717         return 1;
718     }
719     return 0;
720 }
721 #endif
722 
723 /* Note: we convert the 64 bit args to 32 bit and do some alignment
724    and endian swap. Maybe it would be better to do the alignment
725    and endian swap in tcg_reg_alloc_call(). */
726 void tcg_gen_callN(TCGContext *s, void *func, TCGArg ret,
727                    int nargs, TCGArg *args)
728 {
729     int i, real_args, nb_rets, pi, pi_first;
730     unsigned sizemask, flags;
731     TCGHelperInfo *info;
732 
733     info = g_hash_table_lookup(s->helpers, (gpointer)func);
734     flags = info->flags;
735     sizemask = info->sizemask;
736 
737 #if defined(__sparc__) && !defined(__arch64__) \
738     && !defined(CONFIG_TCG_INTERPRETER)
739     /* We have 64-bit values in one register, but need to pass as two
740        separate parameters.  Split them.  */
741     int orig_sizemask = sizemask;
742     int orig_nargs = nargs;
743     TCGv_i64 retl, reth;
744 
745     TCGV_UNUSED_I64(retl);
746     TCGV_UNUSED_I64(reth);
747     if (sizemask != 0) {
748         TCGArg *split_args = __builtin_alloca(sizeof(TCGArg) * nargs * 2);
749         for (i = real_args = 0; i < nargs; ++i) {
750             int is_64bit = sizemask & (1 << (i+1)*2);
751             if (is_64bit) {
752                 TCGv_i64 orig = MAKE_TCGV_I64(args[i]);
753                 TCGv_i32 h = tcg_temp_new_i32();
754                 TCGv_i32 l = tcg_temp_new_i32();
755                 tcg_gen_extr_i64_i32(l, h, orig);
756                 split_args[real_args++] = GET_TCGV_I32(h);
757                 split_args[real_args++] = GET_TCGV_I32(l);
758             } else {
759                 split_args[real_args++] = args[i];
760             }
761         }
762         nargs = real_args;
763         args = split_args;
764         sizemask = 0;
765     }
766 #elif defined(TCG_TARGET_EXTEND_ARGS) && TCG_TARGET_REG_BITS == 64
767     for (i = 0; i < nargs; ++i) {
768         int is_64bit = sizemask & (1 << (i+1)*2);
769         int is_signed = sizemask & (2 << (i+1)*2);
770         if (!is_64bit) {
771             TCGv_i64 temp = tcg_temp_new_i64();
772             TCGv_i64 orig = MAKE_TCGV_I64(args[i]);
773             if (is_signed) {
774                 tcg_gen_ext32s_i64(temp, orig);
775             } else {
776                 tcg_gen_ext32u_i64(temp, orig);
777             }
778             args[i] = GET_TCGV_I64(temp);
779         }
780     }
781 #endif /* TCG_TARGET_EXTEND_ARGS */
782 
783     pi_first = pi = s->gen_next_parm_idx;
784     if (ret != TCG_CALL_DUMMY_ARG) {
785 #if defined(__sparc__) && !defined(__arch64__) \
786     && !defined(CONFIG_TCG_INTERPRETER)
787         if (orig_sizemask & 1) {
788             /* The 32-bit ABI is going to return the 64-bit value in
789                the %o0/%o1 register pair.  Prepare for this by using
790                two return temporaries, and reassemble below.  */
791             retl = tcg_temp_new_i64();
792             reth = tcg_temp_new_i64();
793             s->gen_opparam_buf[pi++] = GET_TCGV_I64(reth);
794             s->gen_opparam_buf[pi++] = GET_TCGV_I64(retl);
795             nb_rets = 2;
796         } else {
797             s->gen_opparam_buf[pi++] = ret;
798             nb_rets = 1;
799         }
800 #else
801         if (TCG_TARGET_REG_BITS < 64 && (sizemask & 1)) {
802 #ifdef HOST_WORDS_BIGENDIAN
803             s->gen_opparam_buf[pi++] = ret + 1;
804             s->gen_opparam_buf[pi++] = ret;
805 #else
806             s->gen_opparam_buf[pi++] = ret;
807             s->gen_opparam_buf[pi++] = ret + 1;
808 #endif
809             nb_rets = 2;
810         } else {
811             s->gen_opparam_buf[pi++] = ret;
812             nb_rets = 1;
813         }
814 #endif
815     } else {
816         nb_rets = 0;
817     }
818     real_args = 0;
819     for (i = 0; i < nargs; i++) {
820         int is_64bit = sizemask & (1 << (i+1)*2);
821         if (TCG_TARGET_REG_BITS < 64 && is_64bit) {
822 #ifdef TCG_TARGET_CALL_ALIGN_ARGS
823             /* some targets want aligned 64 bit args */
824             if (real_args & 1) {
825                 s->gen_opparam_buf[pi++] = TCG_CALL_DUMMY_ARG;
826                 real_args++;
827             }
828 #endif
829            /* If stack grows up, then we will be placing successive
830               arguments at lower addresses, which means we need to
831               reverse the order compared to how we would normally
832               treat either big or little-endian.  For those arguments
833               that will wind up in registers, this still works for
834               HPPA (the only current STACK_GROWSUP target) since the
835               argument registers are *also* allocated in decreasing
836               order.  If another such target is added, this logic may
837               have to get more complicated to differentiate between
838               stack arguments and register arguments.  */
839 #if defined(HOST_WORDS_BIGENDIAN) != defined(TCG_TARGET_STACK_GROWSUP)
840             s->gen_opparam_buf[pi++] = args[i] + 1;
841             s->gen_opparam_buf[pi++] = args[i];
842 #else
843             s->gen_opparam_buf[pi++] = args[i];
844             s->gen_opparam_buf[pi++] = args[i] + 1;
845 #endif
846             real_args += 2;
847             continue;
848         }
849 
850         s->gen_opparam_buf[pi++] = args[i];
851         real_args++;
852     }
853     s->gen_opparam_buf[pi++] = (uintptr_t)func;
854     s->gen_opparam_buf[pi++] = flags;
855 
856     i = s->gen_next_op_idx;
857     tcg_debug_assert(i < OPC_BUF_SIZE);
858     tcg_debug_assert(pi <= OPPARAM_BUF_SIZE);
859 
860     /* Set links for sequential allocation during translation.  */
861     s->gen_op_buf[i] = (TCGOp){
862         .opc = INDEX_op_call,
863         .callo = nb_rets,
864         .calli = real_args,
865         .args = pi_first,
866         .prev = i - 1,
867         .next = i + 1
868     };
869 
870     /* Make sure the calli field didn't overflow.  */
871     tcg_debug_assert(s->gen_op_buf[i].calli == real_args);
872 
873     s->gen_op_buf[0].prev = i;
874     s->gen_next_op_idx = i + 1;
875     s->gen_next_parm_idx = pi;
876 
877 #if defined(__sparc__) && !defined(__arch64__) \
878     && !defined(CONFIG_TCG_INTERPRETER)
879     /* Free all of the parts we allocated above.  */
880     for (i = real_args = 0; i < orig_nargs; ++i) {
881         int is_64bit = orig_sizemask & (1 << (i+1)*2);
882         if (is_64bit) {
883             TCGv_i32 h = MAKE_TCGV_I32(args[real_args++]);
884             TCGv_i32 l = MAKE_TCGV_I32(args[real_args++]);
885             tcg_temp_free_i32(h);
886             tcg_temp_free_i32(l);
887         } else {
888             real_args++;
889         }
890     }
891     if (orig_sizemask & 1) {
892         /* The 32-bit ABI returned two 32-bit pieces.  Re-assemble them.
893            Note that describing these as TCGv_i64 eliminates an unnecessary
894            zero-extension that tcg_gen_concat_i32_i64 would create.  */
895         tcg_gen_concat32_i64(MAKE_TCGV_I64(ret), retl, reth);
896         tcg_temp_free_i64(retl);
897         tcg_temp_free_i64(reth);
898     }
899 #elif defined(TCG_TARGET_EXTEND_ARGS) && TCG_TARGET_REG_BITS == 64
900     for (i = 0; i < nargs; ++i) {
901         int is_64bit = sizemask & (1 << (i+1)*2);
902         if (!is_64bit) {
903             TCGv_i64 temp = MAKE_TCGV_I64(args[i]);
904             tcg_temp_free_i64(temp);
905         }
906     }
907 #endif /* TCG_TARGET_EXTEND_ARGS */
908 }
909 
910 static void tcg_reg_alloc_start(TCGContext *s)
911 {
912     int i;
913     TCGTemp *ts;
914     for(i = 0; i < s->nb_globals; i++) {
915         ts = &s->temps[i];
916         if (ts->fixed_reg) {
917             ts->val_type = TEMP_VAL_REG;
918         } else {
919             ts->val_type = TEMP_VAL_MEM;
920         }
921     }
922     for(i = s->nb_globals; i < s->nb_temps; i++) {
923         ts = &s->temps[i];
924         if (ts->temp_local) {
925             ts->val_type = TEMP_VAL_MEM;
926         } else {
927             ts->val_type = TEMP_VAL_DEAD;
928         }
929         ts->mem_allocated = 0;
930         ts->fixed_reg = 0;
931     }
932 
933     memset(s->reg_to_temp, 0, sizeof(s->reg_to_temp));
934 }
935 
936 static char *tcg_get_arg_str_ptr(TCGContext *s, char *buf, int buf_size,
937                                  TCGTemp *ts)
938 {
939     int idx = temp_idx(s, ts);
940 
941     if (idx < s->nb_globals) {
942         pstrcpy(buf, buf_size, ts->name);
943     } else if (ts->temp_local) {
944         snprintf(buf, buf_size, "loc%d", idx - s->nb_globals);
945     } else {
946         snprintf(buf, buf_size, "tmp%d", idx - s->nb_globals);
947     }
948     return buf;
949 }
950 
951 static char *tcg_get_arg_str_idx(TCGContext *s, char *buf,
952                                  int buf_size, int idx)
953 {
954     tcg_debug_assert(idx >= 0 && idx < s->nb_temps);
955     return tcg_get_arg_str_ptr(s, buf, buf_size, &s->temps[idx]);
956 }
957 
958 /* Find helper name.  */
959 static inline const char *tcg_find_helper(TCGContext *s, uintptr_t val)
960 {
961     const char *ret = NULL;
962     if (s->helpers) {
963         TCGHelperInfo *info = g_hash_table_lookup(s->helpers, (gpointer)val);
964         if (info) {
965             ret = info->name;
966         }
967     }
968     return ret;
969 }
970 
971 static const char * const cond_name[] =
972 {
973     [TCG_COND_NEVER] = "never",
974     [TCG_COND_ALWAYS] = "always",
975     [TCG_COND_EQ] = "eq",
976     [TCG_COND_NE] = "ne",
977     [TCG_COND_LT] = "lt",
978     [TCG_COND_GE] = "ge",
979     [TCG_COND_LE] = "le",
980     [TCG_COND_GT] = "gt",
981     [TCG_COND_LTU] = "ltu",
982     [TCG_COND_GEU] = "geu",
983     [TCG_COND_LEU] = "leu",
984     [TCG_COND_GTU] = "gtu"
985 };
986 
987 static const char * const ldst_name[] =
988 {
989     [MO_UB]   = "ub",
990     [MO_SB]   = "sb",
991     [MO_LEUW] = "leuw",
992     [MO_LESW] = "lesw",
993     [MO_LEUL] = "leul",
994     [MO_LESL] = "lesl",
995     [MO_LEQ]  = "leq",
996     [MO_BEUW] = "beuw",
997     [MO_BESW] = "besw",
998     [MO_BEUL] = "beul",
999     [MO_BESL] = "besl",
1000     [MO_BEQ]  = "beq",
1001 };
1002 
1003 static const char * const alignment_name[(MO_AMASK >> MO_ASHIFT) + 1] = {
1004 #ifdef ALIGNED_ONLY
1005     [MO_UNALN >> MO_ASHIFT]    = "un+",
1006     [MO_ALIGN >> MO_ASHIFT]    = "",
1007 #else
1008     [MO_UNALN >> MO_ASHIFT]    = "",
1009     [MO_ALIGN >> MO_ASHIFT]    = "al+",
1010 #endif
1011     [MO_ALIGN_2 >> MO_ASHIFT]  = "al2+",
1012     [MO_ALIGN_4 >> MO_ASHIFT]  = "al4+",
1013     [MO_ALIGN_8 >> MO_ASHIFT]  = "al8+",
1014     [MO_ALIGN_16 >> MO_ASHIFT] = "al16+",
1015     [MO_ALIGN_32 >> MO_ASHIFT] = "al32+",
1016     [MO_ALIGN_64 >> MO_ASHIFT] = "al64+",
1017 };
1018 
1019 void tcg_dump_ops(TCGContext *s)
1020 {
1021     char buf[128];
1022     TCGOp *op;
1023     int oi;
1024 
1025     for (oi = s->gen_op_buf[0].next; oi != 0; oi = op->next) {
1026         int i, k, nb_oargs, nb_iargs, nb_cargs;
1027         const TCGOpDef *def;
1028         const TCGArg *args;
1029         TCGOpcode c;
1030         int col = 0;
1031 
1032         op = &s->gen_op_buf[oi];
1033         c = op->opc;
1034         def = &tcg_op_defs[c];
1035         args = &s->gen_opparam_buf[op->args];
1036 
1037         if (c == INDEX_op_insn_start) {
1038             col += qemu_log("%s ----", oi != s->gen_op_buf[0].next ? "\n" : "");
1039 
1040             for (i = 0; i < TARGET_INSN_START_WORDS; ++i) {
1041                 target_ulong a;
1042 #if TARGET_LONG_BITS > TCG_TARGET_REG_BITS
1043                 a = ((target_ulong)args[i * 2 + 1] << 32) | args[i * 2];
1044 #else
1045                 a = args[i];
1046 #endif
1047                 col += qemu_log(" " TARGET_FMT_lx, a);
1048             }
1049         } else if (c == INDEX_op_call) {
1050             /* variable number of arguments */
1051             nb_oargs = op->callo;
1052             nb_iargs = op->calli;
1053             nb_cargs = def->nb_cargs;
1054 
1055             /* function name, flags, out args */
1056             col += qemu_log(" %s %s,$0x%" TCG_PRIlx ",$%d", def->name,
1057                             tcg_find_helper(s, args[nb_oargs + nb_iargs]),
1058                             args[nb_oargs + nb_iargs + 1], nb_oargs);
1059             for (i = 0; i < nb_oargs; i++) {
1060                 col += qemu_log(",%s", tcg_get_arg_str_idx(s, buf, sizeof(buf),
1061                                                            args[i]));
1062             }
1063             for (i = 0; i < nb_iargs; i++) {
1064                 TCGArg arg = args[nb_oargs + i];
1065                 const char *t = "<dummy>";
1066                 if (arg != TCG_CALL_DUMMY_ARG) {
1067                     t = tcg_get_arg_str_idx(s, buf, sizeof(buf), arg);
1068                 }
1069                 col += qemu_log(",%s", t);
1070             }
1071         } else {
1072             col += qemu_log(" %s ", def->name);
1073 
1074             nb_oargs = def->nb_oargs;
1075             nb_iargs = def->nb_iargs;
1076             nb_cargs = def->nb_cargs;
1077 
1078             k = 0;
1079             for (i = 0; i < nb_oargs; i++) {
1080                 if (k != 0) {
1081                     col += qemu_log(",");
1082                 }
1083                 col += qemu_log("%s", tcg_get_arg_str_idx(s, buf, sizeof(buf),
1084                                                           args[k++]));
1085             }
1086             for (i = 0; i < nb_iargs; i++) {
1087                 if (k != 0) {
1088                     col += qemu_log(",");
1089                 }
1090                 col += qemu_log("%s", tcg_get_arg_str_idx(s, buf, sizeof(buf),
1091                                                           args[k++]));
1092             }
1093             switch (c) {
1094             case INDEX_op_brcond_i32:
1095             case INDEX_op_setcond_i32:
1096             case INDEX_op_movcond_i32:
1097             case INDEX_op_brcond2_i32:
1098             case INDEX_op_setcond2_i32:
1099             case INDEX_op_brcond_i64:
1100             case INDEX_op_setcond_i64:
1101             case INDEX_op_movcond_i64:
1102                 if (args[k] < ARRAY_SIZE(cond_name) && cond_name[args[k]]) {
1103                     col += qemu_log(",%s", cond_name[args[k++]]);
1104                 } else {
1105                     col += qemu_log(",$0x%" TCG_PRIlx, args[k++]);
1106                 }
1107                 i = 1;
1108                 break;
1109             case INDEX_op_qemu_ld_i32:
1110             case INDEX_op_qemu_st_i32:
1111             case INDEX_op_qemu_ld_i64:
1112             case INDEX_op_qemu_st_i64:
1113                 {
1114                     TCGMemOpIdx oi = args[k++];
1115                     TCGMemOp op = get_memop(oi);
1116                     unsigned ix = get_mmuidx(oi);
1117 
1118                     if (op & ~(MO_AMASK | MO_BSWAP | MO_SSIZE)) {
1119                         col += qemu_log(",$0x%x,%u", op, ix);
1120                     } else {
1121                         const char *s_al, *s_op;
1122                         s_al = alignment_name[(op & MO_AMASK) >> MO_ASHIFT];
1123                         s_op = ldst_name[op & (MO_BSWAP | MO_SSIZE)];
1124                         col += qemu_log(",%s%s,%u", s_al, s_op, ix);
1125                     }
1126                     i = 1;
1127                 }
1128                 break;
1129             default:
1130                 i = 0;
1131                 break;
1132             }
1133             switch (c) {
1134             case INDEX_op_set_label:
1135             case INDEX_op_br:
1136             case INDEX_op_brcond_i32:
1137             case INDEX_op_brcond_i64:
1138             case INDEX_op_brcond2_i32:
1139                 col += qemu_log("%s$L%d", k ? "," : "", arg_label(args[k])->id);
1140                 i++, k++;
1141                 break;
1142             default:
1143                 break;
1144             }
1145             for (; i < nb_cargs; i++, k++) {
1146                 col += qemu_log("%s$0x%" TCG_PRIlx, k ? "," : "", args[k]);
1147             }
1148         }
1149         if (op->life) {
1150             unsigned life = op->life;
1151 
1152             for (; col < 48; ++col) {
1153                 putc(' ', qemu_logfile);
1154             }
1155 
1156             if (life & (SYNC_ARG * 3)) {
1157                 qemu_log("  sync:");
1158                 for (i = 0; i < 2; ++i) {
1159                     if (life & (SYNC_ARG << i)) {
1160                         qemu_log(" %d", i);
1161                     }
1162                 }
1163             }
1164             life /= DEAD_ARG;
1165             if (life) {
1166                 qemu_log("  dead:");
1167                 for (i = 0; life; ++i, life >>= 1) {
1168                     if (life & 1) {
1169                         qemu_log(" %d", i);
1170                     }
1171                 }
1172             }
1173         }
1174         qemu_log("\n");
1175     }
1176 }
1177 
1178 /* we give more priority to constraints with less registers */
1179 static int get_constraint_priority(const TCGOpDef *def, int k)
1180 {
1181     const TCGArgConstraint *arg_ct;
1182 
1183     int i, n;
1184     arg_ct = &def->args_ct[k];
1185     if (arg_ct->ct & TCG_CT_ALIAS) {
1186         /* an alias is equivalent to a single register */
1187         n = 1;
1188     } else {
1189         if (!(arg_ct->ct & TCG_CT_REG))
1190             return 0;
1191         n = 0;
1192         for(i = 0; i < TCG_TARGET_NB_REGS; i++) {
1193             if (tcg_regset_test_reg(arg_ct->u.regs, i))
1194                 n++;
1195         }
1196     }
1197     return TCG_TARGET_NB_REGS - n + 1;
1198 }
1199 
1200 /* sort from highest priority to lowest */
1201 static void sort_constraints(TCGOpDef *def, int start, int n)
1202 {
1203     int i, j, p1, p2, tmp;
1204 
1205     for(i = 0; i < n; i++)
1206         def->sorted_args[start + i] = start + i;
1207     if (n <= 1)
1208         return;
1209     for(i = 0; i < n - 1; i++) {
1210         for(j = i + 1; j < n; j++) {
1211             p1 = get_constraint_priority(def, def->sorted_args[start + i]);
1212             p2 = get_constraint_priority(def, def->sorted_args[start + j]);
1213             if (p1 < p2) {
1214                 tmp = def->sorted_args[start + i];
1215                 def->sorted_args[start + i] = def->sorted_args[start + j];
1216                 def->sorted_args[start + j] = tmp;
1217             }
1218         }
1219     }
1220 }
1221 
1222 void tcg_add_target_add_op_defs(const TCGTargetOpDef *tdefs)
1223 {
1224     TCGOpcode op;
1225     TCGOpDef *def;
1226     const char *ct_str;
1227     int i, nb_args;
1228 
1229     for(;;) {
1230         if (tdefs->op == (TCGOpcode)-1)
1231             break;
1232         op = tdefs->op;
1233         tcg_debug_assert((unsigned)op < NB_OPS);
1234         def = &tcg_op_defs[op];
1235 #if defined(CONFIG_DEBUG_TCG)
1236         /* Duplicate entry in op definitions? */
1237         tcg_debug_assert(!def->used);
1238         def->used = 1;
1239 #endif
1240         nb_args = def->nb_iargs + def->nb_oargs;
1241         for(i = 0; i < nb_args; i++) {
1242             ct_str = tdefs->args_ct_str[i];
1243             /* Incomplete TCGTargetOpDef entry? */
1244             tcg_debug_assert(ct_str != NULL);
1245             tcg_regset_clear(def->args_ct[i].u.regs);
1246             def->args_ct[i].ct = 0;
1247             if (ct_str[0] >= '0' && ct_str[0] <= '9') {
1248                 int oarg;
1249                 oarg = ct_str[0] - '0';
1250                 tcg_debug_assert(oarg < def->nb_oargs);
1251                 tcg_debug_assert(def->args_ct[oarg].ct & TCG_CT_REG);
1252                 /* TCG_CT_ALIAS is for the output arguments. The input
1253                    argument is tagged with TCG_CT_IALIAS. */
1254                 def->args_ct[i] = def->args_ct[oarg];
1255                 def->args_ct[oarg].ct = TCG_CT_ALIAS;
1256                 def->args_ct[oarg].alias_index = i;
1257                 def->args_ct[i].ct |= TCG_CT_IALIAS;
1258                 def->args_ct[i].alias_index = oarg;
1259             } else {
1260                 for(;;) {
1261                     if (*ct_str == '\0')
1262                         break;
1263                     switch(*ct_str) {
1264                     case 'i':
1265                         def->args_ct[i].ct |= TCG_CT_CONST;
1266                         ct_str++;
1267                         break;
1268                     default:
1269                         if (target_parse_constraint(&def->args_ct[i], &ct_str) < 0) {
1270                             fprintf(stderr, "Invalid constraint '%s' for arg %d of operation '%s'\n",
1271                                     ct_str, i, def->name);
1272                             exit(1);
1273                         }
1274                     }
1275                 }
1276             }
1277         }
1278 
1279         /* TCGTargetOpDef entry with too much information? */
1280         tcg_debug_assert(i == TCG_MAX_OP_ARGS || tdefs->args_ct_str[i] == NULL);
1281 
1282         /* sort the constraints (XXX: this is just an heuristic) */
1283         sort_constraints(def, 0, def->nb_oargs);
1284         sort_constraints(def, def->nb_oargs, def->nb_iargs);
1285 
1286 #if 0
1287         {
1288             int i;
1289 
1290             printf("%s: sorted=", def->name);
1291             for(i = 0; i < def->nb_oargs + def->nb_iargs; i++)
1292                 printf(" %d", def->sorted_args[i]);
1293             printf("\n");
1294         }
1295 #endif
1296         tdefs++;
1297     }
1298 
1299 #if defined(CONFIG_DEBUG_TCG)
1300     i = 0;
1301     for (op = 0; op < tcg_op_defs_max; op++) {
1302         const TCGOpDef *def = &tcg_op_defs[op];
1303         if (def->flags & TCG_OPF_NOT_PRESENT) {
1304             /* Wrong entry in op definitions? */
1305             if (def->used) {
1306                 fprintf(stderr, "Invalid op definition for %s\n", def->name);
1307                 i = 1;
1308             }
1309         } else {
1310             /* Missing entry in op definitions? */
1311             if (!def->used) {
1312                 fprintf(stderr, "Missing op definition for %s\n", def->name);
1313                 i = 1;
1314             }
1315         }
1316     }
1317     if (i == 1) {
1318         tcg_abort();
1319     }
1320 #endif
1321 }
1322 
1323 void tcg_op_remove(TCGContext *s, TCGOp *op)
1324 {
1325     int next = op->next;
1326     int prev = op->prev;
1327 
1328     /* We should never attempt to remove the list terminator.  */
1329     tcg_debug_assert(op != &s->gen_op_buf[0]);
1330 
1331     s->gen_op_buf[next].prev = prev;
1332     s->gen_op_buf[prev].next = next;
1333 
1334     memset(op, 0, sizeof(*op));
1335 
1336 #ifdef CONFIG_PROFILER
1337     s->del_op_count++;
1338 #endif
1339 }
1340 
1341 TCGOp *tcg_op_insert_before(TCGContext *s, TCGOp *old_op,
1342                             TCGOpcode opc, int nargs)
1343 {
1344     int oi = s->gen_next_op_idx;
1345     int pi = s->gen_next_parm_idx;
1346     int prev = old_op->prev;
1347     int next = old_op - s->gen_op_buf;
1348     TCGOp *new_op;
1349 
1350     tcg_debug_assert(oi < OPC_BUF_SIZE);
1351     tcg_debug_assert(pi + nargs <= OPPARAM_BUF_SIZE);
1352     s->gen_next_op_idx = oi + 1;
1353     s->gen_next_parm_idx = pi + nargs;
1354 
1355     new_op = &s->gen_op_buf[oi];
1356     *new_op = (TCGOp){
1357         .opc = opc,
1358         .args = pi,
1359         .prev = prev,
1360         .next = next
1361     };
1362     s->gen_op_buf[prev].next = oi;
1363     old_op->prev = oi;
1364 
1365     return new_op;
1366 }
1367 
1368 TCGOp *tcg_op_insert_after(TCGContext *s, TCGOp *old_op,
1369                            TCGOpcode opc, int nargs)
1370 {
1371     int oi = s->gen_next_op_idx;
1372     int pi = s->gen_next_parm_idx;
1373     int prev = old_op - s->gen_op_buf;
1374     int next = old_op->next;
1375     TCGOp *new_op;
1376 
1377     tcg_debug_assert(oi < OPC_BUF_SIZE);
1378     tcg_debug_assert(pi + nargs <= OPPARAM_BUF_SIZE);
1379     s->gen_next_op_idx = oi + 1;
1380     s->gen_next_parm_idx = pi + nargs;
1381 
1382     new_op = &s->gen_op_buf[oi];
1383     *new_op = (TCGOp){
1384         .opc = opc,
1385         .args = pi,
1386         .prev = prev,
1387         .next = next
1388     };
1389     s->gen_op_buf[next].prev = oi;
1390     old_op->next = oi;
1391 
1392     return new_op;
1393 }
1394 
1395 #define TS_DEAD  1
1396 #define TS_MEM   2
1397 
1398 #define IS_DEAD_ARG(n)   (arg_life & (DEAD_ARG << (n)))
1399 #define NEED_SYNC_ARG(n) (arg_life & (SYNC_ARG << (n)))
1400 
1401 /* liveness analysis: end of function: all temps are dead, and globals
1402    should be in memory. */
1403 static inline void tcg_la_func_end(TCGContext *s, uint8_t *temp_state)
1404 {
1405     memset(temp_state, TS_DEAD | TS_MEM, s->nb_globals);
1406     memset(temp_state + s->nb_globals, TS_DEAD, s->nb_temps - s->nb_globals);
1407 }
1408 
1409 /* liveness analysis: end of basic block: all temps are dead, globals
1410    and local temps should be in memory. */
1411 static inline void tcg_la_bb_end(TCGContext *s, uint8_t *temp_state)
1412 {
1413     int i, n;
1414 
1415     tcg_la_func_end(s, temp_state);
1416     for (i = s->nb_globals, n = s->nb_temps; i < n; i++) {
1417         if (s->temps[i].temp_local) {
1418             temp_state[i] |= TS_MEM;
1419         }
1420     }
1421 }
1422 
1423 /* Liveness analysis : update the opc_arg_life array to tell if a
1424    given input arguments is dead. Instructions updating dead
1425    temporaries are removed. */
1426 static void liveness_pass_1(TCGContext *s, uint8_t *temp_state)
1427 {
1428     int nb_globals = s->nb_globals;
1429     int oi, oi_prev;
1430 
1431     tcg_la_func_end(s, temp_state);
1432 
1433     for (oi = s->gen_op_buf[0].prev; oi != 0; oi = oi_prev) {
1434         int i, nb_iargs, nb_oargs;
1435         TCGOpcode opc_new, opc_new2;
1436         bool have_opc_new2;
1437         TCGLifeData arg_life = 0;
1438         TCGArg arg;
1439 
1440         TCGOp * const op = &s->gen_op_buf[oi];
1441         TCGArg * const args = &s->gen_opparam_buf[op->args];
1442         TCGOpcode opc = op->opc;
1443         const TCGOpDef *def = &tcg_op_defs[opc];
1444 
1445         oi_prev = op->prev;
1446 
1447         switch (opc) {
1448         case INDEX_op_call:
1449             {
1450                 int call_flags;
1451 
1452                 nb_oargs = op->callo;
1453                 nb_iargs = op->calli;
1454                 call_flags = args[nb_oargs + nb_iargs + 1];
1455 
1456                 /* pure functions can be removed if their result is unused */
1457                 if (call_flags & TCG_CALL_NO_SIDE_EFFECTS) {
1458                     for (i = 0; i < nb_oargs; i++) {
1459                         arg = args[i];
1460                         if (temp_state[arg] != TS_DEAD) {
1461                             goto do_not_remove_call;
1462                         }
1463                     }
1464                     goto do_remove;
1465                 } else {
1466                 do_not_remove_call:
1467 
1468                     /* output args are dead */
1469                     for (i = 0; i < nb_oargs; i++) {
1470                         arg = args[i];
1471                         if (temp_state[arg] & TS_DEAD) {
1472                             arg_life |= DEAD_ARG << i;
1473                         }
1474                         if (temp_state[arg] & TS_MEM) {
1475                             arg_life |= SYNC_ARG << i;
1476                         }
1477                         temp_state[arg] = TS_DEAD;
1478                     }
1479 
1480                     if (!(call_flags & (TCG_CALL_NO_WRITE_GLOBALS |
1481                                         TCG_CALL_NO_READ_GLOBALS))) {
1482                         /* globals should go back to memory */
1483                         memset(temp_state, TS_DEAD | TS_MEM, nb_globals);
1484                     } else if (!(call_flags & TCG_CALL_NO_READ_GLOBALS)) {
1485                         /* globals should be synced to memory */
1486                         for (i = 0; i < nb_globals; i++) {
1487                             temp_state[i] |= TS_MEM;
1488                         }
1489                     }
1490 
1491                     /* record arguments that die in this helper */
1492                     for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
1493                         arg = args[i];
1494                         if (arg != TCG_CALL_DUMMY_ARG) {
1495                             if (temp_state[arg] & TS_DEAD) {
1496                                 arg_life |= DEAD_ARG << i;
1497                             }
1498                         }
1499                     }
1500                     /* input arguments are live for preceding opcodes */
1501                     for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
1502                         arg = args[i];
1503                         if (arg != TCG_CALL_DUMMY_ARG) {
1504                             temp_state[arg] &= ~TS_DEAD;
1505                         }
1506                     }
1507                 }
1508             }
1509             break;
1510         case INDEX_op_insn_start:
1511             break;
1512         case INDEX_op_discard:
1513             /* mark the temporary as dead */
1514             temp_state[args[0]] = TS_DEAD;
1515             break;
1516 
1517         case INDEX_op_add2_i32:
1518             opc_new = INDEX_op_add_i32;
1519             goto do_addsub2;
1520         case INDEX_op_sub2_i32:
1521             opc_new = INDEX_op_sub_i32;
1522             goto do_addsub2;
1523         case INDEX_op_add2_i64:
1524             opc_new = INDEX_op_add_i64;
1525             goto do_addsub2;
1526         case INDEX_op_sub2_i64:
1527             opc_new = INDEX_op_sub_i64;
1528         do_addsub2:
1529             nb_iargs = 4;
1530             nb_oargs = 2;
1531             /* Test if the high part of the operation is dead, but not
1532                the low part.  The result can be optimized to a simple
1533                add or sub.  This happens often for x86_64 guest when the
1534                cpu mode is set to 32 bit.  */
1535             if (temp_state[args[1]] == TS_DEAD) {
1536                 if (temp_state[args[0]] == TS_DEAD) {
1537                     goto do_remove;
1538                 }
1539                 /* Replace the opcode and adjust the args in place,
1540                    leaving 3 unused args at the end.  */
1541                 op->opc = opc = opc_new;
1542                 args[1] = args[2];
1543                 args[2] = args[4];
1544                 /* Fall through and mark the single-word operation live.  */
1545                 nb_iargs = 2;
1546                 nb_oargs = 1;
1547             }
1548             goto do_not_remove;
1549 
1550         case INDEX_op_mulu2_i32:
1551             opc_new = INDEX_op_mul_i32;
1552             opc_new2 = INDEX_op_muluh_i32;
1553             have_opc_new2 = TCG_TARGET_HAS_muluh_i32;
1554             goto do_mul2;
1555         case INDEX_op_muls2_i32:
1556             opc_new = INDEX_op_mul_i32;
1557             opc_new2 = INDEX_op_mulsh_i32;
1558             have_opc_new2 = TCG_TARGET_HAS_mulsh_i32;
1559             goto do_mul2;
1560         case INDEX_op_mulu2_i64:
1561             opc_new = INDEX_op_mul_i64;
1562             opc_new2 = INDEX_op_muluh_i64;
1563             have_opc_new2 = TCG_TARGET_HAS_muluh_i64;
1564             goto do_mul2;
1565         case INDEX_op_muls2_i64:
1566             opc_new = INDEX_op_mul_i64;
1567             opc_new2 = INDEX_op_mulsh_i64;
1568             have_opc_new2 = TCG_TARGET_HAS_mulsh_i64;
1569             goto do_mul2;
1570         do_mul2:
1571             nb_iargs = 2;
1572             nb_oargs = 2;
1573             if (temp_state[args[1]] == TS_DEAD) {
1574                 if (temp_state[args[0]] == TS_DEAD) {
1575                     /* Both parts of the operation are dead.  */
1576                     goto do_remove;
1577                 }
1578                 /* The high part of the operation is dead; generate the low. */
1579                 op->opc = opc = opc_new;
1580                 args[1] = args[2];
1581                 args[2] = args[3];
1582             } else if (temp_state[args[0]] == TS_DEAD && have_opc_new2) {
1583                 /* The low part of the operation is dead; generate the high. */
1584                 op->opc = opc = opc_new2;
1585                 args[0] = args[1];
1586                 args[1] = args[2];
1587                 args[2] = args[3];
1588             } else {
1589                 goto do_not_remove;
1590             }
1591             /* Mark the single-word operation live.  */
1592             nb_oargs = 1;
1593             goto do_not_remove;
1594 
1595         default:
1596             /* XXX: optimize by hardcoding common cases (e.g. triadic ops) */
1597             nb_iargs = def->nb_iargs;
1598             nb_oargs = def->nb_oargs;
1599 
1600             /* Test if the operation can be removed because all
1601                its outputs are dead. We assume that nb_oargs == 0
1602                implies side effects */
1603             if (!(def->flags & TCG_OPF_SIDE_EFFECTS) && nb_oargs != 0) {
1604                 for (i = 0; i < nb_oargs; i++) {
1605                     if (temp_state[args[i]] != TS_DEAD) {
1606                         goto do_not_remove;
1607                     }
1608                 }
1609             do_remove:
1610                 tcg_op_remove(s, op);
1611             } else {
1612             do_not_remove:
1613                 /* output args are dead */
1614                 for (i = 0; i < nb_oargs; i++) {
1615                     arg = args[i];
1616                     if (temp_state[arg] & TS_DEAD) {
1617                         arg_life |= DEAD_ARG << i;
1618                     }
1619                     if (temp_state[arg] & TS_MEM) {
1620                         arg_life |= SYNC_ARG << i;
1621                     }
1622                     temp_state[arg] = TS_DEAD;
1623                 }
1624 
1625                 /* if end of basic block, update */
1626                 if (def->flags & TCG_OPF_BB_END) {
1627                     tcg_la_bb_end(s, temp_state);
1628                 } else if (def->flags & TCG_OPF_SIDE_EFFECTS) {
1629                     /* globals should be synced to memory */
1630                     for (i = 0; i < nb_globals; i++) {
1631                         temp_state[i] |= TS_MEM;
1632                     }
1633                 }
1634 
1635                 /* record arguments that die in this opcode */
1636                 for (i = nb_oargs; i < nb_oargs + nb_iargs; i++) {
1637                     arg = args[i];
1638                     if (temp_state[arg] & TS_DEAD) {
1639                         arg_life |= DEAD_ARG << i;
1640                     }
1641                 }
1642                 /* input arguments are live for preceding opcodes */
1643                 for (i = nb_oargs; i < nb_oargs + nb_iargs; i++) {
1644                     temp_state[args[i]] &= ~TS_DEAD;
1645                 }
1646             }
1647             break;
1648         }
1649         op->life = arg_life;
1650     }
1651 }
1652 
1653 /* Liveness analysis: Convert indirect regs to direct temporaries.  */
1654 static bool liveness_pass_2(TCGContext *s, uint8_t *temp_state)
1655 {
1656     int nb_globals = s->nb_globals;
1657     int16_t *dir_temps;
1658     int i, oi, oi_next;
1659     bool changes = false;
1660 
1661     dir_temps = tcg_malloc(nb_globals * sizeof(int16_t));
1662     memset(dir_temps, 0, nb_globals * sizeof(int16_t));
1663 
1664     /* Create a temporary for each indirect global.  */
1665     for (i = 0; i < nb_globals; ++i) {
1666         TCGTemp *its = &s->temps[i];
1667         if (its->indirect_reg) {
1668             TCGTemp *dts = tcg_temp_alloc(s);
1669             dts->type = its->type;
1670             dts->base_type = its->base_type;
1671             dir_temps[i] = temp_idx(s, dts);
1672         }
1673     }
1674 
1675     memset(temp_state, TS_DEAD, nb_globals);
1676 
1677     for (oi = s->gen_op_buf[0].next; oi != 0; oi = oi_next) {
1678         TCGOp *op = &s->gen_op_buf[oi];
1679         TCGArg *args = &s->gen_opparam_buf[op->args];
1680         TCGOpcode opc = op->opc;
1681         const TCGOpDef *def = &tcg_op_defs[opc];
1682         TCGLifeData arg_life = op->life;
1683         int nb_iargs, nb_oargs, call_flags;
1684         TCGArg arg, dir;
1685 
1686         oi_next = op->next;
1687 
1688         if (opc == INDEX_op_call) {
1689             nb_oargs = op->callo;
1690             nb_iargs = op->calli;
1691             call_flags = args[nb_oargs + nb_iargs + 1];
1692         } else {
1693             nb_iargs = def->nb_iargs;
1694             nb_oargs = def->nb_oargs;
1695 
1696             /* Set flags similar to how calls require.  */
1697             if (def->flags & TCG_OPF_BB_END) {
1698                 /* Like writing globals: save_globals */
1699                 call_flags = 0;
1700             } else if (def->flags & TCG_OPF_SIDE_EFFECTS) {
1701                 /* Like reading globals: sync_globals */
1702                 call_flags = TCG_CALL_NO_WRITE_GLOBALS;
1703             } else {
1704                 /* No effect on globals.  */
1705                 call_flags = (TCG_CALL_NO_READ_GLOBALS |
1706                               TCG_CALL_NO_WRITE_GLOBALS);
1707             }
1708         }
1709 
1710         /* Make sure that input arguments are available.  */
1711         for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
1712             arg = args[i];
1713             /* Note this unsigned test catches TCG_CALL_ARG_DUMMY too.  */
1714             if (arg < nb_globals) {
1715                 dir = dir_temps[arg];
1716                 if (dir != 0 && temp_state[arg] == TS_DEAD) {
1717                     TCGTemp *its = &s->temps[arg];
1718                     TCGOpcode lopc = (its->type == TCG_TYPE_I32
1719                                       ? INDEX_op_ld_i32
1720                                       : INDEX_op_ld_i64);
1721                     TCGOp *lop = tcg_op_insert_before(s, op, lopc, 3);
1722                     TCGArg *largs = &s->gen_opparam_buf[lop->args];
1723 
1724                     largs[0] = dir;
1725                     largs[1] = temp_idx(s, its->mem_base);
1726                     largs[2] = its->mem_offset;
1727 
1728                     /* Loaded, but synced with memory.  */
1729                     temp_state[arg] = TS_MEM;
1730                 }
1731             }
1732         }
1733 
1734         /* Perform input replacement, and mark inputs that became dead.
1735            No action is required except keeping temp_state up to date
1736            so that we reload when needed.  */
1737         for (i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
1738             arg = args[i];
1739             if (arg < nb_globals) {
1740                 dir = dir_temps[arg];
1741                 if (dir != 0) {
1742                     args[i] = dir;
1743                     changes = true;
1744                     if (IS_DEAD_ARG(i)) {
1745                         temp_state[arg] = TS_DEAD;
1746                     }
1747                 }
1748             }
1749         }
1750 
1751         /* Liveness analysis should ensure that the following are
1752            all correct, for call sites and basic block end points.  */
1753         if (call_flags & TCG_CALL_NO_READ_GLOBALS) {
1754             /* Nothing to do */
1755         } else if (call_flags & TCG_CALL_NO_WRITE_GLOBALS) {
1756             for (i = 0; i < nb_globals; ++i) {
1757                 /* Liveness should see that globals are synced back,
1758                    that is, either TS_DEAD or TS_MEM.  */
1759                 tcg_debug_assert(dir_temps[i] == 0
1760                                  || temp_state[i] != 0);
1761             }
1762         } else {
1763             for (i = 0; i < nb_globals; ++i) {
1764                 /* Liveness should see that globals are saved back,
1765                    that is, TS_DEAD, waiting to be reloaded.  */
1766                 tcg_debug_assert(dir_temps[i] == 0
1767                                  || temp_state[i] == TS_DEAD);
1768             }
1769         }
1770 
1771         /* Outputs become available.  */
1772         for (i = 0; i < nb_oargs; i++) {
1773             arg = args[i];
1774             if (arg >= nb_globals) {
1775                 continue;
1776             }
1777             dir = dir_temps[arg];
1778             if (dir == 0) {
1779                 continue;
1780             }
1781             args[i] = dir;
1782             changes = true;
1783 
1784             /* The output is now live and modified.  */
1785             temp_state[arg] = 0;
1786 
1787             /* Sync outputs upon their last write.  */
1788             if (NEED_SYNC_ARG(i)) {
1789                 TCGTemp *its = &s->temps[arg];
1790                 TCGOpcode sopc = (its->type == TCG_TYPE_I32
1791                                   ? INDEX_op_st_i32
1792                                   : INDEX_op_st_i64);
1793                 TCGOp *sop = tcg_op_insert_after(s, op, sopc, 3);
1794                 TCGArg *sargs = &s->gen_opparam_buf[sop->args];
1795 
1796                 sargs[0] = dir;
1797                 sargs[1] = temp_idx(s, its->mem_base);
1798                 sargs[2] = its->mem_offset;
1799 
1800                 temp_state[arg] = TS_MEM;
1801             }
1802             /* Drop outputs that are dead.  */
1803             if (IS_DEAD_ARG(i)) {
1804                 temp_state[arg] = TS_DEAD;
1805             }
1806         }
1807     }
1808 
1809     return changes;
1810 }
1811 
1812 #ifdef CONFIG_DEBUG_TCG
1813 static void dump_regs(TCGContext *s)
1814 {
1815     TCGTemp *ts;
1816     int i;
1817     char buf[64];
1818 
1819     for(i = 0; i < s->nb_temps; i++) {
1820         ts = &s->temps[i];
1821         printf("  %10s: ", tcg_get_arg_str_idx(s, buf, sizeof(buf), i));
1822         switch(ts->val_type) {
1823         case TEMP_VAL_REG:
1824             printf("%s", tcg_target_reg_names[ts->reg]);
1825             break;
1826         case TEMP_VAL_MEM:
1827             printf("%d(%s)", (int)ts->mem_offset,
1828                    tcg_target_reg_names[ts->mem_base->reg]);
1829             break;
1830         case TEMP_VAL_CONST:
1831             printf("$0x%" TCG_PRIlx, ts->val);
1832             break;
1833         case TEMP_VAL_DEAD:
1834             printf("D");
1835             break;
1836         default:
1837             printf("???");
1838             break;
1839         }
1840         printf("\n");
1841     }
1842 
1843     for(i = 0; i < TCG_TARGET_NB_REGS; i++) {
1844         if (s->reg_to_temp[i] != NULL) {
1845             printf("%s: %s\n",
1846                    tcg_target_reg_names[i],
1847                    tcg_get_arg_str_ptr(s, buf, sizeof(buf), s->reg_to_temp[i]));
1848         }
1849     }
1850 }
1851 
1852 static void check_regs(TCGContext *s)
1853 {
1854     int reg;
1855     int k;
1856     TCGTemp *ts;
1857     char buf[64];
1858 
1859     for (reg = 0; reg < TCG_TARGET_NB_REGS; reg++) {
1860         ts = s->reg_to_temp[reg];
1861         if (ts != NULL) {
1862             if (ts->val_type != TEMP_VAL_REG || ts->reg != reg) {
1863                 printf("Inconsistency for register %s:\n",
1864                        tcg_target_reg_names[reg]);
1865                 goto fail;
1866             }
1867         }
1868     }
1869     for (k = 0; k < s->nb_temps; k++) {
1870         ts = &s->temps[k];
1871         if (ts->val_type == TEMP_VAL_REG && !ts->fixed_reg
1872             && s->reg_to_temp[ts->reg] != ts) {
1873             printf("Inconsistency for temp %s:\n",
1874                    tcg_get_arg_str_ptr(s, buf, sizeof(buf), ts));
1875         fail:
1876             printf("reg state:\n");
1877             dump_regs(s);
1878             tcg_abort();
1879         }
1880     }
1881 }
1882 #endif
1883 
1884 static void temp_allocate_frame(TCGContext *s, int temp)
1885 {
1886     TCGTemp *ts;
1887     ts = &s->temps[temp];
1888 #if !(defined(__sparc__) && TCG_TARGET_REG_BITS == 64)
1889     /* Sparc64 stack is accessed with offset of 2047 */
1890     s->current_frame_offset = (s->current_frame_offset +
1891                                (tcg_target_long)sizeof(tcg_target_long) - 1) &
1892         ~(sizeof(tcg_target_long) - 1);
1893 #endif
1894     if (s->current_frame_offset + (tcg_target_long)sizeof(tcg_target_long) >
1895         s->frame_end) {
1896         tcg_abort();
1897     }
1898     ts->mem_offset = s->current_frame_offset;
1899     ts->mem_base = s->frame_temp;
1900     ts->mem_allocated = 1;
1901     s->current_frame_offset += sizeof(tcg_target_long);
1902 }
1903 
1904 static void temp_load(TCGContext *, TCGTemp *, TCGRegSet, TCGRegSet);
1905 
1906 /* Mark a temporary as free or dead.  If 'free_or_dead' is negative,
1907    mark it free; otherwise mark it dead.  */
1908 static void temp_free_or_dead(TCGContext *s, TCGTemp *ts, int free_or_dead)
1909 {
1910     if (ts->fixed_reg) {
1911         return;
1912     }
1913     if (ts->val_type == TEMP_VAL_REG) {
1914         s->reg_to_temp[ts->reg] = NULL;
1915     }
1916     ts->val_type = (free_or_dead < 0
1917                     || ts->temp_local
1918                     || temp_idx(s, ts) < s->nb_globals
1919                     ? TEMP_VAL_MEM : TEMP_VAL_DEAD);
1920 }
1921 
1922 /* Mark a temporary as dead.  */
1923 static inline void temp_dead(TCGContext *s, TCGTemp *ts)
1924 {
1925     temp_free_or_dead(s, ts, 1);
1926 }
1927 
1928 /* Sync a temporary to memory. 'allocated_regs' is used in case a temporary
1929    registers needs to be allocated to store a constant.  If 'free_or_dead'
1930    is non-zero, subsequently release the temporary; if it is positive, the
1931    temp is dead; if it is negative, the temp is free.  */
1932 static void temp_sync(TCGContext *s, TCGTemp *ts,
1933                       TCGRegSet allocated_regs, int free_or_dead)
1934 {
1935     if (ts->fixed_reg) {
1936         return;
1937     }
1938     if (!ts->mem_coherent) {
1939         if (!ts->mem_allocated) {
1940             temp_allocate_frame(s, temp_idx(s, ts));
1941         }
1942         switch (ts->val_type) {
1943         case TEMP_VAL_CONST:
1944             /* If we're going to free the temp immediately, then we won't
1945                require it later in a register, so attempt to store the
1946                constant to memory directly.  */
1947             if (free_or_dead
1948                 && tcg_out_sti(s, ts->type, ts->val,
1949                                ts->mem_base->reg, ts->mem_offset)) {
1950                 break;
1951             }
1952             temp_load(s, ts, tcg_target_available_regs[ts->type],
1953                       allocated_regs);
1954             /* fallthrough */
1955 
1956         case TEMP_VAL_REG:
1957             tcg_out_st(s, ts->type, ts->reg,
1958                        ts->mem_base->reg, ts->mem_offset);
1959             break;
1960 
1961         case TEMP_VAL_MEM:
1962             break;
1963 
1964         case TEMP_VAL_DEAD:
1965         default:
1966             tcg_abort();
1967         }
1968         ts->mem_coherent = 1;
1969     }
1970     if (free_or_dead) {
1971         temp_free_or_dead(s, ts, free_or_dead);
1972     }
1973 }
1974 
1975 /* free register 'reg' by spilling the corresponding temporary if necessary */
1976 static void tcg_reg_free(TCGContext *s, TCGReg reg, TCGRegSet allocated_regs)
1977 {
1978     TCGTemp *ts = s->reg_to_temp[reg];
1979     if (ts != NULL) {
1980         temp_sync(s, ts, allocated_regs, -1);
1981     }
1982 }
1983 
1984 /* Allocate a register belonging to reg1 & ~reg2 */
1985 static TCGReg tcg_reg_alloc(TCGContext *s, TCGRegSet desired_regs,
1986                             TCGRegSet allocated_regs, bool rev)
1987 {
1988     int i, n = ARRAY_SIZE(tcg_target_reg_alloc_order);
1989     const int *order;
1990     TCGReg reg;
1991     TCGRegSet reg_ct;
1992 
1993     tcg_regset_andnot(reg_ct, desired_regs, allocated_regs);
1994     order = rev ? indirect_reg_alloc_order : tcg_target_reg_alloc_order;
1995 
1996     /* first try free registers */
1997     for(i = 0; i < n; i++) {
1998         reg = order[i];
1999         if (tcg_regset_test_reg(reg_ct, reg) && s->reg_to_temp[reg] == NULL)
2000             return reg;
2001     }
2002 
2003     /* XXX: do better spill choice */
2004     for(i = 0; i < n; i++) {
2005         reg = order[i];
2006         if (tcg_regset_test_reg(reg_ct, reg)) {
2007             tcg_reg_free(s, reg, allocated_regs);
2008             return reg;
2009         }
2010     }
2011 
2012     tcg_abort();
2013 }
2014 
2015 /* Make sure the temporary is in a register.  If needed, allocate the register
2016    from DESIRED while avoiding ALLOCATED.  */
2017 static void temp_load(TCGContext *s, TCGTemp *ts, TCGRegSet desired_regs,
2018                       TCGRegSet allocated_regs)
2019 {
2020     TCGReg reg;
2021 
2022     switch (ts->val_type) {
2023     case TEMP_VAL_REG:
2024         return;
2025     case TEMP_VAL_CONST:
2026         reg = tcg_reg_alloc(s, desired_regs, allocated_regs, ts->indirect_base);
2027         tcg_out_movi(s, ts->type, reg, ts->val);
2028         ts->mem_coherent = 0;
2029         break;
2030     case TEMP_VAL_MEM:
2031         reg = tcg_reg_alloc(s, desired_regs, allocated_regs, ts->indirect_base);
2032         tcg_out_ld(s, ts->type, reg, ts->mem_base->reg, ts->mem_offset);
2033         ts->mem_coherent = 1;
2034         break;
2035     case TEMP_VAL_DEAD:
2036     default:
2037         tcg_abort();
2038     }
2039     ts->reg = reg;
2040     ts->val_type = TEMP_VAL_REG;
2041     s->reg_to_temp[reg] = ts;
2042 }
2043 
2044 /* Save a temporary to memory. 'allocated_regs' is used in case a
2045    temporary registers needs to be allocated to store a constant.  */
2046 static void temp_save(TCGContext *s, TCGTemp *ts, TCGRegSet allocated_regs)
2047 {
2048     /* The liveness analysis already ensures that globals are back
2049        in memory. Keep an tcg_debug_assert for safety. */
2050     tcg_debug_assert(ts->val_type == TEMP_VAL_MEM || ts->fixed_reg);
2051 }
2052 
2053 /* save globals to their canonical location and assume they can be
2054    modified be the following code. 'allocated_regs' is used in case a
2055    temporary registers needs to be allocated to store a constant. */
2056 static void save_globals(TCGContext *s, TCGRegSet allocated_regs)
2057 {
2058     int i;
2059 
2060     for (i = 0; i < s->nb_globals; i++) {
2061         temp_save(s, &s->temps[i], allocated_regs);
2062     }
2063 }
2064 
2065 /* sync globals to their canonical location and assume they can be
2066    read by the following code. 'allocated_regs' is used in case a
2067    temporary registers needs to be allocated to store a constant. */
2068 static void sync_globals(TCGContext *s, TCGRegSet allocated_regs)
2069 {
2070     int i;
2071 
2072     for (i = 0; i < s->nb_globals; i++) {
2073         TCGTemp *ts = &s->temps[i];
2074         tcg_debug_assert(ts->val_type != TEMP_VAL_REG
2075                          || ts->fixed_reg
2076                          || ts->mem_coherent);
2077     }
2078 }
2079 
2080 /* at the end of a basic block, we assume all temporaries are dead and
2081    all globals are stored at their canonical location. */
2082 static void tcg_reg_alloc_bb_end(TCGContext *s, TCGRegSet allocated_regs)
2083 {
2084     int i;
2085 
2086     for (i = s->nb_globals; i < s->nb_temps; i++) {
2087         TCGTemp *ts = &s->temps[i];
2088         if (ts->temp_local) {
2089             temp_save(s, ts, allocated_regs);
2090         } else {
2091             /* The liveness analysis already ensures that temps are dead.
2092                Keep an tcg_debug_assert for safety. */
2093             tcg_debug_assert(ts->val_type == TEMP_VAL_DEAD);
2094         }
2095     }
2096 
2097     save_globals(s, allocated_regs);
2098 }
2099 
2100 static void tcg_reg_alloc_do_movi(TCGContext *s, TCGTemp *ots,
2101                                   tcg_target_ulong val, TCGLifeData arg_life)
2102 {
2103     if (ots->fixed_reg) {
2104         /* For fixed registers, we do not do any constant propagation.  */
2105         tcg_out_movi(s, ots->type, ots->reg, val);
2106         return;
2107     }
2108 
2109     /* The movi is not explicitly generated here.  */
2110     if (ots->val_type == TEMP_VAL_REG) {
2111         s->reg_to_temp[ots->reg] = NULL;
2112     }
2113     ots->val_type = TEMP_VAL_CONST;
2114     ots->val = val;
2115     ots->mem_coherent = 0;
2116     if (NEED_SYNC_ARG(0)) {
2117         temp_sync(s, ots, s->reserved_regs, IS_DEAD_ARG(0));
2118     } else if (IS_DEAD_ARG(0)) {
2119         temp_dead(s, ots);
2120     }
2121 }
2122 
2123 static void tcg_reg_alloc_movi(TCGContext *s, const TCGArg *args,
2124                                TCGLifeData arg_life)
2125 {
2126     TCGTemp *ots = &s->temps[args[0]];
2127     tcg_target_ulong val = args[1];
2128 
2129     tcg_reg_alloc_do_movi(s, ots, val, arg_life);
2130 }
2131 
2132 static void tcg_reg_alloc_mov(TCGContext *s, const TCGOpDef *def,
2133                               const TCGArg *args, TCGLifeData arg_life)
2134 {
2135     TCGRegSet allocated_regs;
2136     TCGTemp *ts, *ots;
2137     TCGType otype, itype;
2138 
2139     tcg_regset_set(allocated_regs, s->reserved_regs);
2140     ots = &s->temps[args[0]];
2141     ts = &s->temps[args[1]];
2142 
2143     /* Note that otype != itype for no-op truncation.  */
2144     otype = ots->type;
2145     itype = ts->type;
2146 
2147     if (ts->val_type == TEMP_VAL_CONST) {
2148         /* propagate constant or generate sti */
2149         tcg_target_ulong val = ts->val;
2150         if (IS_DEAD_ARG(1)) {
2151             temp_dead(s, ts);
2152         }
2153         tcg_reg_alloc_do_movi(s, ots, val, arg_life);
2154         return;
2155     }
2156 
2157     /* If the source value is in memory we're going to be forced
2158        to have it in a register in order to perform the copy.  Copy
2159        the SOURCE value into its own register first, that way we
2160        don't have to reload SOURCE the next time it is used. */
2161     if (ts->val_type == TEMP_VAL_MEM) {
2162         temp_load(s, ts, tcg_target_available_regs[itype], allocated_regs);
2163     }
2164 
2165     tcg_debug_assert(ts->val_type == TEMP_VAL_REG);
2166     if (IS_DEAD_ARG(0) && !ots->fixed_reg) {
2167         /* mov to a non-saved dead register makes no sense (even with
2168            liveness analysis disabled). */
2169         tcg_debug_assert(NEED_SYNC_ARG(0));
2170         if (!ots->mem_allocated) {
2171             temp_allocate_frame(s, args[0]);
2172         }
2173         tcg_out_st(s, otype, ts->reg, ots->mem_base->reg, ots->mem_offset);
2174         if (IS_DEAD_ARG(1)) {
2175             temp_dead(s, ts);
2176         }
2177         temp_dead(s, ots);
2178     } else {
2179         if (IS_DEAD_ARG(1) && !ts->fixed_reg && !ots->fixed_reg) {
2180             /* the mov can be suppressed */
2181             if (ots->val_type == TEMP_VAL_REG) {
2182                 s->reg_to_temp[ots->reg] = NULL;
2183             }
2184             ots->reg = ts->reg;
2185             temp_dead(s, ts);
2186         } else {
2187             if (ots->val_type != TEMP_VAL_REG) {
2188                 /* When allocating a new register, make sure to not spill the
2189                    input one. */
2190                 tcg_regset_set_reg(allocated_regs, ts->reg);
2191                 ots->reg = tcg_reg_alloc(s, tcg_target_available_regs[otype],
2192                                          allocated_regs, ots->indirect_base);
2193             }
2194             tcg_out_mov(s, otype, ots->reg, ts->reg);
2195         }
2196         ots->val_type = TEMP_VAL_REG;
2197         ots->mem_coherent = 0;
2198         s->reg_to_temp[ots->reg] = ots;
2199         if (NEED_SYNC_ARG(0)) {
2200             temp_sync(s, ots, allocated_regs, 0);
2201         }
2202     }
2203 }
2204 
2205 static void tcg_reg_alloc_op(TCGContext *s,
2206                              const TCGOpDef *def, TCGOpcode opc,
2207                              const TCGArg *args, TCGLifeData arg_life)
2208 {
2209     TCGRegSet allocated_regs;
2210     int i, k, nb_iargs, nb_oargs;
2211     TCGReg reg;
2212     TCGArg arg;
2213     const TCGArgConstraint *arg_ct;
2214     TCGTemp *ts;
2215     TCGArg new_args[TCG_MAX_OP_ARGS];
2216     int const_args[TCG_MAX_OP_ARGS];
2217 
2218     nb_oargs = def->nb_oargs;
2219     nb_iargs = def->nb_iargs;
2220 
2221     /* copy constants */
2222     memcpy(new_args + nb_oargs + nb_iargs,
2223            args + nb_oargs + nb_iargs,
2224            sizeof(TCGArg) * def->nb_cargs);
2225 
2226     /* satisfy input constraints */
2227     tcg_regset_set(allocated_regs, s->reserved_regs);
2228     for(k = 0; k < nb_iargs; k++) {
2229         i = def->sorted_args[nb_oargs + k];
2230         arg = args[i];
2231         arg_ct = &def->args_ct[i];
2232         ts = &s->temps[arg];
2233 
2234         if (ts->val_type == TEMP_VAL_CONST
2235             && tcg_target_const_match(ts->val, ts->type, arg_ct)) {
2236             /* constant is OK for instruction */
2237             const_args[i] = 1;
2238             new_args[i] = ts->val;
2239             goto iarg_end;
2240         }
2241 
2242         temp_load(s, ts, arg_ct->u.regs, allocated_regs);
2243 
2244         if (arg_ct->ct & TCG_CT_IALIAS) {
2245             if (ts->fixed_reg) {
2246                 /* if fixed register, we must allocate a new register
2247                    if the alias is not the same register */
2248                 if (arg != args[arg_ct->alias_index])
2249                     goto allocate_in_reg;
2250             } else {
2251                 /* if the input is aliased to an output and if it is
2252                    not dead after the instruction, we must allocate
2253                    a new register and move it */
2254                 if (!IS_DEAD_ARG(i)) {
2255                     goto allocate_in_reg;
2256                 }
2257                 /* check if the current register has already been allocated
2258                    for another input aliased to an output */
2259                 int k2, i2;
2260                 for (k2 = 0 ; k2 < k ; k2++) {
2261                     i2 = def->sorted_args[nb_oargs + k2];
2262                     if ((def->args_ct[i2].ct & TCG_CT_IALIAS) &&
2263                         (new_args[i2] == ts->reg)) {
2264                         goto allocate_in_reg;
2265                     }
2266                 }
2267             }
2268         }
2269         reg = ts->reg;
2270         if (tcg_regset_test_reg(arg_ct->u.regs, reg)) {
2271             /* nothing to do : the constraint is satisfied */
2272         } else {
2273         allocate_in_reg:
2274             /* allocate a new register matching the constraint
2275                and move the temporary register into it */
2276             reg = tcg_reg_alloc(s, arg_ct->u.regs, allocated_regs,
2277                                 ts->indirect_base);
2278             tcg_out_mov(s, ts->type, reg, ts->reg);
2279         }
2280         new_args[i] = reg;
2281         const_args[i] = 0;
2282         tcg_regset_set_reg(allocated_regs, reg);
2283     iarg_end: ;
2284     }
2285 
2286     /* mark dead temporaries and free the associated registers */
2287     for (i = nb_oargs; i < nb_oargs + nb_iargs; i++) {
2288         if (IS_DEAD_ARG(i)) {
2289             temp_dead(s, &s->temps[args[i]]);
2290         }
2291     }
2292 
2293     if (def->flags & TCG_OPF_BB_END) {
2294         tcg_reg_alloc_bb_end(s, allocated_regs);
2295     } else {
2296         if (def->flags & TCG_OPF_CALL_CLOBBER) {
2297             /* XXX: permit generic clobber register list ? */
2298             for (i = 0; i < TCG_TARGET_NB_REGS; i++) {
2299                 if (tcg_regset_test_reg(tcg_target_call_clobber_regs, i)) {
2300                     tcg_reg_free(s, i, allocated_regs);
2301                 }
2302             }
2303         }
2304         if (def->flags & TCG_OPF_SIDE_EFFECTS) {
2305             /* sync globals if the op has side effects and might trigger
2306                an exception. */
2307             sync_globals(s, allocated_regs);
2308         }
2309 
2310         /* satisfy the output constraints */
2311         tcg_regset_set(allocated_regs, s->reserved_regs);
2312         for(k = 0; k < nb_oargs; k++) {
2313             i = def->sorted_args[k];
2314             arg = args[i];
2315             arg_ct = &def->args_ct[i];
2316             ts = &s->temps[arg];
2317             if (arg_ct->ct & TCG_CT_ALIAS) {
2318                 reg = new_args[arg_ct->alias_index];
2319             } else {
2320                 /* if fixed register, we try to use it */
2321                 reg = ts->reg;
2322                 if (ts->fixed_reg &&
2323                     tcg_regset_test_reg(arg_ct->u.regs, reg)) {
2324                     goto oarg_end;
2325                 }
2326                 reg = tcg_reg_alloc(s, arg_ct->u.regs, allocated_regs,
2327                                     ts->indirect_base);
2328             }
2329             tcg_regset_set_reg(allocated_regs, reg);
2330             /* if a fixed register is used, then a move will be done afterwards */
2331             if (!ts->fixed_reg) {
2332                 if (ts->val_type == TEMP_VAL_REG) {
2333                     s->reg_to_temp[ts->reg] = NULL;
2334                 }
2335                 ts->val_type = TEMP_VAL_REG;
2336                 ts->reg = reg;
2337                 /* temp value is modified, so the value kept in memory is
2338                    potentially not the same */
2339                 ts->mem_coherent = 0;
2340                 s->reg_to_temp[reg] = ts;
2341             }
2342         oarg_end:
2343             new_args[i] = reg;
2344         }
2345     }
2346 
2347     /* emit instruction */
2348     tcg_out_op(s, opc, new_args, const_args);
2349 
2350     /* move the outputs in the correct register if needed */
2351     for(i = 0; i < nb_oargs; i++) {
2352         ts = &s->temps[args[i]];
2353         reg = new_args[i];
2354         if (ts->fixed_reg && ts->reg != reg) {
2355             tcg_out_mov(s, ts->type, ts->reg, reg);
2356         }
2357         if (NEED_SYNC_ARG(i)) {
2358             temp_sync(s, ts, allocated_regs, IS_DEAD_ARG(i));
2359         } else if (IS_DEAD_ARG(i)) {
2360             temp_dead(s, ts);
2361         }
2362     }
2363 }
2364 
2365 #ifdef TCG_TARGET_STACK_GROWSUP
2366 #define STACK_DIR(x) (-(x))
2367 #else
2368 #define STACK_DIR(x) (x)
2369 #endif
2370 
2371 static void tcg_reg_alloc_call(TCGContext *s, int nb_oargs, int nb_iargs,
2372                                const TCGArg * const args, TCGLifeData arg_life)
2373 {
2374     int flags, nb_regs, i;
2375     TCGReg reg;
2376     TCGArg arg;
2377     TCGTemp *ts;
2378     intptr_t stack_offset;
2379     size_t call_stack_size;
2380     tcg_insn_unit *func_addr;
2381     int allocate_args;
2382     TCGRegSet allocated_regs;
2383 
2384     func_addr = (tcg_insn_unit *)(intptr_t)args[nb_oargs + nb_iargs];
2385     flags = args[nb_oargs + nb_iargs + 1];
2386 
2387     nb_regs = ARRAY_SIZE(tcg_target_call_iarg_regs);
2388     if (nb_regs > nb_iargs) {
2389         nb_regs = nb_iargs;
2390     }
2391 
2392     /* assign stack slots first */
2393     call_stack_size = (nb_iargs - nb_regs) * sizeof(tcg_target_long);
2394     call_stack_size = (call_stack_size + TCG_TARGET_STACK_ALIGN - 1) &
2395         ~(TCG_TARGET_STACK_ALIGN - 1);
2396     allocate_args = (call_stack_size > TCG_STATIC_CALL_ARGS_SIZE);
2397     if (allocate_args) {
2398         /* XXX: if more than TCG_STATIC_CALL_ARGS_SIZE is needed,
2399            preallocate call stack */
2400         tcg_abort();
2401     }
2402 
2403     stack_offset = TCG_TARGET_CALL_STACK_OFFSET;
2404     for(i = nb_regs; i < nb_iargs; i++) {
2405         arg = args[nb_oargs + i];
2406 #ifdef TCG_TARGET_STACK_GROWSUP
2407         stack_offset -= sizeof(tcg_target_long);
2408 #endif
2409         if (arg != TCG_CALL_DUMMY_ARG) {
2410             ts = &s->temps[arg];
2411             temp_load(s, ts, tcg_target_available_regs[ts->type],
2412                       s->reserved_regs);
2413             tcg_out_st(s, ts->type, ts->reg, TCG_REG_CALL_STACK, stack_offset);
2414         }
2415 #ifndef TCG_TARGET_STACK_GROWSUP
2416         stack_offset += sizeof(tcg_target_long);
2417 #endif
2418     }
2419 
2420     /* assign input registers */
2421     tcg_regset_set(allocated_regs, s->reserved_regs);
2422     for(i = 0; i < nb_regs; i++) {
2423         arg = args[nb_oargs + i];
2424         if (arg != TCG_CALL_DUMMY_ARG) {
2425             ts = &s->temps[arg];
2426             reg = tcg_target_call_iarg_regs[i];
2427             tcg_reg_free(s, reg, allocated_regs);
2428 
2429             if (ts->val_type == TEMP_VAL_REG) {
2430                 if (ts->reg != reg) {
2431                     tcg_out_mov(s, ts->type, reg, ts->reg);
2432                 }
2433             } else {
2434                 TCGRegSet arg_set;
2435 
2436                 tcg_regset_clear(arg_set);
2437                 tcg_regset_set_reg(arg_set, reg);
2438                 temp_load(s, ts, arg_set, allocated_regs);
2439             }
2440 
2441             tcg_regset_set_reg(allocated_regs, reg);
2442         }
2443     }
2444 
2445     /* mark dead temporaries and free the associated registers */
2446     for(i = nb_oargs; i < nb_iargs + nb_oargs; i++) {
2447         if (IS_DEAD_ARG(i)) {
2448             temp_dead(s, &s->temps[args[i]]);
2449         }
2450     }
2451 
2452     /* clobber call registers */
2453     for (i = 0; i < TCG_TARGET_NB_REGS; i++) {
2454         if (tcg_regset_test_reg(tcg_target_call_clobber_regs, i)) {
2455             tcg_reg_free(s, i, allocated_regs);
2456         }
2457     }
2458 
2459     /* Save globals if they might be written by the helper, sync them if
2460        they might be read. */
2461     if (flags & TCG_CALL_NO_READ_GLOBALS) {
2462         /* Nothing to do */
2463     } else if (flags & TCG_CALL_NO_WRITE_GLOBALS) {
2464         sync_globals(s, allocated_regs);
2465     } else {
2466         save_globals(s, allocated_regs);
2467     }
2468 
2469     tcg_out_call(s, func_addr);
2470 
2471     /* assign output registers and emit moves if needed */
2472     for(i = 0; i < nb_oargs; i++) {
2473         arg = args[i];
2474         ts = &s->temps[arg];
2475         reg = tcg_target_call_oarg_regs[i];
2476         tcg_debug_assert(s->reg_to_temp[reg] == NULL);
2477 
2478         if (ts->fixed_reg) {
2479             if (ts->reg != reg) {
2480                 tcg_out_mov(s, ts->type, ts->reg, reg);
2481             }
2482         } else {
2483             if (ts->val_type == TEMP_VAL_REG) {
2484                 s->reg_to_temp[ts->reg] = NULL;
2485             }
2486             ts->val_type = TEMP_VAL_REG;
2487             ts->reg = reg;
2488             ts->mem_coherent = 0;
2489             s->reg_to_temp[reg] = ts;
2490             if (NEED_SYNC_ARG(i)) {
2491                 temp_sync(s, ts, allocated_regs, IS_DEAD_ARG(i));
2492             } else if (IS_DEAD_ARG(i)) {
2493                 temp_dead(s, ts);
2494             }
2495         }
2496     }
2497 }
2498 
2499 #ifdef CONFIG_PROFILER
2500 
2501 static int64_t tcg_table_op_count[NB_OPS];
2502 
2503 void tcg_dump_op_count(FILE *f, fprintf_function cpu_fprintf)
2504 {
2505     int i;
2506 
2507     for (i = 0; i < NB_OPS; i++) {
2508         cpu_fprintf(f, "%s %" PRId64 "\n", tcg_op_defs[i].name,
2509                     tcg_table_op_count[i]);
2510     }
2511 }
2512 #else
2513 void tcg_dump_op_count(FILE *f, fprintf_function cpu_fprintf)
2514 {
2515     cpu_fprintf(f, "[TCG profiler not compiled]\n");
2516 }
2517 #endif
2518 
2519 
2520 int tcg_gen_code(TCGContext *s, TranslationBlock *tb)
2521 {
2522     int i, oi, oi_next, num_insns;
2523 
2524 #ifdef CONFIG_PROFILER
2525     {
2526         int n;
2527 
2528         n = s->gen_op_buf[0].prev + 1;
2529         s->op_count += n;
2530         if (n > s->op_count_max) {
2531             s->op_count_max = n;
2532         }
2533 
2534         n = s->nb_temps;
2535         s->temp_count += n;
2536         if (n > s->temp_count_max) {
2537             s->temp_count_max = n;
2538         }
2539     }
2540 #endif
2541 
2542 #ifdef DEBUG_DISAS
2543     if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP)
2544                  && qemu_log_in_addr_range(tb->pc))) {
2545         qemu_log("OP:\n");
2546         tcg_dump_ops(s);
2547         qemu_log("\n");
2548     }
2549 #endif
2550 
2551 #ifdef CONFIG_PROFILER
2552     s->opt_time -= profile_getclock();
2553 #endif
2554 
2555 #ifdef USE_TCG_OPTIMIZATIONS
2556     tcg_optimize(s);
2557 #endif
2558 
2559 #ifdef CONFIG_PROFILER
2560     s->opt_time += profile_getclock();
2561     s->la_time -= profile_getclock();
2562 #endif
2563 
2564     {
2565         uint8_t *temp_state = tcg_malloc(s->nb_temps + s->nb_indirects);
2566 
2567         liveness_pass_1(s, temp_state);
2568 
2569         if (s->nb_indirects > 0) {
2570 #ifdef DEBUG_DISAS
2571             if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP_IND)
2572                          && qemu_log_in_addr_range(tb->pc))) {
2573                 qemu_log("OP before indirect lowering:\n");
2574                 tcg_dump_ops(s);
2575                 qemu_log("\n");
2576             }
2577 #endif
2578             /* Replace indirect temps with direct temps.  */
2579             if (liveness_pass_2(s, temp_state)) {
2580                 /* If changes were made, re-run liveness.  */
2581                 liveness_pass_1(s, temp_state);
2582             }
2583         }
2584     }
2585 
2586 #ifdef CONFIG_PROFILER
2587     s->la_time += profile_getclock();
2588 #endif
2589 
2590 #ifdef DEBUG_DISAS
2591     if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP_OPT)
2592                  && qemu_log_in_addr_range(tb->pc))) {
2593         qemu_log("OP after optimization and liveness analysis:\n");
2594         tcg_dump_ops(s);
2595         qemu_log("\n");
2596     }
2597 #endif
2598 
2599     tcg_reg_alloc_start(s);
2600 
2601     s->code_buf = tb->tc_ptr;
2602     s->code_ptr = tb->tc_ptr;
2603 
2604     tcg_out_tb_init(s);
2605 
2606     num_insns = -1;
2607     for (oi = s->gen_op_buf[0].next; oi != 0; oi = oi_next) {
2608         TCGOp * const op = &s->gen_op_buf[oi];
2609         TCGArg * const args = &s->gen_opparam_buf[op->args];
2610         TCGOpcode opc = op->opc;
2611         const TCGOpDef *def = &tcg_op_defs[opc];
2612         TCGLifeData arg_life = op->life;
2613 
2614         oi_next = op->next;
2615 #ifdef CONFIG_PROFILER
2616         tcg_table_op_count[opc]++;
2617 #endif
2618 
2619         switch (opc) {
2620         case INDEX_op_mov_i32:
2621         case INDEX_op_mov_i64:
2622             tcg_reg_alloc_mov(s, def, args, arg_life);
2623             break;
2624         case INDEX_op_movi_i32:
2625         case INDEX_op_movi_i64:
2626             tcg_reg_alloc_movi(s, args, arg_life);
2627             break;
2628         case INDEX_op_insn_start:
2629             if (num_insns >= 0) {
2630                 s->gen_insn_end_off[num_insns] = tcg_current_code_size(s);
2631             }
2632             num_insns++;
2633             for (i = 0; i < TARGET_INSN_START_WORDS; ++i) {
2634                 target_ulong a;
2635 #if TARGET_LONG_BITS > TCG_TARGET_REG_BITS
2636                 a = ((target_ulong)args[i * 2 + 1] << 32) | args[i * 2];
2637 #else
2638                 a = args[i];
2639 #endif
2640                 s->gen_insn_data[num_insns][i] = a;
2641             }
2642             break;
2643         case INDEX_op_discard:
2644             temp_dead(s, &s->temps[args[0]]);
2645             break;
2646         case INDEX_op_set_label:
2647             tcg_reg_alloc_bb_end(s, s->reserved_regs);
2648             tcg_out_label(s, arg_label(args[0]), s->code_ptr);
2649             break;
2650         case INDEX_op_call:
2651             tcg_reg_alloc_call(s, op->callo, op->calli, args, arg_life);
2652             break;
2653         default:
2654             /* Sanity check that we've not introduced any unhandled opcodes. */
2655             if (def->flags & TCG_OPF_NOT_PRESENT) {
2656                 tcg_abort();
2657             }
2658             /* Note: in order to speed up the code, it would be much
2659                faster to have specialized register allocator functions for
2660                some common argument patterns */
2661             tcg_reg_alloc_op(s, def, opc, args, arg_life);
2662             break;
2663         }
2664 #ifdef CONFIG_DEBUG_TCG
2665         check_regs(s);
2666 #endif
2667         /* Test for (pending) buffer overflow.  The assumption is that any
2668            one operation beginning below the high water mark cannot overrun
2669            the buffer completely.  Thus we can test for overflow after
2670            generating code without having to check during generation.  */
2671         if (unlikely((void *)s->code_ptr > s->code_gen_highwater)) {
2672             return -1;
2673         }
2674     }
2675     tcg_debug_assert(num_insns >= 0);
2676     s->gen_insn_end_off[num_insns] = tcg_current_code_size(s);
2677 
2678     /* Generate TB finalization at the end of block */
2679     if (!tcg_out_tb_finalize(s)) {
2680         return -1;
2681     }
2682 
2683     /* flush instruction cache */
2684     flush_icache_range((uintptr_t)s->code_buf, (uintptr_t)s->code_ptr);
2685 
2686     return tcg_current_code_size(s);
2687 }
2688 
2689 #ifdef CONFIG_PROFILER
2690 void tcg_dump_info(FILE *f, fprintf_function cpu_fprintf)
2691 {
2692     TCGContext *s = &tcg_ctx;
2693     int64_t tb_count = s->tb_count;
2694     int64_t tb_div_count = tb_count ? tb_count : 1;
2695     int64_t tot = s->interm_time + s->code_time;
2696 
2697     cpu_fprintf(f, "JIT cycles          %" PRId64 " (%0.3f s at 2.4 GHz)\n",
2698                 tot, tot / 2.4e9);
2699     cpu_fprintf(f, "translated TBs      %" PRId64 " (aborted=%" PRId64 " %0.1f%%)\n",
2700                 tb_count, s->tb_count1 - tb_count,
2701                 (double)(s->tb_count1 - s->tb_count)
2702                 / (s->tb_count1 ? s->tb_count1 : 1) * 100.0);
2703     cpu_fprintf(f, "avg ops/TB          %0.1f max=%d\n",
2704                 (double)s->op_count / tb_div_count, s->op_count_max);
2705     cpu_fprintf(f, "deleted ops/TB      %0.2f\n",
2706                 (double)s->del_op_count / tb_div_count);
2707     cpu_fprintf(f, "avg temps/TB        %0.2f max=%d\n",
2708                 (double)s->temp_count / tb_div_count, s->temp_count_max);
2709     cpu_fprintf(f, "avg host code/TB    %0.1f\n",
2710                 (double)s->code_out_len / tb_div_count);
2711     cpu_fprintf(f, "avg search data/TB  %0.1f\n",
2712                 (double)s->search_out_len / tb_div_count);
2713 
2714     cpu_fprintf(f, "cycles/op           %0.1f\n",
2715                 s->op_count ? (double)tot / s->op_count : 0);
2716     cpu_fprintf(f, "cycles/in byte      %0.1f\n",
2717                 s->code_in_len ? (double)tot / s->code_in_len : 0);
2718     cpu_fprintf(f, "cycles/out byte     %0.1f\n",
2719                 s->code_out_len ? (double)tot / s->code_out_len : 0);
2720     cpu_fprintf(f, "cycles/search byte     %0.1f\n",
2721                 s->search_out_len ? (double)tot / s->search_out_len : 0);
2722     if (tot == 0) {
2723         tot = 1;
2724     }
2725     cpu_fprintf(f, "  gen_interm time   %0.1f%%\n",
2726                 (double)s->interm_time / tot * 100.0);
2727     cpu_fprintf(f, "  gen_code time     %0.1f%%\n",
2728                 (double)s->code_time / tot * 100.0);
2729     cpu_fprintf(f, "optim./code time    %0.1f%%\n",
2730                 (double)s->opt_time / (s->code_time ? s->code_time : 1)
2731                 * 100.0);
2732     cpu_fprintf(f, "liveness/code time  %0.1f%%\n",
2733                 (double)s->la_time / (s->code_time ? s->code_time : 1) * 100.0);
2734     cpu_fprintf(f, "cpu_restore count   %" PRId64 "\n",
2735                 s->restore_count);
2736     cpu_fprintf(f, "  avg cycles        %0.1f\n",
2737                 s->restore_count ? (double)s->restore_time / s->restore_count : 0);
2738 }
2739 #else
2740 void tcg_dump_info(FILE *f, fprintf_function cpu_fprintf)
2741 {
2742     cpu_fprintf(f, "[TCG profiler not compiled]\n");
2743 }
2744 #endif
2745 
2746 #ifdef ELF_HOST_MACHINE
2747 /* In order to use this feature, the backend needs to do three things:
2748 
2749    (1) Define ELF_HOST_MACHINE to indicate both what value to
2750        put into the ELF image and to indicate support for the feature.
2751 
2752    (2) Define tcg_register_jit.  This should create a buffer containing
2753        the contents of a .debug_frame section that describes the post-
2754        prologue unwind info for the tcg machine.
2755 
2756    (3) Call tcg_register_jit_int, with the constructed .debug_frame.
2757 */
2758 
2759 /* Begin GDB interface.  THE FOLLOWING MUST MATCH GDB DOCS.  */
2760 typedef enum {
2761     JIT_NOACTION = 0,
2762     JIT_REGISTER_FN,
2763     JIT_UNREGISTER_FN
2764 } jit_actions_t;
2765 
2766 struct jit_code_entry {
2767     struct jit_code_entry *next_entry;
2768     struct jit_code_entry *prev_entry;
2769     const void *symfile_addr;
2770     uint64_t symfile_size;
2771 };
2772 
2773 struct jit_descriptor {
2774     uint32_t version;
2775     uint32_t action_flag;
2776     struct jit_code_entry *relevant_entry;
2777     struct jit_code_entry *first_entry;
2778 };
2779 
2780 void __jit_debug_register_code(void) __attribute__((noinline));
2781 void __jit_debug_register_code(void)
2782 {
2783     asm("");
2784 }
2785 
2786 /* Must statically initialize the version, because GDB may check
2787    the version before we can set it.  */
2788 struct jit_descriptor __jit_debug_descriptor = { 1, 0, 0, 0 };
2789 
2790 /* End GDB interface.  */
2791 
2792 static int find_string(const char *strtab, const char *str)
2793 {
2794     const char *p = strtab + 1;
2795 
2796     while (1) {
2797         if (strcmp(p, str) == 0) {
2798             return p - strtab;
2799         }
2800         p += strlen(p) + 1;
2801     }
2802 }
2803 
2804 static void tcg_register_jit_int(void *buf_ptr, size_t buf_size,
2805                                  const void *debug_frame,
2806                                  size_t debug_frame_size)
2807 {
2808     struct __attribute__((packed)) DebugInfo {
2809         uint32_t  len;
2810         uint16_t  version;
2811         uint32_t  abbrev;
2812         uint8_t   ptr_size;
2813         uint8_t   cu_die;
2814         uint16_t  cu_lang;
2815         uintptr_t cu_low_pc;
2816         uintptr_t cu_high_pc;
2817         uint8_t   fn_die;
2818         char      fn_name[16];
2819         uintptr_t fn_low_pc;
2820         uintptr_t fn_high_pc;
2821         uint8_t   cu_eoc;
2822     };
2823 
2824     struct ElfImage {
2825         ElfW(Ehdr) ehdr;
2826         ElfW(Phdr) phdr;
2827         ElfW(Shdr) shdr[7];
2828         ElfW(Sym)  sym[2];
2829         struct DebugInfo di;
2830         uint8_t    da[24];
2831         char       str[80];
2832     };
2833 
2834     struct ElfImage *img;
2835 
2836     static const struct ElfImage img_template = {
2837         .ehdr = {
2838             .e_ident[EI_MAG0] = ELFMAG0,
2839             .e_ident[EI_MAG1] = ELFMAG1,
2840             .e_ident[EI_MAG2] = ELFMAG2,
2841             .e_ident[EI_MAG3] = ELFMAG3,
2842             .e_ident[EI_CLASS] = ELF_CLASS,
2843             .e_ident[EI_DATA] = ELF_DATA,
2844             .e_ident[EI_VERSION] = EV_CURRENT,
2845             .e_type = ET_EXEC,
2846             .e_machine = ELF_HOST_MACHINE,
2847             .e_version = EV_CURRENT,
2848             .e_phoff = offsetof(struct ElfImage, phdr),
2849             .e_shoff = offsetof(struct ElfImage, shdr),
2850             .e_ehsize = sizeof(ElfW(Shdr)),
2851             .e_phentsize = sizeof(ElfW(Phdr)),
2852             .e_phnum = 1,
2853             .e_shentsize = sizeof(ElfW(Shdr)),
2854             .e_shnum = ARRAY_SIZE(img->shdr),
2855             .e_shstrndx = ARRAY_SIZE(img->shdr) - 1,
2856 #ifdef ELF_HOST_FLAGS
2857             .e_flags = ELF_HOST_FLAGS,
2858 #endif
2859 #ifdef ELF_OSABI
2860             .e_ident[EI_OSABI] = ELF_OSABI,
2861 #endif
2862         },
2863         .phdr = {
2864             .p_type = PT_LOAD,
2865             .p_flags = PF_X,
2866         },
2867         .shdr = {
2868             [0] = { .sh_type = SHT_NULL },
2869             /* Trick: The contents of code_gen_buffer are not present in
2870                this fake ELF file; that got allocated elsewhere.  Therefore
2871                we mark .text as SHT_NOBITS (similar to .bss) so that readers
2872                will not look for contents.  We can record any address.  */
2873             [1] = { /* .text */
2874                 .sh_type = SHT_NOBITS,
2875                 .sh_flags = SHF_EXECINSTR | SHF_ALLOC,
2876             },
2877             [2] = { /* .debug_info */
2878                 .sh_type = SHT_PROGBITS,
2879                 .sh_offset = offsetof(struct ElfImage, di),
2880                 .sh_size = sizeof(struct DebugInfo),
2881             },
2882             [3] = { /* .debug_abbrev */
2883                 .sh_type = SHT_PROGBITS,
2884                 .sh_offset = offsetof(struct ElfImage, da),
2885                 .sh_size = sizeof(img->da),
2886             },
2887             [4] = { /* .debug_frame */
2888                 .sh_type = SHT_PROGBITS,
2889                 .sh_offset = sizeof(struct ElfImage),
2890             },
2891             [5] = { /* .symtab */
2892                 .sh_type = SHT_SYMTAB,
2893                 .sh_offset = offsetof(struct ElfImage, sym),
2894                 .sh_size = sizeof(img->sym),
2895                 .sh_info = 1,
2896                 .sh_link = ARRAY_SIZE(img->shdr) - 1,
2897                 .sh_entsize = sizeof(ElfW(Sym)),
2898             },
2899             [6] = { /* .strtab */
2900                 .sh_type = SHT_STRTAB,
2901                 .sh_offset = offsetof(struct ElfImage, str),
2902                 .sh_size = sizeof(img->str),
2903             }
2904         },
2905         .sym = {
2906             [1] = { /* code_gen_buffer */
2907                 .st_info = ELF_ST_INFO(STB_GLOBAL, STT_FUNC),
2908                 .st_shndx = 1,
2909             }
2910         },
2911         .di = {
2912             .len = sizeof(struct DebugInfo) - 4,
2913             .version = 2,
2914             .ptr_size = sizeof(void *),
2915             .cu_die = 1,
2916             .cu_lang = 0x8001,  /* DW_LANG_Mips_Assembler */
2917             .fn_die = 2,
2918             .fn_name = "code_gen_buffer"
2919         },
2920         .da = {
2921             1,          /* abbrev number (the cu) */
2922             0x11, 1,    /* DW_TAG_compile_unit, has children */
2923             0x13, 0x5,  /* DW_AT_language, DW_FORM_data2 */
2924             0x11, 0x1,  /* DW_AT_low_pc, DW_FORM_addr */
2925             0x12, 0x1,  /* DW_AT_high_pc, DW_FORM_addr */
2926             0, 0,       /* end of abbrev */
2927             2,          /* abbrev number (the fn) */
2928             0x2e, 0,    /* DW_TAG_subprogram, no children */
2929             0x3, 0x8,   /* DW_AT_name, DW_FORM_string */
2930             0x11, 0x1,  /* DW_AT_low_pc, DW_FORM_addr */
2931             0x12, 0x1,  /* DW_AT_high_pc, DW_FORM_addr */
2932             0, 0,       /* end of abbrev */
2933             0           /* no more abbrev */
2934         },
2935         .str = "\0" ".text\0" ".debug_info\0" ".debug_abbrev\0"
2936                ".debug_frame\0" ".symtab\0" ".strtab\0" "code_gen_buffer",
2937     };
2938 
2939     /* We only need a single jit entry; statically allocate it.  */
2940     static struct jit_code_entry one_entry;
2941 
2942     uintptr_t buf = (uintptr_t)buf_ptr;
2943     size_t img_size = sizeof(struct ElfImage) + debug_frame_size;
2944     DebugFrameHeader *dfh;
2945 
2946     img = g_malloc(img_size);
2947     *img = img_template;
2948 
2949     img->phdr.p_vaddr = buf;
2950     img->phdr.p_paddr = buf;
2951     img->phdr.p_memsz = buf_size;
2952 
2953     img->shdr[1].sh_name = find_string(img->str, ".text");
2954     img->shdr[1].sh_addr = buf;
2955     img->shdr[1].sh_size = buf_size;
2956 
2957     img->shdr[2].sh_name = find_string(img->str, ".debug_info");
2958     img->shdr[3].sh_name = find_string(img->str, ".debug_abbrev");
2959 
2960     img->shdr[4].sh_name = find_string(img->str, ".debug_frame");
2961     img->shdr[4].sh_size = debug_frame_size;
2962 
2963     img->shdr[5].sh_name = find_string(img->str, ".symtab");
2964     img->shdr[6].sh_name = find_string(img->str, ".strtab");
2965 
2966     img->sym[1].st_name = find_string(img->str, "code_gen_buffer");
2967     img->sym[1].st_value = buf;
2968     img->sym[1].st_size = buf_size;
2969 
2970     img->di.cu_low_pc = buf;
2971     img->di.cu_high_pc = buf + buf_size;
2972     img->di.fn_low_pc = buf;
2973     img->di.fn_high_pc = buf + buf_size;
2974 
2975     dfh = (DebugFrameHeader *)(img + 1);
2976     memcpy(dfh, debug_frame, debug_frame_size);
2977     dfh->fde.func_start = buf;
2978     dfh->fde.func_len = buf_size;
2979 
2980 #ifdef DEBUG_JIT
2981     /* Enable this block to be able to debug the ELF image file creation.
2982        One can use readelf, objdump, or other inspection utilities.  */
2983     {
2984         FILE *f = fopen("/tmp/qemu.jit", "w+b");
2985         if (f) {
2986             if (fwrite(img, img_size, 1, f) != img_size) {
2987                 /* Avoid stupid unused return value warning for fwrite.  */
2988             }
2989             fclose(f);
2990         }
2991     }
2992 #endif
2993 
2994     one_entry.symfile_addr = img;
2995     one_entry.symfile_size = img_size;
2996 
2997     __jit_debug_descriptor.action_flag = JIT_REGISTER_FN;
2998     __jit_debug_descriptor.relevant_entry = &one_entry;
2999     __jit_debug_descriptor.first_entry = &one_entry;
3000     __jit_debug_register_code();
3001 }
3002 #else
3003 /* No support for the feature.  Provide the entry point expected by exec.c,
3004    and implement the internal function we declared earlier.  */
3005 
3006 static void tcg_register_jit_int(void *buf, size_t size,
3007                                  const void *debug_frame,
3008                                  size_t debug_frame_size)
3009 {
3010 }
3011 
3012 void tcg_register_jit(void *buf, size_t buf_size)
3013 {
3014 }
3015 #endif /* ELF_HOST_MACHINE */
3016