xref: /openbmc/qemu/tcg/tci.c (revision 6e2e2e8a)
1 /*
2  * Tiny Code Interpreter for QEMU
3  *
4  * Copyright (c) 2009, 2011, 2016 Stefan Weil
5  *
6  * This program is free software: you can redistribute it and/or modify
7  * it under the terms of the GNU General Public License as published by
8  * the Free Software Foundation, either version 2 of the License, or
9  * (at your option) any later version.
10  *
11  * This program is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
14  * GNU General Public License for more details.
15  *
16  * You should have received a copy of the GNU General Public License
17  * along with this program.  If not, see <http://www.gnu.org/licenses/>.
18  */
19 
20 #include "qemu/osdep.h"
21 
22 /* Enable TCI assertions only when debugging TCG (and without NDEBUG defined).
23  * Without assertions, the interpreter runs much faster. */
24 #if defined(CONFIG_DEBUG_TCG)
25 # define tci_assert(cond) assert(cond)
26 #else
27 # define tci_assert(cond) ((void)0)
28 #endif
29 
30 #include "qemu-common.h"
31 #include "tcg/tcg.h"           /* MAX_OPC_PARAM_IARGS */
32 #include "exec/cpu_ldst.h"
33 #include "tcg/tcg-op.h"
34 
35 /* Marker for missing code. */
36 #define TODO() \
37     do { \
38         fprintf(stderr, "TODO %s:%u: %s()\n", \
39                 __FILE__, __LINE__, __func__); \
40         tcg_abort(); \
41     } while (0)
42 
43 #if MAX_OPC_PARAM_IARGS != 6
44 # error Fix needed, number of supported input arguments changed!
45 #endif
46 #if TCG_TARGET_REG_BITS == 32
47 typedef uint64_t (*helper_function)(tcg_target_ulong, tcg_target_ulong,
48                                     tcg_target_ulong, tcg_target_ulong,
49                                     tcg_target_ulong, tcg_target_ulong,
50                                     tcg_target_ulong, tcg_target_ulong,
51                                     tcg_target_ulong, tcg_target_ulong,
52                                     tcg_target_ulong, tcg_target_ulong);
53 #else
54 typedef uint64_t (*helper_function)(tcg_target_ulong, tcg_target_ulong,
55                                     tcg_target_ulong, tcg_target_ulong,
56                                     tcg_target_ulong, tcg_target_ulong);
57 #endif
58 
59 static tcg_target_ulong tci_read_reg(const tcg_target_ulong *regs, TCGReg index)
60 {
61     tci_assert(index < TCG_TARGET_NB_REGS);
62     return regs[index];
63 }
64 
65 #if TCG_TARGET_HAS_ext8s_i32 || TCG_TARGET_HAS_ext8s_i64
66 static int8_t tci_read_reg8s(const tcg_target_ulong *regs, TCGReg index)
67 {
68     return (int8_t)tci_read_reg(regs, index);
69 }
70 #endif
71 
72 #if TCG_TARGET_HAS_ext16s_i32 || TCG_TARGET_HAS_ext16s_i64
73 static int16_t tci_read_reg16s(const tcg_target_ulong *regs, TCGReg index)
74 {
75     return (int16_t)tci_read_reg(regs, index);
76 }
77 #endif
78 
79 #if TCG_TARGET_REG_BITS == 64
80 static int32_t tci_read_reg32s(const tcg_target_ulong *regs, TCGReg index)
81 {
82     return (int32_t)tci_read_reg(regs, index);
83 }
84 #endif
85 
86 static uint8_t tci_read_reg8(const tcg_target_ulong *regs, TCGReg index)
87 {
88     return (uint8_t)tci_read_reg(regs, index);
89 }
90 
91 static uint16_t tci_read_reg16(const tcg_target_ulong *regs, TCGReg index)
92 {
93     return (uint16_t)tci_read_reg(regs, index);
94 }
95 
96 static uint32_t tci_read_reg32(const tcg_target_ulong *regs, TCGReg index)
97 {
98     return (uint32_t)tci_read_reg(regs, index);
99 }
100 
101 #if TCG_TARGET_REG_BITS == 64
102 static uint64_t tci_read_reg64(const tcg_target_ulong *regs, TCGReg index)
103 {
104     return tci_read_reg(regs, index);
105 }
106 #endif
107 
108 static void
109 tci_write_reg(tcg_target_ulong *regs, TCGReg index, tcg_target_ulong value)
110 {
111     tci_assert(index < TCG_TARGET_NB_REGS);
112     tci_assert(index != TCG_AREG0);
113     tci_assert(index != TCG_REG_CALL_STACK);
114     regs[index] = value;
115 }
116 
117 #if TCG_TARGET_REG_BITS == 64
118 static void
119 tci_write_reg32s(tcg_target_ulong *regs, TCGReg index, int32_t value)
120 {
121     tci_write_reg(regs, index, value);
122 }
123 #endif
124 
125 static void tci_write_reg8(tcg_target_ulong *regs, TCGReg index, uint8_t value)
126 {
127     tci_write_reg(regs, index, value);
128 }
129 
130 static void
131 tci_write_reg16(tcg_target_ulong *regs, TCGReg index, uint16_t value)
132 {
133     tci_write_reg(regs, index, value);
134 }
135 
136 static void
137 tci_write_reg32(tcg_target_ulong *regs, TCGReg index, uint32_t value)
138 {
139     tci_write_reg(regs, index, value);
140 }
141 
142 #if TCG_TARGET_REG_BITS == 32
143 static void tci_write_reg64(tcg_target_ulong *regs, uint32_t high_index,
144                             uint32_t low_index, uint64_t value)
145 {
146     tci_write_reg(regs, low_index, value);
147     tci_write_reg(regs, high_index, value >> 32);
148 }
149 #elif TCG_TARGET_REG_BITS == 64
150 static void
151 tci_write_reg64(tcg_target_ulong *regs, TCGReg index, uint64_t value)
152 {
153     tci_write_reg(regs, index, value);
154 }
155 #endif
156 
157 #if TCG_TARGET_REG_BITS == 32
158 /* Create a 64 bit value from two 32 bit values. */
159 static uint64_t tci_uint64(uint32_t high, uint32_t low)
160 {
161     return ((uint64_t)high << 32) + low;
162 }
163 #endif
164 
165 /* Read constant (native size) from bytecode. */
166 static tcg_target_ulong tci_read_i(uint8_t **tb_ptr)
167 {
168     tcg_target_ulong value = *(tcg_target_ulong *)(*tb_ptr);
169     *tb_ptr += sizeof(value);
170     return value;
171 }
172 
173 /* Read unsigned constant (32 bit) from bytecode. */
174 static uint32_t tci_read_i32(uint8_t **tb_ptr)
175 {
176     uint32_t value = *(uint32_t *)(*tb_ptr);
177     *tb_ptr += sizeof(value);
178     return value;
179 }
180 
181 /* Read signed constant (32 bit) from bytecode. */
182 static int32_t tci_read_s32(uint8_t **tb_ptr)
183 {
184     int32_t value = *(int32_t *)(*tb_ptr);
185     *tb_ptr += sizeof(value);
186     return value;
187 }
188 
189 #if TCG_TARGET_REG_BITS == 64
190 /* Read constant (64 bit) from bytecode. */
191 static uint64_t tci_read_i64(uint8_t **tb_ptr)
192 {
193     uint64_t value = *(uint64_t *)(*tb_ptr);
194     *tb_ptr += sizeof(value);
195     return value;
196 }
197 #endif
198 
199 /* Read indexed register (native size) from bytecode. */
200 static tcg_target_ulong
201 tci_read_r(const tcg_target_ulong *regs, uint8_t **tb_ptr)
202 {
203     tcg_target_ulong value = tci_read_reg(regs, **tb_ptr);
204     *tb_ptr += 1;
205     return value;
206 }
207 
208 /* Read indexed register (8 bit) from bytecode. */
209 static uint8_t tci_read_r8(const tcg_target_ulong *regs, uint8_t **tb_ptr)
210 {
211     uint8_t value = tci_read_reg8(regs, **tb_ptr);
212     *tb_ptr += 1;
213     return value;
214 }
215 
216 #if TCG_TARGET_HAS_ext8s_i32 || TCG_TARGET_HAS_ext8s_i64
217 /* Read indexed register (8 bit signed) from bytecode. */
218 static int8_t tci_read_r8s(const tcg_target_ulong *regs, uint8_t **tb_ptr)
219 {
220     int8_t value = tci_read_reg8s(regs, **tb_ptr);
221     *tb_ptr += 1;
222     return value;
223 }
224 #endif
225 
226 /* Read indexed register (16 bit) from bytecode. */
227 static uint16_t tci_read_r16(const tcg_target_ulong *regs, uint8_t **tb_ptr)
228 {
229     uint16_t value = tci_read_reg16(regs, **tb_ptr);
230     *tb_ptr += 1;
231     return value;
232 }
233 
234 #if TCG_TARGET_HAS_ext16s_i32 || TCG_TARGET_HAS_ext16s_i64
235 /* Read indexed register (16 bit signed) from bytecode. */
236 static int16_t tci_read_r16s(const tcg_target_ulong *regs, uint8_t **tb_ptr)
237 {
238     int16_t value = tci_read_reg16s(regs, **tb_ptr);
239     *tb_ptr += 1;
240     return value;
241 }
242 #endif
243 
244 /* Read indexed register (32 bit) from bytecode. */
245 static uint32_t tci_read_r32(const tcg_target_ulong *regs, uint8_t **tb_ptr)
246 {
247     uint32_t value = tci_read_reg32(regs, **tb_ptr);
248     *tb_ptr += 1;
249     return value;
250 }
251 
252 #if TCG_TARGET_REG_BITS == 32
253 /* Read two indexed registers (2 * 32 bit) from bytecode. */
254 static uint64_t tci_read_r64(const tcg_target_ulong *regs, uint8_t **tb_ptr)
255 {
256     uint32_t low = tci_read_r32(regs, tb_ptr);
257     return tci_uint64(tci_read_r32(regs, tb_ptr), low);
258 }
259 #elif TCG_TARGET_REG_BITS == 64
260 /* Read indexed register (32 bit signed) from bytecode. */
261 static int32_t tci_read_r32s(const tcg_target_ulong *regs, uint8_t **tb_ptr)
262 {
263     int32_t value = tci_read_reg32s(regs, **tb_ptr);
264     *tb_ptr += 1;
265     return value;
266 }
267 
268 /* Read indexed register (64 bit) from bytecode. */
269 static uint64_t tci_read_r64(const tcg_target_ulong *regs, uint8_t **tb_ptr)
270 {
271     uint64_t value = tci_read_reg64(regs, **tb_ptr);
272     *tb_ptr += 1;
273     return value;
274 }
275 #endif
276 
277 /* Read indexed register(s) with target address from bytecode. */
278 static target_ulong
279 tci_read_ulong(const tcg_target_ulong *regs, uint8_t **tb_ptr)
280 {
281     target_ulong taddr = tci_read_r(regs, tb_ptr);
282 #if TARGET_LONG_BITS > TCG_TARGET_REG_BITS
283     taddr += (uint64_t)tci_read_r(regs, tb_ptr) << 32;
284 #endif
285     return taddr;
286 }
287 
288 /* Read indexed register or constant (native size) from bytecode. */
289 static tcg_target_ulong
290 tci_read_ri(const tcg_target_ulong *regs, uint8_t **tb_ptr)
291 {
292     tcg_target_ulong value;
293     TCGReg r = **tb_ptr;
294     *tb_ptr += 1;
295     if (r == TCG_CONST) {
296         value = tci_read_i(tb_ptr);
297     } else {
298         value = tci_read_reg(regs, r);
299     }
300     return value;
301 }
302 
303 /* Read indexed register or constant (32 bit) from bytecode. */
304 static uint32_t tci_read_ri32(const tcg_target_ulong *regs, uint8_t **tb_ptr)
305 {
306     uint32_t value;
307     TCGReg r = **tb_ptr;
308     *tb_ptr += 1;
309     if (r == TCG_CONST) {
310         value = tci_read_i32(tb_ptr);
311     } else {
312         value = tci_read_reg32(regs, r);
313     }
314     return value;
315 }
316 
317 #if TCG_TARGET_REG_BITS == 32
318 /* Read two indexed registers or constants (2 * 32 bit) from bytecode. */
319 static uint64_t tci_read_ri64(const tcg_target_ulong *regs, uint8_t **tb_ptr)
320 {
321     uint32_t low = tci_read_ri32(regs, tb_ptr);
322     return tci_uint64(tci_read_ri32(regs, tb_ptr), low);
323 }
324 #elif TCG_TARGET_REG_BITS == 64
325 /* Read indexed register or constant (64 bit) from bytecode. */
326 static uint64_t tci_read_ri64(const tcg_target_ulong *regs, uint8_t **tb_ptr)
327 {
328     uint64_t value;
329     TCGReg r = **tb_ptr;
330     *tb_ptr += 1;
331     if (r == TCG_CONST) {
332         value = tci_read_i64(tb_ptr);
333     } else {
334         value = tci_read_reg64(regs, r);
335     }
336     return value;
337 }
338 #endif
339 
340 static tcg_target_ulong tci_read_label(uint8_t **tb_ptr)
341 {
342     tcg_target_ulong label = tci_read_i(tb_ptr);
343     tci_assert(label != 0);
344     return label;
345 }
346 
347 static bool tci_compare32(uint32_t u0, uint32_t u1, TCGCond condition)
348 {
349     bool result = false;
350     int32_t i0 = u0;
351     int32_t i1 = u1;
352     switch (condition) {
353     case TCG_COND_EQ:
354         result = (u0 == u1);
355         break;
356     case TCG_COND_NE:
357         result = (u0 != u1);
358         break;
359     case TCG_COND_LT:
360         result = (i0 < i1);
361         break;
362     case TCG_COND_GE:
363         result = (i0 >= i1);
364         break;
365     case TCG_COND_LE:
366         result = (i0 <= i1);
367         break;
368     case TCG_COND_GT:
369         result = (i0 > i1);
370         break;
371     case TCG_COND_LTU:
372         result = (u0 < u1);
373         break;
374     case TCG_COND_GEU:
375         result = (u0 >= u1);
376         break;
377     case TCG_COND_LEU:
378         result = (u0 <= u1);
379         break;
380     case TCG_COND_GTU:
381         result = (u0 > u1);
382         break;
383     default:
384         TODO();
385     }
386     return result;
387 }
388 
389 static bool tci_compare64(uint64_t u0, uint64_t u1, TCGCond condition)
390 {
391     bool result = false;
392     int64_t i0 = u0;
393     int64_t i1 = u1;
394     switch (condition) {
395     case TCG_COND_EQ:
396         result = (u0 == u1);
397         break;
398     case TCG_COND_NE:
399         result = (u0 != u1);
400         break;
401     case TCG_COND_LT:
402         result = (i0 < i1);
403         break;
404     case TCG_COND_GE:
405         result = (i0 >= i1);
406         break;
407     case TCG_COND_LE:
408         result = (i0 <= i1);
409         break;
410     case TCG_COND_GT:
411         result = (i0 > i1);
412         break;
413     case TCG_COND_LTU:
414         result = (u0 < u1);
415         break;
416     case TCG_COND_GEU:
417         result = (u0 >= u1);
418         break;
419     case TCG_COND_LEU:
420         result = (u0 <= u1);
421         break;
422     case TCG_COND_GTU:
423         result = (u0 > u1);
424         break;
425     default:
426         TODO();
427     }
428     return result;
429 }
430 
431 #ifdef CONFIG_SOFTMMU
432 # define qemu_ld_ub \
433     helper_ret_ldub_mmu(env, taddr, oi, (uintptr_t)tb_ptr)
434 # define qemu_ld_leuw \
435     helper_le_lduw_mmu(env, taddr, oi, (uintptr_t)tb_ptr)
436 # define qemu_ld_leul \
437     helper_le_ldul_mmu(env, taddr, oi, (uintptr_t)tb_ptr)
438 # define qemu_ld_leq \
439     helper_le_ldq_mmu(env, taddr, oi, (uintptr_t)tb_ptr)
440 # define qemu_ld_beuw \
441     helper_be_lduw_mmu(env, taddr, oi, (uintptr_t)tb_ptr)
442 # define qemu_ld_beul \
443     helper_be_ldul_mmu(env, taddr, oi, (uintptr_t)tb_ptr)
444 # define qemu_ld_beq \
445     helper_be_ldq_mmu(env, taddr, oi, (uintptr_t)tb_ptr)
446 # define qemu_st_b(X) \
447     helper_ret_stb_mmu(env, taddr, X, oi, (uintptr_t)tb_ptr)
448 # define qemu_st_lew(X) \
449     helper_le_stw_mmu(env, taddr, X, oi, (uintptr_t)tb_ptr)
450 # define qemu_st_lel(X) \
451     helper_le_stl_mmu(env, taddr, X, oi, (uintptr_t)tb_ptr)
452 # define qemu_st_leq(X) \
453     helper_le_stq_mmu(env, taddr, X, oi, (uintptr_t)tb_ptr)
454 # define qemu_st_bew(X) \
455     helper_be_stw_mmu(env, taddr, X, oi, (uintptr_t)tb_ptr)
456 # define qemu_st_bel(X) \
457     helper_be_stl_mmu(env, taddr, X, oi, (uintptr_t)tb_ptr)
458 # define qemu_st_beq(X) \
459     helper_be_stq_mmu(env, taddr, X, oi, (uintptr_t)tb_ptr)
460 #else
461 # define qemu_ld_ub      ldub_p(g2h(taddr))
462 # define qemu_ld_leuw    lduw_le_p(g2h(taddr))
463 # define qemu_ld_leul    (uint32_t)ldl_le_p(g2h(taddr))
464 # define qemu_ld_leq     ldq_le_p(g2h(taddr))
465 # define qemu_ld_beuw    lduw_be_p(g2h(taddr))
466 # define qemu_ld_beul    (uint32_t)ldl_be_p(g2h(taddr))
467 # define qemu_ld_beq     ldq_be_p(g2h(taddr))
468 # define qemu_st_b(X)    stb_p(g2h(taddr), X)
469 # define qemu_st_lew(X)  stw_le_p(g2h(taddr), X)
470 # define qemu_st_lel(X)  stl_le_p(g2h(taddr), X)
471 # define qemu_st_leq(X)  stq_le_p(g2h(taddr), X)
472 # define qemu_st_bew(X)  stw_be_p(g2h(taddr), X)
473 # define qemu_st_bel(X)  stl_be_p(g2h(taddr), X)
474 # define qemu_st_beq(X)  stq_be_p(g2h(taddr), X)
475 #endif
476 
477 /* Interpret pseudo code in tb. */
478 uintptr_t tcg_qemu_tb_exec(CPUArchState *env, uint8_t *tb_ptr)
479 {
480     tcg_target_ulong regs[TCG_TARGET_NB_REGS];
481     long tcg_temps[CPU_TEMP_BUF_NLONGS];
482     uintptr_t sp_value = (uintptr_t)(tcg_temps + CPU_TEMP_BUF_NLONGS);
483     uintptr_t ret = 0;
484 
485     regs[TCG_AREG0] = (tcg_target_ulong)env;
486     regs[TCG_REG_CALL_STACK] = sp_value;
487     tci_assert(tb_ptr);
488 
489     for (;;) {
490         TCGOpcode opc = tb_ptr[0];
491 #if defined(CONFIG_DEBUG_TCG) && !defined(NDEBUG)
492         uint8_t op_size = tb_ptr[1];
493         uint8_t *old_code_ptr = tb_ptr;
494 #endif
495         tcg_target_ulong t0;
496         tcg_target_ulong t1;
497         tcg_target_ulong t2;
498         tcg_target_ulong label;
499         TCGCond condition;
500         target_ulong taddr;
501         uint8_t tmp8;
502         uint16_t tmp16;
503         uint32_t tmp32;
504         uint64_t tmp64;
505 #if TCG_TARGET_REG_BITS == 32
506         uint64_t v64;
507 #endif
508         TCGMemOpIdx oi;
509 
510 #if defined(GETPC)
511         tci_tb_ptr = (uintptr_t)tb_ptr;
512 #endif
513 
514         /* Skip opcode and size entry. */
515         tb_ptr += 2;
516 
517         switch (opc) {
518         case INDEX_op_call:
519             t0 = tci_read_ri(regs, &tb_ptr);
520 #if TCG_TARGET_REG_BITS == 32
521             tmp64 = ((helper_function)t0)(tci_read_reg(regs, TCG_REG_R0),
522                                           tci_read_reg(regs, TCG_REG_R1),
523                                           tci_read_reg(regs, TCG_REG_R2),
524                                           tci_read_reg(regs, TCG_REG_R3),
525                                           tci_read_reg(regs, TCG_REG_R5),
526                                           tci_read_reg(regs, TCG_REG_R6),
527                                           tci_read_reg(regs, TCG_REG_R7),
528                                           tci_read_reg(regs, TCG_REG_R8),
529                                           tci_read_reg(regs, TCG_REG_R9),
530                                           tci_read_reg(regs, TCG_REG_R10),
531                                           tci_read_reg(regs, TCG_REG_R11),
532                                           tci_read_reg(regs, TCG_REG_R12));
533             tci_write_reg(regs, TCG_REG_R0, tmp64);
534             tci_write_reg(regs, TCG_REG_R1, tmp64 >> 32);
535 #else
536             tmp64 = ((helper_function)t0)(tci_read_reg(regs, TCG_REG_R0),
537                                           tci_read_reg(regs, TCG_REG_R1),
538                                           tci_read_reg(regs, TCG_REG_R2),
539                                           tci_read_reg(regs, TCG_REG_R3),
540                                           tci_read_reg(regs, TCG_REG_R5),
541                                           tci_read_reg(regs, TCG_REG_R6));
542             tci_write_reg(regs, TCG_REG_R0, tmp64);
543 #endif
544             break;
545         case INDEX_op_br:
546             label = tci_read_label(&tb_ptr);
547             tci_assert(tb_ptr == old_code_ptr + op_size);
548             tb_ptr = (uint8_t *)label;
549             continue;
550         case INDEX_op_setcond_i32:
551             t0 = *tb_ptr++;
552             t1 = tci_read_r32(regs, &tb_ptr);
553             t2 = tci_read_ri32(regs, &tb_ptr);
554             condition = *tb_ptr++;
555             tci_write_reg32(regs, t0, tci_compare32(t1, t2, condition));
556             break;
557 #if TCG_TARGET_REG_BITS == 32
558         case INDEX_op_setcond2_i32:
559             t0 = *tb_ptr++;
560             tmp64 = tci_read_r64(regs, &tb_ptr);
561             v64 = tci_read_ri64(regs, &tb_ptr);
562             condition = *tb_ptr++;
563             tci_write_reg32(regs, t0, tci_compare64(tmp64, v64, condition));
564             break;
565 #elif TCG_TARGET_REG_BITS == 64
566         case INDEX_op_setcond_i64:
567             t0 = *tb_ptr++;
568             t1 = tci_read_r64(regs, &tb_ptr);
569             t2 = tci_read_ri64(regs, &tb_ptr);
570             condition = *tb_ptr++;
571             tci_write_reg64(regs, t0, tci_compare64(t1, t2, condition));
572             break;
573 #endif
574         case INDEX_op_mov_i32:
575             t0 = *tb_ptr++;
576             t1 = tci_read_r32(regs, &tb_ptr);
577             tci_write_reg32(regs, t0, t1);
578             break;
579         case INDEX_op_movi_i32:
580             t0 = *tb_ptr++;
581             t1 = tci_read_i32(&tb_ptr);
582             tci_write_reg32(regs, t0, t1);
583             break;
584 
585             /* Load/store operations (32 bit). */
586 
587         case INDEX_op_ld8u_i32:
588             t0 = *tb_ptr++;
589             t1 = tci_read_r(regs, &tb_ptr);
590             t2 = tci_read_s32(&tb_ptr);
591             tci_write_reg8(regs, t0, *(uint8_t *)(t1 + t2));
592             break;
593         case INDEX_op_ld8s_i32:
594             TODO();
595             break;
596         case INDEX_op_ld16u_i32:
597             TODO();
598             break;
599         case INDEX_op_ld16s_i32:
600             TODO();
601             break;
602         case INDEX_op_ld_i32:
603             t0 = *tb_ptr++;
604             t1 = tci_read_r(regs, &tb_ptr);
605             t2 = tci_read_s32(&tb_ptr);
606             tci_write_reg32(regs, t0, *(uint32_t *)(t1 + t2));
607             break;
608         case INDEX_op_st8_i32:
609             t0 = tci_read_r8(regs, &tb_ptr);
610             t1 = tci_read_r(regs, &tb_ptr);
611             t2 = tci_read_s32(&tb_ptr);
612             *(uint8_t *)(t1 + t2) = t0;
613             break;
614         case INDEX_op_st16_i32:
615             t0 = tci_read_r16(regs, &tb_ptr);
616             t1 = tci_read_r(regs, &tb_ptr);
617             t2 = tci_read_s32(&tb_ptr);
618             *(uint16_t *)(t1 + t2) = t0;
619             break;
620         case INDEX_op_st_i32:
621             t0 = tci_read_r32(regs, &tb_ptr);
622             t1 = tci_read_r(regs, &tb_ptr);
623             t2 = tci_read_s32(&tb_ptr);
624             tci_assert(t1 != sp_value || (int32_t)t2 < 0);
625             *(uint32_t *)(t1 + t2) = t0;
626             break;
627 
628             /* Arithmetic operations (32 bit). */
629 
630         case INDEX_op_add_i32:
631             t0 = *tb_ptr++;
632             t1 = tci_read_ri32(regs, &tb_ptr);
633             t2 = tci_read_ri32(regs, &tb_ptr);
634             tci_write_reg32(regs, t0, t1 + t2);
635             break;
636         case INDEX_op_sub_i32:
637             t0 = *tb_ptr++;
638             t1 = tci_read_ri32(regs, &tb_ptr);
639             t2 = tci_read_ri32(regs, &tb_ptr);
640             tci_write_reg32(regs, t0, t1 - t2);
641             break;
642         case INDEX_op_mul_i32:
643             t0 = *tb_ptr++;
644             t1 = tci_read_ri32(regs, &tb_ptr);
645             t2 = tci_read_ri32(regs, &tb_ptr);
646             tci_write_reg32(regs, t0, t1 * t2);
647             break;
648 #if TCG_TARGET_HAS_div_i32
649         case INDEX_op_div_i32:
650             t0 = *tb_ptr++;
651             t1 = tci_read_ri32(regs, &tb_ptr);
652             t2 = tci_read_ri32(regs, &tb_ptr);
653             tci_write_reg32(regs, t0, (int32_t)t1 / (int32_t)t2);
654             break;
655         case INDEX_op_divu_i32:
656             t0 = *tb_ptr++;
657             t1 = tci_read_ri32(regs, &tb_ptr);
658             t2 = tci_read_ri32(regs, &tb_ptr);
659             tci_write_reg32(regs, t0, t1 / t2);
660             break;
661         case INDEX_op_rem_i32:
662             t0 = *tb_ptr++;
663             t1 = tci_read_ri32(regs, &tb_ptr);
664             t2 = tci_read_ri32(regs, &tb_ptr);
665             tci_write_reg32(regs, t0, (int32_t)t1 % (int32_t)t2);
666             break;
667         case INDEX_op_remu_i32:
668             t0 = *tb_ptr++;
669             t1 = tci_read_ri32(regs, &tb_ptr);
670             t2 = tci_read_ri32(regs, &tb_ptr);
671             tci_write_reg32(regs, t0, t1 % t2);
672             break;
673 #elif TCG_TARGET_HAS_div2_i32
674         case INDEX_op_div2_i32:
675         case INDEX_op_divu2_i32:
676             TODO();
677             break;
678 #endif
679         case INDEX_op_and_i32:
680             t0 = *tb_ptr++;
681             t1 = tci_read_ri32(regs, &tb_ptr);
682             t2 = tci_read_ri32(regs, &tb_ptr);
683             tci_write_reg32(regs, t0, t1 & t2);
684             break;
685         case INDEX_op_or_i32:
686             t0 = *tb_ptr++;
687             t1 = tci_read_ri32(regs, &tb_ptr);
688             t2 = tci_read_ri32(regs, &tb_ptr);
689             tci_write_reg32(regs, t0, t1 | t2);
690             break;
691         case INDEX_op_xor_i32:
692             t0 = *tb_ptr++;
693             t1 = tci_read_ri32(regs, &tb_ptr);
694             t2 = tci_read_ri32(regs, &tb_ptr);
695             tci_write_reg32(regs, t0, t1 ^ t2);
696             break;
697 
698             /* Shift/rotate operations (32 bit). */
699 
700         case INDEX_op_shl_i32:
701             t0 = *tb_ptr++;
702             t1 = tci_read_ri32(regs, &tb_ptr);
703             t2 = tci_read_ri32(regs, &tb_ptr);
704             tci_write_reg32(regs, t0, t1 << (t2 & 31));
705             break;
706         case INDEX_op_shr_i32:
707             t0 = *tb_ptr++;
708             t1 = tci_read_ri32(regs, &tb_ptr);
709             t2 = tci_read_ri32(regs, &tb_ptr);
710             tci_write_reg32(regs, t0, t1 >> (t2 & 31));
711             break;
712         case INDEX_op_sar_i32:
713             t0 = *tb_ptr++;
714             t1 = tci_read_ri32(regs, &tb_ptr);
715             t2 = tci_read_ri32(regs, &tb_ptr);
716             tci_write_reg32(regs, t0, ((int32_t)t1 >> (t2 & 31)));
717             break;
718 #if TCG_TARGET_HAS_rot_i32
719         case INDEX_op_rotl_i32:
720             t0 = *tb_ptr++;
721             t1 = tci_read_ri32(regs, &tb_ptr);
722             t2 = tci_read_ri32(regs, &tb_ptr);
723             tci_write_reg32(regs, t0, rol32(t1, t2 & 31));
724             break;
725         case INDEX_op_rotr_i32:
726             t0 = *tb_ptr++;
727             t1 = tci_read_ri32(regs, &tb_ptr);
728             t2 = tci_read_ri32(regs, &tb_ptr);
729             tci_write_reg32(regs, t0, ror32(t1, t2 & 31));
730             break;
731 #endif
732 #if TCG_TARGET_HAS_deposit_i32
733         case INDEX_op_deposit_i32:
734             t0 = *tb_ptr++;
735             t1 = tci_read_r32(regs, &tb_ptr);
736             t2 = tci_read_r32(regs, &tb_ptr);
737             tmp16 = *tb_ptr++;
738             tmp8 = *tb_ptr++;
739             tmp32 = (((1 << tmp8) - 1) << tmp16);
740             tci_write_reg32(regs, t0, (t1 & ~tmp32) | ((t2 << tmp16) & tmp32));
741             break;
742 #endif
743         case INDEX_op_brcond_i32:
744             t0 = tci_read_r32(regs, &tb_ptr);
745             t1 = tci_read_ri32(regs, &tb_ptr);
746             condition = *tb_ptr++;
747             label = tci_read_label(&tb_ptr);
748             if (tci_compare32(t0, t1, condition)) {
749                 tci_assert(tb_ptr == old_code_ptr + op_size);
750                 tb_ptr = (uint8_t *)label;
751                 continue;
752             }
753             break;
754 #if TCG_TARGET_REG_BITS == 32
755         case INDEX_op_add2_i32:
756             t0 = *tb_ptr++;
757             t1 = *tb_ptr++;
758             tmp64 = tci_read_r64(regs, &tb_ptr);
759             tmp64 += tci_read_r64(regs, &tb_ptr);
760             tci_write_reg64(regs, t1, t0, tmp64);
761             break;
762         case INDEX_op_sub2_i32:
763             t0 = *tb_ptr++;
764             t1 = *tb_ptr++;
765             tmp64 = tci_read_r64(regs, &tb_ptr);
766             tmp64 -= tci_read_r64(regs, &tb_ptr);
767             tci_write_reg64(regs, t1, t0, tmp64);
768             break;
769         case INDEX_op_brcond2_i32:
770             tmp64 = tci_read_r64(regs, &tb_ptr);
771             v64 = tci_read_ri64(regs, &tb_ptr);
772             condition = *tb_ptr++;
773             label = tci_read_label(&tb_ptr);
774             if (tci_compare64(tmp64, v64, condition)) {
775                 tci_assert(tb_ptr == old_code_ptr + op_size);
776                 tb_ptr = (uint8_t *)label;
777                 continue;
778             }
779             break;
780         case INDEX_op_mulu2_i32:
781             t0 = *tb_ptr++;
782             t1 = *tb_ptr++;
783             t2 = tci_read_r32(regs, &tb_ptr);
784             tmp64 = tci_read_r32(regs, &tb_ptr);
785             tci_write_reg64(regs, t1, t0, t2 * tmp64);
786             break;
787 #endif /* TCG_TARGET_REG_BITS == 32 */
788 #if TCG_TARGET_HAS_ext8s_i32
789         case INDEX_op_ext8s_i32:
790             t0 = *tb_ptr++;
791             t1 = tci_read_r8s(regs, &tb_ptr);
792             tci_write_reg32(regs, t0, t1);
793             break;
794 #endif
795 #if TCG_TARGET_HAS_ext16s_i32
796         case INDEX_op_ext16s_i32:
797             t0 = *tb_ptr++;
798             t1 = tci_read_r16s(regs, &tb_ptr);
799             tci_write_reg32(regs, t0, t1);
800             break;
801 #endif
802 #if TCG_TARGET_HAS_ext8u_i32
803         case INDEX_op_ext8u_i32:
804             t0 = *tb_ptr++;
805             t1 = tci_read_r8(regs, &tb_ptr);
806             tci_write_reg32(regs, t0, t1);
807             break;
808 #endif
809 #if TCG_TARGET_HAS_ext16u_i32
810         case INDEX_op_ext16u_i32:
811             t0 = *tb_ptr++;
812             t1 = tci_read_r16(regs, &tb_ptr);
813             tci_write_reg32(regs, t0, t1);
814             break;
815 #endif
816 #if TCG_TARGET_HAS_bswap16_i32
817         case INDEX_op_bswap16_i32:
818             t0 = *tb_ptr++;
819             t1 = tci_read_r16(regs, &tb_ptr);
820             tci_write_reg32(regs, t0, bswap16(t1));
821             break;
822 #endif
823 #if TCG_TARGET_HAS_bswap32_i32
824         case INDEX_op_bswap32_i32:
825             t0 = *tb_ptr++;
826             t1 = tci_read_r32(regs, &tb_ptr);
827             tci_write_reg32(regs, t0, bswap32(t1));
828             break;
829 #endif
830 #if TCG_TARGET_HAS_not_i32
831         case INDEX_op_not_i32:
832             t0 = *tb_ptr++;
833             t1 = tci_read_r32(regs, &tb_ptr);
834             tci_write_reg32(regs, t0, ~t1);
835             break;
836 #endif
837 #if TCG_TARGET_HAS_neg_i32
838         case INDEX_op_neg_i32:
839             t0 = *tb_ptr++;
840             t1 = tci_read_r32(regs, &tb_ptr);
841             tci_write_reg32(regs, t0, -t1);
842             break;
843 #endif
844 #if TCG_TARGET_REG_BITS == 64
845         case INDEX_op_mov_i64:
846             t0 = *tb_ptr++;
847             t1 = tci_read_r64(regs, &tb_ptr);
848             tci_write_reg64(regs, t0, t1);
849             break;
850         case INDEX_op_movi_i64:
851             t0 = *tb_ptr++;
852             t1 = tci_read_i64(&tb_ptr);
853             tci_write_reg64(regs, t0, t1);
854             break;
855 
856             /* Load/store operations (64 bit). */
857 
858         case INDEX_op_ld8u_i64:
859             t0 = *tb_ptr++;
860             t1 = tci_read_r(regs, &tb_ptr);
861             t2 = tci_read_s32(&tb_ptr);
862             tci_write_reg8(regs, t0, *(uint8_t *)(t1 + t2));
863             break;
864         case INDEX_op_ld8s_i64:
865             TODO();
866             break;
867         case INDEX_op_ld16u_i64:
868             t0 = *tb_ptr++;
869             t1 = tci_read_r(regs, &tb_ptr);
870             t2 = tci_read_s32(&tb_ptr);
871             tci_write_reg16(regs, t0, *(uint16_t *)(t1 + t2));
872             break;
873         case INDEX_op_ld16s_i64:
874             TODO();
875             break;
876         case INDEX_op_ld32u_i64:
877             t0 = *tb_ptr++;
878             t1 = tci_read_r(regs, &tb_ptr);
879             t2 = tci_read_s32(&tb_ptr);
880             tci_write_reg32(regs, t0, *(uint32_t *)(t1 + t2));
881             break;
882         case INDEX_op_ld32s_i64:
883             t0 = *tb_ptr++;
884             t1 = tci_read_r(regs, &tb_ptr);
885             t2 = tci_read_s32(&tb_ptr);
886             tci_write_reg32s(regs, t0, *(int32_t *)(t1 + t2));
887             break;
888         case INDEX_op_ld_i64:
889             t0 = *tb_ptr++;
890             t1 = tci_read_r(regs, &tb_ptr);
891             t2 = tci_read_s32(&tb_ptr);
892             tci_write_reg64(regs, t0, *(uint64_t *)(t1 + t2));
893             break;
894         case INDEX_op_st8_i64:
895             t0 = tci_read_r8(regs, &tb_ptr);
896             t1 = tci_read_r(regs, &tb_ptr);
897             t2 = tci_read_s32(&tb_ptr);
898             *(uint8_t *)(t1 + t2) = t0;
899             break;
900         case INDEX_op_st16_i64:
901             t0 = tci_read_r16(regs, &tb_ptr);
902             t1 = tci_read_r(regs, &tb_ptr);
903             t2 = tci_read_s32(&tb_ptr);
904             *(uint16_t *)(t1 + t2) = t0;
905             break;
906         case INDEX_op_st32_i64:
907             t0 = tci_read_r32(regs, &tb_ptr);
908             t1 = tci_read_r(regs, &tb_ptr);
909             t2 = tci_read_s32(&tb_ptr);
910             *(uint32_t *)(t1 + t2) = t0;
911             break;
912         case INDEX_op_st_i64:
913             t0 = tci_read_r64(regs, &tb_ptr);
914             t1 = tci_read_r(regs, &tb_ptr);
915             t2 = tci_read_s32(&tb_ptr);
916             tci_assert(t1 != sp_value || (int32_t)t2 < 0);
917             *(uint64_t *)(t1 + t2) = t0;
918             break;
919 
920             /* Arithmetic operations (64 bit). */
921 
922         case INDEX_op_add_i64:
923             t0 = *tb_ptr++;
924             t1 = tci_read_ri64(regs, &tb_ptr);
925             t2 = tci_read_ri64(regs, &tb_ptr);
926             tci_write_reg64(regs, t0, t1 + t2);
927             break;
928         case INDEX_op_sub_i64:
929             t0 = *tb_ptr++;
930             t1 = tci_read_ri64(regs, &tb_ptr);
931             t2 = tci_read_ri64(regs, &tb_ptr);
932             tci_write_reg64(regs, t0, t1 - t2);
933             break;
934         case INDEX_op_mul_i64:
935             t0 = *tb_ptr++;
936             t1 = tci_read_ri64(regs, &tb_ptr);
937             t2 = tci_read_ri64(regs, &tb_ptr);
938             tci_write_reg64(regs, t0, t1 * t2);
939             break;
940 #if TCG_TARGET_HAS_div_i64
941         case INDEX_op_div_i64:
942         case INDEX_op_divu_i64:
943         case INDEX_op_rem_i64:
944         case INDEX_op_remu_i64:
945             TODO();
946             break;
947 #elif TCG_TARGET_HAS_div2_i64
948         case INDEX_op_div2_i64:
949         case INDEX_op_divu2_i64:
950             TODO();
951             break;
952 #endif
953         case INDEX_op_and_i64:
954             t0 = *tb_ptr++;
955             t1 = tci_read_ri64(regs, &tb_ptr);
956             t2 = tci_read_ri64(regs, &tb_ptr);
957             tci_write_reg64(regs, t0, t1 & t2);
958             break;
959         case INDEX_op_or_i64:
960             t0 = *tb_ptr++;
961             t1 = tci_read_ri64(regs, &tb_ptr);
962             t2 = tci_read_ri64(regs, &tb_ptr);
963             tci_write_reg64(regs, t0, t1 | t2);
964             break;
965         case INDEX_op_xor_i64:
966             t0 = *tb_ptr++;
967             t1 = tci_read_ri64(regs, &tb_ptr);
968             t2 = tci_read_ri64(regs, &tb_ptr);
969             tci_write_reg64(regs, t0, t1 ^ t2);
970             break;
971 
972             /* Shift/rotate operations (64 bit). */
973 
974         case INDEX_op_shl_i64:
975             t0 = *tb_ptr++;
976             t1 = tci_read_ri64(regs, &tb_ptr);
977             t2 = tci_read_ri64(regs, &tb_ptr);
978             tci_write_reg64(regs, t0, t1 << (t2 & 63));
979             break;
980         case INDEX_op_shr_i64:
981             t0 = *tb_ptr++;
982             t1 = tci_read_ri64(regs, &tb_ptr);
983             t2 = tci_read_ri64(regs, &tb_ptr);
984             tci_write_reg64(regs, t0, t1 >> (t2 & 63));
985             break;
986         case INDEX_op_sar_i64:
987             t0 = *tb_ptr++;
988             t1 = tci_read_ri64(regs, &tb_ptr);
989             t2 = tci_read_ri64(regs, &tb_ptr);
990             tci_write_reg64(regs, t0, ((int64_t)t1 >> (t2 & 63)));
991             break;
992 #if TCG_TARGET_HAS_rot_i64
993         case INDEX_op_rotl_i64:
994             t0 = *tb_ptr++;
995             t1 = tci_read_ri64(regs, &tb_ptr);
996             t2 = tci_read_ri64(regs, &tb_ptr);
997             tci_write_reg64(regs, t0, rol64(t1, t2 & 63));
998             break;
999         case INDEX_op_rotr_i64:
1000             t0 = *tb_ptr++;
1001             t1 = tci_read_ri64(regs, &tb_ptr);
1002             t2 = tci_read_ri64(regs, &tb_ptr);
1003             tci_write_reg64(regs, t0, ror64(t1, t2 & 63));
1004             break;
1005 #endif
1006 #if TCG_TARGET_HAS_deposit_i64
1007         case INDEX_op_deposit_i64:
1008             t0 = *tb_ptr++;
1009             t1 = tci_read_r64(regs, &tb_ptr);
1010             t2 = tci_read_r64(regs, &tb_ptr);
1011             tmp16 = *tb_ptr++;
1012             tmp8 = *tb_ptr++;
1013             tmp64 = (((1ULL << tmp8) - 1) << tmp16);
1014             tci_write_reg64(regs, t0, (t1 & ~tmp64) | ((t2 << tmp16) & tmp64));
1015             break;
1016 #endif
1017         case INDEX_op_brcond_i64:
1018             t0 = tci_read_r64(regs, &tb_ptr);
1019             t1 = tci_read_ri64(regs, &tb_ptr);
1020             condition = *tb_ptr++;
1021             label = tci_read_label(&tb_ptr);
1022             if (tci_compare64(t0, t1, condition)) {
1023                 tci_assert(tb_ptr == old_code_ptr + op_size);
1024                 tb_ptr = (uint8_t *)label;
1025                 continue;
1026             }
1027             break;
1028 #if TCG_TARGET_HAS_ext8u_i64
1029         case INDEX_op_ext8u_i64:
1030             t0 = *tb_ptr++;
1031             t1 = tci_read_r8(regs, &tb_ptr);
1032             tci_write_reg64(regs, t0, t1);
1033             break;
1034 #endif
1035 #if TCG_TARGET_HAS_ext8s_i64
1036         case INDEX_op_ext8s_i64:
1037             t0 = *tb_ptr++;
1038             t1 = tci_read_r8s(regs, &tb_ptr);
1039             tci_write_reg64(regs, t0, t1);
1040             break;
1041 #endif
1042 #if TCG_TARGET_HAS_ext16s_i64
1043         case INDEX_op_ext16s_i64:
1044             t0 = *tb_ptr++;
1045             t1 = tci_read_r16s(regs, &tb_ptr);
1046             tci_write_reg64(regs, t0, t1);
1047             break;
1048 #endif
1049 #if TCG_TARGET_HAS_ext16u_i64
1050         case INDEX_op_ext16u_i64:
1051             t0 = *tb_ptr++;
1052             t1 = tci_read_r16(regs, &tb_ptr);
1053             tci_write_reg64(regs, t0, t1);
1054             break;
1055 #endif
1056 #if TCG_TARGET_HAS_ext32s_i64
1057         case INDEX_op_ext32s_i64:
1058 #endif
1059         case INDEX_op_ext_i32_i64:
1060             t0 = *tb_ptr++;
1061             t1 = tci_read_r32s(regs, &tb_ptr);
1062             tci_write_reg64(regs, t0, t1);
1063             break;
1064 #if TCG_TARGET_HAS_ext32u_i64
1065         case INDEX_op_ext32u_i64:
1066 #endif
1067         case INDEX_op_extu_i32_i64:
1068             t0 = *tb_ptr++;
1069             t1 = tci_read_r32(regs, &tb_ptr);
1070             tci_write_reg64(regs, t0, t1);
1071             break;
1072 #if TCG_TARGET_HAS_bswap16_i64
1073         case INDEX_op_bswap16_i64:
1074             t0 = *tb_ptr++;
1075             t1 = tci_read_r16(regs, &tb_ptr);
1076             tci_write_reg64(regs, t0, bswap16(t1));
1077             break;
1078 #endif
1079 #if TCG_TARGET_HAS_bswap32_i64
1080         case INDEX_op_bswap32_i64:
1081             t0 = *tb_ptr++;
1082             t1 = tci_read_r32(regs, &tb_ptr);
1083             tci_write_reg64(regs, t0, bswap32(t1));
1084             break;
1085 #endif
1086 #if TCG_TARGET_HAS_bswap64_i64
1087         case INDEX_op_bswap64_i64:
1088             t0 = *tb_ptr++;
1089             t1 = tci_read_r64(regs, &tb_ptr);
1090             tci_write_reg64(regs, t0, bswap64(t1));
1091             break;
1092 #endif
1093 #if TCG_TARGET_HAS_not_i64
1094         case INDEX_op_not_i64:
1095             t0 = *tb_ptr++;
1096             t1 = tci_read_r64(regs, &tb_ptr);
1097             tci_write_reg64(regs, t0, ~t1);
1098             break;
1099 #endif
1100 #if TCG_TARGET_HAS_neg_i64
1101         case INDEX_op_neg_i64:
1102             t0 = *tb_ptr++;
1103             t1 = tci_read_r64(regs, &tb_ptr);
1104             tci_write_reg64(regs, t0, -t1);
1105             break;
1106 #endif
1107 #endif /* TCG_TARGET_REG_BITS == 64 */
1108 
1109             /* QEMU specific operations. */
1110 
1111         case INDEX_op_exit_tb:
1112             ret = *(uint64_t *)tb_ptr;
1113             goto exit;
1114             break;
1115         case INDEX_op_goto_tb:
1116             /* Jump address is aligned */
1117             tb_ptr = QEMU_ALIGN_PTR_UP(tb_ptr, 4);
1118             t0 = qatomic_read((int32_t *)tb_ptr);
1119             tb_ptr += sizeof(int32_t);
1120             tci_assert(tb_ptr == old_code_ptr + op_size);
1121             tb_ptr += (int32_t)t0;
1122             continue;
1123         case INDEX_op_qemu_ld_i32:
1124             t0 = *tb_ptr++;
1125             taddr = tci_read_ulong(regs, &tb_ptr);
1126             oi = tci_read_i(&tb_ptr);
1127             switch (get_memop(oi) & (MO_BSWAP | MO_SSIZE)) {
1128             case MO_UB:
1129                 tmp32 = qemu_ld_ub;
1130                 break;
1131             case MO_SB:
1132                 tmp32 = (int8_t)qemu_ld_ub;
1133                 break;
1134             case MO_LEUW:
1135                 tmp32 = qemu_ld_leuw;
1136                 break;
1137             case MO_LESW:
1138                 tmp32 = (int16_t)qemu_ld_leuw;
1139                 break;
1140             case MO_LEUL:
1141                 tmp32 = qemu_ld_leul;
1142                 break;
1143             case MO_BEUW:
1144                 tmp32 = qemu_ld_beuw;
1145                 break;
1146             case MO_BESW:
1147                 tmp32 = (int16_t)qemu_ld_beuw;
1148                 break;
1149             case MO_BEUL:
1150                 tmp32 = qemu_ld_beul;
1151                 break;
1152             default:
1153                 tcg_abort();
1154             }
1155             tci_write_reg(regs, t0, tmp32);
1156             break;
1157         case INDEX_op_qemu_ld_i64:
1158             t0 = *tb_ptr++;
1159             if (TCG_TARGET_REG_BITS == 32) {
1160                 t1 = *tb_ptr++;
1161             }
1162             taddr = tci_read_ulong(regs, &tb_ptr);
1163             oi = tci_read_i(&tb_ptr);
1164             switch (get_memop(oi) & (MO_BSWAP | MO_SSIZE)) {
1165             case MO_UB:
1166                 tmp64 = qemu_ld_ub;
1167                 break;
1168             case MO_SB:
1169                 tmp64 = (int8_t)qemu_ld_ub;
1170                 break;
1171             case MO_LEUW:
1172                 tmp64 = qemu_ld_leuw;
1173                 break;
1174             case MO_LESW:
1175                 tmp64 = (int16_t)qemu_ld_leuw;
1176                 break;
1177             case MO_LEUL:
1178                 tmp64 = qemu_ld_leul;
1179                 break;
1180             case MO_LESL:
1181                 tmp64 = (int32_t)qemu_ld_leul;
1182                 break;
1183             case MO_LEQ:
1184                 tmp64 = qemu_ld_leq;
1185                 break;
1186             case MO_BEUW:
1187                 tmp64 = qemu_ld_beuw;
1188                 break;
1189             case MO_BESW:
1190                 tmp64 = (int16_t)qemu_ld_beuw;
1191                 break;
1192             case MO_BEUL:
1193                 tmp64 = qemu_ld_beul;
1194                 break;
1195             case MO_BESL:
1196                 tmp64 = (int32_t)qemu_ld_beul;
1197                 break;
1198             case MO_BEQ:
1199                 tmp64 = qemu_ld_beq;
1200                 break;
1201             default:
1202                 tcg_abort();
1203             }
1204             tci_write_reg(regs, t0, tmp64);
1205             if (TCG_TARGET_REG_BITS == 32) {
1206                 tci_write_reg(regs, t1, tmp64 >> 32);
1207             }
1208             break;
1209         case INDEX_op_qemu_st_i32:
1210             t0 = tci_read_r(regs, &tb_ptr);
1211             taddr = tci_read_ulong(regs, &tb_ptr);
1212             oi = tci_read_i(&tb_ptr);
1213             switch (get_memop(oi) & (MO_BSWAP | MO_SIZE)) {
1214             case MO_UB:
1215                 qemu_st_b(t0);
1216                 break;
1217             case MO_LEUW:
1218                 qemu_st_lew(t0);
1219                 break;
1220             case MO_LEUL:
1221                 qemu_st_lel(t0);
1222                 break;
1223             case MO_BEUW:
1224                 qemu_st_bew(t0);
1225                 break;
1226             case MO_BEUL:
1227                 qemu_st_bel(t0);
1228                 break;
1229             default:
1230                 tcg_abort();
1231             }
1232             break;
1233         case INDEX_op_qemu_st_i64:
1234             tmp64 = tci_read_r64(regs, &tb_ptr);
1235             taddr = tci_read_ulong(regs, &tb_ptr);
1236             oi = tci_read_i(&tb_ptr);
1237             switch (get_memop(oi) & (MO_BSWAP | MO_SIZE)) {
1238             case MO_UB:
1239                 qemu_st_b(tmp64);
1240                 break;
1241             case MO_LEUW:
1242                 qemu_st_lew(tmp64);
1243                 break;
1244             case MO_LEUL:
1245                 qemu_st_lel(tmp64);
1246                 break;
1247             case MO_LEQ:
1248                 qemu_st_leq(tmp64);
1249                 break;
1250             case MO_BEUW:
1251                 qemu_st_bew(tmp64);
1252                 break;
1253             case MO_BEUL:
1254                 qemu_st_bel(tmp64);
1255                 break;
1256             case MO_BEQ:
1257                 qemu_st_beq(tmp64);
1258                 break;
1259             default:
1260                 tcg_abort();
1261             }
1262             break;
1263         case INDEX_op_mb:
1264             /* Ensure ordering for all kinds */
1265             smp_mb();
1266             break;
1267         default:
1268             TODO();
1269             break;
1270         }
1271         tci_assert(tb_ptr == old_code_ptr + op_size);
1272     }
1273 exit:
1274     return ret;
1275 }
1276