Lines Matching +full:0 +full:x48

34 	do { prog = emit_code(prog, bytes, len); } while (0)
42 do { EMIT1(b1); EMIT(off, 4); } while (0)
44 do { EMIT2(b1, b2); EMIT(off, 4); } while (0)
46 do { EMIT3(b1, b2, b3); EMIT(off, 4); } while (0)
48 do { EMIT4(b1, b2, b3, b4); EMIT(off, 4); } while (0)
68 * 211: 74 7d je 0x290
69 * 213: 48 8b 77 00 mov rsi,QWORD PTR [rdi+0x0]
72 * 28c: 74 17 je 0x2a5
73 * 28e: e9 7f ff ff ff jmp 0x212
74 * 293: bf 03 00 00 00 mov edi,0x3
75 * Note that insn at 0x211 is 2-byte cond jump insn for offset 0x7d (-125)
76 * and insn at 0x28e is 5-byte jmp insn with offset -129.
81 * 211: 0f 84 80 00 00 00 je 0x297
82 * 217: 48 8b 77 00 mov rsi,QWORD PTR [rdi+0x0]
85 * 290: 74 1a je 0x2ac
86 * 292: eb 84 jmp 0x218
87 * 294: bf 03 00 00 00 mov edi,0x3
88 * Note that insn at 0x211 is 6-byte cond jump insn now since its offset
89 * becomes 0x80 based on previous round (0x293 - 0x213 = 0x80).
90 * At the same time, insn at 0x292 is a 2-byte insn since its offset is
97 * cycle in the above. In the above example je offset <= 0x7c should work.
99 * For other cases, je <-> je needs offset <= 0x7b to avoid no convergence
100 * issue. For jmp <-> je and jmp <-> jmp cases, jmp offset <= 0x7c should
104 * to maximum 123 (0x7b). This way, the jit pass can eventually converge.
125 EMIT3(add_2mod(0x48, DST, SRC), 0x89, add_2reg(0xC0, DST, SRC)); \
126 } while (0)
139 return 0; in bpf_size_to_x86_bytes()
144 * Add 0x10 (and an extra 0x0f) to generate far jumps (. + s32)
146 #define X86_JB 0x72
147 #define X86_JAE 0x73
148 #define X86_JE 0x74
149 #define X86_JNE 0x75
150 #define X86_JBE 0x76
151 #define X86_JA 0x77
152 #define X86_JL 0x7C
153 #define X86_JGE 0x7D
154 #define X86_JLE 0x7E
155 #define X86_JG 0x7F
172 [BPF_REG_0] = 0, /* RAX */
177 [BPF_REG_5] = 0, /* R8 */
266 [BPF_ADD] = 0x01,
267 [BPF_SUB] = 0x29,
268 [BPF_AND] = 0x21,
269 [BPF_OR] = 0x09,
270 [BPF_XOR] = 0x31,
271 [BPF_LSH] = 0xE0,
272 [BPF_RSH] = 0xE8,
273 [BPF_ARSH] = 0xF8,
279 memset(area, 0xcc, size); in jit_fill_hole()
284 return IS_ERR_OR_NULL(text_poke_set(dst, 0xcc, len)); in bpf_arch_text_invalidate()
312 if (callee_regs_used[0]) in push_callee_regs()
313 EMIT1(0x53); /* push rbx */ in push_callee_regs()
315 EMIT2(0x41, 0x55); /* push r13 */ in push_callee_regs()
317 EMIT2(0x41, 0x56); /* push r14 */ in push_callee_regs()
319 EMIT2(0x41, 0x57); /* push r15 */ in push_callee_regs()
328 EMIT2(0x41, 0x5F); /* pop r15 */ in pop_callee_regs()
330 EMIT2(0x41, 0x5E); /* pop r14 */ in pop_callee_regs()
332 EMIT2(0x41, 0x5D); /* pop r13 */ in pop_callee_regs()
333 if (callee_regs_used[0]) in pop_callee_regs()
334 EMIT1(0x5B); /* pop rbx */ in pop_callee_regs()
356 EMIT2(0x31, 0xC0); /* xor eax, eax */ in emit_prologue()
358 EMIT2(0x66, 0x90); /* nop2 */ in emit_prologue()
360 EMIT1(0x55); /* push rbp */ in emit_prologue()
361 EMIT3(0x48, 0x89, 0xE5); /* mov rbp, rsp */ in emit_prologue()
368 EMIT3_off32(0x48, 0x81, 0xEC, round_up(stack_depth, 8)); in emit_prologue()
370 EMIT1(0x50); /* push rax */ in emit_prologue()
386 return 0; in emit_patch()
391 return emit_patch(pprog, func, ip, 0xE8); in emit_call()
398 return emit_patch(pprog, func, ip, 0xE8); in emit_rsb_call()
403 return emit_patch(pprog, func, ip, 0xE9); in emit_jump()
442 ret = 0; in __bpf_arch_text_poke()
467 #define EMIT_LFENCE() EMIT3(0x0F, 0xAE, 0xE8)
475 EMIT2(0xFF, 0xE0 + reg); in emit_indirect_jump()
483 EMIT2(0xFF, 0xE0 + reg); /* jmp *%\reg */ in emit_indirect_jump()
485 EMIT1(0xCC); /* int3 */ in emit_indirect_jump()
498 EMIT1(0xC3); /* ret */ in emit_return()
500 EMIT1(0xCC); /* int3 */ in emit_return()
538 EMIT2(0x89, 0xD2); /* mov edx, edx */ in emit_bpf_tail_call_indirect()
539 EMIT3(0x39, 0x56, /* cmp dword ptr [rsi + 16], edx */ in emit_bpf_tail_call_indirect()
549 EMIT2_off32(0x8B, 0x85, tcc_off); /* mov eax, dword ptr [rbp - tcc_off] */ in emit_bpf_tail_call_indirect()
550 EMIT3(0x83, 0xF8, MAX_TAIL_CALL_CNT); /* cmp eax, MAX_TAIL_CALL_CNT */ in emit_bpf_tail_call_indirect()
554 EMIT3(0x83, 0xC0, 0x01); /* add eax, 1 */ in emit_bpf_tail_call_indirect()
555 EMIT2_off32(0x89, 0x85, tcc_off); /* mov dword ptr [rbp - tcc_off], eax */ in emit_bpf_tail_call_indirect()
558 EMIT4_off32(0x48, 0x8B, 0x8C, 0xD6, /* mov rcx, [rsi + rdx * 8 + offsetof(...)] */ in emit_bpf_tail_call_indirect()
565 EMIT3(0x48, 0x85, 0xC9); /* test rcx,rcx */ in emit_bpf_tail_call_indirect()
572 EMIT1(0x58); /* pop rax */ in emit_bpf_tail_call_indirect()
574 EMIT3_off32(0x48, 0x81, 0xC4, /* add rsp, sd */ in emit_bpf_tail_call_indirect()
578 EMIT4(0x48, 0x8B, 0x49, /* mov rcx, qword ptr [rcx + 32] */ in emit_bpf_tail_call_indirect()
580 EMIT4(0x48, 0x83, 0xC1, /* add rcx, X86_TAIL_CALL_OFFSET */ in emit_bpf_tail_call_indirect()
607 EMIT2_off32(0x8B, 0x85, tcc_off); /* mov eax, dword ptr [rbp - tcc_off] */ in emit_bpf_tail_call_direct()
608 EMIT3(0x83, 0xF8, MAX_TAIL_CALL_CNT); /* cmp eax, MAX_TAIL_CALL_CNT */ in emit_bpf_tail_call_direct()
612 EMIT3(0x83, 0xC0, 0x01); /* add eax, 1 */ in emit_bpf_tail_call_direct()
613 EMIT2_off32(0x89, 0x85, tcc_off); /* mov dword ptr [rbp - tcc_off], eax */ in emit_bpf_tail_call_direct()
624 EMIT1(0x58); /* pop rax */ in emit_bpf_tail_call_direct()
626 EMIT3_off32(0x48, 0x81, 0xC4, round_up(stack_depth, 8)); in emit_bpf_tail_call_direct()
644 for (i = 0; i < prog->aux->size_poke_tab; i++) { in bpf_tail_call_direct_fixup()
662 BUG_ON(ret < 0); in bpf_tail_call_direct_fixup()
667 BUG_ON(ret < 0); in bpf_tail_call_direct_fixup()
684 if (sign_propagate && (s32)imm32 < 0) { in emit_mov_imm32()
686 b1 = add_1mod(0x48, dst_reg); in emit_mov_imm32()
687 b2 = 0xC7; in emit_mov_imm32()
688 b3 = 0xC0; in emit_mov_imm32()
697 if (imm32 == 0) { in emit_mov_imm32()
699 EMIT1(add_2mod(0x40, dst_reg, dst_reg)); in emit_mov_imm32()
700 b2 = 0x31; /* xor */ in emit_mov_imm32()
701 b3 = 0xC0; in emit_mov_imm32()
708 EMIT1(add_1mod(0x40, dst_reg)); in emit_mov_imm32()
709 EMIT1_off32(add_1reg(0xB8, dst_reg), imm32); in emit_mov_imm32()
729 EMIT2(add_1mod(0x48, dst_reg), add_1reg(0xB8, dst_reg)); in emit_mov_imm64()
747 EMIT1(add_2mod(0x40, dst_reg, src_reg)); in emit_mov_reg()
748 EMIT2(0x89, add_2reg(0xC0, dst_reg, src_reg)); in emit_mov_reg()
762 EMIT4(add_2mod(0x48, src_reg, dst_reg), 0x0f, 0xbe, in emit_movsx_reg()
763 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
765 EMIT4(add_2mod(0x48, src_reg, dst_reg), 0x0f, 0xbf, in emit_movsx_reg()
766 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
768 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x63, in emit_movsx_reg()
769 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
773 EMIT4(add_2mod(0x40, src_reg, dst_reg), 0x0f, 0xbe, in emit_movsx_reg()
774 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
777 EMIT1(add_2mod(0x40, src_reg, dst_reg)); in emit_movsx_reg()
778 EMIT3(add_2mod(0x0f, src_reg, dst_reg), 0xbf, in emit_movsx_reg()
779 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
794 * If off == 0 we could skip this and save one extra byte, but in emit_insn_suffix()
798 EMIT2(add_2reg(0x40, ptr_reg, val_reg), off); in emit_insn_suffix()
801 EMIT1_off32(add_2reg(0x80, ptr_reg, val_reg), off); in emit_insn_suffix()
814 EMIT1(add_2mod(0x48, dst_reg, src_reg)); in maybe_emit_mod()
816 EMIT1(add_2mod(0x40, dst_reg, src_reg)); in maybe_emit_mod()
828 EMIT1(add_1mod(0x48, reg)); in maybe_emit_1mod()
830 EMIT1(add_1mod(0x40, reg)); in maybe_emit_1mod()
842 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xB6); in emit_ldx()
846 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xB7); in emit_ldx()
849 /* Emit 'mov eax, dword ptr [rax+0x14]' */ in emit_ldx()
851 EMIT2(add_2mod(0x40, src_reg, dst_reg), 0x8B); in emit_ldx()
853 EMIT1(0x8B); in emit_ldx()
856 /* Emit 'mov rax, qword ptr [rax+0x14]' */ in emit_ldx()
857 EMIT2(add_2mod(0x48, src_reg, dst_reg), 0x8B); in emit_ldx()
872 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xBE); in emit_ldsx()
876 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xBF); in emit_ldsx()
879 /* Emit 'movsx rax, dword ptr [rax+0x14]' */ in emit_ldsx()
880 EMIT2(add_2mod(0x48, src_reg, dst_reg), 0x63); in emit_ldsx()
897 EMIT2(add_2mod(0x40, dst_reg, src_reg), 0x88); in emit_stx()
899 EMIT1(0x88); in emit_stx()
903 EMIT3(0x66, add_2mod(0x40, dst_reg, src_reg), 0x89); in emit_stx()
905 EMIT2(0x66, 0x89); in emit_stx()
909 EMIT2(add_2mod(0x40, dst_reg, src_reg), 0x89); in emit_stx()
911 EMIT1(0x89); in emit_stx()
914 EMIT2(add_2mod(0x48, dst_reg, src_reg), 0x89); in emit_stx()
926 EMIT1(0xF0); /* lock prefix */ in emit_atomic()
941 EMIT2(0x0F, 0xC1); in emit_atomic()
945 EMIT1(0x87); in emit_atomic()
949 EMIT2(0x0F, 0xB1); in emit_atomic()
959 return 0; in emit_atomic()
967 *(unsigned long *)((void *)regs + reg) = 0; in ex_handler_bpf()
968 regs->ip += x->fixup & 0xff; in ex_handler_bpf()
981 regs_used[0] = true; in detect_reg_usage()
996 while (len > 0) { in emit_nops()
1002 for (i = 0; i < noplen; i++) in emit_nops()
1015 * m: opcode map select, encoding escape bytes e.g. 0x0f38
1019 * pp: opcode prefix (none, 0x66, 0xf2 or 0xf3)
1025 const u8 b0 = 0xc4; /* first byte of 3-byte VEX prefix */ in emit_3vex()
1037 * 7 0 in emit_3vex()
1042 b1 = (!r << 7) | (!x << 6) | (!b << 5) | (m & 0x1f); in emit_3vex()
1046 * 7 0 in emit_3vex()
1051 b2 = (w << 7) | ((~vvvv & 0xf) << 3) | (l << 2) | (pp & 3); in emit_3vex()
1062 u8 m = 2; /* escape code 0f38 */ in emit_shiftx()
1065 EMIT2(0xf7, add_2reg(0xC0, dst_reg, dst_reg)); in emit_shiftx()
1073 EMIT3_off32(0x48, 0x8B, 0x85, -round_up(stack, 8) - 8)
1085 int i, excnt = 0; in do_jit()
1086 int ilen, proglen = 0; in do_jit()
1098 bpf_prog->aux->func_idx != 0); in do_jit()
1105 addrs[0] = proglen; in do_jit()
1112 u8 b2 = 0, b3 = 0; in do_jit()
1135 EMIT2(b2, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1140 if (insn->off == 0) in do_jit()
1155 EMIT2(0xF7, add_1reg(0xD8, dst_reg)); in do_jit()
1177 b3 = 0xC0; in do_jit()
1178 b2 = 0x05; in do_jit()
1181 b3 = 0xE8; in do_jit()
1182 b2 = 0x2D; in do_jit()
1185 b3 = 0xE0; in do_jit()
1186 b2 = 0x25; in do_jit()
1189 b3 = 0xC8; in do_jit()
1190 b2 = 0x0D; in do_jit()
1193 b3 = 0xF0; in do_jit()
1194 b2 = 0x35; in do_jit()
1199 EMIT3(0x83, add_1reg(b3, dst_reg), imm32); in do_jit()
1203 EMIT2_off32(0x81, add_1reg(b3, dst_reg), imm32); in do_jit()
1213 emit_mov_imm64(&prog, dst_reg, insn[1].imm, insn[0].imm); in do_jit()
1230 EMIT1(0x50); /* push rax */ in do_jit()
1232 EMIT1(0x52); /* push rdx */ in do_jit()
1243 EMIT3_off32(0x49, 0xC7, 0xC3, imm32); in do_jit()
1251 if (insn->off == 0) { in do_jit()
1256 EMIT2(0x31, 0xd2); in do_jit()
1260 EMIT2(0xF7, add_1reg(0xF0, src_reg)); in do_jit()
1263 EMIT1(0x99); /* cdq */ in do_jit()
1265 EMIT2(0x48, 0x99); /* cqo */ in do_jit()
1269 EMIT2(0xF7, add_1reg(0xF8, src_reg)); in do_jit()
1282 EMIT1(0x5A); /* pop rdx */ in do_jit()
1284 EMIT1(0x58); /* pop rax */ in do_jit()
1295 EMIT3(0x6B, add_2reg(0xC0, dst_reg, dst_reg), in do_jit()
1299 EMIT2_off32(0x69, in do_jit()
1300 add_2reg(0xC0, dst_reg, dst_reg), in do_jit()
1310 EMIT3(0x0F, 0xAF, add_2reg(0xC0, src_reg, dst_reg)); in do_jit()
1325 EMIT2(0xD1, add_1reg(b3, dst_reg)); in do_jit()
1327 EMIT3(0xC1, add_1reg(b3, dst_reg), imm32); in do_jit()
1344 op = 1; /* prefix 0x66 */ in do_jit()
1347 op = 3; /* prefix 0xf2 */ in do_jit()
1350 op = 2; /* prefix 0xf3 */ in do_jit()
1366 EMIT1(0x51); /* push rcx */ in do_jit()
1377 EMIT2(0xD3, add_1reg(b3, dst_reg)); in do_jit()
1384 EMIT1(0x59); /* pop rcx */ in do_jit()
1394 EMIT1(0x66); in do_jit()
1396 EMIT1(0x41); in do_jit()
1397 EMIT3(0xC1, add_1reg(0xC8, dst_reg), 8); in do_jit()
1401 EMIT3(0x45, 0x0F, 0xB7); in do_jit()
1403 EMIT2(0x0F, 0xB7); in do_jit()
1404 EMIT1(add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1409 EMIT2(0x41, 0x0F); in do_jit()
1411 EMIT1(0x0F); in do_jit()
1412 EMIT1(add_1reg(0xC8, dst_reg)); in do_jit()
1416 EMIT3(add_1mod(0x48, dst_reg), 0x0F, in do_jit()
1417 add_1reg(0xC8, dst_reg)); in do_jit()
1430 EMIT3(0x45, 0x0F, 0xB7); in do_jit()
1432 EMIT2(0x0F, 0xB7); in do_jit()
1433 EMIT1(add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1438 EMIT1(0x45); in do_jit()
1439 EMIT2(0x89, add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1455 EMIT2(0x41, 0xC6); in do_jit()
1457 EMIT1(0xC6); in do_jit()
1461 EMIT3(0x66, 0x41, 0xC7); in do_jit()
1463 EMIT2(0x66, 0xC7); in do_jit()
1467 EMIT2(0x41, 0xC7); in do_jit()
1469 EMIT1(0xC7); in do_jit()
1472 EMIT2(add_1mod(0x48, dst_reg), 0xC7); in do_jit()
1475 EMIT2(add_1reg(0x40, dst_reg), insn->off); in do_jit()
1477 EMIT1_off32(add_1reg(0x80, dst_reg), insn->off); in do_jit()
1529 EMIT2_off32(0x81, add_1reg(0xC0, AUX_REG), insn->off); in do_jit()
1534 EMIT2(0x29, add_2reg(0xC0, AUX_REG, BPF_REG_AX)); in do_jit()
1542 EMIT2(0x39, add_2reg(0xC0, AUX_REG, BPF_REG_AX)); in do_jit()
1545 EMIT2(X86_JA, 0); in do_jit()
1549 emit_mov_imm32(&prog, false, dst_reg, 0); in do_jit()
1551 EMIT2(0xEB, 0); in do_jit()
1601 * End result: x86 insn "mov rbx, qword ptr [rax+0x14]" in do_jit()
1641 add_2reg(0xC0, AUX_REG, real_src_reg)); in do_jit()
1726 EMIT2(0x39, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1734 EMIT2(0x85, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1742 EMIT2_off32(0xF7, add_1reg(0xC0, dst_reg), imm32); in do_jit()
1766 if (imm32 == 0) { in do_jit()
1769 EMIT2(0x85, add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1778 EMIT3(0x83, add_1reg(0xF8, dst_reg), imm32); in do_jit()
1780 EMIT2_off32(0x81, add_1reg(0xF8, dst_reg), imm32); in do_jit()
1835 * "nops" is 0. in do_jit()
1845 if (nops != 0 && nops != 4) { in do_jit()
1854 EMIT2_off32(0x0F, jmp_cond + 0x10, jmp_offset); in do_jit()
1891 * to pad any extra byte (0 byte). in do_jit()
1899 if (nops != 0 && nops != 2 && nops != 5) { in do_jit()
1916 * jmp, there is nothing to pad (0 byte). in do_jit()
1924 if (nops != 0 && nops != 3) { in do_jit()
1931 EMIT2(0xEB, jmp_offset); in do_jit()
1933 EMIT1_off32(0xE9, jmp_offset); in do_jit()
1949 EMIT1(0xC9); /* leave */ in do_jit()
2036 /* mov DWORD PTR [rbp + off], 0 */ in clean_stack_garbage()
2038 EMIT2_off32(0xC7, 0x85, off); in clean_stack_garbage()
2040 EMIT3(0xC7, 0x45, off); in clean_stack_garbage()
2041 EMIT(0, 4); in clean_stack_garbage()
2049 int i, arg_regs, nr_used_regs = 0; in get_nr_used_regs()
2051 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in get_nr_used_regs()
2066 int arg_regs, first_off = 0, nr_regs = 0, nr_stack_slots = 0; in save_args()
2071 * mov QWORD PTR [rbp-0x10],rdi in save_args()
2072 * mov QWORD PTR [rbp-0x8],rsi in save_args()
2074 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in save_args()
2105 for (j = 0; j < arg_regs; j++) { in save_args()
2107 nr_stack_slots * 8 + 0x18); in save_args()
2127 for (j = 0; j < arg_regs; j++) { in save_args()
2143 int i, j, arg_regs, nr_regs = 0; in restore_regs()
2147 * EMIT4(0x48, 0x8B, 0x7D, 0xF0); mov rdi,QWORD PTR [rbp-0x10] in restore_regs()
2148 * EMIT4(0x48, 0x8B, 0x75, 0xF8); mov rsi,QWORD PTR [rbp-0x8] in restore_regs()
2152 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in restore_regs()
2155 for (j = 0; j < arg_regs; j++) { in restore_regs()
2198 EMIT3_off32(0x48, 0x8D, 0xB5, -run_ctx_off); in invoke_bpf_prog()
2200 EMIT4(0x48, 0x8D, 0x75, -run_ctx_off); in invoke_bpf_prog()
2207 /* if (__bpf_prog_enter*(prog) == 0) in invoke_bpf_prog()
2210 EMIT3(0x48, 0x85, 0xC0); /* test rax,rax */ in invoke_bpf_prog()
2217 EMIT3_off32(0x48, 0x8D, 0xBD, -stack_size); in invoke_bpf_prog()
2219 EMIT4(0x48, 0x8D, 0x7D, -stack_size); in invoke_bpf_prog()
2241 jmp_insn[0] = X86_JE; in invoke_bpf_prog()
2250 EMIT3_off32(0x48, 0x8D, 0x95, -run_ctx_off); in invoke_bpf_prog()
2252 EMIT4(0x48, 0x8D, 0x55, -run_ctx_off); in invoke_bpf_prog()
2257 return 0; in invoke_bpf_prog()
2281 EMIT2_off32(0x0F, jmp_cond + 0x10, offset); in emit_cond_near_jump()
2283 return 0; in emit_cond_near_jump()
2293 for (i = 0; i < tl->nr_links; i++) { in invoke_bpf()
2299 return 0; in invoke_bpf()
2310 * Set this to 0 to avoid confusing the program. in invoke_bpf_mod_ret()
2312 emit_mov_imm32(&prog, false, BPF_REG_0, 0); in invoke_bpf_mod_ret()
2314 for (i = 0; i < tl->nr_links; i++) { in invoke_bpf_mod_ret()
2319 * if (*(u64 *)(rbp - 8) != 0) in invoke_bpf_mod_ret()
2322 /* cmp QWORD PTR [rbp - 0x8], 0x0 */ in invoke_bpf_mod_ret()
2323 EMIT4(0x48, 0x83, 0x7d, 0xf8); EMIT1(0x00); in invoke_bpf_mod_ret()
2335 return 0; in invoke_bpf_mod_ret()
2403 int i, ret, nr_regs = m->nr_args, stack_size = 0; in arch_prepare_bpf_trampoline()
2414 for (i = 0; i < m->nr_args; i++) in arch_prepare_bpf_trampoline()
2427 * RBP + 0 [ RBP ] in arch_prepare_bpf_trampoline()
2471 stack_size += (sizeof(struct bpf_tramp_run_ctx) + 7) & ~0x7; in arch_prepare_bpf_trampoline()
2483 stack_size += (stack_size % 16) ? 0 : 8; in arch_prepare_bpf_trampoline()
2505 EMIT1(0x55); /* push rbp */ in arch_prepare_bpf_trampoline()
2506 EMIT3(0x48, 0x89, 0xE5); /* mov rbp, rsp */ in arch_prepare_bpf_trampoline()
2509 EMIT3_off32(0x48, 0x81, 0xEC, stack_size); in arch_prepare_bpf_trampoline()
2512 EMIT4(0x48, 0x83, 0xEC, stack_size); in arch_prepare_bpf_trampoline()
2514 EMIT1(0x50); /* push rax */ in arch_prepare_bpf_trampoline()
2522 emit_mov_imm64(&prog, BPF_REG_0, 0, (u32) nr_regs); in arch_prepare_bpf_trampoline()
2575 EMIT2(0xff, 0xd3); /* call *rbx */ in arch_prepare_bpf_trampoline()
2600 for (i = 0; i < fmod_ret->nr_links; i++) in arch_prepare_bpf_trampoline()
2637 EMIT1(0xC9); /* leave */ in arch_prepare_bpf_trampoline()
2640 EMIT4(0x48, 0x83, 0xC4, 8); /* add rsp, 8 */ in arch_prepare_bpf_trampoline()
2664 EMIT1(add_1mod(0x48, BPF_REG_3)); /* cmp rdx,func */ in emit_bpf_dispatcher()
2667 EMIT2_off32(0x81, add_1reg(0xF8, BPF_REG_3), in emit_bpf_dispatcher()
2678 return 0; in emit_bpf_dispatcher()
2685 EMIT1(add_1mod(0x48, BPF_REG_3)); /* cmp rdx,func */ in emit_bpf_dispatcher()
2688 EMIT2_off32(0x81, add_1reg(0xF8, BPF_REG_3), progs[a + pivot]); in emit_bpf_dispatcher()
2693 EMIT2_off32(0x0F, X86_JG + 0x10, 0); in emit_bpf_dispatcher()
2695 EMIT2(X86_JG, 0); in emit_bpf_dispatcher()
2719 return 0; in emit_bpf_dispatcher()
2731 return 0; in cmp_ips()
2738 sort(funcs, num_funcs, sizeof(funcs[0]), cmp_ips, NULL); in arch_prepare_bpf_dispatcher()
2739 return emit_bpf_dispatcher(&prog, 0, num_funcs - 1, funcs, image, buf); in arch_prepare_bpf_dispatcher()
2760 int proglen, oldproglen = 0; in bpf_int_jit_compile()
2817 for (proglen = 0, i = 0; i <= prog->len; i++) { in bpf_int_jit_compile()
2830 for (pass = 0; pass < MAX_PASSES || image; pass++) { in bpf_int_jit_compile()
2834 if (proglen <= 0) { in bpf_int_jit_compile()
2846 prog->jited = 0; in bpf_int_jit_compile()
2847 prog->jited_len = 0; in bpf_int_jit_compile()
2995 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
3001 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
3008 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
3018 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()