Lines Matching refs:src_reg
249 insn->src_reg == 0; in bpf_helper_call()
255 insn->src_reg == BPF_PSEUDO_CALL; in bpf_pseudo_call()
261 insn->src_reg == BPF_PSEUDO_KFUNC_CALL; in bpf_pseudo_kfunc_call()
3072 insn[i].src_reg == 0 && in check_subprogs()
3233 if (insn->src_reg == BPF_PSEUDO_CALL) in is_reg64()
3313 return insn->src_reg; in insn_def_regno()
3472 if (insn->src_reg != BPF_PSEUDO_KFUNC_CALL) in disasm_kfunc_name()
3662 u32 sreg = insn->src_reg; in backtrack_insn()
3726 if (insn->src_reg != BPF_REG_FP) in backtrack_insn()
3843 if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL && insn->imm == 0) in backtrack_insn()
6959 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_atomic()
6982 if (is_pointer_value(env, insn->src_reg)) { in check_atomic()
6983 verbose(env, "R%d leaks addr into mem\n", insn->src_reg); in check_atomic()
7001 load_reg = insn->src_reg; in check_atomic()
9282 insn->src_reg == 0 && in push_callback_call()
12245 mark_reg_unknown(env, regs, insn->src_reg); in sanitize_speculative_path()
12367 u32 dst = insn->dst_reg, src = insn->src_reg; in sanitize_err()
12701 struct bpf_reg_state *src_reg) in scalar32_min_max_add() argument
12703 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_add()
12704 s32 smax_val = src_reg->s32_max_value; in scalar32_min_max_add()
12705 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_add()
12706 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_add()
12727 struct bpf_reg_state *src_reg) in scalar_min_max_add() argument
12729 s64 smin_val = src_reg->smin_value; in scalar_min_max_add()
12730 s64 smax_val = src_reg->smax_value; in scalar_min_max_add()
12731 u64 umin_val = src_reg->umin_value; in scalar_min_max_add()
12732 u64 umax_val = src_reg->umax_value; in scalar_min_max_add()
12753 struct bpf_reg_state *src_reg) in scalar32_min_max_sub() argument
12755 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_sub()
12756 s32 smax_val = src_reg->s32_max_value; in scalar32_min_max_sub()
12757 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_sub()
12758 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_sub()
12781 struct bpf_reg_state *src_reg) in scalar_min_max_sub() argument
12783 s64 smin_val = src_reg->smin_value; in scalar_min_max_sub()
12784 s64 smax_val = src_reg->smax_value; in scalar_min_max_sub()
12785 u64 umin_val = src_reg->umin_value; in scalar_min_max_sub()
12786 u64 umax_val = src_reg->umax_value; in scalar_min_max_sub()
12809 struct bpf_reg_state *src_reg) in scalar32_min_max_mul() argument
12811 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_mul()
12812 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_mul()
12813 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_mul()
12841 struct bpf_reg_state *src_reg) in scalar_min_max_mul() argument
12843 s64 smin_val = src_reg->smin_value; in scalar_min_max_mul()
12844 u64 umin_val = src_reg->umin_value; in scalar_min_max_mul()
12845 u64 umax_val = src_reg->umax_value; in scalar_min_max_mul()
12873 struct bpf_reg_state *src_reg) in scalar32_min_max_and() argument
12875 bool src_known = tnum_subreg_is_const(src_reg->var_off); in scalar32_min_max_and()
12878 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_and()
12879 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_and()
12907 struct bpf_reg_state *src_reg) in scalar_min_max_and() argument
12909 bool src_known = tnum_is_const(src_reg->var_off); in scalar_min_max_and()
12911 s64 smin_val = src_reg->smin_value; in scalar_min_max_and()
12912 u64 umax_val = src_reg->umax_value; in scalar_min_max_and()
12942 struct bpf_reg_state *src_reg) in scalar32_min_max_or() argument
12944 bool src_known = tnum_subreg_is_const(src_reg->var_off); in scalar32_min_max_or()
12947 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_or()
12948 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_or()
12976 struct bpf_reg_state *src_reg) in scalar_min_max_or() argument
12978 bool src_known = tnum_is_const(src_reg->var_off); in scalar_min_max_or()
12980 s64 smin_val = src_reg->smin_value; in scalar_min_max_or()
12981 u64 umin_val = src_reg->umin_value; in scalar_min_max_or()
13011 struct bpf_reg_state *src_reg) in scalar32_min_max_xor() argument
13013 bool src_known = tnum_subreg_is_const(src_reg->var_off); in scalar32_min_max_xor()
13016 s32 smin_val = src_reg->s32_min_value; in scalar32_min_max_xor()
13040 struct bpf_reg_state *src_reg) in scalar_min_max_xor() argument
13042 bool src_known = tnum_is_const(src_reg->var_off); in scalar_min_max_xor()
13044 s64 smin_val = src_reg->smin_value; in scalar_min_max_xor()
13089 struct bpf_reg_state *src_reg) in scalar32_min_max_lsh() argument
13091 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_lsh()
13092 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_lsh()
13137 struct bpf_reg_state *src_reg) in scalar_min_max_lsh() argument
13139 u64 umax_val = src_reg->umax_value; in scalar_min_max_lsh()
13140 u64 umin_val = src_reg->umin_value; in scalar_min_max_lsh()
13152 struct bpf_reg_state *src_reg) in scalar32_min_max_rsh() argument
13155 u32 umax_val = src_reg->u32_max_value; in scalar32_min_max_rsh()
13156 u32 umin_val = src_reg->u32_min_value; in scalar32_min_max_rsh()
13184 struct bpf_reg_state *src_reg) in scalar_min_max_rsh() argument
13186 u64 umax_val = src_reg->umax_value; in scalar_min_max_rsh()
13187 u64 umin_val = src_reg->umin_value; in scalar_min_max_rsh()
13218 struct bpf_reg_state *src_reg) in scalar32_min_max_arsh() argument
13220 u64 umin_val = src_reg->u32_min_value; in scalar32_min_max_arsh()
13241 struct bpf_reg_state *src_reg) in scalar_min_max_arsh() argument
13243 u64 umin_val = src_reg->umin_value; in scalar_min_max_arsh()
13274 struct bpf_reg_state src_reg) in adjust_scalar_min_max_vals() argument
13287 smin_val = src_reg.smin_value; in adjust_scalar_min_max_vals()
13288 smax_val = src_reg.smax_value; in adjust_scalar_min_max_vals()
13289 umin_val = src_reg.umin_value; in adjust_scalar_min_max_vals()
13290 umax_val = src_reg.umax_value; in adjust_scalar_min_max_vals()
13292 s32_min_val = src_reg.s32_min_value; in adjust_scalar_min_max_vals()
13293 s32_max_val = src_reg.s32_max_value; in adjust_scalar_min_max_vals()
13294 u32_min_val = src_reg.u32_min_value; in adjust_scalar_min_max_vals()
13295 u32_max_val = src_reg.u32_max_value; in adjust_scalar_min_max_vals()
13298 src_known = tnum_subreg_is_const(src_reg.var_off); in adjust_scalar_min_max_vals()
13309 src_known = tnum_is_const(src_reg.var_off); in adjust_scalar_min_max_vals()
13349 scalar32_min_max_add(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13350 scalar_min_max_add(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13351 dst_reg->var_off = tnum_add(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
13354 scalar32_min_max_sub(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13355 scalar_min_max_sub(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13356 dst_reg->var_off = tnum_sub(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
13359 dst_reg->var_off = tnum_mul(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
13360 scalar32_min_max_mul(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13361 scalar_min_max_mul(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13364 dst_reg->var_off = tnum_and(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
13365 scalar32_min_max_and(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13366 scalar_min_max_and(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13369 dst_reg->var_off = tnum_or(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
13370 scalar32_min_max_or(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13371 scalar_min_max_or(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13374 dst_reg->var_off = tnum_xor(dst_reg->var_off, src_reg.var_off); in adjust_scalar_min_max_vals()
13375 scalar32_min_max_xor(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13376 scalar_min_max_xor(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13387 scalar32_min_max_lsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13389 scalar_min_max_lsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13400 scalar32_min_max_rsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13402 scalar_min_max_rsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13413 scalar32_min_max_arsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13415 scalar_min_max_arsh(dst_reg, &src_reg); in adjust_scalar_min_max_vals()
13437 struct bpf_reg_state *regs = state->regs, *dst_reg, *src_reg; in adjust_reg_min_max_vals() local
13443 src_reg = NULL; in adjust_reg_min_max_vals()
13452 src_reg = ®s[insn->src_reg]; in adjust_reg_min_max_vals()
13453 if (src_reg->type != SCALAR_VALUE) { in adjust_reg_min_max_vals()
13476 src_reg, dst_reg); in adjust_reg_min_max_vals()
13480 err = mark_chain_precision(env, insn->src_reg); in adjust_reg_min_max_vals()
13484 dst_reg, src_reg); in adjust_reg_min_max_vals()
13487 err = mark_chain_precision(env, insn->src_reg); in adjust_reg_min_max_vals()
13497 src_reg = &off_reg; in adjust_reg_min_max_vals()
13500 ptr_reg, src_reg); in adjust_reg_min_max_vals()
13509 if (WARN_ON(!src_reg)) { in adjust_reg_min_max_vals()
13514 return adjust_scalar_min_max_vals(env, insn, dst_reg, *src_reg); in adjust_reg_min_max_vals()
13527 insn->src_reg != BPF_REG_0 || in check_alu_op()
13533 if (insn->src_reg != BPF_REG_0 || insn->off != 0 || in check_alu_op()
13580 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_alu_op()
13584 if (insn->src_reg != BPF_REG_0 || insn->off != 0) { in check_alu_op()
13596 struct bpf_reg_state *src_reg = regs + insn->src_reg; in check_alu_op() local
13598 bool need_id = src_reg->type == SCALAR_VALUE && !src_reg->id && in check_alu_op()
13599 !tnum_is_const(src_reg->var_off); in check_alu_op()
13611 src_reg->id = ++env->id_gen; in check_alu_op()
13612 copy_register_state(dst_reg, src_reg); in check_alu_op()
13617 if (is_pointer_value(env, insn->src_reg)) { in check_alu_op()
13620 insn->src_reg); in check_alu_op()
13622 } else if (src_reg->type == SCALAR_VALUE) { in check_alu_op()
13625 no_sext = src_reg->umax_value < (1ULL << (insn->off - 1)); in check_alu_op()
13627 src_reg->id = ++env->id_gen; in check_alu_op()
13628 copy_register_state(dst_reg, src_reg); in check_alu_op()
13640 if (is_pointer_value(env, insn->src_reg)) { in check_alu_op()
13643 insn->src_reg); in check_alu_op()
13645 } else if (src_reg->type == SCALAR_VALUE) { in check_alu_op()
13647 bool is_src_reg_u32 = src_reg->umax_value <= U32_MAX; in check_alu_op()
13650 src_reg->id = ++env->id_gen; in check_alu_op()
13651 copy_register_state(dst_reg, src_reg); in check_alu_op()
13662 bool no_sext = src_reg->umax_value < (1ULL << (insn->off - 1)); in check_alu_op()
13665 src_reg->id = ++env->id_gen; in check_alu_op()
13666 copy_register_state(dst_reg, src_reg); in check_alu_op()
13709 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_alu_op()
13713 if (insn->src_reg != BPF_REG_0 || insn->off > 1 || in check_alu_op()
14042 struct bpf_reg_state *src_reg, in is_pkt_ptr_branch_taken() argument
14047 if (src_reg->type == PTR_TO_PACKET_END) { in is_pkt_ptr_branch_taken()
14050 pkt = src_reg; in is_pkt_ptr_branch_taken()
14260 static void __reg_combine_min_max(struct bpf_reg_state *src_reg, in __reg_combine_min_max() argument
14263 src_reg->umin_value = dst_reg->umin_value = max(src_reg->umin_value, in __reg_combine_min_max()
14265 src_reg->umax_value = dst_reg->umax_value = min(src_reg->umax_value, in __reg_combine_min_max()
14267 src_reg->smin_value = dst_reg->smin_value = max(src_reg->smin_value, in __reg_combine_min_max()
14269 src_reg->smax_value = dst_reg->smax_value = min(src_reg->smax_value, in __reg_combine_min_max()
14271 src_reg->var_off = dst_reg->var_off = tnum_intersect(src_reg->var_off, in __reg_combine_min_max()
14273 reg_bounds_sync(src_reg); in __reg_combine_min_max()
14365 struct bpf_reg_state *src_reg, in try_match_pkt_pointers() argument
14379 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
14381 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
14387 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
14389 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
14391 find_good_pkt_pointers(other_branch, src_reg, in try_match_pkt_pointers()
14392 src_reg->type, true); in try_match_pkt_pointers()
14393 mark_pkt_end(this_branch, insn->src_reg, false); in try_match_pkt_pointers()
14400 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
14402 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
14408 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
14410 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
14412 find_good_pkt_pointers(this_branch, src_reg, in try_match_pkt_pointers()
14413 src_reg->type, false); in try_match_pkt_pointers()
14414 mark_pkt_end(other_branch, insn->src_reg, true); in try_match_pkt_pointers()
14421 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
14423 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
14429 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
14431 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
14433 find_good_pkt_pointers(other_branch, src_reg, in try_match_pkt_pointers()
14434 src_reg->type, false); in try_match_pkt_pointers()
14435 mark_pkt_end(this_branch, insn->src_reg, true); in try_match_pkt_pointers()
14442 src_reg->type == PTR_TO_PACKET_END) || in try_match_pkt_pointers()
14444 reg_is_init_pkt_pointer(src_reg, PTR_TO_PACKET))) { in try_match_pkt_pointers()
14450 src_reg->type == PTR_TO_PACKET) || in try_match_pkt_pointers()
14452 src_reg->type == PTR_TO_PACKET_META)) { in try_match_pkt_pointers()
14454 find_good_pkt_pointers(this_branch, src_reg, in try_match_pkt_pointers()
14455 src_reg->type, true); in try_match_pkt_pointers()
14456 mark_pkt_end(other_branch, insn->src_reg, false); in try_match_pkt_pointers()
14489 struct bpf_reg_state *dst_reg, *other_branch_regs, *src_reg = NULL; in check_cond_jmp_op() local
14515 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_cond_jmp_op()
14519 src_reg = ®s[insn->src_reg]; in check_cond_jmp_op()
14520 if (!(reg_is_pkt_pointer_any(dst_reg) && reg_is_pkt_pointer_any(src_reg)) && in check_cond_jmp_op()
14521 is_pointer_value(env, insn->src_reg)) { in check_cond_jmp_op()
14523 insn->src_reg); in check_cond_jmp_op()
14527 if (insn->src_reg != BPF_REG_0) { in check_cond_jmp_op()
14537 } else if (src_reg->type == SCALAR_VALUE && in check_cond_jmp_op()
14538 is_jmp32 && tnum_is_const(tnum_subreg(src_reg->var_off))) { in check_cond_jmp_op()
14540 tnum_subreg(src_reg->var_off).value, in check_cond_jmp_op()
14543 } else if (src_reg->type == SCALAR_VALUE && in check_cond_jmp_op()
14544 !is_jmp32 && tnum_is_const(src_reg->var_off)) { in check_cond_jmp_op()
14546 src_reg->var_off.value, in check_cond_jmp_op()
14551 pred = is_branch_taken(src_reg, in check_cond_jmp_op()
14557 pred = is_branch_taken(src_reg, in check_cond_jmp_op()
14562 reg_is_pkt_pointer_any(src_reg) && in check_cond_jmp_op()
14564 pred = is_pkt_ptr_branch_taken(dst_reg, src_reg, opcode); in check_cond_jmp_op()
14574 !__is_pointer_value(false, src_reg)) in check_cond_jmp_op()
14575 err = mark_chain_precision(env, insn->src_reg); in check_cond_jmp_op()
14622 struct bpf_reg_state *src_reg = ®s[insn->src_reg]; in check_cond_jmp_op() local
14625 src_reg->type == SCALAR_VALUE) { in check_cond_jmp_op()
14626 if (tnum_is_const(src_reg->var_off) || in check_cond_jmp_op()
14628 tnum_is_const(tnum_subreg(src_reg->var_off)))) in check_cond_jmp_op()
14631 src_reg->var_off.value, in check_cond_jmp_op()
14632 tnum_subreg(src_reg->var_off).value, in check_cond_jmp_op()
14637 reg_set_min_max_inv(&other_branch_regs[insn->src_reg], in check_cond_jmp_op()
14638 src_reg, in check_cond_jmp_op()
14645 reg_combine_min_max(&other_branch_regs[insn->src_reg], in check_cond_jmp_op()
14647 src_reg, dst_reg, opcode); in check_cond_jmp_op()
14648 if (src_reg->id && in check_cond_jmp_op()
14649 !WARN_ON_ONCE(src_reg->id != other_branch_regs[insn->src_reg].id)) { in check_cond_jmp_op()
14650 find_equal_scalars(this_branch, src_reg); in check_cond_jmp_op()
14651 find_equal_scalars(other_branch, &other_branch_regs[insn->src_reg]); in check_cond_jmp_op()
14680 __is_pointer_value(false, src_reg) && __is_pointer_value(false, dst_reg) && in check_cond_jmp_op()
14681 type_may_be_null(src_reg->type) != type_may_be_null(dst_reg->type) && in check_cond_jmp_op()
14682 base_type(src_reg->type) != PTR_TO_BTF_ID && in check_cond_jmp_op()
14697 if (type_may_be_null(src_reg->type)) in check_cond_jmp_op()
14698 mark_ptr_not_null_reg(&eq_branch_regs[insn->src_reg]); in check_cond_jmp_op()
14718 } else if (!try_match_pkt_pointers(insn, dst_reg, ®s[insn->src_reg], in check_cond_jmp_op()
14753 if (insn->src_reg == 0) { in check_ld_imm()
14767 if (insn->src_reg == BPF_PSEUDO_BTF_ID) { in check_ld_imm()
14784 if (insn->src_reg == BPF_PSEUDO_FUNC) { in check_ld_imm()
14806 if (insn->src_reg == BPF_PSEUDO_MAP_VALUE || in check_ld_imm()
14807 insn->src_reg == BPF_PSEUDO_MAP_IDX_VALUE) { in check_ld_imm()
14812 } else if (insn->src_reg == BPF_PSEUDO_MAP_FD || in check_ld_imm()
14813 insn->src_reg == BPF_PSEUDO_MAP_IDX) { in check_ld_imm()
14869 (mode == BPF_ABS && insn->src_reg != BPF_REG_0)) { in check_ld_abs()
14907 err = check_reg_arg(env, insn->src_reg, SRC_OP); in check_ld_abs()
15272 if (insn->src_reg == 0 && insn->imm == BPF_FUNC_timer_set_callback) in visit_insn()
15294 if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL) { in visit_insn()
15314 return visit_func_call_insn(t, insns, env, insn->src_reg == BPF_PSEUDO_CALL); in visit_insn()
17103 err = check_reg_arg(env, insn->src_reg, SRC_OP); in do_check()
17111 src_reg_type = regs[insn->src_reg].type; in do_check()
17116 err = check_mem_access(env, env->insn_idx, insn->src_reg, in do_check()
17143 err = check_reg_arg(env, insn->src_reg, SRC_OP); in do_check()
17156 BPF_WRITE, insn->src_reg, false, false); in do_check()
17167 insn->src_reg != BPF_REG_0) { in do_check()
17194 (insn->src_reg != BPF_PSEUDO_KFUNC_CALL in do_check()
17196 (insn->src_reg != BPF_REG_0 && in do_check()
17197 insn->src_reg != BPF_PSEUDO_CALL && in do_check()
17198 insn->src_reg != BPF_PSEUDO_KFUNC_CALL) || in do_check()
17206 if ((insn->src_reg == BPF_REG_0 && insn->imm != BPF_FUNC_spin_unlock) || in do_check()
17207 (insn->src_reg == BPF_PSEUDO_CALL) || in do_check()
17208 (insn->src_reg == BPF_PSEUDO_KFUNC_CALL && in do_check()
17214 if (insn->src_reg == BPF_PSEUDO_CALL) in do_check()
17216 else if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL) in do_check()
17226 insn->src_reg != BPF_REG_0 || in do_check()
17243 insn->src_reg != BPF_REG_0 || in do_check()
17622 insn[1].dst_reg != 0 || insn[1].src_reg != 0 || in resolve_pseudo_ldimm64()
17628 if (insn[0].src_reg == 0) in resolve_pseudo_ldimm64()
17632 if (insn[0].src_reg == BPF_PSEUDO_BTF_ID) { in resolve_pseudo_ldimm64()
17640 if (insn[0].src_reg == BPF_PSEUDO_FUNC) { in resolve_pseudo_ldimm64()
17649 switch (insn[0].src_reg) { in resolve_pseudo_ldimm64()
17663 switch (insn[0].src_reg) { in resolve_pseudo_ldimm64()
17694 if (insn[0].src_reg == BPF_PSEUDO_MAP_FD || in resolve_pseudo_ldimm64()
17695 insn[0].src_reg == BPF_PSEUDO_MAP_IDX) { in resolve_pseudo_ldimm64()
17805 if (insn->src_reg == BPF_PSEUDO_FUNC) in convert_pseudo_ld_imm64()
17807 insn->src_reg = 0; in convert_pseudo_ld_imm64()
18245 zext_patch[1].src_reg = load_reg; in opt_subreg_zext_lo32_rnd_hi32()
18912 BPF_JNE | BPF_K, insn->src_reg, in do_misc_fixups()
18921 BPF_JEQ | BPF_K, insn->src_reg, in do_misc_fixups()
18981 off_reg = issrc ? insn->src_reg : insn->dst_reg; in do_misc_fixups()
18995 *patch++ = BPF_MOV64_REG(insn->dst_reg, insn->src_reg); in do_misc_fixups()
18996 insn->src_reg = BPF_REG_AX; in do_misc_fixups()
19017 if (insn->src_reg == BPF_PSEUDO_CALL) in do_misc_fixups()
19019 if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL) { in do_misc_fixups()
19490 insn->src_reg == 0 && in is_bpf_loop_call()