/openbmc/linux/net/core/ |
H A D | gso.c | 16 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_eth_gso_segment() local 22 segs = ptype->callbacks.gso_segment(skb, features); in skb_eth_gso_segment() 28 return segs; in skb_eth_gso_segment() 40 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_mac_gso_segment() local 53 segs = ptype->callbacks.gso_segment(skb, features); in skb_mac_gso_segment() 61 return segs; in skb_mac_gso_segment() 91 struct sk_buff *segs; in __skb_gso_segment() local 124 segs = skb_mac_gso_segment(skb, features); in __skb_gso_segment() 126 if (segs != skb && unlikely(skb_needs_check(skb, tx_path) && !IS_ERR(segs))) in __skb_gso_segment() 129 return segs; in __skb_gso_segment()
|
/openbmc/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_flow.c | 714 hdrs = prof->segs[i].hdrs; in ice_flow_proc_seg_hdrs() 1272 if (segs[i].hdrs != p->segs[i].hdrs || in ice_flow_find_prof_conds() 1274 segs[i].match != p->segs[i].match)) in ice_flow_find_prof_conds() 1378 memcpy(¶ms->prof->segs[i], &segs[i], sizeof(*segs)); in ice_flow_add_prof_sync() 1533 if (!segs) in ice_flow_add_prof() 2066 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_add_rss_cfg_sync() 2067 if (!segs) in ice_add_rss_cfg_sync() 2145 kfree(segs); in ice_add_rss_cfg_sync() 2200 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_rem_rss_cfg_sync() 2201 if (!segs) in ice_rem_rss_cfg_sync() [all …]
|
/openbmc/linux/net/sctp/ |
H A D | offload.c | 42 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local 68 segs = NULL; in sctp_gso_segment() 72 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); in sctp_gso_segment() 73 if (IS_ERR(segs)) in sctp_gso_segment() 78 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment() 87 return segs; in sctp_gso_segment()
|
/openbmc/qemu/bsd-user/i386/ |
H A D | target_arch_reg.h | 57 regs->r_fs = env->segs[R_FS].selector & 0xffff; in target_copy_regs() 58 regs->r_es = env->segs[R_ES].selector & 0xffff; in target_copy_regs() 59 regs->r_ds = env->segs[R_DS].selector & 0xffff; in target_copy_regs() 73 regs->r_cs = env->segs[R_CS].selector & 0xffff; in target_copy_regs() 78 regs->r_ss = env->segs[R_SS].selector & 0xffff; in target_copy_regs() 79 regs->r_gs = env->segs[R_GS].selector & 0xffff; in target_copy_regs()
|
/openbmc/linux/net/ipv4/ |
H A D | udp_offload.c | 98 skb = segs; in __skb_udp_tunnel_segment() 150 return segs; in __skb_udp_tunnel_segment() 188 return segs; in skb_udp_tunnel_segment() 225 seg = segs; in __udpv4_gso_segment_list_csum() 233 return segs; in __udpv4_gso_segment_list_csum() 247 return segs; in __udpv4_gso_segment_list_csum() 302 return segs; in __udp_gso_segment() 309 if (skb_is_gso(segs)) in __udp_gso_segment() 312 seg = segs; in __udp_gso_segment() 372 return segs; in __udp_gso_segment() [all …]
|
H A D | tcp_offload.c | 59 struct sk_buff *segs = ERR_PTR(-EINVAL); in tcp_gso_segment() local 91 segs = NULL; in tcp_gso_segment() 100 segs = skb_segment(skb, features); in tcp_gso_segment() 101 if (IS_ERR(segs)) in tcp_gso_segment() 105 segs->ooo_okay = ooo_okay; in tcp_gso_segment() 111 if (skb_is_gso(segs)) in tcp_gso_segment() 112 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment() 116 skb = segs; in tcp_gso_segment() 121 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment() 178 return segs; in tcp_gso_segment()
|
H A D | gre_offload.c | 21 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local 58 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment() 59 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment() 65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 69 skb = segs; in gre_gso_segment() 119 return segs; in gre_gso_segment()
|
/openbmc/qemu/linux-user/ |
H A D | vm86.c | 217 if (env->segs[R_CS].selector == TARGET_BIOSSEG) in do_int() 231 ssp = env->segs[R_SS].selector << 4; in do_int() 277 csp = env->segs[R_CS].selector << 4; in handle_vm86_fault() 280 ssp = env->segs[R_SS].selector << 4; in handle_vm86_fault() 284 env->segs[R_CS].selector, env->eip); in handle_vm86_fault() 431 ts->vm86_saved_regs.cs = env->segs[R_CS].selector; in do_vm86() 432 ts->vm86_saved_regs.ss = env->segs[R_SS].selector; in do_vm86() 433 ts->vm86_saved_regs.ds = env->segs[R_DS].selector; in do_vm86() 434 ts->vm86_saved_regs.es = env->segs[R_ES].selector; in do_vm86() 435 ts->vm86_saved_regs.fs = env->segs[R_FS].selector; in do_vm86() [all …]
|
/openbmc/qemu/target/i386/ |
H A D | arch_dump.c | 74 regs.cs = env->segs[R_CS].selector; in x86_64_write_elf64_note() 75 regs.ss = env->segs[R_SS].selector; in x86_64_write_elf64_note() 76 regs.fs_base = env->segs[R_FS].base; in x86_64_write_elf64_note() 77 regs.gs_base = env->segs[R_GS].base; in x86_64_write_elf64_note() 78 regs.ds = env->segs[R_DS].selector; in x86_64_write_elf64_note() 79 regs.es = env->segs[R_ES].selector; in x86_64_write_elf64_note() 80 regs.fs = env->segs[R_FS].selector; in x86_64_write_elf64_note() 81 regs.gs = env->segs[R_GS].selector; in x86_64_write_elf64_note() 308 copy_segment(&s->cs, &env->segs[R_CS]); in qemu_get_cpustate() 309 copy_segment(&s->ds, &env->segs[R_DS]); in qemu_get_cpustate() [all …]
|
H A D | gdbstub.c | 153 return gdb_get_reg32(mem_buf, env->segs[R_CS].selector); in x86_cpu_gdb_read_register() 155 return gdb_get_reg32(mem_buf, env->segs[R_SS].selector); in x86_cpu_gdb_read_register() 157 return gdb_get_reg32(mem_buf, env->segs[R_DS].selector); in x86_cpu_gdb_read_register() 159 return gdb_get_reg32(mem_buf, env->segs[R_ES].selector); in x86_cpu_gdb_read_register() 161 return gdb_get_reg32(mem_buf, env->segs[R_FS].selector); in x86_cpu_gdb_read_register() 163 return gdb_get_reg32(mem_buf, env->segs[R_GS].selector); in x86_cpu_gdb_read_register() 165 return gdb_read_reg_cs64(env->hflags, mem_buf, env->segs[R_FS].base); in x86_cpu_gdb_read_register() 167 return gdb_read_reg_cs64(env->hflags, mem_buf, env->segs[R_GS].base); in x86_cpu_gdb_read_register() 226 if (selector != env->segs[sreg].selector) { in x86_cpu_gdb_load_seg() 320 return gdb_write_reg_cs64(env->hflags, mem_buf, &env->segs[R_FS].base); in x86_cpu_gdb_write_register() [all …]
|
H A D | machine.c | 243 env->segs[R_CS].flags &= ~(env->segs[R_CS].flags & DESC_DPL_MASK); in cpu_pre_save() 244 env->segs[R_DS].flags &= ~(env->segs[R_DS].flags & DESC_DPL_MASK); in cpu_pre_save() 245 env->segs[R_ES].flags &= ~(env->segs[R_ES].flags & DESC_DPL_MASK); in cpu_pre_save() 246 env->segs[R_FS].flags &= ~(env->segs[R_FS].flags & DESC_DPL_MASK); in cpu_pre_save() 247 env->segs[R_GS].flags &= ~(env->segs[R_GS].flags & DESC_DPL_MASK); in cpu_pre_save() 248 env->segs[R_SS].flags &= ~(env->segs[R_SS].flags & DESC_DPL_MASK); in cpu_pre_save() 341 env->segs[R_CS].flags &= ~(env->segs[R_CS].flags & DESC_DPL_MASK); in cpu_post_load() 342 env->segs[R_DS].flags &= ~(env->segs[R_DS].flags & DESC_DPL_MASK); in cpu_post_load() 343 env->segs[R_ES].flags &= ~(env->segs[R_ES].flags & DESC_DPL_MASK); in cpu_post_load() 344 env->segs[R_FS].flags &= ~(env->segs[R_FS].flags & DESC_DPL_MASK); in cpu_post_load() [all …]
|
/openbmc/qemu/bsd-user/x86_64/ |
H A D | target_arch_reg.h | 80 regs->r_fs = env->segs[R_FS].selector & 0xffff; in target_copy_regs() 81 regs->r_gs = env->segs[R_GS].selector & 0xffff; in target_copy_regs() 83 regs->r_es = env->segs[R_ES].selector & 0xffff; in target_copy_regs() 84 regs->r_ds = env->segs[R_DS].selector & 0xffff; in target_copy_regs() 86 regs->r_cs = env->segs[R_CS].selector & 0xffff; in target_copy_regs() 89 regs->r_ss = env->segs[R_SS].selector & 0xffff; in target_copy_regs()
|
/openbmc/linux/net/mpls/ |
H A D | mpls_gso.c | 23 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local 51 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment() 52 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment() 57 skb = segs; in mpls_gso_segment() 73 return segs; in mpls_gso_segment()
|
/openbmc/qemu/target/i386/nvmm/ |
H A D | nvmm-all.c | 117 nvmm_set_segment(&state->segs[NVMM_X64_SEG_CS], &env->segs[R_CS]); in nvmm_set_registers() 118 nvmm_set_segment(&state->segs[NVMM_X64_SEG_DS], &env->segs[R_DS]); in nvmm_set_registers() 119 nvmm_set_segment(&state->segs[NVMM_X64_SEG_ES], &env->segs[R_ES]); in nvmm_set_registers() 120 nvmm_set_segment(&state->segs[NVMM_X64_SEG_FS], &env->segs[R_FS]); in nvmm_set_registers() 121 nvmm_set_segment(&state->segs[NVMM_X64_SEG_GS], &env->segs[R_GS]); in nvmm_set_registers() 122 nvmm_set_segment(&state->segs[NVMM_X64_SEG_SS], &env->segs[R_SS]); in nvmm_set_registers() 269 nvmm_get_segment(&env->segs[R_ES], &state->segs[NVMM_X64_SEG_ES]); in nvmm_get_registers() 270 nvmm_get_segment(&env->segs[R_CS], &state->segs[NVMM_X64_SEG_CS]); in nvmm_get_registers() 271 nvmm_get_segment(&env->segs[R_SS], &state->segs[NVMM_X64_SEG_SS]); in nvmm_get_registers() 272 nvmm_get_segment(&env->segs[R_DS], &state->segs[NVMM_X64_SEG_DS]); in nvmm_get_registers() [all …]
|
/openbmc/qemu/target/i386/tcg/ |
H A D | seg_helper.c | 737 ssp = env->segs[R_SS].base; in do_interrupt_protected() 1058 ssp = env->segs[R_SS].base; in do_interrupt_real() 1064 old_cs = env->segs[R_CS].selector; in do_interrupt_real() 1519 ssp = env->segs[R_SS].base; in helper_lcall_real() 1530 env->segs[R_CS].selector = new_cs; in helper_lcall_real() 1598 ssp = env->segs[R_SS].base; in helper_lcall_protected() 1785 ssp = env->segs[R_SS].base; in helper_lcall_protected() 1840 ssp = env->segs[R_SS].base; in helper_iret_real() 1854 env->segs[R_CS].selector = new_cs; in helper_iret_real() 1884 e2 = env->segs[seg_reg].flags; in validate_seg() [all …]
|
/openbmc/qemu/dump/ |
H A D | win_dump.c | 309 .SegEs = env->segs[0].selector, in patch_and_save_context() 310 .SegCs = env->segs[1].selector, in patch_and_save_context() 311 .SegSs = env->segs[2].selector, in patch_and_save_context() 312 .SegDs = env->segs[3].selector, in patch_and_save_context() 313 .SegFs = env->segs[4].selector, in patch_and_save_context() 314 .SegGs = env->segs[5].selector, in patch_and_save_context() 350 .SegEs = env->segs[0].selector, in patch_and_save_context() 351 .SegCs = env->segs[1].selector, in patch_and_save_context() 352 .SegSs = env->segs[2].selector, in patch_and_save_context() 353 .SegDs = env->segs[3].selector, in patch_and_save_context() [all …]
|
/openbmc/linux/net/nsh/ |
H A D | nsh.c | 81 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local 110 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment() 111 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment() 117 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment() 126 return segs; in nsh_gso_segment()
|
/openbmc/linux/tools/lib/bpf/ |
H A D | usdt.c | 427 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); in parse_elf_segs() 431 *segs = tmp; in parse_elf_segs() 432 seg = *segs + *seg_cnt; in parse_elf_segs() 446 qsort(*segs, *seg_cnt, sizeof(**segs), cmp_elf_segs); in parse_elf_segs() 510 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); in parse_vma_segs() 516 *segs = tmp; in parse_vma_segs() 517 seg = *segs + *seg_cnt; in parse_vma_segs() 533 qsort(*segs, *seg_cnt, sizeof(**segs), cmp_elf_segs); in parse_vma_segs() 583 struct elf_seg *segs = NULL, *vma_segs = NULL; in collect_usdt_targets() local 670 seg = find_elf_seg(segs, seg_cnt, usdt_abs_ip); in collect_usdt_targets() [all …]
|
/openbmc/qemu/target/i386/hvf/ |
H A D | x86hvf.c | 104 hvf_set_segment(cs, &seg, &env->segs[R_CS], false); in hvf_put_segments() 107 hvf_set_segment(cs, &seg, &env->segs[R_DS], false); in hvf_put_segments() 110 hvf_set_segment(cs, &seg, &env->segs[R_ES], false); in hvf_put_segments() 113 hvf_set_segment(cs, &seg, &env->segs[R_SS], false); in hvf_put_segments() 116 hvf_set_segment(cs, &seg, &env->segs[R_FS], false); in hvf_put_segments() 175 hvf_get_segment(&env->segs[R_CS], &seg); in hvf_get_segments() 178 hvf_get_segment(&env->segs[R_DS], &seg); in hvf_get_segments() 181 hvf_get_segment(&env->segs[R_ES], &seg); in hvf_get_segments() 184 hvf_get_segment(&env->segs[R_FS], &seg); in hvf_get_segments() 187 hvf_get_segment(&env->segs[R_GS], &seg); in hvf_get_segments() [all …]
|
/openbmc/linux/include/uapi/linux/ |
H A D | elf-fdpic.h | 30 struct elf32_fdpic_loadseg segs[]; member 45 struct elf64_fdpic_loadseg segs[]; member
|
/openbmc/linux/net/ipv6/ |
H A D | ip6_offload.c | 112 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local 139 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment() 154 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment() 155 if (!segs) in ipv6_gso_segment() 159 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment() 162 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 164 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment() 179 kfree_skb_list(segs); in ipv6_gso_segment() 194 return segs; in ipv6_gso_segment()
|
/openbmc/linux/net/sched/ |
H A D | sch_tbf.c | 209 struct sk_buff *segs, *nskb; in tbf_segment() local 214 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment() 216 if (IS_ERR_OR_NULL(segs)) in tbf_segment() 220 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment() 221 skb_mark_not_on_list(segs); in tbf_segment() 222 qdisc_skb_cb(segs)->pkt_len = segs->len; in tbf_segment() 223 len += segs->len; in tbf_segment() 224 ret = qdisc_enqueue(segs, q->qdisc, to_free); in tbf_segment()
|
/openbmc/linux/drivers/infiniband/sw/rdmavt/ |
H A D | mr.c | 379 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr() 380 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr() 569 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page() 570 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page() 787 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok() 788 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok() 797 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok() 894 while (off >= mr->map[m]->segs[n].length) { in rvt_rkey_ok() 895 off -= mr->map[m]->segs[n].length; in rvt_rkey_ok() 904 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_rkey_ok() [all …]
|
/openbmc/linux/include/linux/ |
H A D | blk-integrity.h | 69 unsigned int segs) in blk_queue_max_integrity_segments() argument 71 q->limits.max_integrity_segments = segs; in blk_queue_max_integrity_segments() 154 unsigned int segs) in blk_queue_max_integrity_segments() argument
|
/openbmc/qemu/contrib/elf2dmp/ |
H A D | pdb.c | 40 for (ptr = r->segs; (ptr < r->segs + r->segs_size); ) { in get_seg_by_num() 108 g_free(r->segs); in pdb_exit_segments() 165 r->segs = pdb_ds_read_file(r, stream_idx); in pdb_init_segments() 166 if (!r->segs) { in pdb_init_segments()
|