/openbmc/linux/net/ipv4/ |
H A D | tcp_offload.c | 21 skb_shinfo(skb)->tx_flags |= SKBTX_SW_TSTAMP; in tcp_gso_tstamp() 22 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp() 34 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4)) in tcp4_gso_segment() 82 mss = skb_shinfo(skb)->gso_size; in tcp_gso_segment() 112 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment() 256 mss = skb_shinfo(p)->gso_size; in tcp_gro_receive() 263 flush |= (mss != skb_shinfo(skb)->gso_size); in tcp_gro_receive() 307 skb_shinfo(skb)->gso_segs = NAPI_GRO_CB(skb)->count; in tcp_gro_complete() 310 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tcp_gro_complete() 338 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV4; in tcp4_gro_complete() [all …]
|
H A D | udp_offload.c | 42 if (skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) in __skb_udp_tunnel_segment() 126 uh->len = htons(skb_shinfo(skb)->gso_size + in __skb_udp_tunnel_segment() 254 unsigned int mss = skb_shinfo(skb)->gso_size; in __udp_gso_segment_list() 277 mss = skb_shinfo(gso_skb)->gso_size; in __udp_gso_segment() 310 mss *= skb_shinfo(segs)->gso_segs; in __udp_gso_segment() 316 skb_shinfo(seg)->tskey = skb_shinfo(gso_skb)->tskey; in __udp_gso_segment() 317 skb_shinfo(seg)->tx_flags |= in __udp_gso_segment() 386 (skb_shinfo(skb)->gso_type & in udp4_ufo_fragment() 401 mss = skb_shinfo(skb)->gso_size; in udp4_ufo_fragment() 442 skb_shinfo(p)->frag_list = skb; in skb_gro_receive_list() [all …]
|
H A D | gre_offload.c | 45 need_csum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_GRE_CSUM); in gre_gso_segment() 65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 103 skb_shinfo(skb)->gso_size; in gre_gso_segment() 245 skb_shinfo(skb)->gso_type = SKB_GSO_GRE; in gre_gro_complete()
|
/openbmc/linux/net/core/ |
H A D | skbuff.c | 1942 skb_shinfo(new)->gso_size = skb_shinfo(old)->gso_size; in skb_copy_header() 1943 skb_shinfo(new)->gso_segs = skb_shinfo(old)->gso_segs; in skb_copy_header() 1944 skb_shinfo(new)->gso_type = skb_shinfo(old)->gso_type; in skb_copy_header() 2048 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; in __pskb_copy_fclone() 2055 skb_shinfo(n)->frag_list = skb_shinfo(skb)->frag_list; in __pskb_copy_fclone() 3638 skb_shinfo(to)->frags[j] = skb_shinfo(from)->frags[i]; in skb_zerocopy() 3878 skb_shinfo(skb1)->frags[i] = skb_shinfo(skb)->frags[i]; in skb_split_inside_header() 3880 skb_shinfo(skb1)->nr_frags = skb_shinfo(skb)->nr_frags; in skb_split_inside_header() 3905 skb_shinfo(skb1)->frags[k] = skb_shinfo(skb)->frags[i]; in skb_split_no_header() 4658 skb_shinfo(nskb)->flags |= skb_shinfo(head_skb)->flags & in skb_segment() [all …]
|
H A D | tso.c | 53 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_build_data() 54 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data() 79 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_start() 80 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
|
H A D | gro.c | 98 struct skb_shared_info *pinfo, *skbinfo = skb_shinfo(skb); in skb_gro_receive() 135 pinfo = skb_shinfo(lp); in skb_gro_receive() 215 skb_shinfo(p)->frag_list = skb; in skb_gro_receive() 247 skb_shinfo(skb)->gso_size = 0; in napi_gro_complete() 373 const struct skb_shared_info *pinfo = skb_shinfo(skb); in skb_gro_reset_offset() 393 struct skb_shared_info *pinfo = skb_shinfo(skb); in gro_pull_from_frag0() 474 NAPI_GRO_CB(skb)->count = skb_shinfo(skb)->gso_segs; in dev_gro_receive() 477 (skb_shinfo(skb)->gso_type & SKB_GSO_DODGY)) in dev_gro_receive() 528 skb_shinfo(skb)->gso_size = skb_gro_len(skb); in dev_gro_receive() 635 skb_shinfo(skb)->gso_type = 0; in napi_reuse_skb() [all …]
|
/openbmc/linux/drivers/net/ethernet/sfc/siena/ |
H A D | tx.h | 28 if (skb_shinfo(skb)->gso_segs > 1 && in efx_tx_csum_type_skb() 29 !(skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) && in efx_tx_csum_type_skb() 30 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM)) in efx_tx_csum_type_skb()
|
/openbmc/linux/drivers/net/ethernet/sfc/ |
H A D | tx.h | 35 if (skb_shinfo(skb)->gso_segs > 1 && in efx_tx_csum_type_skb() 36 !(skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) && in efx_tx_csum_type_skb() 37 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM)) in efx_tx_csum_type_skb()
|
H A D | tx_tso.c | 291 bool is_last = st->out_len <= skb_shinfo(skb)->gso_size; in tso_start_new_packet() 295 st->packet_space = skb_shinfo(skb)->gso_size; in tso_start_new_packet() 341 st->seqnum += skb_shinfo(skb)->gso_size; in tso_start_new_packet() 386 EFX_WARN_ON_ONCE_PARANOID(skb_shinfo(skb)->nr_frags < 1); in efx_enqueue_skb_tso() 389 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso() 408 if (++frag_i >= skb_shinfo(skb)->nr_frags) in efx_enqueue_skb_tso() 412 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso()
|
H A D | ef100_tx.c | 67 mss = skb_shinfo(skb)->gso_size; in ef100_tx_can_tso() 77 if (skb_shinfo(skb)->gso_segs > nic_data->tso_max_payload_num_segs) { in ef100_tx_can_tso() 190 bool gso_partial = skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL; in ef100_make_tso_desc() 195 u32 mss = skb_shinfo(skb)->gso_size; in ef100_make_tso_desc() 203 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCP_FIXEDID) in ef100_make_tso_desc() 216 if (skb_shinfo(skb)->gso_type & in ef100_make_tso_desc() 224 outer_csum = skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM; in ef100_make_tso_desc() 392 segments = skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 0; in __ef100_enqueue_skb()
|
/openbmc/linux/drivers/net/xen-netback/ |
H A D | netback.c | 484 shinfo = skb_shinfo(nskb); in xenvif_get_requests() 703 skb_shinfo(skb)->destructor_arg = in xenvif_fill_frags() 1033 skb_shinfo(skb)->nr_frags = ret; in xenvif_tx_build_gops() 1045 skb_shinfo(skb)->nr_frags = 0; in xenvif_tx_build_gops() 1061 skb_shinfo(skb)->nr_frags = 0; in xenvif_tx_build_gops() 1164 skb_shinfo(skb)->nr_frags = i; in xenvif_handle_frag_list() 1190 skb_shinfo(skb)->nr_frags = 0; in xenvif_tx_submit() 1193 skb_shinfo(skb)->frag_list; in xenvif_tx_submit() 1194 skb_shinfo(nskb)->nr_frags = 0; in xenvif_tx_submit() 1253 mss = skb_shinfo(skb)->gso_size; in xenvif_tx_submit() [all …]
|
/openbmc/linux/net/ipv6/ |
H A D | udp_offload.c | 33 if (skb->encapsulation && skb_shinfo(skb)->gso_type & in udp6_ufo_fragment() 40 if (!(skb_shinfo(skb)->gso_type & (SKB_GSO_UDP | SKB_GSO_UDP_L4))) in udp6_ufo_fragment() 46 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) in udp6_ufo_fragment() 49 mss = skb_shinfo(skb)->gso_size; in udp6_ufo_fragment() 175 skb_shinfo(skb)->gso_type |= (SKB_GSO_FRAGLIST|SKB_GSO_UDP_L4); in udp6_gro_complete() 176 skb_shinfo(skb)->gso_segs = NAPI_GRO_CB(skb)->count; in udp6_gro_complete()
|
H A D | ip6_offload.c | 144 skb_shinfo(skb)->gso_type & (SKB_GSO_IPXIP4 | SKB_GSO_IPXIP6)) in ipv6_gso_segment() 146 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP); in ipv6_gso_segment() 149 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP); in ipv6_gso_segment() 162 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 167 payload_len = skb_shinfo(skb)->gso_size + in ipv6_gso_segment() 408 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP4; in sit_gro_complete() 415 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP6; in ip6ip6_gro_complete() 422 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP6; in ip4ip6_gro_complete() 438 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_IPXIP4)) in sit_gso_segment() 447 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_IPXIP6)) in ip4ip6_gso_segment() [all …]
|
H A D | tcpv6_offload.c | 37 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV6; in tcp6_gro_complete() 48 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6)) in tcp6_gso_segment()
|
/openbmc/linux/include/linux/ |
H A D | udp.h | 137 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) { in udp_cmsg_recv() 138 gso_size = skb_shinfo(skb)->gso_size; in udp_cmsg_recv() 166 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4 && in udp_unexpected_gso() 170 if (skb_shinfo(skb)->gso_type & SKB_GSO_FRAGLIST && in udp_unexpected_gso() 180 !(skb_shinfo(skb)->gso_type & in udp_unexpected_gso()
|
H A D | skbuff.h | 1660 return &skb_shinfo(skb)->hwtstamps; in skb_hwtstamps() 1693 skb_shinfo(skb)->destructor_arg = uarg; in skb_zcopy_init() 1694 skb_shinfo(skb)->flags |= uarg->flags; in skb_zcopy_init() 2492 skb_shinfo(skb)->nr_frags = i + 1; in skb_fill_page_desc() 3958 skb_shinfo(skb)->frag_list = NULL; in skb_frag_list_init() 4252 return skb_shinfo(skb)->meta_len; in skb_metadata_len() 4311 skb_shinfo(skb)->meta_len = meta_len; in skb_metadata_set() 4898 return skb_shinfo(skb)->gso_size; in skb_is_gso() 4921 skb_shinfo(skb)->gso_size = 0; in skb_gso_reset() 4922 skb_shinfo(skb)->gso_segs = 0; in skb_gso_reset() [all …]
|
/openbmc/linux/net/openvswitch/ |
H A D | openvswitch_trace.h | 52 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 53 __entry->gso_size = skb_shinfo(skb)->gso_size; 54 __entry->gso_type = skb_shinfo(skb)->gso_type; 122 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 123 __entry->gso_size = skb_shinfo(skb)->gso_size; 124 __entry->gso_type = skb_shinfo(skb)->gso_type;
|
/openbmc/linux/drivers/net/ethernet/hisilicon/hns3/ |
H A D | hns3_trace.h | 35 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 36 __entry->gso_size = skb_shinfo(skb)->gso_size; 37 __entry->gso_segs = skb_shinfo(skb)->gso_segs; 38 __entry->gso_type = skb_shinfo(skb)->gso_type; 43 hns3_shinfo_pack(skb_shinfo(skb), __entry->size);
|
/openbmc/linux/include/trace/events/ |
H A D | net.h | 56 __entry->tx_flags = skb_shinfo(skb)->tx_flags; 57 __entry->gso_size = skb_shinfo(skb)->gso_size; 58 __entry->gso_segs = skb_shinfo(skb)->gso_segs; 59 __entry->gso_type = skb_shinfo(skb)->gso_type; 214 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 215 __entry->gso_size = skb_shinfo(skb)->gso_size; 216 __entry->gso_type = skb_shinfo(skb)->gso_type;
|
/openbmc/linux/net/tls/ |
H A D | tls_strp.c | 31 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_anchor_free() 52 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in tls_strp_skb_copy() 53 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in tls_strp_skb_copy() 146 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_msg_hold() 193 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_flush_anchor_copy() 217 frag = &skb_shinfo(skb)->frags[skb->len / PAGE_SIZE]; in tls_strp_copyin_frag() 298 shinfo = skb_shinfo(skb); in tls_strp_copyin_skb() 397 shinfo = skb_shinfo(strp->anchor); in tls_strp_read_copy() 432 first = skb_shinfo(strp->anchor)->frag_list; in tls_strp_check_queue_ok() 468 skb_shinfo(strp->anchor)->frag_list = first; in tls_strp_load_anchor_with_queue()
|
/openbmc/linux/include/net/ |
H A D | tso.h | 23 return skb_shinfo(skb)->gso_segs * 2 + skb_shinfo(skb)->nr_frags; in tso_count_descs()
|
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 96 frag = &skb_shinfo(skb)->frags[seg - 1]; in otx2_dma_map_skb_frag() 204 if (likely(!skb_shinfo(skb)->nr_frags)) { in otx2_skb_add_frag() 666 if (skb_shinfo(skb)->gso_size) { in otx2_sqe_add_ext() 669 ext->lso_mps = skb_shinfo(skb)->gso_size; in otx2_sqe_add_ext() 802 int num_segs = skb_shinfo(skb)->nr_frags + 1; in otx2_dma_map_tso_skb() 842 frag = &skb_shinfo(skb)->frags[seg]; in otx2_tso_frag_dma_addr() 992 if (!skb_shinfo(skb)->gso_size) in otx2_get_sqe_count() 1000 return skb_shinfo(skb)->gso_segs; in otx2_get_sqe_count() 1090 if (unlikely(!skb_shinfo(skb)->gso_size && in otx2_set_txtstamp() 1159 num_segs = skb_shinfo(skb)->nr_frags + 1; in otx2_sq_append_skb() [all …]
|
/openbmc/linux/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cxgb4_ptp.h | 57 return skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP; in cxgb4_xmit_with_hwtstamp() 62 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in cxgb4_xmit_hwtstamp_pending()
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_tx.c | 156 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) in mlx5e_tx_get_gso_ihs() 164 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) { in mlx5e_tx_get_gso_ihs() 204 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mlx5e_txwqe_build_dsegs() 205 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mlx5e_txwqe_build_dsegs() 278 .mss = cpu_to_be16(skb_shinfo(skb)->gso_size), in mlx5e_sq_xmit_prepare() 285 stats->packets += skb_shinfo(skb)->gso_segs; in mlx5e_sq_xmit_prepare() 342 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) in mlx5e_tx_skb_update_hwts_flags() 343 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in mlx5e_tx_skb_update_hwts_flags() 402 (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP))) { in mlx5e_txwqe_complete() 661 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) in mlx5e_cqe_ts_id_eseg() [all …]
|
/openbmc/linux/drivers/staging/octeon/ |
H A D | ethernet-tx.c | 189 if (unlikely(skb_shinfo(skb)->nr_frags > 5)) { in cvm_oct_xmit() 261 if (skb_shinfo(skb)->nr_frags == 0) { in cvm_oct_xmit() 270 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in cvm_oct_xmit() 271 skb_frag_t *fs = skb_shinfo(skb)->frags + i; in cvm_oct_xmit() 280 hw_buffer.s.size = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit() 281 pko_command.s.segs = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit() 323 if (unlikely(skb_shinfo(skb)->nr_frags)) { in cvm_oct_xmit()
|