Lines Matching +full:xo +full:- +full:1

1 // SPDX-License-Identifier: GPL-2.0-only
34 struct xfrm_offload *xo; in esp4_gro_receive() local
45 xo = xfrm_offload(skb); in esp4_gro_receive()
46 if (!xo || !(xo->flags & CRYPTO_DONE)) { in esp4_gro_receive()
52 if (sp->len == XFRM_MAX_DEPTH) in esp4_gro_receive()
55 x = xfrm_state_lookup(dev_net(skb->dev), skb->mark, in esp4_gro_receive()
56 (xfrm_address_t *)&ip_hdr(skb)->daddr, in esp4_gro_receive()
61 skb->mark = xfrm_smark_get(skb->mark, x); in esp4_gro_receive()
63 sp->xvec[sp->len++] = x; in esp4_gro_receive()
64 sp->olen++; in esp4_gro_receive()
66 xo = xfrm_offload(skb); in esp4_gro_receive()
67 if (!xo) in esp4_gro_receive()
71 xo->flags |= XFRM_GRO; in esp4_gro_receive()
73 XFRM_TUNNEL_SKB_CB(skb)->tunnel.ip4 = NULL; in esp4_gro_receive()
74 XFRM_SPI_SKB_CB(skb)->family = AF_INET; in esp4_gro_receive()
75 XFRM_SPI_SKB_CB(skb)->daddroff = offsetof(struct iphdr, daddr); in esp4_gro_receive()
76 XFRM_SPI_SKB_CB(skb)->seq = seq; in esp4_gro_receive()
80 xfrm_input(skb, IPPROTO_ESP, spi, -2); in esp4_gro_receive()
82 return ERR_PTR(-EINPROGRESS); in esp4_gro_receive()
87 NAPI_GRO_CB(skb)->same_flow = 0; in esp4_gro_receive()
88 NAPI_GRO_CB(skb)->flush = 1; in esp4_gro_receive()
97 struct xfrm_offload *xo = xfrm_offload(skb); in esp4_gso_encap() local
98 int proto = iph->protocol; in esp4_gso_encap()
100 skb_push(skb, -skb_network_offset(skb)); in esp4_gso_encap()
104 esph->spi = x->id.spi; in esp4_gso_encap()
105 esph->seq_no = htonl(XFRM_SKB_CB(skb)->seq.output.low); in esp4_gso_encap()
107 xo->proto = proto; in esp4_gso_encap()
114 __be16 type = x->inner_mode.family == AF_INET6 ? htons(ETH_P_IPV6) in xfrm4_tunnel_gso_segment()
125 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment()
126 struct xfrm_offload *xo = xfrm_offload(skb); in xfrm4_transport_gso_segment() local
128 skb->transport_header += x->props.header_len; in xfrm4_transport_gso_segment()
129 ops = rcu_dereference(inet_offloads[xo->proto]); in xfrm4_transport_gso_segment()
130 if (likely(ops && ops->callbacks.gso_segment)) in xfrm4_transport_gso_segment()
131 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment()
140 struct xfrm_offload *xo = xfrm_offload(skb); in xfrm4_beet_gso_segment() local
141 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment()
143 u8 proto = xo->proto; in xfrm4_beet_gso_segment()
145 skb->transport_header += x->props.header_len; in xfrm4_beet_gso_segment()
147 if (x->sel.family != AF_INET6) { in xfrm4_beet_gso_segment()
150 (struct ip_beet_phdr *)skb->data; in xfrm4_beet_gso_segment()
152 skb->transport_header += ph->hdrlen * 8; in xfrm4_beet_gso_segment()
153 proto = ph->nexthdr; in xfrm4_beet_gso_segment()
155 skb->transport_header -= IPV4_BEET_PHMAXLEN; in xfrm4_beet_gso_segment()
160 skb->transport_header += in xfrm4_beet_gso_segment()
163 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV4; in xfrm4_beet_gso_segment()
167 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP4; in xfrm4_beet_gso_segment()
171 if (likely(ops && ops->callbacks.gso_segment)) in xfrm4_beet_gso_segment()
172 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment()
181 switch (x->outer_mode.encap) { in xfrm4_outer_mode_gso_segment()
190 return ERR_PTR(-EOPNOTSUPP); in xfrm4_outer_mode_gso_segment()
200 struct xfrm_offload *xo = xfrm_offload(skb); in esp4_gso_segment() local
203 if (!xo) in esp4_gso_segment()
204 return ERR_PTR(-EINVAL); in esp4_gso_segment()
206 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_ESP)) in esp4_gso_segment()
207 return ERR_PTR(-EINVAL); in esp4_gso_segment()
210 x = sp->xvec[sp->len - 1]; in esp4_gso_segment()
211 aead = x->data; in esp4_gso_segment()
214 if (esph->spi != x->id.spi) in esp4_gso_segment()
215 return ERR_PTR(-EINVAL); in esp4_gso_segment()
218 return ERR_PTR(-EINVAL); in esp4_gso_segment()
222 skb->encap_hdr_csum = 1; in esp4_gso_segment()
224 if ((!(skb->dev->gso_partial_features & NETIF_F_HW_ESP) && in esp4_gso_segment()
225 !(features & NETIF_F_HW_ESP)) || x->xso.dev != skb->dev) in esp4_gso_segment()
229 !(skb->dev->gso_partial_features & NETIF_F_HW_ESP_TX_CSUM)) in esp4_gso_segment()
233 xo->flags |= XFRM_GSO_SEGMENT; in esp4_gso_segment()
240 struct crypto_aead *aead = x->data; in esp_input_tail()
241 struct xfrm_offload *xo = xfrm_offload(skb); in esp_input_tail() local
244 return -EINVAL; in esp_input_tail()
246 if (!(xo->flags & CRYPTO_DONE)) in esp_input_tail()
247 skb->ip_summed = CHECKSUM_NONE; in esp_input_tail()
257 struct xfrm_offload *xo; in esp_xmit() local
266 xo = xfrm_offload(skb); in esp_xmit()
268 if (!xo) in esp_xmit()
269 return -EINVAL; in esp_xmit()
272 !(skb->dev->gso_partial_features & NETIF_F_HW_ESP)) || in esp_xmit()
273 x->xso.dev != skb->dev) { in esp_xmit()
274 xo->flags |= CRYPTO_FALLBACK; in esp_xmit()
278 esp.proto = xo->proto; in esp_xmit()
282 aead = x->data; in esp_xmit()
289 esp.clen = ALIGN(skb->len + 2 + esp.tfclen, blksize); in esp_xmit()
290 esp.plen = esp.clen - skb->len - esp.tfclen; in esp_xmit()
302 seq = xo->seq.low; in esp_xmit()
305 esph->spi = x->id.spi; in esp_xmit()
307 skb_push(skb, -skb_network_offset(skb)); in esp_xmit()
309 if (xo->flags & XFRM_GSO_SEGMENT) { in esp_xmit()
310 esph->seq_no = htonl(seq); in esp_xmit()
313 xo->seq.low++; in esp_xmit()
315 xo->seq.low += skb_shinfo(skb)->gso_segs; in esp_xmit()
318 if (xo->seq.low < seq) in esp_xmit()
319 xo->seq.hi++; in esp_xmit()
321 esp.seqno = cpu_to_be64(seq + ((u64)xo->seq.hi << 32)); in esp_xmit()
323 ip_hdr(skb)->tot_len = htons(skb->len); in esp_xmit()
328 return -ENOMEM; in esp_xmit()
330 xo = xfrm_offload(skb); in esp_xmit()
331 if (!xo) in esp_xmit()
332 return -EINVAL; in esp_xmit()
334 xo->flags |= XFRM_XMIT; in esp_xmit()
344 if (skb_needs_linearize(skb, skb->dev->features) && in esp_xmit()
346 return -ENOMEM; in esp_xmit()
369 return -EAGAIN; in esp4_offload_init()