xref: /openbmc/linux/net/ipv6/tcp_ipv6.c (revision cfc905f1)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *	TCP over IPv6
4  *	Linux INET6 implementation
5  *
6  *	Authors:
7  *	Pedro Roque		<roque@di.fc.ul.pt>
8  *
9  *	Based on:
10  *	linux/net/ipv4/tcp.c
11  *	linux/net/ipv4/tcp_input.c
12  *	linux/net/ipv4/tcp_output.c
13  *
14  *	Fixes:
15  *	Hideaki YOSHIFUJI	:	sin6_scope_id support
16  *	YOSHIFUJI Hideaki @USAGI and:	Support IPV6_V6ONLY socket option, which
17  *	Alexey Kuznetsov		allow both IPv4 and IPv6 sockets to bind
18  *					a single port at the same time.
19  *	YOSHIFUJI Hideaki @USAGI:	convert /proc/net/tcp6 to seq_file.
20  */
21 
22 #include <linux/bottom_half.h>
23 #include <linux/module.h>
24 #include <linux/errno.h>
25 #include <linux/types.h>
26 #include <linux/socket.h>
27 #include <linux/sockios.h>
28 #include <linux/net.h>
29 #include <linux/jiffies.h>
30 #include <linux/in.h>
31 #include <linux/in6.h>
32 #include <linux/netdevice.h>
33 #include <linux/init.h>
34 #include <linux/jhash.h>
35 #include <linux/ipsec.h>
36 #include <linux/times.h>
37 #include <linux/slab.h>
38 #include <linux/uaccess.h>
39 #include <linux/ipv6.h>
40 #include <linux/icmpv6.h>
41 #include <linux/random.h>
42 #include <linux/indirect_call_wrapper.h>
43 
44 #include <net/tcp.h>
45 #include <net/ndisc.h>
46 #include <net/inet6_hashtables.h>
47 #include <net/inet6_connection_sock.h>
48 #include <net/ipv6.h>
49 #include <net/transp_v6.h>
50 #include <net/addrconf.h>
51 #include <net/ip6_route.h>
52 #include <net/ip6_checksum.h>
53 #include <net/inet_ecn.h>
54 #include <net/protocol.h>
55 #include <net/xfrm.h>
56 #include <net/snmp.h>
57 #include <net/dsfield.h>
58 #include <net/timewait_sock.h>
59 #include <net/inet_common.h>
60 #include <net/secure_seq.h>
61 #include <net/busy_poll.h>
62 
63 #include <linux/proc_fs.h>
64 #include <linux/seq_file.h>
65 
66 #include <crypto/hash.h>
67 #include <linux/scatterlist.h>
68 
69 #include <trace/events/tcp.h>
70 
71 static void	tcp_v6_send_reset(const struct sock *sk, struct sk_buff *skb);
72 static void	tcp_v6_reqsk_send_ack(const struct sock *sk, struct sk_buff *skb,
73 				      struct request_sock *req);
74 
75 static int	tcp_v6_do_rcv(struct sock *sk, struct sk_buff *skb);
76 
77 static const struct inet_connection_sock_af_ops ipv6_mapped;
78 const struct inet_connection_sock_af_ops ipv6_specific;
79 #ifdef CONFIG_TCP_MD5SIG
80 static const struct tcp_sock_af_ops tcp_sock_ipv6_specific;
81 static const struct tcp_sock_af_ops tcp_sock_ipv6_mapped_specific;
82 #else
83 static struct tcp_md5sig_key *tcp_v6_md5_do_lookup(const struct sock *sk,
84 						   const struct in6_addr *addr,
85 						   int l3index)
86 {
87 	return NULL;
88 }
89 #endif
90 
91 /* Helper returning the inet6 address from a given tcp socket.
92  * It can be used in TCP stack instead of inet6_sk(sk).
93  * This avoids a dereference and allow compiler optimizations.
94  * It is a specialized version of inet6_sk_generic().
95  */
96 static struct ipv6_pinfo *tcp_inet6_sk(const struct sock *sk)
97 {
98 	unsigned int offset = sizeof(struct tcp6_sock) - sizeof(struct ipv6_pinfo);
99 
100 	return (struct ipv6_pinfo *)(((u8 *)sk) + offset);
101 }
102 
103 static void inet6_sk_rx_dst_set(struct sock *sk, const struct sk_buff *skb)
104 {
105 	struct dst_entry *dst = skb_dst(skb);
106 
107 	if (dst && dst_hold_safe(dst)) {
108 		const struct rt6_info *rt = (const struct rt6_info *)dst;
109 
110 		sk->sk_rx_dst = dst;
111 		inet_sk(sk)->rx_dst_ifindex = skb->skb_iif;
112 		tcp_inet6_sk(sk)->rx_dst_cookie = rt6_get_cookie(rt);
113 	}
114 }
115 
116 static u32 tcp_v6_init_seq(const struct sk_buff *skb)
117 {
118 	return secure_tcpv6_seq(ipv6_hdr(skb)->daddr.s6_addr32,
119 				ipv6_hdr(skb)->saddr.s6_addr32,
120 				tcp_hdr(skb)->dest,
121 				tcp_hdr(skb)->source);
122 }
123 
124 static u32 tcp_v6_init_ts_off(const struct net *net, const struct sk_buff *skb)
125 {
126 	return secure_tcpv6_ts_off(net, ipv6_hdr(skb)->daddr.s6_addr32,
127 				   ipv6_hdr(skb)->saddr.s6_addr32);
128 }
129 
130 static int tcp_v6_pre_connect(struct sock *sk, struct sockaddr *uaddr,
131 			      int addr_len)
132 {
133 	/* This check is replicated from tcp_v6_connect() and intended to
134 	 * prevent BPF program called below from accessing bytes that are out
135 	 * of the bound specified by user in addr_len.
136 	 */
137 	if (addr_len < SIN6_LEN_RFC2133)
138 		return -EINVAL;
139 
140 	sock_owned_by_me(sk);
141 
142 	return BPF_CGROUP_RUN_PROG_INET6_CONNECT(sk, uaddr);
143 }
144 
145 static int tcp_v6_connect(struct sock *sk, struct sockaddr *uaddr,
146 			  int addr_len)
147 {
148 	struct sockaddr_in6 *usin = (struct sockaddr_in6 *) uaddr;
149 	struct inet_sock *inet = inet_sk(sk);
150 	struct inet_connection_sock *icsk = inet_csk(sk);
151 	struct ipv6_pinfo *np = tcp_inet6_sk(sk);
152 	struct tcp_sock *tp = tcp_sk(sk);
153 	struct in6_addr *saddr = NULL, *final_p, final;
154 	struct ipv6_txoptions *opt;
155 	struct flowi6 fl6;
156 	struct dst_entry *dst;
157 	int addr_type;
158 	int err;
159 	struct inet_timewait_death_row *tcp_death_row = &sock_net(sk)->ipv4.tcp_death_row;
160 
161 	if (addr_len < SIN6_LEN_RFC2133)
162 		return -EINVAL;
163 
164 	if (usin->sin6_family != AF_INET6)
165 		return -EAFNOSUPPORT;
166 
167 	memset(&fl6, 0, sizeof(fl6));
168 
169 	if (np->sndflow) {
170 		fl6.flowlabel = usin->sin6_flowinfo&IPV6_FLOWINFO_MASK;
171 		IP6_ECN_flow_init(fl6.flowlabel);
172 		if (fl6.flowlabel&IPV6_FLOWLABEL_MASK) {
173 			struct ip6_flowlabel *flowlabel;
174 			flowlabel = fl6_sock_lookup(sk, fl6.flowlabel);
175 			if (IS_ERR(flowlabel))
176 				return -EINVAL;
177 			fl6_sock_release(flowlabel);
178 		}
179 	}
180 
181 	/*
182 	 *	connect() to INADDR_ANY means loopback (BSD'ism).
183 	 */
184 
185 	if (ipv6_addr_any(&usin->sin6_addr)) {
186 		if (ipv6_addr_v4mapped(&sk->sk_v6_rcv_saddr))
187 			ipv6_addr_set_v4mapped(htonl(INADDR_LOOPBACK),
188 					       &usin->sin6_addr);
189 		else
190 			usin->sin6_addr = in6addr_loopback;
191 	}
192 
193 	addr_type = ipv6_addr_type(&usin->sin6_addr);
194 
195 	if (addr_type & IPV6_ADDR_MULTICAST)
196 		return -ENETUNREACH;
197 
198 	if (addr_type&IPV6_ADDR_LINKLOCAL) {
199 		if (addr_len >= sizeof(struct sockaddr_in6) &&
200 		    usin->sin6_scope_id) {
201 			/* If interface is set while binding, indices
202 			 * must coincide.
203 			 */
204 			if (!sk_dev_equal_l3scope(sk, usin->sin6_scope_id))
205 				return -EINVAL;
206 
207 			sk->sk_bound_dev_if = usin->sin6_scope_id;
208 		}
209 
210 		/* Connect to link-local address requires an interface */
211 		if (!sk->sk_bound_dev_if)
212 			return -EINVAL;
213 	}
214 
215 	if (tp->rx_opt.ts_recent_stamp &&
216 	    !ipv6_addr_equal(&sk->sk_v6_daddr, &usin->sin6_addr)) {
217 		tp->rx_opt.ts_recent = 0;
218 		tp->rx_opt.ts_recent_stamp = 0;
219 		WRITE_ONCE(tp->write_seq, 0);
220 	}
221 
222 	sk->sk_v6_daddr = usin->sin6_addr;
223 	np->flow_label = fl6.flowlabel;
224 
225 	/*
226 	 *	TCP over IPv4
227 	 */
228 
229 	if (addr_type & IPV6_ADDR_MAPPED) {
230 		u32 exthdrlen = icsk->icsk_ext_hdr_len;
231 		struct sockaddr_in sin;
232 
233 		if (__ipv6_only_sock(sk))
234 			return -ENETUNREACH;
235 
236 		sin.sin_family = AF_INET;
237 		sin.sin_port = usin->sin6_port;
238 		sin.sin_addr.s_addr = usin->sin6_addr.s6_addr32[3];
239 
240 		icsk->icsk_af_ops = &ipv6_mapped;
241 		if (sk_is_mptcp(sk))
242 			mptcpv6_handle_mapped(sk, true);
243 		sk->sk_backlog_rcv = tcp_v4_do_rcv;
244 #ifdef CONFIG_TCP_MD5SIG
245 		tp->af_specific = &tcp_sock_ipv6_mapped_specific;
246 #endif
247 
248 		err = tcp_v4_connect(sk, (struct sockaddr *)&sin, sizeof(sin));
249 
250 		if (err) {
251 			icsk->icsk_ext_hdr_len = exthdrlen;
252 			icsk->icsk_af_ops = &ipv6_specific;
253 			if (sk_is_mptcp(sk))
254 				mptcpv6_handle_mapped(sk, false);
255 			sk->sk_backlog_rcv = tcp_v6_do_rcv;
256 #ifdef CONFIG_TCP_MD5SIG
257 			tp->af_specific = &tcp_sock_ipv6_specific;
258 #endif
259 			goto failure;
260 		}
261 		np->saddr = sk->sk_v6_rcv_saddr;
262 
263 		return err;
264 	}
265 
266 	if (!ipv6_addr_any(&sk->sk_v6_rcv_saddr))
267 		saddr = &sk->sk_v6_rcv_saddr;
268 
269 	fl6.flowi6_proto = IPPROTO_TCP;
270 	fl6.daddr = sk->sk_v6_daddr;
271 	fl6.saddr = saddr ? *saddr : np->saddr;
272 	fl6.flowi6_oif = sk->sk_bound_dev_if;
273 	fl6.flowi6_mark = sk->sk_mark;
274 	fl6.fl6_dport = usin->sin6_port;
275 	fl6.fl6_sport = inet->inet_sport;
276 	fl6.flowi6_uid = sk->sk_uid;
277 
278 	opt = rcu_dereference_protected(np->opt, lockdep_sock_is_held(sk));
279 	final_p = fl6_update_dst(&fl6, opt, &final);
280 
281 	security_sk_classify_flow(sk, flowi6_to_flowi(&fl6));
282 
283 	dst = ip6_dst_lookup_flow(sock_net(sk), sk, &fl6, final_p);
284 	if (IS_ERR(dst)) {
285 		err = PTR_ERR(dst);
286 		goto failure;
287 	}
288 
289 	if (!saddr) {
290 		saddr = &fl6.saddr;
291 		sk->sk_v6_rcv_saddr = *saddr;
292 	}
293 
294 	/* set the source address */
295 	np->saddr = *saddr;
296 	inet->inet_rcv_saddr = LOOPBACK4_IPV6;
297 
298 	sk->sk_gso_type = SKB_GSO_TCPV6;
299 	ip6_dst_store(sk, dst, NULL, NULL);
300 
301 	icsk->icsk_ext_hdr_len = 0;
302 	if (opt)
303 		icsk->icsk_ext_hdr_len = opt->opt_flen +
304 					 opt->opt_nflen;
305 
306 	tp->rx_opt.mss_clamp = IPV6_MIN_MTU - sizeof(struct tcphdr) - sizeof(struct ipv6hdr);
307 
308 	inet->inet_dport = usin->sin6_port;
309 
310 	tcp_set_state(sk, TCP_SYN_SENT);
311 	err = inet6_hash_connect(tcp_death_row, sk);
312 	if (err)
313 		goto late_failure;
314 
315 	sk_set_txhash(sk);
316 
317 	if (likely(!tp->repair)) {
318 		if (!tp->write_seq)
319 			WRITE_ONCE(tp->write_seq,
320 				   secure_tcpv6_seq(np->saddr.s6_addr32,
321 						    sk->sk_v6_daddr.s6_addr32,
322 						    inet->inet_sport,
323 						    inet->inet_dport));
324 		tp->tsoffset = secure_tcpv6_ts_off(sock_net(sk),
325 						   np->saddr.s6_addr32,
326 						   sk->sk_v6_daddr.s6_addr32);
327 	}
328 
329 	if (tcp_fastopen_defer_connect(sk, &err))
330 		return err;
331 	if (err)
332 		goto late_failure;
333 
334 	err = tcp_connect(sk);
335 	if (err)
336 		goto late_failure;
337 
338 	return 0;
339 
340 late_failure:
341 	tcp_set_state(sk, TCP_CLOSE);
342 failure:
343 	inet->inet_dport = 0;
344 	sk->sk_route_caps = 0;
345 	return err;
346 }
347 
348 static void tcp_v6_mtu_reduced(struct sock *sk)
349 {
350 	struct dst_entry *dst;
351 
352 	if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE))
353 		return;
354 
355 	dst = inet6_csk_update_pmtu(sk, tcp_sk(sk)->mtu_info);
356 	if (!dst)
357 		return;
358 
359 	if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) {
360 		tcp_sync_mss(sk, dst_mtu(dst));
361 		tcp_simple_retransmit(sk);
362 	}
363 }
364 
365 static int tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
366 		u8 type, u8 code, int offset, __be32 info)
367 {
368 	const struct ipv6hdr *hdr = (const struct ipv6hdr *)skb->data;
369 	const struct tcphdr *th = (struct tcphdr *)(skb->data+offset);
370 	struct net *net = dev_net(skb->dev);
371 	struct request_sock *fastopen;
372 	struct ipv6_pinfo *np;
373 	struct tcp_sock *tp;
374 	__u32 seq, snd_una;
375 	struct sock *sk;
376 	bool fatal;
377 	int err;
378 
379 	sk = __inet6_lookup_established(net, &tcp_hashinfo,
380 					&hdr->daddr, th->dest,
381 					&hdr->saddr, ntohs(th->source),
382 					skb->dev->ifindex, inet6_sdif(skb));
383 
384 	if (!sk) {
385 		__ICMP6_INC_STATS(net, __in6_dev_get(skb->dev),
386 				  ICMP6_MIB_INERRORS);
387 		return -ENOENT;
388 	}
389 
390 	if (sk->sk_state == TCP_TIME_WAIT) {
391 		inet_twsk_put(inet_twsk(sk));
392 		return 0;
393 	}
394 	seq = ntohl(th->seq);
395 	fatal = icmpv6_err_convert(type, code, &err);
396 	if (sk->sk_state == TCP_NEW_SYN_RECV) {
397 		tcp_req_err(sk, seq, fatal);
398 		return 0;
399 	}
400 
401 	bh_lock_sock(sk);
402 	if (sock_owned_by_user(sk) && type != ICMPV6_PKT_TOOBIG)
403 		__NET_INC_STATS(net, LINUX_MIB_LOCKDROPPEDICMPS);
404 
405 	if (sk->sk_state == TCP_CLOSE)
406 		goto out;
407 
408 	if (ipv6_hdr(skb)->hop_limit < tcp_inet6_sk(sk)->min_hopcount) {
409 		__NET_INC_STATS(net, LINUX_MIB_TCPMINTTLDROP);
410 		goto out;
411 	}
412 
413 	tp = tcp_sk(sk);
414 	/* XXX (TFO) - tp->snd_una should be ISN (tcp_create_openreq_child() */
415 	fastopen = rcu_dereference(tp->fastopen_rsk);
416 	snd_una = fastopen ? tcp_rsk(fastopen)->snt_isn : tp->snd_una;
417 	if (sk->sk_state != TCP_LISTEN &&
418 	    !between(seq, snd_una, tp->snd_nxt)) {
419 		__NET_INC_STATS(net, LINUX_MIB_OUTOFWINDOWICMPS);
420 		goto out;
421 	}
422 
423 	np = tcp_inet6_sk(sk);
424 
425 	if (type == NDISC_REDIRECT) {
426 		if (!sock_owned_by_user(sk)) {
427 			struct dst_entry *dst = __sk_dst_check(sk, np->dst_cookie);
428 
429 			if (dst)
430 				dst->ops->redirect(dst, sk, skb);
431 		}
432 		goto out;
433 	}
434 
435 	if (type == ICMPV6_PKT_TOOBIG) {
436 		/* We are not interested in TCP_LISTEN and open_requests
437 		 * (SYN-ACKs send out by Linux are always <576bytes so
438 		 * they should go through unfragmented).
439 		 */
440 		if (sk->sk_state == TCP_LISTEN)
441 			goto out;
442 
443 		if (!ip6_sk_accept_pmtu(sk))
444 			goto out;
445 
446 		tp->mtu_info = ntohl(info);
447 		if (!sock_owned_by_user(sk))
448 			tcp_v6_mtu_reduced(sk);
449 		else if (!test_and_set_bit(TCP_MTU_REDUCED_DEFERRED,
450 					   &sk->sk_tsq_flags))
451 			sock_hold(sk);
452 		goto out;
453 	}
454 
455 
456 	/* Might be for an request_sock */
457 	switch (sk->sk_state) {
458 	case TCP_SYN_SENT:
459 	case TCP_SYN_RECV:
460 		/* Only in fast or simultaneous open. If a fast open socket is
461 		 * is already accepted it is treated as a connected one below.
462 		 */
463 		if (fastopen && !fastopen->sk)
464 			break;
465 
466 		ipv6_icmp_error(sk, skb, err, th->dest, ntohl(info), (u8 *)th);
467 
468 		if (!sock_owned_by_user(sk)) {
469 			sk->sk_err = err;
470 			sk->sk_error_report(sk);		/* Wake people up to see the error (see connect in sock.c) */
471 
472 			tcp_done(sk);
473 		} else
474 			sk->sk_err_soft = err;
475 		goto out;
476 	case TCP_LISTEN:
477 		break;
478 	default:
479 		/* check if this ICMP message allows revert of backoff.
480 		 * (see RFC 6069)
481 		 */
482 		if (!fastopen && type == ICMPV6_DEST_UNREACH &&
483 		    code == ICMPV6_NOROUTE)
484 			tcp_ld_RTO_revert(sk, seq);
485 	}
486 
487 	if (!sock_owned_by_user(sk) && np->recverr) {
488 		sk->sk_err = err;
489 		sk->sk_error_report(sk);
490 	} else
491 		sk->sk_err_soft = err;
492 
493 out:
494 	bh_unlock_sock(sk);
495 	sock_put(sk);
496 	return 0;
497 }
498 
499 
500 static int tcp_v6_send_synack(const struct sock *sk, struct dst_entry *dst,
501 			      struct flowi *fl,
502 			      struct request_sock *req,
503 			      struct tcp_fastopen_cookie *foc,
504 			      enum tcp_synack_type synack_type)
505 {
506 	struct inet_request_sock *ireq = inet_rsk(req);
507 	struct ipv6_pinfo *np = tcp_inet6_sk(sk);
508 	struct ipv6_txoptions *opt;
509 	struct flowi6 *fl6 = &fl->u.ip6;
510 	struct sk_buff *skb;
511 	int err = -ENOMEM;
512 
513 	/* First, grab a route. */
514 	if (!dst && (dst = inet6_csk_route_req(sk, fl6, req,
515 					       IPPROTO_TCP)) == NULL)
516 		goto done;
517 
518 	skb = tcp_make_synack(sk, dst, req, foc, synack_type);
519 
520 	if (skb) {
521 		__tcp_v6_send_check(skb, &ireq->ir_v6_loc_addr,
522 				    &ireq->ir_v6_rmt_addr);
523 
524 		fl6->daddr = ireq->ir_v6_rmt_addr;
525 		if (np->repflow && ireq->pktopts)
526 			fl6->flowlabel = ip6_flowlabel(ipv6_hdr(ireq->pktopts));
527 
528 		rcu_read_lock();
529 		opt = ireq->ipv6_opt;
530 		if (!opt)
531 			opt = rcu_dereference(np->opt);
532 		err = ip6_xmit(sk, skb, fl6, sk->sk_mark, opt, np->tclass,
533 			       sk->sk_priority);
534 		rcu_read_unlock();
535 		err = net_xmit_eval(err);
536 	}
537 
538 done:
539 	return err;
540 }
541 
542 
543 static void tcp_v6_reqsk_destructor(struct request_sock *req)
544 {
545 	kfree(inet_rsk(req)->ipv6_opt);
546 	kfree_skb(inet_rsk(req)->pktopts);
547 }
548 
549 #ifdef CONFIG_TCP_MD5SIG
550 static struct tcp_md5sig_key *tcp_v6_md5_do_lookup(const struct sock *sk,
551 						   const struct in6_addr *addr,
552 						   int l3index)
553 {
554 	return tcp_md5_do_lookup(sk, l3index,
555 				 (union tcp_md5_addr *)addr, AF_INET6);
556 }
557 
558 static struct tcp_md5sig_key *tcp_v6_md5_lookup(const struct sock *sk,
559 						const struct sock *addr_sk)
560 {
561 	int l3index;
562 
563 	l3index = l3mdev_master_ifindex_by_index(sock_net(sk),
564 						 addr_sk->sk_bound_dev_if);
565 	return tcp_v6_md5_do_lookup(sk, &addr_sk->sk_v6_daddr,
566 				    l3index);
567 }
568 
569 static int tcp_v6_parse_md5_keys(struct sock *sk, int optname,
570 				 sockptr_t optval, int optlen)
571 {
572 	struct tcp_md5sig cmd;
573 	struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&cmd.tcpm_addr;
574 	int l3index = 0;
575 	u8 prefixlen;
576 
577 	if (optlen < sizeof(cmd))
578 		return -EINVAL;
579 
580 	if (copy_from_sockptr(&cmd, optval, sizeof(cmd)))
581 		return -EFAULT;
582 
583 	if (sin6->sin6_family != AF_INET6)
584 		return -EINVAL;
585 
586 	if (optname == TCP_MD5SIG_EXT &&
587 	    cmd.tcpm_flags & TCP_MD5SIG_FLAG_PREFIX) {
588 		prefixlen = cmd.tcpm_prefixlen;
589 		if (prefixlen > 128 || (ipv6_addr_v4mapped(&sin6->sin6_addr) &&
590 					prefixlen > 32))
591 			return -EINVAL;
592 	} else {
593 		prefixlen = ipv6_addr_v4mapped(&sin6->sin6_addr) ? 32 : 128;
594 	}
595 
596 	if (optname == TCP_MD5SIG_EXT &&
597 	    cmd.tcpm_flags & TCP_MD5SIG_FLAG_IFINDEX) {
598 		struct net_device *dev;
599 
600 		rcu_read_lock();
601 		dev = dev_get_by_index_rcu(sock_net(sk), cmd.tcpm_ifindex);
602 		if (dev && netif_is_l3_master(dev))
603 			l3index = dev->ifindex;
604 		rcu_read_unlock();
605 
606 		/* ok to reference set/not set outside of rcu;
607 		 * right now device MUST be an L3 master
608 		 */
609 		if (!dev || !l3index)
610 			return -EINVAL;
611 	}
612 
613 	if (!cmd.tcpm_keylen) {
614 		if (ipv6_addr_v4mapped(&sin6->sin6_addr))
615 			return tcp_md5_do_del(sk, (union tcp_md5_addr *)&sin6->sin6_addr.s6_addr32[3],
616 					      AF_INET, prefixlen,
617 					      l3index);
618 		return tcp_md5_do_del(sk, (union tcp_md5_addr *)&sin6->sin6_addr,
619 				      AF_INET6, prefixlen, l3index);
620 	}
621 
622 	if (cmd.tcpm_keylen > TCP_MD5SIG_MAXKEYLEN)
623 		return -EINVAL;
624 
625 	if (ipv6_addr_v4mapped(&sin6->sin6_addr))
626 		return tcp_md5_do_add(sk, (union tcp_md5_addr *)&sin6->sin6_addr.s6_addr32[3],
627 				      AF_INET, prefixlen, l3index,
628 				      cmd.tcpm_key, cmd.tcpm_keylen,
629 				      GFP_KERNEL);
630 
631 	return tcp_md5_do_add(sk, (union tcp_md5_addr *)&sin6->sin6_addr,
632 			      AF_INET6, prefixlen, l3index,
633 			      cmd.tcpm_key, cmd.tcpm_keylen, GFP_KERNEL);
634 }
635 
636 static int tcp_v6_md5_hash_headers(struct tcp_md5sig_pool *hp,
637 				   const struct in6_addr *daddr,
638 				   const struct in6_addr *saddr,
639 				   const struct tcphdr *th, int nbytes)
640 {
641 	struct tcp6_pseudohdr *bp;
642 	struct scatterlist sg;
643 	struct tcphdr *_th;
644 
645 	bp = hp->scratch;
646 	/* 1. TCP pseudo-header (RFC2460) */
647 	bp->saddr = *saddr;
648 	bp->daddr = *daddr;
649 	bp->protocol = cpu_to_be32(IPPROTO_TCP);
650 	bp->len = cpu_to_be32(nbytes);
651 
652 	_th = (struct tcphdr *)(bp + 1);
653 	memcpy(_th, th, sizeof(*th));
654 	_th->check = 0;
655 
656 	sg_init_one(&sg, bp, sizeof(*bp) + sizeof(*th));
657 	ahash_request_set_crypt(hp->md5_req, &sg, NULL,
658 				sizeof(*bp) + sizeof(*th));
659 	return crypto_ahash_update(hp->md5_req);
660 }
661 
662 static int tcp_v6_md5_hash_hdr(char *md5_hash, const struct tcp_md5sig_key *key,
663 			       const struct in6_addr *daddr, struct in6_addr *saddr,
664 			       const struct tcphdr *th)
665 {
666 	struct tcp_md5sig_pool *hp;
667 	struct ahash_request *req;
668 
669 	hp = tcp_get_md5sig_pool();
670 	if (!hp)
671 		goto clear_hash_noput;
672 	req = hp->md5_req;
673 
674 	if (crypto_ahash_init(req))
675 		goto clear_hash;
676 	if (tcp_v6_md5_hash_headers(hp, daddr, saddr, th, th->doff << 2))
677 		goto clear_hash;
678 	if (tcp_md5_hash_key(hp, key))
679 		goto clear_hash;
680 	ahash_request_set_crypt(req, NULL, md5_hash, 0);
681 	if (crypto_ahash_final(req))
682 		goto clear_hash;
683 
684 	tcp_put_md5sig_pool();
685 	return 0;
686 
687 clear_hash:
688 	tcp_put_md5sig_pool();
689 clear_hash_noput:
690 	memset(md5_hash, 0, 16);
691 	return 1;
692 }
693 
694 static int tcp_v6_md5_hash_skb(char *md5_hash,
695 			       const struct tcp_md5sig_key *key,
696 			       const struct sock *sk,
697 			       const struct sk_buff *skb)
698 {
699 	const struct in6_addr *saddr, *daddr;
700 	struct tcp_md5sig_pool *hp;
701 	struct ahash_request *req;
702 	const struct tcphdr *th = tcp_hdr(skb);
703 
704 	if (sk) { /* valid for establish/request sockets */
705 		saddr = &sk->sk_v6_rcv_saddr;
706 		daddr = &sk->sk_v6_daddr;
707 	} else {
708 		const struct ipv6hdr *ip6h = ipv6_hdr(skb);
709 		saddr = &ip6h->saddr;
710 		daddr = &ip6h->daddr;
711 	}
712 
713 	hp = tcp_get_md5sig_pool();
714 	if (!hp)
715 		goto clear_hash_noput;
716 	req = hp->md5_req;
717 
718 	if (crypto_ahash_init(req))
719 		goto clear_hash;
720 
721 	if (tcp_v6_md5_hash_headers(hp, daddr, saddr, th, skb->len))
722 		goto clear_hash;
723 	if (tcp_md5_hash_skb_data(hp, skb, th->doff << 2))
724 		goto clear_hash;
725 	if (tcp_md5_hash_key(hp, key))
726 		goto clear_hash;
727 	ahash_request_set_crypt(req, NULL, md5_hash, 0);
728 	if (crypto_ahash_final(req))
729 		goto clear_hash;
730 
731 	tcp_put_md5sig_pool();
732 	return 0;
733 
734 clear_hash:
735 	tcp_put_md5sig_pool();
736 clear_hash_noput:
737 	memset(md5_hash, 0, 16);
738 	return 1;
739 }
740 
741 #endif
742 
743 static bool tcp_v6_inbound_md5_hash(const struct sock *sk,
744 				    const struct sk_buff *skb,
745 				    int dif, int sdif)
746 {
747 #ifdef CONFIG_TCP_MD5SIG
748 	const __u8 *hash_location = NULL;
749 	struct tcp_md5sig_key *hash_expected;
750 	const struct ipv6hdr *ip6h = ipv6_hdr(skb);
751 	const struct tcphdr *th = tcp_hdr(skb);
752 	int genhash, l3index;
753 	u8 newhash[16];
754 
755 	/* sdif set, means packet ingressed via a device
756 	 * in an L3 domain and dif is set to the l3mdev
757 	 */
758 	l3index = sdif ? dif : 0;
759 
760 	hash_expected = tcp_v6_md5_do_lookup(sk, &ip6h->saddr, l3index);
761 	hash_location = tcp_parse_md5sig_option(th);
762 
763 	/* We've parsed the options - do we have a hash? */
764 	if (!hash_expected && !hash_location)
765 		return false;
766 
767 	if (hash_expected && !hash_location) {
768 		NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMD5NOTFOUND);
769 		return true;
770 	}
771 
772 	if (!hash_expected && hash_location) {
773 		NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMD5UNEXPECTED);
774 		return true;
775 	}
776 
777 	/* check the signature */
778 	genhash = tcp_v6_md5_hash_skb(newhash,
779 				      hash_expected,
780 				      NULL, skb);
781 
782 	if (genhash || memcmp(hash_location, newhash, 16) != 0) {
783 		NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMD5FAILURE);
784 		net_info_ratelimited("MD5 Hash %s for [%pI6c]:%u->[%pI6c]:%u L3 index %d\n",
785 				     genhash ? "failed" : "mismatch",
786 				     &ip6h->saddr, ntohs(th->source),
787 				     &ip6h->daddr, ntohs(th->dest), l3index);
788 		return true;
789 	}
790 #endif
791 	return false;
792 }
793 
794 static void tcp_v6_init_req(struct request_sock *req,
795 			    const struct sock *sk_listener,
796 			    struct sk_buff *skb)
797 {
798 	bool l3_slave = ipv6_l3mdev_skb(TCP_SKB_CB(skb)->header.h6.flags);
799 	struct inet_request_sock *ireq = inet_rsk(req);
800 	const struct ipv6_pinfo *np = tcp_inet6_sk(sk_listener);
801 
802 	ireq->ir_v6_rmt_addr = ipv6_hdr(skb)->saddr;
803 	ireq->ir_v6_loc_addr = ipv6_hdr(skb)->daddr;
804 
805 	/* So that link locals have meaning */
806 	if ((!sk_listener->sk_bound_dev_if || l3_slave) &&
807 	    ipv6_addr_type(&ireq->ir_v6_rmt_addr) & IPV6_ADDR_LINKLOCAL)
808 		ireq->ir_iif = tcp_v6_iif(skb);
809 
810 	if (!TCP_SKB_CB(skb)->tcp_tw_isn &&
811 	    (ipv6_opt_accepted(sk_listener, skb, &TCP_SKB_CB(skb)->header.h6) ||
812 	     np->rxopt.bits.rxinfo ||
813 	     np->rxopt.bits.rxoinfo || np->rxopt.bits.rxhlim ||
814 	     np->rxopt.bits.rxohlim || np->repflow)) {
815 		refcount_inc(&skb->users);
816 		ireq->pktopts = skb;
817 	}
818 }
819 
820 static struct dst_entry *tcp_v6_route_req(const struct sock *sk,
821 					  struct flowi *fl,
822 					  const struct request_sock *req)
823 {
824 	return inet6_csk_route_req(sk, &fl->u.ip6, req, IPPROTO_TCP);
825 }
826 
827 struct request_sock_ops tcp6_request_sock_ops __read_mostly = {
828 	.family		=	AF_INET6,
829 	.obj_size	=	sizeof(struct tcp6_request_sock),
830 	.rtx_syn_ack	=	tcp_rtx_synack,
831 	.send_ack	=	tcp_v6_reqsk_send_ack,
832 	.destructor	=	tcp_v6_reqsk_destructor,
833 	.send_reset	=	tcp_v6_send_reset,
834 	.syn_ack_timeout =	tcp_syn_ack_timeout,
835 };
836 
837 const struct tcp_request_sock_ops tcp_request_sock_ipv6_ops = {
838 	.mss_clamp	=	IPV6_MIN_MTU - sizeof(struct tcphdr) -
839 				sizeof(struct ipv6hdr),
840 #ifdef CONFIG_TCP_MD5SIG
841 	.req_md5_lookup	=	tcp_v6_md5_lookup,
842 	.calc_md5_hash	=	tcp_v6_md5_hash_skb,
843 #endif
844 	.init_req	=	tcp_v6_init_req,
845 #ifdef CONFIG_SYN_COOKIES
846 	.cookie_init_seq =	cookie_v6_init_sequence,
847 #endif
848 	.route_req	=	tcp_v6_route_req,
849 	.init_seq	=	tcp_v6_init_seq,
850 	.init_ts_off	=	tcp_v6_init_ts_off,
851 	.send_synack	=	tcp_v6_send_synack,
852 };
853 
854 static void tcp_v6_send_response(const struct sock *sk, struct sk_buff *skb, u32 seq,
855 				 u32 ack, u32 win, u32 tsval, u32 tsecr,
856 				 int oif, struct tcp_md5sig_key *key, int rst,
857 				 u8 tclass, __be32 label, u32 priority)
858 {
859 	const struct tcphdr *th = tcp_hdr(skb);
860 	struct tcphdr *t1;
861 	struct sk_buff *buff;
862 	struct flowi6 fl6;
863 	struct net *net = sk ? sock_net(sk) : dev_net(skb_dst(skb)->dev);
864 	struct sock *ctl_sk = net->ipv6.tcp_sk;
865 	unsigned int tot_len = sizeof(struct tcphdr);
866 	struct dst_entry *dst;
867 	__be32 *topt;
868 	__u32 mark = 0;
869 
870 	if (tsecr)
871 		tot_len += TCPOLEN_TSTAMP_ALIGNED;
872 #ifdef CONFIG_TCP_MD5SIG
873 	if (key)
874 		tot_len += TCPOLEN_MD5SIG_ALIGNED;
875 #endif
876 
877 	buff = alloc_skb(MAX_HEADER + sizeof(struct ipv6hdr) + tot_len,
878 			 GFP_ATOMIC);
879 	if (!buff)
880 		return;
881 
882 	skb_reserve(buff, MAX_HEADER + sizeof(struct ipv6hdr) + tot_len);
883 
884 	t1 = skb_push(buff, tot_len);
885 	skb_reset_transport_header(buff);
886 
887 	/* Swap the send and the receive. */
888 	memset(t1, 0, sizeof(*t1));
889 	t1->dest = th->source;
890 	t1->source = th->dest;
891 	t1->doff = tot_len / 4;
892 	t1->seq = htonl(seq);
893 	t1->ack_seq = htonl(ack);
894 	t1->ack = !rst || !th->ack;
895 	t1->rst = rst;
896 	t1->window = htons(win);
897 
898 	topt = (__be32 *)(t1 + 1);
899 
900 	if (tsecr) {
901 		*topt++ = htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) |
902 				(TCPOPT_TIMESTAMP << 8) | TCPOLEN_TIMESTAMP);
903 		*topt++ = htonl(tsval);
904 		*topt++ = htonl(tsecr);
905 	}
906 
907 #ifdef CONFIG_TCP_MD5SIG
908 	if (key) {
909 		*topt++ = htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) |
910 				(TCPOPT_MD5SIG << 8) | TCPOLEN_MD5SIG);
911 		tcp_v6_md5_hash_hdr((__u8 *)topt, key,
912 				    &ipv6_hdr(skb)->saddr,
913 				    &ipv6_hdr(skb)->daddr, t1);
914 	}
915 #endif
916 
917 	memset(&fl6, 0, sizeof(fl6));
918 	fl6.daddr = ipv6_hdr(skb)->saddr;
919 	fl6.saddr = ipv6_hdr(skb)->daddr;
920 	fl6.flowlabel = label;
921 
922 	buff->ip_summed = CHECKSUM_PARTIAL;
923 	buff->csum = 0;
924 
925 	__tcp_v6_send_check(buff, &fl6.saddr, &fl6.daddr);
926 
927 	fl6.flowi6_proto = IPPROTO_TCP;
928 	if (rt6_need_strict(&fl6.daddr) && !oif)
929 		fl6.flowi6_oif = tcp_v6_iif(skb);
930 	else {
931 		if (!oif && netif_index_is_l3_master(net, skb->skb_iif))
932 			oif = skb->skb_iif;
933 
934 		fl6.flowi6_oif = oif;
935 	}
936 
937 	if (sk) {
938 		if (sk->sk_state == TCP_TIME_WAIT) {
939 			mark = inet_twsk(sk)->tw_mark;
940 			/* autoflowlabel relies on buff->hash */
941 			skb_set_hash(buff, inet_twsk(sk)->tw_txhash,
942 				     PKT_HASH_TYPE_L4);
943 		} else {
944 			mark = sk->sk_mark;
945 		}
946 		buff->tstamp = tcp_transmit_time(sk);
947 	}
948 	fl6.flowi6_mark = IP6_REPLY_MARK(net, skb->mark) ?: mark;
949 	fl6.fl6_dport = t1->dest;
950 	fl6.fl6_sport = t1->source;
951 	fl6.flowi6_uid = sock_net_uid(net, sk && sk_fullsock(sk) ? sk : NULL);
952 	security_skb_classify_flow(skb, flowi6_to_flowi(&fl6));
953 
954 	/* Pass a socket to ip6_dst_lookup either it is for RST
955 	 * Underlying function will use this to retrieve the network
956 	 * namespace
957 	 */
958 	dst = ip6_dst_lookup_flow(sock_net(ctl_sk), ctl_sk, &fl6, NULL);
959 	if (!IS_ERR(dst)) {
960 		skb_dst_set(buff, dst);
961 		ip6_xmit(ctl_sk, buff, &fl6, fl6.flowi6_mark, NULL, tclass,
962 			 priority);
963 		TCP_INC_STATS(net, TCP_MIB_OUTSEGS);
964 		if (rst)
965 			TCP_INC_STATS(net, TCP_MIB_OUTRSTS);
966 		return;
967 	}
968 
969 	kfree_skb(buff);
970 }
971 
972 static void tcp_v6_send_reset(const struct sock *sk, struct sk_buff *skb)
973 {
974 	const struct tcphdr *th = tcp_hdr(skb);
975 	struct ipv6hdr *ipv6h = ipv6_hdr(skb);
976 	u32 seq = 0, ack_seq = 0;
977 	struct tcp_md5sig_key *key = NULL;
978 #ifdef CONFIG_TCP_MD5SIG
979 	const __u8 *hash_location = NULL;
980 	unsigned char newhash[16];
981 	int genhash;
982 	struct sock *sk1 = NULL;
983 #endif
984 	__be32 label = 0;
985 	u32 priority = 0;
986 	struct net *net;
987 	int oif = 0;
988 
989 	if (th->rst)
990 		return;
991 
992 	/* If sk not NULL, it means we did a successful lookup and incoming
993 	 * route had to be correct. prequeue might have dropped our dst.
994 	 */
995 	if (!sk && !ipv6_unicast_destination(skb))
996 		return;
997 
998 	net = sk ? sock_net(sk) : dev_net(skb_dst(skb)->dev);
999 #ifdef CONFIG_TCP_MD5SIG
1000 	rcu_read_lock();
1001 	hash_location = tcp_parse_md5sig_option(th);
1002 	if (sk && sk_fullsock(sk)) {
1003 		int l3index;
1004 
1005 		/* sdif set, means packet ingressed via a device
1006 		 * in an L3 domain and inet_iif is set to it.
1007 		 */
1008 		l3index = tcp_v6_sdif(skb) ? tcp_v6_iif_l3_slave(skb) : 0;
1009 		key = tcp_v6_md5_do_lookup(sk, &ipv6h->saddr, l3index);
1010 	} else if (hash_location) {
1011 		int dif = tcp_v6_iif_l3_slave(skb);
1012 		int sdif = tcp_v6_sdif(skb);
1013 		int l3index;
1014 
1015 		/*
1016 		 * active side is lost. Try to find listening socket through
1017 		 * source port, and then find md5 key through listening socket.
1018 		 * we are not loose security here:
1019 		 * Incoming packet is checked with md5 hash with finding key,
1020 		 * no RST generated if md5 hash doesn't match.
1021 		 */
1022 		sk1 = inet6_lookup_listener(net,
1023 					   &tcp_hashinfo, NULL, 0,
1024 					   &ipv6h->saddr,
1025 					   th->source, &ipv6h->daddr,
1026 					   ntohs(th->source), dif, sdif);
1027 		if (!sk1)
1028 			goto out;
1029 
1030 		/* sdif set, means packet ingressed via a device
1031 		 * in an L3 domain and dif is set to it.
1032 		 */
1033 		l3index = tcp_v6_sdif(skb) ? dif : 0;
1034 
1035 		key = tcp_v6_md5_do_lookup(sk1, &ipv6h->saddr, l3index);
1036 		if (!key)
1037 			goto out;
1038 
1039 		genhash = tcp_v6_md5_hash_skb(newhash, key, NULL, skb);
1040 		if (genhash || memcmp(hash_location, newhash, 16) != 0)
1041 			goto out;
1042 	}
1043 #endif
1044 
1045 	if (th->ack)
1046 		seq = ntohl(th->ack_seq);
1047 	else
1048 		ack_seq = ntohl(th->seq) + th->syn + th->fin + skb->len -
1049 			  (th->doff << 2);
1050 
1051 	if (sk) {
1052 		oif = sk->sk_bound_dev_if;
1053 		if (sk_fullsock(sk)) {
1054 			const struct ipv6_pinfo *np = tcp_inet6_sk(sk);
1055 
1056 			trace_tcp_send_reset(sk, skb);
1057 			if (np->repflow)
1058 				label = ip6_flowlabel(ipv6h);
1059 			priority = sk->sk_priority;
1060 		}
1061 		if (sk->sk_state == TCP_TIME_WAIT) {
1062 			label = cpu_to_be32(inet_twsk(sk)->tw_flowlabel);
1063 			priority = inet_twsk(sk)->tw_priority;
1064 		}
1065 	} else {
1066 		if (net->ipv6.sysctl.flowlabel_reflect & FLOWLABEL_REFLECT_TCP_RESET)
1067 			label = ip6_flowlabel(ipv6h);
1068 	}
1069 
1070 	tcp_v6_send_response(sk, skb, seq, ack_seq, 0, 0, 0, oif, key, 1, 0,
1071 			     label, priority);
1072 
1073 #ifdef CONFIG_TCP_MD5SIG
1074 out:
1075 	rcu_read_unlock();
1076 #endif
1077 }
1078 
1079 static void tcp_v6_send_ack(const struct sock *sk, struct sk_buff *skb, u32 seq,
1080 			    u32 ack, u32 win, u32 tsval, u32 tsecr, int oif,
1081 			    struct tcp_md5sig_key *key, u8 tclass,
1082 			    __be32 label, u32 priority)
1083 {
1084 	tcp_v6_send_response(sk, skb, seq, ack, win, tsval, tsecr, oif, key, 0,
1085 			     tclass, label, priority);
1086 }
1087 
1088 static void tcp_v6_timewait_ack(struct sock *sk, struct sk_buff *skb)
1089 {
1090 	struct inet_timewait_sock *tw = inet_twsk(sk);
1091 	struct tcp_timewait_sock *tcptw = tcp_twsk(sk);
1092 
1093 	tcp_v6_send_ack(sk, skb, tcptw->tw_snd_nxt, tcptw->tw_rcv_nxt,
1094 			tcptw->tw_rcv_wnd >> tw->tw_rcv_wscale,
1095 			tcp_time_stamp_raw() + tcptw->tw_ts_offset,
1096 			tcptw->tw_ts_recent, tw->tw_bound_dev_if, tcp_twsk_md5_key(tcptw),
1097 			tw->tw_tclass, cpu_to_be32(tw->tw_flowlabel), tw->tw_priority);
1098 
1099 	inet_twsk_put(tw);
1100 }
1101 
1102 static void tcp_v6_reqsk_send_ack(const struct sock *sk, struct sk_buff *skb,
1103 				  struct request_sock *req)
1104 {
1105 	int l3index;
1106 
1107 	l3index = tcp_v6_sdif(skb) ? tcp_v6_iif_l3_slave(skb) : 0;
1108 
1109 	/* sk->sk_state == TCP_LISTEN -> for regular TCP_SYN_RECV
1110 	 * sk->sk_state == TCP_SYN_RECV -> for Fast Open.
1111 	 */
1112 	/* RFC 7323 2.3
1113 	 * The window field (SEG.WND) of every outgoing segment, with the
1114 	 * exception of <SYN> segments, MUST be right-shifted by
1115 	 * Rcv.Wind.Shift bits:
1116 	 */
1117 	tcp_v6_send_ack(sk, skb, (sk->sk_state == TCP_LISTEN) ?
1118 			tcp_rsk(req)->snt_isn + 1 : tcp_sk(sk)->snd_nxt,
1119 			tcp_rsk(req)->rcv_nxt,
1120 			req->rsk_rcv_wnd >> inet_rsk(req)->rcv_wscale,
1121 			tcp_time_stamp_raw() + tcp_rsk(req)->ts_off,
1122 			req->ts_recent, sk->sk_bound_dev_if,
1123 			tcp_v6_md5_do_lookup(sk, &ipv6_hdr(skb)->saddr, l3index),
1124 			0, 0, sk->sk_priority);
1125 }
1126 
1127 
1128 static struct sock *tcp_v6_cookie_check(struct sock *sk, struct sk_buff *skb)
1129 {
1130 #ifdef CONFIG_SYN_COOKIES
1131 	const struct tcphdr *th = tcp_hdr(skb);
1132 
1133 	if (!th->syn)
1134 		sk = cookie_v6_check(sk, skb);
1135 #endif
1136 	return sk;
1137 }
1138 
1139 u16 tcp_v6_get_syncookie(struct sock *sk, struct ipv6hdr *iph,
1140 			 struct tcphdr *th, u32 *cookie)
1141 {
1142 	u16 mss = 0;
1143 #ifdef CONFIG_SYN_COOKIES
1144 	mss = tcp_get_syncookie_mss(&tcp6_request_sock_ops,
1145 				    &tcp_request_sock_ipv6_ops, sk, th);
1146 	if (mss) {
1147 		*cookie = __cookie_v6_init_sequence(iph, th, &mss);
1148 		tcp_synq_overflow(sk);
1149 	}
1150 #endif
1151 	return mss;
1152 }
1153 
1154 static int tcp_v6_conn_request(struct sock *sk, struct sk_buff *skb)
1155 {
1156 	if (skb->protocol == htons(ETH_P_IP))
1157 		return tcp_v4_conn_request(sk, skb);
1158 
1159 	if (!ipv6_unicast_destination(skb))
1160 		goto drop;
1161 
1162 	return tcp_conn_request(&tcp6_request_sock_ops,
1163 				&tcp_request_sock_ipv6_ops, sk, skb);
1164 
1165 drop:
1166 	tcp_listendrop(sk);
1167 	return 0; /* don't send reset */
1168 }
1169 
1170 static void tcp_v6_restore_cb(struct sk_buff *skb)
1171 {
1172 	/* We need to move header back to the beginning if xfrm6_policy_check()
1173 	 * and tcp_v6_fill_cb() are going to be called again.
1174 	 * ip6_datagram_recv_specific_ctl() also expects IP6CB to be there.
1175 	 */
1176 	memmove(IP6CB(skb), &TCP_SKB_CB(skb)->header.h6,
1177 		sizeof(struct inet6_skb_parm));
1178 }
1179 
1180 static struct sock *tcp_v6_syn_recv_sock(const struct sock *sk, struct sk_buff *skb,
1181 					 struct request_sock *req,
1182 					 struct dst_entry *dst,
1183 					 struct request_sock *req_unhash,
1184 					 bool *own_req)
1185 {
1186 	struct inet_request_sock *ireq;
1187 	struct ipv6_pinfo *newnp;
1188 	const struct ipv6_pinfo *np = tcp_inet6_sk(sk);
1189 	struct ipv6_txoptions *opt;
1190 	struct inet_sock *newinet;
1191 	struct tcp_sock *newtp;
1192 	struct sock *newsk;
1193 #ifdef CONFIG_TCP_MD5SIG
1194 	struct tcp_md5sig_key *key;
1195 	int l3index;
1196 #endif
1197 	struct flowi6 fl6;
1198 
1199 	if (skb->protocol == htons(ETH_P_IP)) {
1200 		/*
1201 		 *	v6 mapped
1202 		 */
1203 
1204 		newsk = tcp_v4_syn_recv_sock(sk, skb, req, dst,
1205 					     req_unhash, own_req);
1206 
1207 		if (!newsk)
1208 			return NULL;
1209 
1210 		inet_sk(newsk)->pinet6 = tcp_inet6_sk(newsk);
1211 
1212 		newinet = inet_sk(newsk);
1213 		newnp = tcp_inet6_sk(newsk);
1214 		newtp = tcp_sk(newsk);
1215 
1216 		memcpy(newnp, np, sizeof(struct ipv6_pinfo));
1217 
1218 		newnp->saddr = newsk->sk_v6_rcv_saddr;
1219 
1220 		inet_csk(newsk)->icsk_af_ops = &ipv6_mapped;
1221 		if (sk_is_mptcp(newsk))
1222 			mptcpv6_handle_mapped(newsk, true);
1223 		newsk->sk_backlog_rcv = tcp_v4_do_rcv;
1224 #ifdef CONFIG_TCP_MD5SIG
1225 		newtp->af_specific = &tcp_sock_ipv6_mapped_specific;
1226 #endif
1227 
1228 		newnp->ipv6_mc_list = NULL;
1229 		newnp->ipv6_ac_list = NULL;
1230 		newnp->ipv6_fl_list = NULL;
1231 		newnp->pktoptions  = NULL;
1232 		newnp->opt	   = NULL;
1233 		newnp->mcast_oif   = inet_iif(skb);
1234 		newnp->mcast_hops  = ip_hdr(skb)->ttl;
1235 		newnp->rcv_flowinfo = 0;
1236 		if (np->repflow)
1237 			newnp->flow_label = 0;
1238 
1239 		/*
1240 		 * No need to charge this sock to the relevant IPv6 refcnt debug socks count
1241 		 * here, tcp_create_openreq_child now does this for us, see the comment in
1242 		 * that function for the gory details. -acme
1243 		 */
1244 
1245 		/* It is tricky place. Until this moment IPv4 tcp
1246 		   worked with IPv6 icsk.icsk_af_ops.
1247 		   Sync it now.
1248 		 */
1249 		tcp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie);
1250 
1251 		return newsk;
1252 	}
1253 
1254 	ireq = inet_rsk(req);
1255 
1256 	if (sk_acceptq_is_full(sk))
1257 		goto out_overflow;
1258 
1259 	if (!dst) {
1260 		dst = inet6_csk_route_req(sk, &fl6, req, IPPROTO_TCP);
1261 		if (!dst)
1262 			goto out;
1263 	}
1264 
1265 	newsk = tcp_create_openreq_child(sk, req, skb);
1266 	if (!newsk)
1267 		goto out_nonewsk;
1268 
1269 	/*
1270 	 * No need to charge this sock to the relevant IPv6 refcnt debug socks
1271 	 * count here, tcp_create_openreq_child now does this for us, see the
1272 	 * comment in that function for the gory details. -acme
1273 	 */
1274 
1275 	newsk->sk_gso_type = SKB_GSO_TCPV6;
1276 	ip6_dst_store(newsk, dst, NULL, NULL);
1277 	inet6_sk_rx_dst_set(newsk, skb);
1278 
1279 	inet_sk(newsk)->pinet6 = tcp_inet6_sk(newsk);
1280 
1281 	newtp = tcp_sk(newsk);
1282 	newinet = inet_sk(newsk);
1283 	newnp = tcp_inet6_sk(newsk);
1284 
1285 	memcpy(newnp, np, sizeof(struct ipv6_pinfo));
1286 
1287 	newsk->sk_v6_daddr = ireq->ir_v6_rmt_addr;
1288 	newnp->saddr = ireq->ir_v6_loc_addr;
1289 	newsk->sk_v6_rcv_saddr = ireq->ir_v6_loc_addr;
1290 	newsk->sk_bound_dev_if = ireq->ir_iif;
1291 
1292 	/* Now IPv6 options...
1293 
1294 	   First: no IPv4 options.
1295 	 */
1296 	newinet->inet_opt = NULL;
1297 	newnp->ipv6_mc_list = NULL;
1298 	newnp->ipv6_ac_list = NULL;
1299 	newnp->ipv6_fl_list = NULL;
1300 
1301 	/* Clone RX bits */
1302 	newnp->rxopt.all = np->rxopt.all;
1303 
1304 	newnp->pktoptions = NULL;
1305 	newnp->opt	  = NULL;
1306 	newnp->mcast_oif  = tcp_v6_iif(skb);
1307 	newnp->mcast_hops = ipv6_hdr(skb)->hop_limit;
1308 	newnp->rcv_flowinfo = ip6_flowinfo(ipv6_hdr(skb));
1309 	if (np->repflow)
1310 		newnp->flow_label = ip6_flowlabel(ipv6_hdr(skb));
1311 
1312 	/* Clone native IPv6 options from listening socket (if any)
1313 
1314 	   Yes, keeping reference count would be much more clever,
1315 	   but we make one more one thing there: reattach optmem
1316 	   to newsk.
1317 	 */
1318 	opt = ireq->ipv6_opt;
1319 	if (!opt)
1320 		opt = rcu_dereference(np->opt);
1321 	if (opt) {
1322 		opt = ipv6_dup_options(newsk, opt);
1323 		RCU_INIT_POINTER(newnp->opt, opt);
1324 	}
1325 	inet_csk(newsk)->icsk_ext_hdr_len = 0;
1326 	if (opt)
1327 		inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen +
1328 						    opt->opt_flen;
1329 
1330 	tcp_ca_openreq_child(newsk, dst);
1331 
1332 	tcp_sync_mss(newsk, dst_mtu(dst));
1333 	newtp->advmss = tcp_mss_clamp(tcp_sk(sk), dst_metric_advmss(dst));
1334 
1335 	tcp_initialize_rcv_mss(newsk);
1336 
1337 	newinet->inet_daddr = newinet->inet_saddr = LOOPBACK4_IPV6;
1338 	newinet->inet_rcv_saddr = LOOPBACK4_IPV6;
1339 
1340 #ifdef CONFIG_TCP_MD5SIG
1341 	l3index = l3mdev_master_ifindex_by_index(sock_net(sk), ireq->ir_iif);
1342 
1343 	/* Copy over the MD5 key from the original socket */
1344 	key = tcp_v6_md5_do_lookup(sk, &newsk->sk_v6_daddr, l3index);
1345 	if (key) {
1346 		/* We're using one, so create a matching key
1347 		 * on the newsk structure. If we fail to get
1348 		 * memory, then we end up not copying the key
1349 		 * across. Shucks.
1350 		 */
1351 		tcp_md5_do_add(newsk, (union tcp_md5_addr *)&newsk->sk_v6_daddr,
1352 			       AF_INET6, 128, l3index, key->key, key->keylen,
1353 			       sk_gfp_mask(sk, GFP_ATOMIC));
1354 	}
1355 #endif
1356 
1357 	if (__inet_inherit_port(sk, newsk) < 0) {
1358 		inet_csk_prepare_forced_close(newsk);
1359 		tcp_done(newsk);
1360 		goto out;
1361 	}
1362 	*own_req = inet_ehash_nolisten(newsk, req_to_sk(req_unhash));
1363 	if (*own_req) {
1364 		tcp_move_syn(newtp, req);
1365 
1366 		/* Clone pktoptions received with SYN, if we own the req */
1367 		if (ireq->pktopts) {
1368 			newnp->pktoptions = skb_clone(ireq->pktopts,
1369 						      sk_gfp_mask(sk, GFP_ATOMIC));
1370 			consume_skb(ireq->pktopts);
1371 			ireq->pktopts = NULL;
1372 			if (newnp->pktoptions) {
1373 				tcp_v6_restore_cb(newnp->pktoptions);
1374 				skb_set_owner_r(newnp->pktoptions, newsk);
1375 			}
1376 		}
1377 	}
1378 
1379 	return newsk;
1380 
1381 out_overflow:
1382 	__NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS);
1383 out_nonewsk:
1384 	dst_release(dst);
1385 out:
1386 	tcp_listendrop(sk);
1387 	return NULL;
1388 }
1389 
1390 /* The socket must have it's spinlock held when we get
1391  * here, unless it is a TCP_LISTEN socket.
1392  *
1393  * We have a potential double-lock case here, so even when
1394  * doing backlog processing we use the BH locking scheme.
1395  * This is because we cannot sleep with the original spinlock
1396  * held.
1397  */
1398 static int tcp_v6_do_rcv(struct sock *sk, struct sk_buff *skb)
1399 {
1400 	struct ipv6_pinfo *np = tcp_inet6_sk(sk);
1401 	struct sk_buff *opt_skb = NULL;
1402 	struct tcp_sock *tp;
1403 
1404 	/* Imagine: socket is IPv6. IPv4 packet arrives,
1405 	   goes to IPv4 receive handler and backlogged.
1406 	   From backlog it always goes here. Kerboom...
1407 	   Fortunately, tcp_rcv_established and rcv_established
1408 	   handle them correctly, but it is not case with
1409 	   tcp_v6_hnd_req and tcp_v6_send_reset().   --ANK
1410 	 */
1411 
1412 	if (skb->protocol == htons(ETH_P_IP))
1413 		return tcp_v4_do_rcv(sk, skb);
1414 
1415 	/*
1416 	 *	socket locking is here for SMP purposes as backlog rcv
1417 	 *	is currently called with bh processing disabled.
1418 	 */
1419 
1420 	/* Do Stevens' IPV6_PKTOPTIONS.
1421 
1422 	   Yes, guys, it is the only place in our code, where we
1423 	   may make it not affecting IPv4.
1424 	   The rest of code is protocol independent,
1425 	   and I do not like idea to uglify IPv4.
1426 
1427 	   Actually, all the idea behind IPV6_PKTOPTIONS
1428 	   looks not very well thought. For now we latch
1429 	   options, received in the last packet, enqueued
1430 	   by tcp. Feel free to propose better solution.
1431 					       --ANK (980728)
1432 	 */
1433 	if (np->rxopt.all)
1434 		opt_skb = skb_clone(skb, sk_gfp_mask(sk, GFP_ATOMIC));
1435 
1436 	if (sk->sk_state == TCP_ESTABLISHED) { /* Fast path */
1437 		struct dst_entry *dst = sk->sk_rx_dst;
1438 
1439 		sock_rps_save_rxhash(sk, skb);
1440 		sk_mark_napi_id(sk, skb);
1441 		if (dst) {
1442 			if (inet_sk(sk)->rx_dst_ifindex != skb->skb_iif ||
1443 			    dst->ops->check(dst, np->rx_dst_cookie) == NULL) {
1444 				dst_release(dst);
1445 				sk->sk_rx_dst = NULL;
1446 			}
1447 		}
1448 
1449 		tcp_rcv_established(sk, skb);
1450 		if (opt_skb)
1451 			goto ipv6_pktoptions;
1452 		return 0;
1453 	}
1454 
1455 	if (tcp_checksum_complete(skb))
1456 		goto csum_err;
1457 
1458 	if (sk->sk_state == TCP_LISTEN) {
1459 		struct sock *nsk = tcp_v6_cookie_check(sk, skb);
1460 
1461 		if (!nsk)
1462 			goto discard;
1463 
1464 		if (nsk != sk) {
1465 			if (tcp_child_process(sk, nsk, skb))
1466 				goto reset;
1467 			if (opt_skb)
1468 				__kfree_skb(opt_skb);
1469 			return 0;
1470 		}
1471 	} else
1472 		sock_rps_save_rxhash(sk, skb);
1473 
1474 	if (tcp_rcv_state_process(sk, skb))
1475 		goto reset;
1476 	if (opt_skb)
1477 		goto ipv6_pktoptions;
1478 	return 0;
1479 
1480 reset:
1481 	tcp_v6_send_reset(sk, skb);
1482 discard:
1483 	if (opt_skb)
1484 		__kfree_skb(opt_skb);
1485 	kfree_skb(skb);
1486 	return 0;
1487 csum_err:
1488 	TCP_INC_STATS(sock_net(sk), TCP_MIB_CSUMERRORS);
1489 	TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS);
1490 	goto discard;
1491 
1492 
1493 ipv6_pktoptions:
1494 	/* Do you ask, what is it?
1495 
1496 	   1. skb was enqueued by tcp.
1497 	   2. skb is added to tail of read queue, rather than out of order.
1498 	   3. socket is not in passive state.
1499 	   4. Finally, it really contains options, which user wants to receive.
1500 	 */
1501 	tp = tcp_sk(sk);
1502 	if (TCP_SKB_CB(opt_skb)->end_seq == tp->rcv_nxt &&
1503 	    !((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN))) {
1504 		if (np->rxopt.bits.rxinfo || np->rxopt.bits.rxoinfo)
1505 			np->mcast_oif = tcp_v6_iif(opt_skb);
1506 		if (np->rxopt.bits.rxhlim || np->rxopt.bits.rxohlim)
1507 			np->mcast_hops = ipv6_hdr(opt_skb)->hop_limit;
1508 		if (np->rxopt.bits.rxflow || np->rxopt.bits.rxtclass)
1509 			np->rcv_flowinfo = ip6_flowinfo(ipv6_hdr(opt_skb));
1510 		if (np->repflow)
1511 			np->flow_label = ip6_flowlabel(ipv6_hdr(opt_skb));
1512 		if (ipv6_opt_accepted(sk, opt_skb, &TCP_SKB_CB(opt_skb)->header.h6)) {
1513 			skb_set_owner_r(opt_skb, sk);
1514 			tcp_v6_restore_cb(opt_skb);
1515 			opt_skb = xchg(&np->pktoptions, opt_skb);
1516 		} else {
1517 			__kfree_skb(opt_skb);
1518 			opt_skb = xchg(&np->pktoptions, NULL);
1519 		}
1520 	}
1521 
1522 	kfree_skb(opt_skb);
1523 	return 0;
1524 }
1525 
1526 static void tcp_v6_fill_cb(struct sk_buff *skb, const struct ipv6hdr *hdr,
1527 			   const struct tcphdr *th)
1528 {
1529 	/* This is tricky: we move IP6CB at its correct location into
1530 	 * TCP_SKB_CB(). It must be done after xfrm6_policy_check(), because
1531 	 * _decode_session6() uses IP6CB().
1532 	 * barrier() makes sure compiler won't play aliasing games.
1533 	 */
1534 	memmove(&TCP_SKB_CB(skb)->header.h6, IP6CB(skb),
1535 		sizeof(struct inet6_skb_parm));
1536 	barrier();
1537 
1538 	TCP_SKB_CB(skb)->seq = ntohl(th->seq);
1539 	TCP_SKB_CB(skb)->end_seq = (TCP_SKB_CB(skb)->seq + th->syn + th->fin +
1540 				    skb->len - th->doff*4);
1541 	TCP_SKB_CB(skb)->ack_seq = ntohl(th->ack_seq);
1542 	TCP_SKB_CB(skb)->tcp_flags = tcp_flag_byte(th);
1543 	TCP_SKB_CB(skb)->tcp_tw_isn = 0;
1544 	TCP_SKB_CB(skb)->ip_dsfield = ipv6_get_dsfield(hdr);
1545 	TCP_SKB_CB(skb)->sacked = 0;
1546 	TCP_SKB_CB(skb)->has_rxtstamp =
1547 			skb->tstamp || skb_hwtstamps(skb)->hwtstamp;
1548 }
1549 
1550 INDIRECT_CALLABLE_SCOPE int tcp_v6_rcv(struct sk_buff *skb)
1551 {
1552 	struct sk_buff *skb_to_free;
1553 	int sdif = inet6_sdif(skb);
1554 	int dif = inet6_iif(skb);
1555 	const struct tcphdr *th;
1556 	const struct ipv6hdr *hdr;
1557 	bool refcounted;
1558 	struct sock *sk;
1559 	int ret;
1560 	struct net *net = dev_net(skb->dev);
1561 
1562 	if (skb->pkt_type != PACKET_HOST)
1563 		goto discard_it;
1564 
1565 	/*
1566 	 *	Count it even if it's bad.
1567 	 */
1568 	__TCP_INC_STATS(net, TCP_MIB_INSEGS);
1569 
1570 	if (!pskb_may_pull(skb, sizeof(struct tcphdr)))
1571 		goto discard_it;
1572 
1573 	th = (const struct tcphdr *)skb->data;
1574 
1575 	if (unlikely(th->doff < sizeof(struct tcphdr)/4))
1576 		goto bad_packet;
1577 	if (!pskb_may_pull(skb, th->doff*4))
1578 		goto discard_it;
1579 
1580 	if (skb_checksum_init(skb, IPPROTO_TCP, ip6_compute_pseudo))
1581 		goto csum_error;
1582 
1583 	th = (const struct tcphdr *)skb->data;
1584 	hdr = ipv6_hdr(skb);
1585 
1586 lookup:
1587 	sk = __inet6_lookup_skb(&tcp_hashinfo, skb, __tcp_hdrlen(th),
1588 				th->source, th->dest, inet6_iif(skb), sdif,
1589 				&refcounted);
1590 	if (!sk)
1591 		goto no_tcp_socket;
1592 
1593 process:
1594 	if (sk->sk_state == TCP_TIME_WAIT)
1595 		goto do_time_wait;
1596 
1597 	if (sk->sk_state == TCP_NEW_SYN_RECV) {
1598 		struct request_sock *req = inet_reqsk(sk);
1599 		bool req_stolen = false;
1600 		struct sock *nsk;
1601 
1602 		sk = req->rsk_listener;
1603 		if (tcp_v6_inbound_md5_hash(sk, skb, dif, sdif)) {
1604 			sk_drops_add(sk, skb);
1605 			reqsk_put(req);
1606 			goto discard_it;
1607 		}
1608 		if (tcp_checksum_complete(skb)) {
1609 			reqsk_put(req);
1610 			goto csum_error;
1611 		}
1612 		if (unlikely(sk->sk_state != TCP_LISTEN)) {
1613 			inet_csk_reqsk_queue_drop_and_put(sk, req);
1614 			goto lookup;
1615 		}
1616 		sock_hold(sk);
1617 		refcounted = true;
1618 		nsk = NULL;
1619 		if (!tcp_filter(sk, skb)) {
1620 			th = (const struct tcphdr *)skb->data;
1621 			hdr = ipv6_hdr(skb);
1622 			tcp_v6_fill_cb(skb, hdr, th);
1623 			nsk = tcp_check_req(sk, skb, req, false, &req_stolen);
1624 		}
1625 		if (!nsk) {
1626 			reqsk_put(req);
1627 			if (req_stolen) {
1628 				/* Another cpu got exclusive access to req
1629 				 * and created a full blown socket.
1630 				 * Try to feed this packet to this socket
1631 				 * instead of discarding it.
1632 				 */
1633 				tcp_v6_restore_cb(skb);
1634 				sock_put(sk);
1635 				goto lookup;
1636 			}
1637 			goto discard_and_relse;
1638 		}
1639 		if (nsk == sk) {
1640 			reqsk_put(req);
1641 			tcp_v6_restore_cb(skb);
1642 		} else if (tcp_child_process(sk, nsk, skb)) {
1643 			tcp_v6_send_reset(nsk, skb);
1644 			goto discard_and_relse;
1645 		} else {
1646 			sock_put(sk);
1647 			return 0;
1648 		}
1649 	}
1650 	if (hdr->hop_limit < tcp_inet6_sk(sk)->min_hopcount) {
1651 		__NET_INC_STATS(net, LINUX_MIB_TCPMINTTLDROP);
1652 		goto discard_and_relse;
1653 	}
1654 
1655 	if (!xfrm6_policy_check(sk, XFRM_POLICY_IN, skb))
1656 		goto discard_and_relse;
1657 
1658 	if (tcp_v6_inbound_md5_hash(sk, skb, dif, sdif))
1659 		goto discard_and_relse;
1660 
1661 	if (tcp_filter(sk, skb))
1662 		goto discard_and_relse;
1663 	th = (const struct tcphdr *)skb->data;
1664 	hdr = ipv6_hdr(skb);
1665 	tcp_v6_fill_cb(skb, hdr, th);
1666 
1667 	skb->dev = NULL;
1668 
1669 	if (sk->sk_state == TCP_LISTEN) {
1670 		ret = tcp_v6_do_rcv(sk, skb);
1671 		goto put_and_return;
1672 	}
1673 
1674 	sk_incoming_cpu_update(sk);
1675 
1676 	bh_lock_sock_nested(sk);
1677 	tcp_segs_in(tcp_sk(sk), skb);
1678 	ret = 0;
1679 	if (!sock_owned_by_user(sk)) {
1680 		skb_to_free = sk->sk_rx_skb_cache;
1681 		sk->sk_rx_skb_cache = NULL;
1682 		ret = tcp_v6_do_rcv(sk, skb);
1683 	} else {
1684 		if (tcp_add_backlog(sk, skb))
1685 			goto discard_and_relse;
1686 		skb_to_free = NULL;
1687 	}
1688 	bh_unlock_sock(sk);
1689 	if (skb_to_free)
1690 		__kfree_skb(skb_to_free);
1691 put_and_return:
1692 	if (refcounted)
1693 		sock_put(sk);
1694 	return ret ? -1 : 0;
1695 
1696 no_tcp_socket:
1697 	if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb))
1698 		goto discard_it;
1699 
1700 	tcp_v6_fill_cb(skb, hdr, th);
1701 
1702 	if (tcp_checksum_complete(skb)) {
1703 csum_error:
1704 		__TCP_INC_STATS(net, TCP_MIB_CSUMERRORS);
1705 bad_packet:
1706 		__TCP_INC_STATS(net, TCP_MIB_INERRS);
1707 	} else {
1708 		tcp_v6_send_reset(NULL, skb);
1709 	}
1710 
1711 discard_it:
1712 	kfree_skb(skb);
1713 	return 0;
1714 
1715 discard_and_relse:
1716 	sk_drops_add(sk, skb);
1717 	if (refcounted)
1718 		sock_put(sk);
1719 	goto discard_it;
1720 
1721 do_time_wait:
1722 	if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) {
1723 		inet_twsk_put(inet_twsk(sk));
1724 		goto discard_it;
1725 	}
1726 
1727 	tcp_v6_fill_cb(skb, hdr, th);
1728 
1729 	if (tcp_checksum_complete(skb)) {
1730 		inet_twsk_put(inet_twsk(sk));
1731 		goto csum_error;
1732 	}
1733 
1734 	switch (tcp_timewait_state_process(inet_twsk(sk), skb, th)) {
1735 	case TCP_TW_SYN:
1736 	{
1737 		struct sock *sk2;
1738 
1739 		sk2 = inet6_lookup_listener(dev_net(skb->dev), &tcp_hashinfo,
1740 					    skb, __tcp_hdrlen(th),
1741 					    &ipv6_hdr(skb)->saddr, th->source,
1742 					    &ipv6_hdr(skb)->daddr,
1743 					    ntohs(th->dest),
1744 					    tcp_v6_iif_l3_slave(skb),
1745 					    sdif);
1746 		if (sk2) {
1747 			struct inet_timewait_sock *tw = inet_twsk(sk);
1748 			inet_twsk_deschedule_put(tw);
1749 			sk = sk2;
1750 			tcp_v6_restore_cb(skb);
1751 			refcounted = false;
1752 			goto process;
1753 		}
1754 	}
1755 		/* to ACK */
1756 		fallthrough;
1757 	case TCP_TW_ACK:
1758 		tcp_v6_timewait_ack(sk, skb);
1759 		break;
1760 	case TCP_TW_RST:
1761 		tcp_v6_send_reset(sk, skb);
1762 		inet_twsk_deschedule_put(inet_twsk(sk));
1763 		goto discard_it;
1764 	case TCP_TW_SUCCESS:
1765 		;
1766 	}
1767 	goto discard_it;
1768 }
1769 
1770 INDIRECT_CALLABLE_SCOPE void tcp_v6_early_demux(struct sk_buff *skb)
1771 {
1772 	const struct ipv6hdr *hdr;
1773 	const struct tcphdr *th;
1774 	struct sock *sk;
1775 
1776 	if (skb->pkt_type != PACKET_HOST)
1777 		return;
1778 
1779 	if (!pskb_may_pull(skb, skb_transport_offset(skb) + sizeof(struct tcphdr)))
1780 		return;
1781 
1782 	hdr = ipv6_hdr(skb);
1783 	th = tcp_hdr(skb);
1784 
1785 	if (th->doff < sizeof(struct tcphdr) / 4)
1786 		return;
1787 
1788 	/* Note : We use inet6_iif() here, not tcp_v6_iif() */
1789 	sk = __inet6_lookup_established(dev_net(skb->dev), &tcp_hashinfo,
1790 					&hdr->saddr, th->source,
1791 					&hdr->daddr, ntohs(th->dest),
1792 					inet6_iif(skb), inet6_sdif(skb));
1793 	if (sk) {
1794 		skb->sk = sk;
1795 		skb->destructor = sock_edemux;
1796 		if (sk_fullsock(sk)) {
1797 			struct dst_entry *dst = READ_ONCE(sk->sk_rx_dst);
1798 
1799 			if (dst)
1800 				dst = dst_check(dst, tcp_inet6_sk(sk)->rx_dst_cookie);
1801 			if (dst &&
1802 			    inet_sk(sk)->rx_dst_ifindex == skb->skb_iif)
1803 				skb_dst_set_noref(skb, dst);
1804 		}
1805 	}
1806 }
1807 
1808 static struct timewait_sock_ops tcp6_timewait_sock_ops = {
1809 	.twsk_obj_size	= sizeof(struct tcp6_timewait_sock),
1810 	.twsk_unique	= tcp_twsk_unique,
1811 	.twsk_destructor = tcp_twsk_destructor,
1812 };
1813 
1814 INDIRECT_CALLABLE_SCOPE void tcp_v6_send_check(struct sock *sk, struct sk_buff *skb)
1815 {
1816 	struct ipv6_pinfo *np = inet6_sk(sk);
1817 
1818 	__tcp_v6_send_check(skb, &np->saddr, &sk->sk_v6_daddr);
1819 }
1820 
1821 const struct inet_connection_sock_af_ops ipv6_specific = {
1822 	.queue_xmit	   = inet6_csk_xmit,
1823 	.send_check	   = tcp_v6_send_check,
1824 	.rebuild_header	   = inet6_sk_rebuild_header,
1825 	.sk_rx_dst_set	   = inet6_sk_rx_dst_set,
1826 	.conn_request	   = tcp_v6_conn_request,
1827 	.syn_recv_sock	   = tcp_v6_syn_recv_sock,
1828 	.net_header_len	   = sizeof(struct ipv6hdr),
1829 	.net_frag_header_len = sizeof(struct frag_hdr),
1830 	.setsockopt	   = ipv6_setsockopt,
1831 	.getsockopt	   = ipv6_getsockopt,
1832 	.addr2sockaddr	   = inet6_csk_addr2sockaddr,
1833 	.sockaddr_len	   = sizeof(struct sockaddr_in6),
1834 	.mtu_reduced	   = tcp_v6_mtu_reduced,
1835 };
1836 
1837 #ifdef CONFIG_TCP_MD5SIG
1838 static const struct tcp_sock_af_ops tcp_sock_ipv6_specific = {
1839 	.md5_lookup	=	tcp_v6_md5_lookup,
1840 	.calc_md5_hash	=	tcp_v6_md5_hash_skb,
1841 	.md5_parse	=	tcp_v6_parse_md5_keys,
1842 };
1843 #endif
1844 
1845 /*
1846  *	TCP over IPv4 via INET6 API
1847  */
1848 static const struct inet_connection_sock_af_ops ipv6_mapped = {
1849 	.queue_xmit	   = ip_queue_xmit,
1850 	.send_check	   = tcp_v4_send_check,
1851 	.rebuild_header	   = inet_sk_rebuild_header,
1852 	.sk_rx_dst_set	   = inet_sk_rx_dst_set,
1853 	.conn_request	   = tcp_v6_conn_request,
1854 	.syn_recv_sock	   = tcp_v6_syn_recv_sock,
1855 	.net_header_len	   = sizeof(struct iphdr),
1856 	.setsockopt	   = ipv6_setsockopt,
1857 	.getsockopt	   = ipv6_getsockopt,
1858 	.addr2sockaddr	   = inet6_csk_addr2sockaddr,
1859 	.sockaddr_len	   = sizeof(struct sockaddr_in6),
1860 	.mtu_reduced	   = tcp_v4_mtu_reduced,
1861 };
1862 
1863 #ifdef CONFIG_TCP_MD5SIG
1864 static const struct tcp_sock_af_ops tcp_sock_ipv6_mapped_specific = {
1865 	.md5_lookup	=	tcp_v4_md5_lookup,
1866 	.calc_md5_hash	=	tcp_v4_md5_hash_skb,
1867 	.md5_parse	=	tcp_v6_parse_md5_keys,
1868 };
1869 #endif
1870 
1871 /* NOTE: A lot of things set to zero explicitly by call to
1872  *       sk_alloc() so need not be done here.
1873  */
1874 static int tcp_v6_init_sock(struct sock *sk)
1875 {
1876 	struct inet_connection_sock *icsk = inet_csk(sk);
1877 
1878 	tcp_init_sock(sk);
1879 
1880 	icsk->icsk_af_ops = &ipv6_specific;
1881 
1882 #ifdef CONFIG_TCP_MD5SIG
1883 	tcp_sk(sk)->af_specific = &tcp_sock_ipv6_specific;
1884 #endif
1885 
1886 	return 0;
1887 }
1888 
1889 static void tcp_v6_destroy_sock(struct sock *sk)
1890 {
1891 	tcp_v4_destroy_sock(sk);
1892 	inet6_destroy_sock(sk);
1893 }
1894 
1895 #ifdef CONFIG_PROC_FS
1896 /* Proc filesystem TCPv6 sock list dumping. */
1897 static void get_openreq6(struct seq_file *seq,
1898 			 const struct request_sock *req, int i)
1899 {
1900 	long ttd = req->rsk_timer.expires - jiffies;
1901 	const struct in6_addr *src = &inet_rsk(req)->ir_v6_loc_addr;
1902 	const struct in6_addr *dest = &inet_rsk(req)->ir_v6_rmt_addr;
1903 
1904 	if (ttd < 0)
1905 		ttd = 0;
1906 
1907 	seq_printf(seq,
1908 		   "%4d: %08X%08X%08X%08X:%04X %08X%08X%08X%08X:%04X "
1909 		   "%02X %08X:%08X %02X:%08lX %08X %5u %8d %d %d %pK\n",
1910 		   i,
1911 		   src->s6_addr32[0], src->s6_addr32[1],
1912 		   src->s6_addr32[2], src->s6_addr32[3],
1913 		   inet_rsk(req)->ir_num,
1914 		   dest->s6_addr32[0], dest->s6_addr32[1],
1915 		   dest->s6_addr32[2], dest->s6_addr32[3],
1916 		   ntohs(inet_rsk(req)->ir_rmt_port),
1917 		   TCP_SYN_RECV,
1918 		   0, 0, /* could print option size, but that is af dependent. */
1919 		   1,   /* timers active (only the expire timer) */
1920 		   jiffies_to_clock_t(ttd),
1921 		   req->num_timeout,
1922 		   from_kuid_munged(seq_user_ns(seq),
1923 				    sock_i_uid(req->rsk_listener)),
1924 		   0,  /* non standard timer */
1925 		   0, /* open_requests have no inode */
1926 		   0, req);
1927 }
1928 
1929 static void get_tcp6_sock(struct seq_file *seq, struct sock *sp, int i)
1930 {
1931 	const struct in6_addr *dest, *src;
1932 	__u16 destp, srcp;
1933 	int timer_active;
1934 	unsigned long timer_expires;
1935 	const struct inet_sock *inet = inet_sk(sp);
1936 	const struct tcp_sock *tp = tcp_sk(sp);
1937 	const struct inet_connection_sock *icsk = inet_csk(sp);
1938 	const struct fastopen_queue *fastopenq = &icsk->icsk_accept_queue.fastopenq;
1939 	int rx_queue;
1940 	int state;
1941 
1942 	dest  = &sp->sk_v6_daddr;
1943 	src   = &sp->sk_v6_rcv_saddr;
1944 	destp = ntohs(inet->inet_dport);
1945 	srcp  = ntohs(inet->inet_sport);
1946 
1947 	if (icsk->icsk_pending == ICSK_TIME_RETRANS ||
1948 	    icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT ||
1949 	    icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) {
1950 		timer_active	= 1;
1951 		timer_expires	= icsk->icsk_timeout;
1952 	} else if (icsk->icsk_pending == ICSK_TIME_PROBE0) {
1953 		timer_active	= 4;
1954 		timer_expires	= icsk->icsk_timeout;
1955 	} else if (timer_pending(&sp->sk_timer)) {
1956 		timer_active	= 2;
1957 		timer_expires	= sp->sk_timer.expires;
1958 	} else {
1959 		timer_active	= 0;
1960 		timer_expires = jiffies;
1961 	}
1962 
1963 	state = inet_sk_state_load(sp);
1964 	if (state == TCP_LISTEN)
1965 		rx_queue = READ_ONCE(sp->sk_ack_backlog);
1966 	else
1967 		/* Because we don't lock the socket,
1968 		 * we might find a transient negative value.
1969 		 */
1970 		rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) -
1971 				      READ_ONCE(tp->copied_seq), 0);
1972 
1973 	seq_printf(seq,
1974 		   "%4d: %08X%08X%08X%08X:%04X %08X%08X%08X%08X:%04X "
1975 		   "%02X %08X:%08X %02X:%08lX %08X %5u %8d %lu %d %pK %lu %lu %u %u %d\n",
1976 		   i,
1977 		   src->s6_addr32[0], src->s6_addr32[1],
1978 		   src->s6_addr32[2], src->s6_addr32[3], srcp,
1979 		   dest->s6_addr32[0], dest->s6_addr32[1],
1980 		   dest->s6_addr32[2], dest->s6_addr32[3], destp,
1981 		   state,
1982 		   READ_ONCE(tp->write_seq) - tp->snd_una,
1983 		   rx_queue,
1984 		   timer_active,
1985 		   jiffies_delta_to_clock_t(timer_expires - jiffies),
1986 		   icsk->icsk_retransmits,
1987 		   from_kuid_munged(seq_user_ns(seq), sock_i_uid(sp)),
1988 		   icsk->icsk_probes_out,
1989 		   sock_i_ino(sp),
1990 		   refcount_read(&sp->sk_refcnt), sp,
1991 		   jiffies_to_clock_t(icsk->icsk_rto),
1992 		   jiffies_to_clock_t(icsk->icsk_ack.ato),
1993 		   (icsk->icsk_ack.quick << 1) | inet_csk_in_pingpong_mode(sp),
1994 		   tp->snd_cwnd,
1995 		   state == TCP_LISTEN ?
1996 			fastopenq->max_qlen :
1997 			(tcp_in_initial_slowstart(tp) ? -1 : tp->snd_ssthresh)
1998 		   );
1999 }
2000 
2001 static void get_timewait6_sock(struct seq_file *seq,
2002 			       struct inet_timewait_sock *tw, int i)
2003 {
2004 	long delta = tw->tw_timer.expires - jiffies;
2005 	const struct in6_addr *dest, *src;
2006 	__u16 destp, srcp;
2007 
2008 	dest = &tw->tw_v6_daddr;
2009 	src  = &tw->tw_v6_rcv_saddr;
2010 	destp = ntohs(tw->tw_dport);
2011 	srcp  = ntohs(tw->tw_sport);
2012 
2013 	seq_printf(seq,
2014 		   "%4d: %08X%08X%08X%08X:%04X %08X%08X%08X%08X:%04X "
2015 		   "%02X %08X:%08X %02X:%08lX %08X %5d %8d %d %d %pK\n",
2016 		   i,
2017 		   src->s6_addr32[0], src->s6_addr32[1],
2018 		   src->s6_addr32[2], src->s6_addr32[3], srcp,
2019 		   dest->s6_addr32[0], dest->s6_addr32[1],
2020 		   dest->s6_addr32[2], dest->s6_addr32[3], destp,
2021 		   tw->tw_substate, 0, 0,
2022 		   3, jiffies_delta_to_clock_t(delta), 0, 0, 0, 0,
2023 		   refcount_read(&tw->tw_refcnt), tw);
2024 }
2025 
2026 static int tcp6_seq_show(struct seq_file *seq, void *v)
2027 {
2028 	struct tcp_iter_state *st;
2029 	struct sock *sk = v;
2030 
2031 	if (v == SEQ_START_TOKEN) {
2032 		seq_puts(seq,
2033 			 "  sl  "
2034 			 "local_address                         "
2035 			 "remote_address                        "
2036 			 "st tx_queue rx_queue tr tm->when retrnsmt"
2037 			 "   uid  timeout inode\n");
2038 		goto out;
2039 	}
2040 	st = seq->private;
2041 
2042 	if (sk->sk_state == TCP_TIME_WAIT)
2043 		get_timewait6_sock(seq, v, st->num);
2044 	else if (sk->sk_state == TCP_NEW_SYN_RECV)
2045 		get_openreq6(seq, v, st->num);
2046 	else
2047 		get_tcp6_sock(seq, v, st->num);
2048 out:
2049 	return 0;
2050 }
2051 
2052 static const struct seq_operations tcp6_seq_ops = {
2053 	.show		= tcp6_seq_show,
2054 	.start		= tcp_seq_start,
2055 	.next		= tcp_seq_next,
2056 	.stop		= tcp_seq_stop,
2057 };
2058 
2059 static struct tcp_seq_afinfo tcp6_seq_afinfo = {
2060 	.family		= AF_INET6,
2061 };
2062 
2063 int __net_init tcp6_proc_init(struct net *net)
2064 {
2065 	if (!proc_create_net_data("tcp6", 0444, net->proc_net, &tcp6_seq_ops,
2066 			sizeof(struct tcp_iter_state), &tcp6_seq_afinfo))
2067 		return -ENOMEM;
2068 	return 0;
2069 }
2070 
2071 void tcp6_proc_exit(struct net *net)
2072 {
2073 	remove_proc_entry("tcp6", net->proc_net);
2074 }
2075 #endif
2076 
2077 struct proto tcpv6_prot = {
2078 	.name			= "TCPv6",
2079 	.owner			= THIS_MODULE,
2080 	.close			= tcp_close,
2081 	.pre_connect		= tcp_v6_pre_connect,
2082 	.connect		= tcp_v6_connect,
2083 	.disconnect		= tcp_disconnect,
2084 	.accept			= inet_csk_accept,
2085 	.ioctl			= tcp_ioctl,
2086 	.init			= tcp_v6_init_sock,
2087 	.destroy		= tcp_v6_destroy_sock,
2088 	.shutdown		= tcp_shutdown,
2089 	.setsockopt		= tcp_setsockopt,
2090 	.getsockopt		= tcp_getsockopt,
2091 	.keepalive		= tcp_set_keepalive,
2092 	.recvmsg		= tcp_recvmsg,
2093 	.sendmsg		= tcp_sendmsg,
2094 	.sendpage		= tcp_sendpage,
2095 	.backlog_rcv		= tcp_v6_do_rcv,
2096 	.release_cb		= tcp_release_cb,
2097 	.hash			= inet6_hash,
2098 	.unhash			= inet_unhash,
2099 	.get_port		= inet_csk_get_port,
2100 	.enter_memory_pressure	= tcp_enter_memory_pressure,
2101 	.leave_memory_pressure	= tcp_leave_memory_pressure,
2102 	.stream_memory_free	= tcp_stream_memory_free,
2103 	.sockets_allocated	= &tcp_sockets_allocated,
2104 	.memory_allocated	= &tcp_memory_allocated,
2105 	.memory_pressure	= &tcp_memory_pressure,
2106 	.orphan_count		= &tcp_orphan_count,
2107 	.sysctl_mem		= sysctl_tcp_mem,
2108 	.sysctl_wmem_offset	= offsetof(struct net, ipv4.sysctl_tcp_wmem),
2109 	.sysctl_rmem_offset	= offsetof(struct net, ipv4.sysctl_tcp_rmem),
2110 	.max_header		= MAX_TCP_HEADER,
2111 	.obj_size		= sizeof(struct tcp6_sock),
2112 	.slab_flags		= SLAB_TYPESAFE_BY_RCU,
2113 	.twsk_prot		= &tcp6_timewait_sock_ops,
2114 	.rsk_prot		= &tcp6_request_sock_ops,
2115 	.h.hashinfo		= &tcp_hashinfo,
2116 	.no_autobind		= true,
2117 	.diag_destroy		= tcp_abort,
2118 };
2119 EXPORT_SYMBOL_GPL(tcpv6_prot);
2120 
2121 /* thinking of making this const? Don't.
2122  * early_demux can change based on sysctl.
2123  */
2124 static struct inet6_protocol tcpv6_protocol = {
2125 	.early_demux	=	tcp_v6_early_demux,
2126 	.early_demux_handler =  tcp_v6_early_demux,
2127 	.handler	=	tcp_v6_rcv,
2128 	.err_handler	=	tcp_v6_err,
2129 	.flags		=	INET6_PROTO_NOPOLICY|INET6_PROTO_FINAL,
2130 };
2131 
2132 static struct inet_protosw tcpv6_protosw = {
2133 	.type		=	SOCK_STREAM,
2134 	.protocol	=	IPPROTO_TCP,
2135 	.prot		=	&tcpv6_prot,
2136 	.ops		=	&inet6_stream_ops,
2137 	.flags		=	INET_PROTOSW_PERMANENT |
2138 				INET_PROTOSW_ICSK,
2139 };
2140 
2141 static int __net_init tcpv6_net_init(struct net *net)
2142 {
2143 	return inet_ctl_sock_create(&net->ipv6.tcp_sk, PF_INET6,
2144 				    SOCK_RAW, IPPROTO_TCP, net);
2145 }
2146 
2147 static void __net_exit tcpv6_net_exit(struct net *net)
2148 {
2149 	inet_ctl_sock_destroy(net->ipv6.tcp_sk);
2150 }
2151 
2152 static void __net_exit tcpv6_net_exit_batch(struct list_head *net_exit_list)
2153 {
2154 	inet_twsk_purge(&tcp_hashinfo, AF_INET6);
2155 }
2156 
2157 static struct pernet_operations tcpv6_net_ops = {
2158 	.init	    = tcpv6_net_init,
2159 	.exit	    = tcpv6_net_exit,
2160 	.exit_batch = tcpv6_net_exit_batch,
2161 };
2162 
2163 int __init tcpv6_init(void)
2164 {
2165 	int ret;
2166 
2167 	ret = inet6_add_protocol(&tcpv6_protocol, IPPROTO_TCP);
2168 	if (ret)
2169 		goto out;
2170 
2171 	/* register inet6 protocol */
2172 	ret = inet6_register_protosw(&tcpv6_protosw);
2173 	if (ret)
2174 		goto out_tcpv6_protocol;
2175 
2176 	ret = register_pernet_subsys(&tcpv6_net_ops);
2177 	if (ret)
2178 		goto out_tcpv6_protosw;
2179 
2180 	ret = mptcpv6_init();
2181 	if (ret)
2182 		goto out_tcpv6_pernet_subsys;
2183 
2184 out:
2185 	return ret;
2186 
2187 out_tcpv6_pernet_subsys:
2188 	unregister_pernet_subsys(&tcpv6_net_ops);
2189 out_tcpv6_protosw:
2190 	inet6_unregister_protosw(&tcpv6_protosw);
2191 out_tcpv6_protocol:
2192 	inet6_del_protocol(&tcpv6_protocol, IPPROTO_TCP);
2193 	goto out;
2194 }
2195 
2196 void tcpv6_exit(void)
2197 {
2198 	unregister_pernet_subsys(&tcpv6_net_ops);
2199 	inet6_unregister_protosw(&tcpv6_protosw);
2200 	inet6_del_protocol(&tcpv6_protocol, IPPROTO_TCP);
2201 }
2202