Home
last modified time | relevance | path

Searched refs:sock_net (Results 1 – 25 of 270) sorted by relevance

1234567891011

/openbmc/linux/net/ipv4/
H A Dtcp_fastopen.c322 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check()
339 return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) || in tcp_fastopen_no_cookie()
354 int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen); in tcp_try_fastopen()
360 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen()
379 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
397 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
402 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
406 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
494 struct net *net = sock_net(sk); in tcp_fastopen_active_disable()
496 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout)) in tcp_fastopen_active_disable()
[all …]
H A Dtcp_plb.c28 struct net *net = sock_net(sk); in tcp_plb_update_state()
48 struct net *net = sock_net(sk); in tcp_plb_check_rehash()
83 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPLBREHASH); in tcp_plb_check_rehash()
93 struct net *net = sock_net(sk); in tcp_plb_update_state_upon_rto()
H A Dtcp_timer.c73 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err()
124 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources()
128 if (!check_net(sock_net(sk))) { in tcp_out_of_resources()
144 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */ in tcp_orphan_retries()
160 const struct net *net = sock_net(sk); in tcp_mtu_probing()
233 struct net *net = sock_net(sk); in tcp_write_timeout()
287 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTREHASH); in tcp_write_timeout()
331 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler()
355 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer()
395 max_probes = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_retries2); in tcp_probe_timer()
[all …]
H A Dsyncookies.c336 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v4_check()
345 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v4_check()
349 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v4_check()
353 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v4_check()
356 tsoff = secure_tcp_ts_off(sock_net(sk), in cookie_v4_check()
362 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v4_check()
399 RCU_INIT_POINTER(ireq->ireq_opt, tcp_v4_save_options(sock_net(sk), skb)); in cookie_v4_check()
420 rt = ip_route_output_key(sock_net(sk), &fl4); in cookie_v4_check()
441 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), &rt->dst); in cookie_v4_check()
H A Dudp_diag.c36 struct net *net = sock_net(in_skb->sk); in udp_dump_one()
94 struct net *net = sock_net(skb->sk); in udp_dump()
117 if (!net_eq(sock_net(sk), net)) in udp_dump()
150 udp_dump(sock_net(cb->skb->sk)->ipv4.udp_table, skb, cb, r); in udp_diag_dump()
156 return udp_dump_one(sock_net(cb->skb->sk)->ipv4.udp_table, cb, req); in udp_diag_dump_one()
171 struct net *net = sock_net(in_skb->sk); in __udp_diag_destroy()
228 return __udp_diag_destroy(in_skb, req, sock_net(in_skb->sk)->ipv4.udp_table); in udp_diag_destroy()
H A Dtcp_input.c216 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn()
462 min(sndmem, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[2]))); in tcp_sndbuf_expand()
497 int window = tcp_win_from_space(sk, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])) >> 1; in __tcp_grow_window()
570 int tcp_app_win = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_app_win); in tcp_init_buffer_space()
610 struct net *net = sock_net(sk); in tcp_clamp_window()
762 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf) && in tcp_rcv_space_adjust()
778 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])); in tcp_rcv_space_adjust()
943 rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio); in tcp_update_pacing_rate()
945 rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio); in tcp_update_pacing_rate()
1084 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_max_reordering)); in tcp_check_sack_reordering()
[all …]
H A Dtcp_minisocks.c284 struct net *net = sock_net(sk); in tcp_time_wait()
582 __TCP_INC_STATS(sock_net(sk), TCP_MIB_PASSIVEOPENS); in tcp_create_openreq_child()
615 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, NULL); in tcp_check_req()
657 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req()
746 !tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req()
751 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_check_req()
767 TCP_INC_STATS(sock_net(sk), TCP_MIB_ATTEMPTFAILS); in tcp_check_req()
790 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDEFERACCEPTDROP); in tcp_check_req()
822 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMIGRATEREQFAILURE); in tcp_check_req()
824 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_abort_on_overflow)) { in tcp_check_req()
[all …]
H A Dinet_hashtables.c52 return inet6_ehashfn(sock_net(sk), in sk_ehashfn()
56 return inet_ehashfn(sock_net(sk), in sk_ehashfn()
185 struct net *net = sock_net(sk); in __inet_put_port()
227 struct net *net = sock_net(sk); in __inet_inherit_port()
308 hash = ipv6_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk()
313 hash = ipv4_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk()
325 if (net_eq(sock_net(sk), net) && sk->sk_num == hnum && in compute_score()
553 struct net *net = sock_net(sk); in __inet_check_established()
594 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in __inet_check_established()
628 struct net *net = sock_net(sk); in inet_ehash_lookup_by_sk()
[all …]
H A Ddatagram.c41 if (!oif || netif_index_is_l3_master(sock_net(sk), oif)) in __ip4_datagram_connect()
55 IP_INC_STATS(sock_net(sk), IPSTATS_MIB_OUTNOROUTES); in __ip4_datagram_connect()
119 rt = ip_route_output_ports(sock_net(sk), &fl4, sk, daddr, in ip4_datagram_release_cb()
H A Draw_diag.c98 net = sock_net(in_skb->sk); in raw_diag_dump_one()
142 struct net *net = sock_net(skb->sk); in raw_diag_dump()
165 if (!net_eq(sock_net(sk), net)) in raw_diag_dump()
202 struct net *net = sock_net(in_skb->sk); in raw_diag_destroy()
H A Dudp.c138 return sk->sk_prot->h.udp_table ? : sock_net(sk)->ipv4.udp_table; in udp_get_table_prot()
150 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse()
187 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse2()
210 struct net *net = sock_net(sk); in udp_reuseport_add_sock()
215 if (net_eq(sock_net(sk2), net) && in udp_reuseport_add_sock()
244 struct net *net = sock_net(sk); in udp_lib_get_port()
333 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in udp_lib_get_port()
359 ipv4_portaddr_hash(sock_net(sk), htonl(INADDR_ANY), snum); in udp_v4_get_port()
361 ipv4_portaddr_hash(sock_net(sk), inet_sk(sk)->inet_rcv_saddr, 0); in udp_v4_get_port()
377 if (!net_eq(sock_net(sk), net) || in compute_score()
[all …]
H A Dtcp_output.c83 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPORIGDATASENT, in tcp_event_new_data_sent()
185 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_event_ack_sent()
230 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows)) in tcp_select_initial_window()
241 space = max_t(u32, space, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])); in tcp_select_initial_window()
261 struct net *net = sock_net(sk); in tcp_select_window()
338 bool use_ecn = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn) == 1 || in tcp_ecn_send_syn()
360 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_fallback)) in tcp_ecn_clear_syn()
805 if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_timestamps) && !*md5)) { in tcp_syn_options()
811 if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_window_scaling))) { in tcp_syn_options()
816 if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_sack))) { in tcp_syn_options()
[all …]
/openbmc/linux/net/ipv6/
H A Dsyncookies.c144 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v6_check()
153 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v6_check()
157 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v6_check()
161 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v6_check()
164 tsoff = secure_tcpv6_ts_off(sock_net(sk), in cookie_v6_check()
170 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v6_check()
241 dst = ip6_dst_lookup_flow(sock_net(sk), sk, &fl6, final_p); in cookie_v6_check()
259 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), dst); in cookie_v6_check()
/openbmc/linux/net/mptcp/
H A Dtoken.c179 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in mptcp_token_new_connect()
198 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in mptcp_token_accept()
260 !net_eq(sock_net(sk), net)) in mptcp_token_get_sock()
267 !net_eq(sock_net(sk), net)) { in mptcp_token_get_sock()
316 if (!net_eq(sock_net(sk), net)) in mptcp_token_iter_next()
325 if (!net_eq(sock_net(sk), net)) { in mptcp_token_iter_next()
384 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in mptcp_token_destroy()
H A Doptions.c412 opts->csum_reqd = mptcp_is_checksum_enabled(sock_net(sk)); in mptcp_syn_options()
413 opts->allow_join_id0 = mptcp_allow_join_id0(sock_net(sk)); in mptcp_syn_options()
479 opts->allow_join_id0 = mptcp_allow_join_id0(sock_net(sk)); in mptcp_established_options_mp()
694 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_ADDADDRTX); in mptcp_established_options_add_addr()
699 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_ECHOADDTX); in mptcp_established_options_add_addr()
733 MPTCP_ADD_STATS(sock_net(sk), MPTCP_MIB_RMADDRTX, opts->rm_list.nr); in mptcp_established_options_rm_addr()
777 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPRSTTX); in mptcp_established_options_rst()
801 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPFASTCLOSETX); in mptcp_established_options_fastclose()
823 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPFAILTX); in mptcp_established_options_mp_fail()
1158 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPFASTCLOSERX); in mptcp_incoming_options()
[all …]
H A Dmptcp_diag.c38 net = sock_net(in_skb->sk); in mptcp_diag_dump_one()
83 struct net *net = sock_net(skb->sk); in mptcp_diag_dump_listeners()
111 if (!sk || !net_eq(sock_net(sk), net)) in mptcp_diag_dump_listeners()
157 struct net *net = sock_net(skb->sk); in mptcp_diag_dump()
/openbmc/linux/net/vmw_vsock/
H A Ddiag.c59 net = sock_net(skb->sk); in vsock_diag_dump()
79 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump()
113 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump()
144 struct net *net = sock_net(skb->sk); in vsock_diag_handler_dump()
/openbmc/linux/net/can/
H A Draw.c279 if (!net_eq(dev_net(dev), sock_net(sk))) in raw_notify()
401 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_release()
447 dev = dev_get_by_index(sock_net(sk), addr->can_ifindex); in raw_bind()
463 err = raw_enable_allfilters(sock_net(sk), dev, sk); in raw_bind()
471 err = raw_enable_allfilters(sock_net(sk), NULL, sk); in raw_bind()
483 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_bind()
580 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt()
583 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt()
592 raw_disable_filters(sock_net(sk), dev, sk, ro->filter, in raw_setsockopt()
638 err = raw_enable_errfilter(sock_net(sk), dev, sk, in raw_setsockopt()
[all …]
/openbmc/linux/net/ieee802154/
H A Dsocket.c136 dev_load(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl()
137 dev = dev_get_by_name(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl()
178 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash()
187 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash()
214 dev = ieee802154_get_dev(sock_net(sk), &addr); in raw_bind()
256 dev = dev_getfirstbyhwtype(sock_net(sk), ARPHRD_IEEE802154); in raw_sendmsg()
258 dev = dev_get_by_index(sock_net(sk), sk->sk_bound_dev_if); in raw_sendmsg()
461 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in dgram_hash()
470 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in dgram_unhash()
510 dev = ieee802154_get_dev(sock_net(sk), &haddr); in dgram_bind()
[all …]
/openbmc/linux/drivers/infiniband/core/
H A Dnldev.c1033 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_get_doit()
1058 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_get_doit()
1081 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_set_doit()
1175 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_port_get_doit()
1199 err = fill_port_info(msg, device, port, sock_net(skb->sk)); in nldev_port_get_doit()
1206 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_port_get_doit()
1233 device = ib_device_get_by_index(sock_net(skb->sk), ifindex); in nldev_port_get_dumpit()
1259 if (!nlh || fill_port_info(skb, device, p, sock_net(skb->sk))) { in nldev_port_get_dumpit()
1288 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_res_get_doit()
1312 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_res_get_doit()
[all …]
/openbmc/linux/net/phonet/
H A Dsocket.c79 if (!net_eq(sock_net(sknode), net)) in pn_find_sock_by_sa()
116 if (!net_eq(sock_net(sknode), net)) in pn_deliver_sock_broadcast()
175 if (saddr && phonet_address_lookup(sock_net(sk), saddr)) in pn_socket_bind()
372 dev = dev_get_by_index(sock_net(sk), in pn_socket_ioctl()
375 dev = phonet_device_get(sock_net(sk)); in pn_socket_ioctl()
471 struct net *net = sock_net(sk); in pn_sock_get_port()
526 if (!net_eq(net, sock_net(sknode))) in pn_sock_get_idx()
543 while (sk && !net_eq(net, sock_net(sk))); in pn_sock_get_next()
633 if (!net_eq(sock_net(sk), &init_net)) in pn_sock_bind_res()
/openbmc/linux/net/mctp/
H A Daf_mctp.c130 addr->smctp_network = mctp_default_net(sock_net(sk)); in mctp_sendmsg()
140 dev = dev_get_by_index_rcu(sock_net(sk), extaddr->smctp_ifindex); in mctp_sendmsg()
151 rt = mctp_route_lookup(sock_net(sk), addr->smctp_network, in mctp_sendmsg()
355 struct net *net = sock_net(&msk->sk); in mctp_ioctl_alloctag()
397 struct net *net = sock_net(&msk->sk); in mctp_ioctl_droptag()
497 struct net *net = sock_net(&msk->sk); in mctp_sk_expire_keys()
551 struct net *net = sock_net(sk); in mctp_sk_hash()
563 struct net *net = sock_net(sk); in mctp_sk_unhash()
/openbmc/linux/net/netlink/
H A Ddiag.c94 struct net *net = sock_net(skb->sk); in __netlink_diag_dump()
134 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump()
160 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump()
228 struct net *net = sock_net(skb->sk); in netlink_diag_handler_dump()
/openbmc/linux/net/handshake/
H A Drequest.c95 trace_handshake_destruct(sock_net(sk), req, sk); in handshake_sk_destruct()
244 net = sock_net(req->hr_sk); in handshake_req_submit()
290 struct net *net = sock_net(sk); in handshake_complete()
319 net = sock_net(sk); in handshake_req_cancel()
/openbmc/linux/net/smc/
H A Dsmc_stats.h111 struct net *_net = sock_net(&__smc->sk); \
126 struct net *_net = sock_net(&__smc->sk); \
157 struct net *_net = sock_net(&(_smc)->sk); \
175 struct net *net = sock_net(&(_smc)->sk); \
221 struct net *net = sock_net(&(__smc)->sk); \

1234567891011