/openbmc/linux/net/ipv4/ |
H A D | tcp_westwood.c | 191 w->accounted += tp->mss_cache; in westwood_acked_count() 192 w->cumul_ack = tp->mss_cache; in westwood_acked_count() 195 if (w->cumul_ack > tp->mss_cache) { in westwood_acked_count() 199 w->cumul_ack = tp->mss_cache; in westwood_acked_count() 222 return max_t(u32, (w->bw_est * w->rtt_min) / tp->mss_cache, 2); in tcp_westwood_bw_rttmin()
|
H A D | tcp_nv.c | 349 bytes_acked >= (ca->nv_min_cwnd - 1) * tp->mss_cache && in tcpnv_acked() 355 ca->nv_min_cwnd * tp->mss_cache; in tcpnv_acked() 367 80000ULL * tp->mss_cache); in tcpnv_acked()
|
H A D | tcp_dctcp.c | 231 info->dctcp.dctcp_ab_ecn = tp->mss_cache * in dctcp_get_info() 233 info->dctcp.dctcp_ab_tot = tp->mss_cache * in dctcp_get_info()
|
H A D | tcp_rate.c | 199 tp->write_seq - tp->snd_nxt < tp->mss_cache && in tcp_rate_check_app_limited()
|
H A D | tcp_bbr.c | 246 unsigned int mss = tcp_sk(sk)->mss_cache; in bbr_rate_bytes_per_sec() 314 segs = max_t(u32, bytes / tp->mss_cache, bbr_min_tso_segs(sk)); in bbr_tso_segs_goal() 1115 bw = bw * tp->mss_cache * USEC_PER_SEC >> BW_SCALE; in bbr_get_info()
|
H A D | tcp_output.c | 1818 tp->mss_cache = mss_now; in tcp_sync_mss() 1836 mss_now = tp->mss_cache; in tcp_current_mss() 2215 cong_win = (tcp_snd_cwnd(tp) - in_flight) * tp->mss_cache; in tcp_tso_should_defer() 2220 if (limit >= max_segs * tp->mss_cache) in tcp_tso_should_defer() 2229 u32 chunk = min(tp->snd_wnd, tcp_snd_cwnd(tp) * tp->mss_cache); in tcp_tso_should_defer() 2243 if (limit > tcp_max_tso_deferred_mss(tp) * tp->mss_cache) in tcp_tso_should_defer() 2419 size_needed = probe_size + (tp->reordering + 1) * tp->mss_cache; in tcp_mtu_probe()
|
H A D | tcp_input.c | 437 per_mss = max_t(u32, tp->rx_opt.mss_clamp, tp->mss_cache) + in tcp_sndbuf_expand() 631 unsigned int hint = min_t(unsigned int, tp->advmss, tp->mss_cache); in tcp_initialize_rcv_mss() 926 rate = (u64)tp->mss_cache * ((USEC_PER_SEC / 100) << 3); in tcp_update_pacing_rate() 1025 if (seq_len > tp->mss_cache) in tcp_dsack_seen() 1026 dup_segs = DIV_ROUND_UP(seq_len, tp->mss_cache); in tcp_dsack_seen() 1060 const u32 mss = tp->mss_cache; in tcp_check_sack_reordering() 2791 val = (u64)tcp_snd_cwnd(tp) * tcp_mss_to_mtu(sk, tp->mss_cache); in tcp_mtup_probe_success() 2975 tp->snd_una + tp->reordering * tp->mss_cache); in tcp_force_fast_retransmit() 3446 (tp->snd_una - prior_snd_una) < tp->mss_cache && in tcp_clean_rtx_queue() 3505 sample.in_flight = tp->mss_cache * in tcp_clean_rtx_queue()
|
H A D | tcp.c | 401 rate64 = (u64)rate * tp->mss_cache * USEC_PER_SEC; in tcp_compute_delivery_rate() 445 tp->mss_cache = TCP_MSS_DEFAULT; in tcp_init_sock() 3799 info->tcpi_snd_mss = tp->mss_cache; in tcp_get_info() 3995 val = tp->mss_cache; in do_tcp_getsockopt()
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | tcp_ca_write_sk_pacing.c | 46 ((tp->snd_cwnd * tp->mss_cache * USEC_PER_SEC) << 3) / in BPF_PROG()
|
H A D | test_sock_fields.c | 95 dst->mss_cache = src->mss_cache; in tpcpy()
|
/openbmc/linux/drivers/infiniband/sw/siw/ |
H A D | siw_qp_tx.c | 688 c_tx->tcp_seglen = tp->mss_cache * tp->gso_segs; in siw_update_tcpseg() 691 tp->mss_cache * in siw_update_tcpseg() 694 c_tx->tcp_seglen = tp->mss_cache; in siw_update_tcpseg()
|
/openbmc/linux/include/linux/ |
H A D | tcp.h | 239 u32 mss_cache; /* Cached effective mss, not including SACKS */ member
|
/openbmc/linux/tools/testing/selftests/bpf/prog_tests/ |
H A D | sock_fields.c | 93 tp->rcv_nxt, tp->snd_nxt, tp->snd_una, tp->mss_cache, in print_tp()
|
/openbmc/linux/include/uapi/linux/ |
H A D | bpf.h | 6227 __u32 mss_cache; /* Cached effective mss, not including SACKS */ member 6650 __u32 mss_cache; member
|
/openbmc/linux/tools/include/uapi/linux/ |
H A D | bpf.h | 6230 __u32 mss_cache; /* Cached effective mss, not including SACKS */ member 6653 __u32 mss_cache; member
|
/openbmc/linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls_io.c | 192 FLOWC_PARAM(MSS, tp->mss_cache); in send_tx_flowc_wr()
|
H A D | chtls_cm.c | 153 tp->mss_cache = tp->rx_opt.mss_clamp; in assign_rxopt()
|
/openbmc/linux/include/net/ |
H A D | tcp.h | 2248 tcp_sk(sk)->advmss, tcp_sk(sk)->mss_cache, in tcp_rto_delta_us()
|
/openbmc/linux/net/ipv6/ |
H A D | tcp_ipv6.c | 364 if (tcp_mtu_to_mss(sk, mtu) >= tcp_sk(sk)->mss_cache) in tcp_v6_mtu_reduced()
|
/openbmc/linux/net/core/ |
H A D | filter.c | 7150 case offsetof(struct bpf_tcp_sock, mss_cache): in bpf_tcp_sock_convert_ctx_access() 7151 BPF_TCP_SOCK_GET_COMMON(mss_cache); in bpf_tcp_sock_convert_ctx_access() 10602 case offsetof(struct bpf_sock_ops, mss_cache): in sock_ops_convert_ctx_access() 10603 SOCK_OPS_GET_TCP_SOCK_FIELD(mss_cache); in sock_ops_convert_ctx_access()
|