Lines Matching refs:flag

1002 	int	flag;  member
1028 state->flag |= FLAG_DSACK_TLP; in tcp_dsack_seen()
1042 if (tp->reord_seen && !(state->flag & FLAG_DSACK_TLP)) in tcp_dsack_seen()
1045 state->flag |= FLAG_DSACKING_ACK; in tcp_dsack_seen()
1415 state->flag |= FLAG_ORIG_SACK_ACKED; in tcp_sacktag_one()
1428 state->flag |= FLAG_DATA_SACKED; in tcp_sacktag_one()
1849 state->flag = 0; in tcp_sacktag_write_queue()
2025 return state->flag; in tcp_sacktag_write_queue()
2355 static bool tcp_time_to_recover(struct sock *sk, int flag) in tcp_time_to_recover() argument
2659 void tcp_cwnd_reduction(struct sock *sk, int newly_acked_sacked, int newly_lost, int flag) in tcp_cwnd_reduction() argument
2676 if (flag & FLAG_SND_UNA_ADVANCED && !newly_lost) in tcp_cwnd_reduction()
2729 static void tcp_try_to_open(struct sock *sk, int flag) in tcp_try_to_open() argument
2738 if (flag & FLAG_ECE) in tcp_try_to_open()
2878 static void tcp_process_loss(struct sock *sk, int flag, int num_dupack, in tcp_process_loss() argument
2884 if ((flag & FLAG_SND_UNA_ADVANCED || rcu_access_pointer(tp->fastopen_rsk)) && in tcp_process_loss()
2892 if ((flag & FLAG_ORIG_SACK_ACKED) && in tcp_process_loss()
2897 if (flag & FLAG_DATA_SACKED || num_dupack) in tcp_process_loss()
2899 } else if (flag & FLAG_SND_UNA_ADVANCED && !recovered) { in tcp_process_loss()
2924 tcp_add_reno_sack(sk, num_dupack, flag & FLAG_ECE); in tcp_process_loss()
2925 else if (flag & FLAG_SND_UNA_ADVANCED) in tcp_process_loss()
3009 int fast_rexmit = 0, flag = *ack_flag; in tcp_fastretrans_alert() local
3010 bool ece_ack = flag & FLAG_ECE; in tcp_fastretrans_alert()
3011 bool do_lost = num_dupack || ((flag & FLAG_DATA_SACKED) && in tcp_fastretrans_alert()
3058 if (!(flag & FLAG_SND_UNA_ADVANCED)) { in tcp_fastretrans_alert()
3065 tcp_try_to_open(sk, flag); in tcp_fastretrans_alert()
3069 if (!tcp_time_to_recover(sk, flag)) in tcp_fastretrans_alert()
3078 tcp_process_loss(sk, flag, num_dupack, rexmit); in tcp_fastretrans_alert()
3087 if (flag & FLAG_SND_UNA_ADVANCED) in tcp_fastretrans_alert()
3096 if (!tcp_time_to_recover(sk, flag)) { in tcp_fastretrans_alert()
3097 tcp_try_to_open(sk, flag); in tcp_fastretrans_alert()
3122 static void tcp_update_rtt_min(struct sock *sk, u32 rtt_us, const int flag) in tcp_update_rtt_min() argument
3127 if ((flag & FLAG_ACK_MAYBE_DELAYED) && rtt_us > tcp_min_rtt(tp)) { in tcp_update_rtt_min()
3138 static bool tcp_ack_update_rtt(struct sock *sk, const int flag, in tcp_ack_update_rtt() argument
3159 flag & FLAG_ACKED) { in tcp_ack_update_rtt()
3177 tcp_update_rtt_min(sk, ca_rtt_us, flag); in tcp_ack_update_rtt()
3305 int flag = 0; in tcp_clean_rtx_queue() local
3332 flag |= FLAG_RETRANS_DATA_ACKED; in tcp_clean_rtx_queue()
3342 flag |= FLAG_ORIG_SACK_ACKED; in tcp_clean_rtx_queue()
3368 flag |= FLAG_DATA_ACKED; in tcp_clean_rtx_queue()
3370 flag |= FLAG_SYN_ACKED; in tcp_clean_rtx_queue()
3397 flag |= FLAG_SACK_RENEGING; in tcp_clean_rtx_queue()
3400 if (likely(first_ackt) && !(flag & FLAG_RETRANS_DATA_ACKED)) { in tcp_clean_rtx_queue()
3407 !(flag & (FLAG_CA_ALERT | FLAG_SYN_ACKED))) { in tcp_clean_rtx_queue()
3412 flag |= FLAG_ACK_MAYBE_DELAYED; in tcp_clean_rtx_queue()
3419 rtt_update = tcp_ack_update_rtt(sk, flag, seq_rtt_us, sack_rtt_us, in tcp_clean_rtx_queue()
3422 if (flag & FLAG_ACKED) { in tcp_clean_rtx_queue()
3423 flag |= FLAG_SET_XMIT_TIMER; /* set TLP or RTO timer */ in tcp_clean_rtx_queue()
3438 if (flag & FLAG_RETRANS_DATA_ACKED) in tcp_clean_rtx_queue()
3439 flag &= ~FLAG_ORIG_SACK_ACKED; in tcp_clean_rtx_queue()
3457 flag |= FLAG_SET_XMIT_TIMER; /* set TLP or RTO timer */ in tcp_clean_rtx_queue()
3492 return flag; in tcp_clean_rtx_queue()
3519 static inline bool tcp_ack_is_dubious(const struct sock *sk, const int flag) in tcp_ack_is_dubious() argument
3521 return !(flag & FLAG_NOT_DUP) || (flag & FLAG_CA_ALERT) || in tcp_ack_is_dubious()
3526 static inline bool tcp_may_raise_cwnd(const struct sock *sk, const int flag) in tcp_may_raise_cwnd() argument
3536 return flag & FLAG_FORWARD_PROGRESS; in tcp_may_raise_cwnd()
3538 return flag & FLAG_DATA_ACKED; in tcp_may_raise_cwnd()
3547 int flag, const struct rate_sample *rs) in tcp_cong_control() argument
3558 tcp_cwnd_reduction(sk, acked_sacked, rs->losses, flag); in tcp_cong_control()
3559 } else if (tcp_may_raise_cwnd(sk, flag)) { in tcp_cong_control()
3607 int flag = 0; in tcp_ack_update_window() local
3614 flag |= FLAG_WIN_UPDATE; in tcp_ack_update_window()
3638 return flag; in tcp_ack_update_window()
3742 static void tcp_process_tlp_ack(struct sock *sk, u32 ack, int flag) in tcp_process_tlp_ack() argument
3752 } else if (flag & FLAG_DSACK_TLP) { in tcp_process_tlp_ack()
3765 } else if (!(flag & (FLAG_SND_UNA_ADVANCED | in tcp_process_tlp_ack()
3802 static u32 tcp_newly_delivered(struct sock *sk, u32 prior_delivered, int flag) in tcp_newly_delivered() argument
3810 if (flag & FLAG_ECE) in tcp_newly_delivered()
3817 static int tcp_ack(struct sock *sk, const struct sk_buff *skb, int flag) in tcp_ack() argument
3851 if (!(flag & FLAG_NO_CHALLENGE_ACK)) in tcp_ack()
3865 flag |= FLAG_SND_UNA_ADVANCED; in tcp_ack()
3881 if (flag & FLAG_UPDATE_TS_RECENT) in tcp_ack()
3884 if ((flag & (FLAG_SLOWPATH | FLAG_SND_UNA_ADVANCED)) == in tcp_ack()
3892 flag |= FLAG_WIN_UPDATE; in tcp_ack()
3901 flag |= FLAG_DATA; in tcp_ack()
3905 flag |= tcp_ack_update_window(sk, skb, ack, ack_seq); in tcp_ack()
3908 flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una, in tcp_ack()
3912 flag |= FLAG_ECE; in tcp_ack()
3918 flag & FLAG_ECE); in tcp_ack()
3920 if (flag & FLAG_WIN_UPDATE) in tcp_ack()
3945 flag |= tcp_clean_rtx_queue(sk, skb, prior_fack, prior_snd_una, in tcp_ack()
3946 &sack_state, flag & FLAG_ECE); in tcp_ack()
3951 tcp_process_tlp_ack(sk, ack, flag); in tcp_ack()
3953 if (tcp_ack_is_dubious(sk, flag)) { in tcp_ack()
3954 if (!(flag & (FLAG_SND_UNA_ADVANCED | in tcp_ack()
3958 if (!(flag & FLAG_DATA)) in tcp_ack()
3961 tcp_fastretrans_alert(sk, prior_snd_una, num_dupack, &flag, in tcp_ack()
3966 if (flag & FLAG_SET_XMIT_TIMER) in tcp_ack()
3969 if ((flag & FLAG_FORWARD_PROGRESS) || !(flag & FLAG_NOT_DUP)) in tcp_ack()
3972 delivered = tcp_newly_delivered(sk, delivered, flag); in tcp_ack()
3974 rs.is_ack_delayed = !!(flag & FLAG_ACK_MAYBE_DELAYED); in tcp_ack()
3976 tcp_cong_control(sk, ack, delivered, flag, sack_state.rate); in tcp_ack()
3982 if (flag & FLAG_DSACKING_ACK) { in tcp_ack()
3983 tcp_fastretrans_alert(sk, prior_snd_una, num_dupack, &flag, in tcp_ack()
3985 tcp_newly_delivered(sk, delivered, flag); in tcp_ack()
3994 tcp_process_tlp_ack(sk, ack, flag); in tcp_ack()
4002 flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una, in tcp_ack()
4004 tcp_fastretrans_alert(sk, prior_snd_una, num_dupack, &flag, in tcp_ack()
4006 tcp_newly_delivered(sk, delivered, flag); in tcp_ack()