diff options
| author | Eric Dumazet <edumazet@google.com> | 2012-05-16 23:15:34 +0000 | 
|---|---|---|
| committer | David S. Miller <davem@davemloft.net> | 2012-05-17 14:59:59 -0400 | 
| commit | a2a385d627e1549da4b43a8b3dfe370589766e1c (patch) | |
| tree | d61e9913497c6c14406032f6a0822738707f1abf /net/ipv4/tcp_output.c | |
| parent | e005d193d55ee5f757b13306112d8c23aac27a88 (diff) | |
| download | olio-linux-3.10-a2a385d627e1549da4b43a8b3dfe370589766e1c.tar.xz olio-linux-3.10-a2a385d627e1549da4b43a8b3dfe370589766e1c.zip  | |
tcp: bool conversions
bool conversions where possible.
__inline__ -> inline
space cleanups
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net/ipv4/tcp_output.c')
| -rw-r--r-- | net/ipv4/tcp_output.c | 75 | 
1 files changed, 38 insertions, 37 deletions
diff --git a/net/ipv4/tcp_output.c b/net/ipv4/tcp_output.c index 1a630825c45..803cbfe82fb 100644 --- a/net/ipv4/tcp_output.c +++ b/net/ipv4/tcp_output.c @@ -370,7 +370,7 @@ static void tcp_init_nondata_skb(struct sk_buff *skb, u32 seq, u8 flags)  	TCP_SKB_CB(skb)->end_seq = seq;  } -static inline int tcp_urg_mode(const struct tcp_sock *tp) +static inline bool tcp_urg_mode(const struct tcp_sock *tp)  {  	return tp->snd_una != tp->snd_up;  } @@ -1391,20 +1391,20 @@ static int tcp_init_tso_segs(const struct sock *sk, struct sk_buff *skb,  }  /* Minshall's variant of the Nagle send check. */ -static inline int tcp_minshall_check(const struct tcp_sock *tp) +static inline bool tcp_minshall_check(const struct tcp_sock *tp)  {  	return after(tp->snd_sml, tp->snd_una) &&  		!after(tp->snd_sml, tp->snd_nxt);  } -/* Return 0, if packet can be sent now without violation Nagle's rules: +/* Return false, if packet can be sent now without violation Nagle's rules:   * 1. It is full sized.   * 2. Or it contains FIN. (already checked by caller)   * 3. Or TCP_CORK is not set, and TCP_NODELAY is set.   * 4. Or TCP_CORK is not set, and all sent packets are ACKed.   *    With Minshall's modification: all sent small packets are ACKed.   */ -static inline int tcp_nagle_check(const struct tcp_sock *tp, +static inline bool tcp_nagle_check(const struct tcp_sock *tp,  				  const struct sk_buff *skb,  				  unsigned int mss_now, int nonagle)  { @@ -1413,11 +1413,11 @@ static inline int tcp_nagle_check(const struct tcp_sock *tp,  		 (!nonagle && tp->packets_out && tcp_minshall_check(tp)));  } -/* Return non-zero if the Nagle test allows this packet to be +/* Return true if the Nagle test allows this packet to be   * sent now.   */ -static inline int tcp_nagle_test(const struct tcp_sock *tp, const struct sk_buff *skb, -				 unsigned int cur_mss, int nonagle) +static inline bool tcp_nagle_test(const struct tcp_sock *tp, const struct sk_buff *skb, +				  unsigned int cur_mss, int nonagle)  {  	/* Nagle rule does not apply to frames, which sit in the middle of the  	 * write_queue (they have no chances to get new data). @@ -1426,24 +1426,25 @@ static inline int tcp_nagle_test(const struct tcp_sock *tp, const struct sk_buff  	 * argument based upon the location of SKB in the send queue.  	 */  	if (nonagle & TCP_NAGLE_PUSH) -		return 1; +		return true;  	/* Don't use the nagle rule for urgent data (or for the final FIN).  	 * Nagle can be ignored during F-RTO too (see RFC4138).  	 */  	if (tcp_urg_mode(tp) || (tp->frto_counter == 2) ||  	    (TCP_SKB_CB(skb)->tcp_flags & TCPHDR_FIN)) -		return 1; +		return true;  	if (!tcp_nagle_check(tp, skb, cur_mss, nonagle)) -		return 1; +		return true; -	return 0; +	return false;  }  /* Does at least the first segment of SKB fit into the send window? */ -static inline int tcp_snd_wnd_test(const struct tcp_sock *tp, const struct sk_buff *skb, -				   unsigned int cur_mss) +static bool tcp_snd_wnd_test(const struct tcp_sock *tp, +			     const struct sk_buff *skb, +			     unsigned int cur_mss)  {  	u32 end_seq = TCP_SKB_CB(skb)->end_seq; @@ -1476,7 +1477,7 @@ static unsigned int tcp_snd_test(const struct sock *sk, struct sk_buff *skb,  }  /* Test if sending is allowed right now. */ -int tcp_may_send_now(struct sock *sk) +bool tcp_may_send_now(struct sock *sk)  {  	const struct tcp_sock *tp = tcp_sk(sk);  	struct sk_buff *skb = tcp_send_head(sk); @@ -1546,7 +1547,7 @@ static int tso_fragment(struct sock *sk, struct sk_buff *skb, unsigned int len,   *   * This algorithm is from John Heffner.   */ -static int tcp_tso_should_defer(struct sock *sk, struct sk_buff *skb) +static bool tcp_tso_should_defer(struct sock *sk, struct sk_buff *skb)  {  	struct tcp_sock *tp = tcp_sk(sk);  	const struct inet_connection_sock *icsk = inet_csk(sk); @@ -1606,11 +1607,11 @@ static int tcp_tso_should_defer(struct sock *sk, struct sk_buff *skb)  	/* Ok, it looks like it is advisable to defer.  */  	tp->tso_deferred = 1 | (jiffies << 1); -	return 1; +	return true;  send_now:  	tp->tso_deferred = 0; -	return 0; +	return false;  }  /* Create a new MTU probe if we are ready. @@ -1752,11 +1753,11 @@ static int tcp_mtu_probe(struct sock *sk)   * snd_up-64k-mss .. snd_up cannot be large. However, taking into   * account rare use of URG, this is not a big flaw.   * - * Returns 1, if no segments are in flight and we have queued segments, but - * cannot send anything now because of SWS or another problem. + * Returns true, if no segments are in flight and we have queued segments, + * but cannot send anything now because of SWS or another problem.   */ -static int tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle, -			  int push_one, gfp_t gfp) +static bool tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle, +			   int push_one, gfp_t gfp)  {  	struct tcp_sock *tp = tcp_sk(sk);  	struct sk_buff *skb; @@ -1770,7 +1771,7 @@ static int tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle,  		/* Do MTU probing. */  		result = tcp_mtu_probe(sk);  		if (!result) { -			return 0; +			return false;  		} else if (result > 0) {  			sent_pkts = 1;  		} @@ -1829,7 +1830,7 @@ static int tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle,  	if (likely(sent_pkts)) {  		tcp_cwnd_validate(sk); -		return 0; +		return false;  	}  	return !tp->packets_out && tcp_send_head(sk);  } @@ -2028,22 +2029,22 @@ static void tcp_collapse_retrans(struct sock *sk, struct sk_buff *skb)  }  /* Check if coalescing SKBs is legal. */ -static int tcp_can_collapse(const struct sock *sk, const struct sk_buff *skb) +static bool tcp_can_collapse(const struct sock *sk, const struct sk_buff *skb)  {  	if (tcp_skb_pcount(skb) > 1) -		return 0; +		return false;  	/* TODO: SACK collapsing could be used to remove this condition */  	if (skb_shinfo(skb)->nr_frags != 0) -		return 0; +		return false;  	if (skb_cloned(skb)) -		return 0; +		return false;  	if (skb == tcp_send_head(sk)) -		return 0; +		return false;  	/* Some heurestics for collapsing over SACK'd could be invented */  	if (TCP_SKB_CB(skb)->sacked & TCPCB_SACKED_ACKED) -		return 0; +		return false; -	return 1; +	return true;  }  /* Collapse packets in the retransmit queue to make to create @@ -2054,7 +2055,7 @@ static void tcp_retrans_try_collapse(struct sock *sk, struct sk_buff *to,  {  	struct tcp_sock *tp = tcp_sk(sk);  	struct sk_buff *skb = to, *tmp; -	int first = 1; +	bool first = true;  	if (!sysctl_tcp_retrans_collapse)  		return; @@ -2068,7 +2069,7 @@ static void tcp_retrans_try_collapse(struct sock *sk, struct sk_buff *to,  		space -= skb->len;  		if (first) { -			first = 0; +			first = false;  			continue;  		} @@ -2208,18 +2209,18 @@ int tcp_retransmit_skb(struct sock *sk, struct sk_buff *skb)  /* Check if we forward retransmits are possible in the current   * window/congestion state.   */ -static int tcp_can_forward_retransmit(struct sock *sk) +static bool tcp_can_forward_retransmit(struct sock *sk)  {  	const struct inet_connection_sock *icsk = inet_csk(sk);  	const struct tcp_sock *tp = tcp_sk(sk);  	/* Forward retransmissions are possible only during Recovery. */  	if (icsk->icsk_ca_state != TCP_CA_Recovery) -		return 0; +		return false;  	/* No forward retransmissions in Reno are possible. */  	if (tcp_is_reno(tp)) -		return 0; +		return false;  	/* Yeah, we have to make difficult choice between forward transmission  	 * and retransmission... Both ways have their merits... @@ -2230,9 +2231,9 @@ static int tcp_can_forward_retransmit(struct sock *sk)  	 */  	if (tcp_may_send_now(sk)) -		return 0; +		return false; -	return 1; +	return true;  }  /* This gets called after a retransmit timeout, and the initially  |