Lines Matching refs:mss_now

61 static bool tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle,
1442 static void tcp_set_skb_tso_segs(struct sk_buff *skb, unsigned int mss_now) in tcp_set_skb_tso_segs() argument
1444 if (skb->len <= mss_now) { in tcp_set_skb_tso_segs()
1451 tcp_skb_pcount_set(skb, DIV_ROUND_UP(skb->len, mss_now)); in tcp_set_skb_tso_segs()
1452 TCP_SKB_CB(skb)->tcp_gso_size = mss_now; in tcp_set_skb_tso_segs()
1532 unsigned int mss_now, gfp_t gfp) in tcp_fragment() argument
1599 tcp_set_skb_tso_segs(skb, mss_now); in tcp_fragment()
1600 tcp_set_skb_tso_segs(buff, mss_now); in tcp_fragment()
1697 int mss_now; in __tcp_mtu_to_mss() local
1702 mss_now = pmtu - icsk->icsk_af_ops->net_header_len - sizeof(struct tcphdr); in __tcp_mtu_to_mss()
1709 mss_now -= icsk->icsk_af_ops->net_frag_header_len; in __tcp_mtu_to_mss()
1713 if (mss_now > tp->rx_opt.mss_clamp) in __tcp_mtu_to_mss()
1714 mss_now = tp->rx_opt.mss_clamp; in __tcp_mtu_to_mss()
1717 mss_now -= icsk->icsk_ext_hdr_len; in __tcp_mtu_to_mss()
1720 mss_now = max(mss_now, in __tcp_mtu_to_mss()
1722 return mss_now; in __tcp_mtu_to_mss()
1800 int mss_now; in tcp_sync_mss() local
1805 mss_now = tcp_mtu_to_mss(sk, pmtu); in tcp_sync_mss()
1806 mss_now = tcp_bound_to_half_wnd(tp, mss_now); in tcp_sync_mss()
1811 mss_now = min(mss_now, tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low)); in tcp_sync_mss()
1812 tp->mss_cache = mss_now; in tcp_sync_mss()
1814 return mss_now; in tcp_sync_mss()
1825 u32 mss_now; in tcp_current_mss() local
1830 mss_now = tp->mss_cache; in tcp_current_mss()
1835 mss_now = tcp_sync_mss(sk, mtu); in tcp_current_mss()
1846 mss_now -= delta; in tcp_current_mss()
1849 return mss_now; in tcp_current_mss()
1938 static void tcp_minshall_update(struct tcp_sock *tp, unsigned int mss_now, in tcp_minshall_update() argument
1941 if (skb->len < tcp_skb_pcount(skb) * mss_now) in tcp_minshall_update()
1963 static u32 tcp_tso_autosize(const struct sock *sk, unsigned int mss_now, in tcp_tso_autosize() argument
1977 segs = max_t(u32, bytes / mss_now, min_tso_segs); in tcp_tso_autosize()
1985 static u32 tcp_tso_segs(struct sock *sk, unsigned int mss_now) in tcp_tso_segs() argument
1994 tso_segs = tcp_tso_autosize(sk, mss_now, min_tso); in tcp_tso_segs()
2001 unsigned int mss_now, in tcp_mss_split_point() argument
2009 max_len = mss_now * max_segs; in tcp_mss_split_point()
2019 partial = needed % mss_now; in tcp_mss_split_point()
2059 static int tcp_init_tso_segs(struct sk_buff *skb, unsigned int mss_now) in tcp_init_tso_segs() argument
2063 if (!tso_segs || (tso_segs > 1 && tcp_skb_mss(skb) != mss_now)) { in tcp_init_tso_segs()
2064 tcp_set_skb_tso_segs(skb, mss_now); in tcp_init_tso_segs()
2117 unsigned int mss_now, gfp_t gfp) in tso_fragment() argument
2126 skb, len, mss_now, gfp); in tso_fragment()
2158 tcp_set_skb_tso_segs(skb, mss_now); in tso_fragment()
2159 tcp_set_skb_tso_segs(buff, mss_now); in tso_fragment()
2336 int mss_now; in tcp_mtu_probe() local
2355 mss_now = tcp_current_mss(sk); in tcp_mtu_probe()
2436 tcp_set_skb_tso_segs(skb, mss_now); in tcp_mtu_probe()
2599 static bool tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle, in tcp_write_xmit() argument
2623 max_segs = tcp_tso_segs(sk, mss_now); in tcp_write_xmit()
2631 tcp_init_tso_segs(skb, mss_now); in tcp_write_xmit()
2638 tso_segs = tcp_init_tso_segs(skb, mss_now); in tcp_write_xmit()
2650 if (unlikely(!tcp_snd_wnd_test(tp, skb, mss_now))) { in tcp_write_xmit()
2656 if (unlikely(!tcp_nagle_test(tp, skb, mss_now, in tcp_write_xmit()
2667 limit = mss_now; in tcp_write_xmit()
2669 limit = tcp_mss_split_point(sk, skb, mss_now, in tcp_write_xmit()
2676 unlikely(tso_fragment(sk, skb, limit, mss_now, gfp))) in tcp_write_xmit()
2699 tcp_minshall_update(tp, mss_now, skb); in tcp_write_xmit()
2878 void tcp_push_one(struct sock *sk, unsigned int mss_now) in tcp_push_one() argument
2882 BUG_ON(!skb || skb->len < mss_now); in tcp_push_one()
2884 tcp_write_xmit(sk, mss_now, TCP_NAGLE_PUSH, 1, sk->sk_allocation); in tcp_push_one()