1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * SUCS NET3:
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Generic stream handling routines. These are generic for most
6*4882a593Smuzhiyun * protocols. Even IP. Tonight 8-).
7*4882a593Smuzhiyun * This is used because TCP, LLC (others too) layer all have mostly
8*4882a593Smuzhiyun * identical sendmsg() and recvmsg() code.
9*4882a593Smuzhiyun * So we (will) share it here.
10*4882a593Smuzhiyun *
11*4882a593Smuzhiyun * Authors: Arnaldo Carvalho de Melo <acme@conectiva.com.br>
12*4882a593Smuzhiyun * (from old tcp.c code)
13*4882a593Smuzhiyun * Alan Cox <alan@lxorguk.ukuu.org.uk> (Borrowed comments 8-))
14*4882a593Smuzhiyun */
15*4882a593Smuzhiyun
16*4882a593Smuzhiyun #include <linux/module.h>
17*4882a593Smuzhiyun #include <linux/sched/signal.h>
18*4882a593Smuzhiyun #include <linux/net.h>
19*4882a593Smuzhiyun #include <linux/signal.h>
20*4882a593Smuzhiyun #include <linux/tcp.h>
21*4882a593Smuzhiyun #include <linux/wait.h>
22*4882a593Smuzhiyun #include <net/sock.h>
23*4882a593Smuzhiyun
24*4882a593Smuzhiyun /**
25*4882a593Smuzhiyun * sk_stream_write_space - stream socket write_space callback.
26*4882a593Smuzhiyun * @sk: socket
27*4882a593Smuzhiyun *
28*4882a593Smuzhiyun * FIXME: write proper description
29*4882a593Smuzhiyun */
sk_stream_write_space(struct sock * sk)30*4882a593Smuzhiyun void sk_stream_write_space(struct sock *sk)
31*4882a593Smuzhiyun {
32*4882a593Smuzhiyun struct socket *sock = sk->sk_socket;
33*4882a593Smuzhiyun struct socket_wq *wq;
34*4882a593Smuzhiyun
35*4882a593Smuzhiyun if (__sk_stream_is_writeable(sk, 1) && sock) {
36*4882a593Smuzhiyun clear_bit(SOCK_NOSPACE, &sock->flags);
37*4882a593Smuzhiyun
38*4882a593Smuzhiyun rcu_read_lock();
39*4882a593Smuzhiyun wq = rcu_dereference(sk->sk_wq);
40*4882a593Smuzhiyun if (skwq_has_sleeper(wq))
41*4882a593Smuzhiyun wake_up_interruptible_poll(&wq->wait, EPOLLOUT |
42*4882a593Smuzhiyun EPOLLWRNORM | EPOLLWRBAND);
43*4882a593Smuzhiyun if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN))
44*4882a593Smuzhiyun sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT);
45*4882a593Smuzhiyun rcu_read_unlock();
46*4882a593Smuzhiyun }
47*4882a593Smuzhiyun }
48*4882a593Smuzhiyun
49*4882a593Smuzhiyun /**
50*4882a593Smuzhiyun * sk_stream_wait_connect - Wait for a socket to get into the connected state
51*4882a593Smuzhiyun * @sk: sock to wait on
52*4882a593Smuzhiyun * @timeo_p: for how long to wait
53*4882a593Smuzhiyun *
54*4882a593Smuzhiyun * Must be called with the socket locked.
55*4882a593Smuzhiyun */
sk_stream_wait_connect(struct sock * sk,long * timeo_p)56*4882a593Smuzhiyun int sk_stream_wait_connect(struct sock *sk, long *timeo_p)
57*4882a593Smuzhiyun {
58*4882a593Smuzhiyun DEFINE_WAIT_FUNC(wait, woken_wake_function);
59*4882a593Smuzhiyun struct task_struct *tsk = current;
60*4882a593Smuzhiyun int done;
61*4882a593Smuzhiyun
62*4882a593Smuzhiyun do {
63*4882a593Smuzhiyun int err = sock_error(sk);
64*4882a593Smuzhiyun if (err)
65*4882a593Smuzhiyun return err;
66*4882a593Smuzhiyun if ((1 << sk->sk_state) & ~(TCPF_SYN_SENT | TCPF_SYN_RECV))
67*4882a593Smuzhiyun return -EPIPE;
68*4882a593Smuzhiyun if (!*timeo_p)
69*4882a593Smuzhiyun return -EAGAIN;
70*4882a593Smuzhiyun if (signal_pending(tsk))
71*4882a593Smuzhiyun return sock_intr_errno(*timeo_p);
72*4882a593Smuzhiyun
73*4882a593Smuzhiyun add_wait_queue(sk_sleep(sk), &wait);
74*4882a593Smuzhiyun sk->sk_write_pending++;
75*4882a593Smuzhiyun done = sk_wait_event(sk, timeo_p,
76*4882a593Smuzhiyun !sk->sk_err &&
77*4882a593Smuzhiyun !((1 << sk->sk_state) &
78*4882a593Smuzhiyun ~(TCPF_ESTABLISHED | TCPF_CLOSE_WAIT)), &wait);
79*4882a593Smuzhiyun remove_wait_queue(sk_sleep(sk), &wait);
80*4882a593Smuzhiyun sk->sk_write_pending--;
81*4882a593Smuzhiyun } while (!done);
82*4882a593Smuzhiyun return 0;
83*4882a593Smuzhiyun }
84*4882a593Smuzhiyun EXPORT_SYMBOL(sk_stream_wait_connect);
85*4882a593Smuzhiyun
86*4882a593Smuzhiyun /**
87*4882a593Smuzhiyun * sk_stream_closing - Return 1 if we still have things to send in our buffers.
88*4882a593Smuzhiyun * @sk: socket to verify
89*4882a593Smuzhiyun */
sk_stream_closing(struct sock * sk)90*4882a593Smuzhiyun static inline int sk_stream_closing(struct sock *sk)
91*4882a593Smuzhiyun {
92*4882a593Smuzhiyun return (1 << sk->sk_state) &
93*4882a593Smuzhiyun (TCPF_FIN_WAIT1 | TCPF_CLOSING | TCPF_LAST_ACK);
94*4882a593Smuzhiyun }
95*4882a593Smuzhiyun
sk_stream_wait_close(struct sock * sk,long timeout)96*4882a593Smuzhiyun void sk_stream_wait_close(struct sock *sk, long timeout)
97*4882a593Smuzhiyun {
98*4882a593Smuzhiyun if (timeout) {
99*4882a593Smuzhiyun DEFINE_WAIT_FUNC(wait, woken_wake_function);
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun add_wait_queue(sk_sleep(sk), &wait);
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun do {
104*4882a593Smuzhiyun if (sk_wait_event(sk, &timeout, !sk_stream_closing(sk), &wait))
105*4882a593Smuzhiyun break;
106*4882a593Smuzhiyun } while (!signal_pending(current) && timeout);
107*4882a593Smuzhiyun
108*4882a593Smuzhiyun remove_wait_queue(sk_sleep(sk), &wait);
109*4882a593Smuzhiyun }
110*4882a593Smuzhiyun }
111*4882a593Smuzhiyun EXPORT_SYMBOL(sk_stream_wait_close);
112*4882a593Smuzhiyun
113*4882a593Smuzhiyun /**
114*4882a593Smuzhiyun * sk_stream_wait_memory - Wait for more memory for a socket
115*4882a593Smuzhiyun * @sk: socket to wait for memory
116*4882a593Smuzhiyun * @timeo_p: for how long
117*4882a593Smuzhiyun */
sk_stream_wait_memory(struct sock * sk,long * timeo_p)118*4882a593Smuzhiyun int sk_stream_wait_memory(struct sock *sk, long *timeo_p)
119*4882a593Smuzhiyun {
120*4882a593Smuzhiyun int err = 0;
121*4882a593Smuzhiyun long vm_wait = 0;
122*4882a593Smuzhiyun long current_timeo = *timeo_p;
123*4882a593Smuzhiyun DEFINE_WAIT_FUNC(wait, woken_wake_function);
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun if (sk_stream_memory_free(sk))
126*4882a593Smuzhiyun current_timeo = vm_wait = (prandom_u32() % (HZ / 5)) + 2;
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun add_wait_queue(sk_sleep(sk), &wait);
129*4882a593Smuzhiyun
130*4882a593Smuzhiyun while (1) {
131*4882a593Smuzhiyun sk_set_bit(SOCKWQ_ASYNC_NOSPACE, sk);
132*4882a593Smuzhiyun
133*4882a593Smuzhiyun if (sk->sk_err || (sk->sk_shutdown & SEND_SHUTDOWN))
134*4882a593Smuzhiyun goto do_error;
135*4882a593Smuzhiyun if (!*timeo_p)
136*4882a593Smuzhiyun goto do_eagain;
137*4882a593Smuzhiyun if (signal_pending(current))
138*4882a593Smuzhiyun goto do_interrupted;
139*4882a593Smuzhiyun sk_clear_bit(SOCKWQ_ASYNC_NOSPACE, sk);
140*4882a593Smuzhiyun if (sk_stream_memory_free(sk) && !vm_wait)
141*4882a593Smuzhiyun break;
142*4882a593Smuzhiyun
143*4882a593Smuzhiyun set_bit(SOCK_NOSPACE, &sk->sk_socket->flags);
144*4882a593Smuzhiyun sk->sk_write_pending++;
145*4882a593Smuzhiyun sk_wait_event(sk, ¤t_timeo, sk->sk_err ||
146*4882a593Smuzhiyun (sk->sk_shutdown & SEND_SHUTDOWN) ||
147*4882a593Smuzhiyun (sk_stream_memory_free(sk) &&
148*4882a593Smuzhiyun !vm_wait), &wait);
149*4882a593Smuzhiyun sk->sk_write_pending--;
150*4882a593Smuzhiyun
151*4882a593Smuzhiyun if (vm_wait) {
152*4882a593Smuzhiyun vm_wait -= current_timeo;
153*4882a593Smuzhiyun current_timeo = *timeo_p;
154*4882a593Smuzhiyun if (current_timeo != MAX_SCHEDULE_TIMEOUT &&
155*4882a593Smuzhiyun (current_timeo -= vm_wait) < 0)
156*4882a593Smuzhiyun current_timeo = 0;
157*4882a593Smuzhiyun vm_wait = 0;
158*4882a593Smuzhiyun }
159*4882a593Smuzhiyun *timeo_p = current_timeo;
160*4882a593Smuzhiyun }
161*4882a593Smuzhiyun out:
162*4882a593Smuzhiyun if (!sock_flag(sk, SOCK_DEAD))
163*4882a593Smuzhiyun remove_wait_queue(sk_sleep(sk), &wait);
164*4882a593Smuzhiyun return err;
165*4882a593Smuzhiyun
166*4882a593Smuzhiyun do_error:
167*4882a593Smuzhiyun err = -EPIPE;
168*4882a593Smuzhiyun goto out;
169*4882a593Smuzhiyun do_eagain:
170*4882a593Smuzhiyun /* Make sure that whenever EAGAIN is returned, EPOLLOUT event can
171*4882a593Smuzhiyun * be generated later.
172*4882a593Smuzhiyun * When TCP receives ACK packets that make room, tcp_check_space()
173*4882a593Smuzhiyun * only calls tcp_new_space() if SOCK_NOSPACE is set.
174*4882a593Smuzhiyun */
175*4882a593Smuzhiyun set_bit(SOCK_NOSPACE, &sk->sk_socket->flags);
176*4882a593Smuzhiyun err = -EAGAIN;
177*4882a593Smuzhiyun goto out;
178*4882a593Smuzhiyun do_interrupted:
179*4882a593Smuzhiyun err = sock_intr_errno(*timeo_p);
180*4882a593Smuzhiyun goto out;
181*4882a593Smuzhiyun }
182*4882a593Smuzhiyun EXPORT_SYMBOL(sk_stream_wait_memory);
183*4882a593Smuzhiyun
sk_stream_error(struct sock * sk,int flags,int err)184*4882a593Smuzhiyun int sk_stream_error(struct sock *sk, int flags, int err)
185*4882a593Smuzhiyun {
186*4882a593Smuzhiyun if (err == -EPIPE)
187*4882a593Smuzhiyun err = sock_error(sk) ? : -EPIPE;
188*4882a593Smuzhiyun if (err == -EPIPE && !(flags & MSG_NOSIGNAL))
189*4882a593Smuzhiyun send_sig(SIGPIPE, current, 0);
190*4882a593Smuzhiyun return err;
191*4882a593Smuzhiyun }
192*4882a593Smuzhiyun EXPORT_SYMBOL(sk_stream_error);
193*4882a593Smuzhiyun
sk_stream_kill_queues(struct sock * sk)194*4882a593Smuzhiyun void sk_stream_kill_queues(struct sock *sk)
195*4882a593Smuzhiyun {
196*4882a593Smuzhiyun /* First the read buffer. */
197*4882a593Smuzhiyun __skb_queue_purge(&sk->sk_receive_queue);
198*4882a593Smuzhiyun
199*4882a593Smuzhiyun /* Next, the write queue. */
200*4882a593Smuzhiyun WARN_ON(!skb_queue_empty(&sk->sk_write_queue));
201*4882a593Smuzhiyun
202*4882a593Smuzhiyun /* Account for returned memory. */
203*4882a593Smuzhiyun sk_mem_reclaim(sk);
204*4882a593Smuzhiyun
205*4882a593Smuzhiyun WARN_ON(sk->sk_wmem_queued);
206*4882a593Smuzhiyun WARN_ON(sk->sk_forward_alloc);
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun /* It is _impossible_ for the backlog to contain anything
209*4882a593Smuzhiyun * when we get here. All user references to this socket
210*4882a593Smuzhiyun * have gone away, only the net layer knows can touch it.
211*4882a593Smuzhiyun */
212*4882a593Smuzhiyun }
213*4882a593Smuzhiyun EXPORT_SYMBOL(sk_stream_kill_queues);
214