xref: /OK3568_Linux_fs/kernel/net/sunrpc/socklib.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * linux/net/sunrpc/socklib.c
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * Common socket helper routines for RPC client and server
6*4882a593Smuzhiyun  *
7*4882a593Smuzhiyun  * Copyright (C) 1995, 1996 Olaf Kirch <okir@monad.swb.de>
8*4882a593Smuzhiyun  */
9*4882a593Smuzhiyun 
10*4882a593Smuzhiyun #include <linux/compiler.h>
11*4882a593Smuzhiyun #include <linux/netdevice.h>
12*4882a593Smuzhiyun #include <linux/gfp.h>
13*4882a593Smuzhiyun #include <linux/skbuff.h>
14*4882a593Smuzhiyun #include <linux/types.h>
15*4882a593Smuzhiyun #include <linux/pagemap.h>
16*4882a593Smuzhiyun #include <linux/udp.h>
17*4882a593Smuzhiyun #include <linux/sunrpc/msg_prot.h>
18*4882a593Smuzhiyun #include <linux/sunrpc/xdr.h>
19*4882a593Smuzhiyun #include <linux/export.h>
20*4882a593Smuzhiyun 
21*4882a593Smuzhiyun #include "socklib.h"
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun /*
24*4882a593Smuzhiyun  * Helper structure for copying from an sk_buff.
25*4882a593Smuzhiyun  */
26*4882a593Smuzhiyun struct xdr_skb_reader {
27*4882a593Smuzhiyun 	struct sk_buff	*skb;
28*4882a593Smuzhiyun 	unsigned int	offset;
29*4882a593Smuzhiyun 	size_t		count;
30*4882a593Smuzhiyun 	__wsum		csum;
31*4882a593Smuzhiyun };
32*4882a593Smuzhiyun 
33*4882a593Smuzhiyun typedef size_t (*xdr_skb_read_actor)(struct xdr_skb_reader *desc, void *to,
34*4882a593Smuzhiyun 				     size_t len);
35*4882a593Smuzhiyun 
36*4882a593Smuzhiyun /**
37*4882a593Smuzhiyun  * xdr_skb_read_bits - copy some data bits from skb to internal buffer
38*4882a593Smuzhiyun  * @desc: sk_buff copy helper
39*4882a593Smuzhiyun  * @to: copy destination
40*4882a593Smuzhiyun  * @len: number of bytes to copy
41*4882a593Smuzhiyun  *
42*4882a593Smuzhiyun  * Possibly called several times to iterate over an sk_buff and copy
43*4882a593Smuzhiyun  * data out of it.
44*4882a593Smuzhiyun  */
45*4882a593Smuzhiyun static size_t
xdr_skb_read_bits(struct xdr_skb_reader * desc,void * to,size_t len)46*4882a593Smuzhiyun xdr_skb_read_bits(struct xdr_skb_reader *desc, void *to, size_t len)
47*4882a593Smuzhiyun {
48*4882a593Smuzhiyun 	if (len > desc->count)
49*4882a593Smuzhiyun 		len = desc->count;
50*4882a593Smuzhiyun 	if (unlikely(skb_copy_bits(desc->skb, desc->offset, to, len)))
51*4882a593Smuzhiyun 		return 0;
52*4882a593Smuzhiyun 	desc->count -= len;
53*4882a593Smuzhiyun 	desc->offset += len;
54*4882a593Smuzhiyun 	return len;
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun 
57*4882a593Smuzhiyun /**
58*4882a593Smuzhiyun  * xdr_skb_read_and_csum_bits - copy and checksum from skb to buffer
59*4882a593Smuzhiyun  * @desc: sk_buff copy helper
60*4882a593Smuzhiyun  * @to: copy destination
61*4882a593Smuzhiyun  * @len: number of bytes to copy
62*4882a593Smuzhiyun  *
63*4882a593Smuzhiyun  * Same as skb_read_bits, but calculate a checksum at the same time.
64*4882a593Smuzhiyun  */
xdr_skb_read_and_csum_bits(struct xdr_skb_reader * desc,void * to,size_t len)65*4882a593Smuzhiyun static size_t xdr_skb_read_and_csum_bits(struct xdr_skb_reader *desc, void *to, size_t len)
66*4882a593Smuzhiyun {
67*4882a593Smuzhiyun 	unsigned int pos;
68*4882a593Smuzhiyun 	__wsum csum2;
69*4882a593Smuzhiyun 
70*4882a593Smuzhiyun 	if (len > desc->count)
71*4882a593Smuzhiyun 		len = desc->count;
72*4882a593Smuzhiyun 	pos = desc->offset;
73*4882a593Smuzhiyun 	csum2 = skb_copy_and_csum_bits(desc->skb, pos, to, len);
74*4882a593Smuzhiyun 	desc->csum = csum_block_add(desc->csum, csum2, pos);
75*4882a593Smuzhiyun 	desc->count -= len;
76*4882a593Smuzhiyun 	desc->offset += len;
77*4882a593Smuzhiyun 	return len;
78*4882a593Smuzhiyun }
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun /**
81*4882a593Smuzhiyun  * xdr_partial_copy_from_skb - copy data out of an skb
82*4882a593Smuzhiyun  * @xdr: target XDR buffer
83*4882a593Smuzhiyun  * @base: starting offset
84*4882a593Smuzhiyun  * @desc: sk_buff copy helper
85*4882a593Smuzhiyun  * @copy_actor: virtual method for copying data
86*4882a593Smuzhiyun  *
87*4882a593Smuzhiyun  */
88*4882a593Smuzhiyun static ssize_t
xdr_partial_copy_from_skb(struct xdr_buf * xdr,unsigned int base,struct xdr_skb_reader * desc,xdr_skb_read_actor copy_actor)89*4882a593Smuzhiyun xdr_partial_copy_from_skb(struct xdr_buf *xdr, unsigned int base, struct xdr_skb_reader *desc, xdr_skb_read_actor copy_actor)
90*4882a593Smuzhiyun {
91*4882a593Smuzhiyun 	struct page	**ppage = xdr->pages;
92*4882a593Smuzhiyun 	unsigned int	len, pglen = xdr->page_len;
93*4882a593Smuzhiyun 	ssize_t		copied = 0;
94*4882a593Smuzhiyun 	size_t		ret;
95*4882a593Smuzhiyun 
96*4882a593Smuzhiyun 	len = xdr->head[0].iov_len;
97*4882a593Smuzhiyun 	if (base < len) {
98*4882a593Smuzhiyun 		len -= base;
99*4882a593Smuzhiyun 		ret = copy_actor(desc, (char *)xdr->head[0].iov_base + base, len);
100*4882a593Smuzhiyun 		copied += ret;
101*4882a593Smuzhiyun 		if (ret != len || !desc->count)
102*4882a593Smuzhiyun 			goto out;
103*4882a593Smuzhiyun 		base = 0;
104*4882a593Smuzhiyun 	} else
105*4882a593Smuzhiyun 		base -= len;
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun 	if (unlikely(pglen == 0))
108*4882a593Smuzhiyun 		goto copy_tail;
109*4882a593Smuzhiyun 	if (unlikely(base >= pglen)) {
110*4882a593Smuzhiyun 		base -= pglen;
111*4882a593Smuzhiyun 		goto copy_tail;
112*4882a593Smuzhiyun 	}
113*4882a593Smuzhiyun 	if (base || xdr->page_base) {
114*4882a593Smuzhiyun 		pglen -= base;
115*4882a593Smuzhiyun 		base += xdr->page_base;
116*4882a593Smuzhiyun 		ppage += base >> PAGE_SHIFT;
117*4882a593Smuzhiyun 		base &= ~PAGE_MASK;
118*4882a593Smuzhiyun 	}
119*4882a593Smuzhiyun 	do {
120*4882a593Smuzhiyun 		char *kaddr;
121*4882a593Smuzhiyun 
122*4882a593Smuzhiyun 		/* ACL likes to be lazy in allocating pages - ACLs
123*4882a593Smuzhiyun 		 * are small by default but can get huge. */
124*4882a593Smuzhiyun 		if ((xdr->flags & XDRBUF_SPARSE_PAGES) && *ppage == NULL) {
125*4882a593Smuzhiyun 			*ppage = alloc_page(GFP_NOWAIT | __GFP_NOWARN);
126*4882a593Smuzhiyun 			if (unlikely(*ppage == NULL)) {
127*4882a593Smuzhiyun 				if (copied == 0)
128*4882a593Smuzhiyun 					copied = -ENOMEM;
129*4882a593Smuzhiyun 				goto out;
130*4882a593Smuzhiyun 			}
131*4882a593Smuzhiyun 		}
132*4882a593Smuzhiyun 
133*4882a593Smuzhiyun 		len = PAGE_SIZE;
134*4882a593Smuzhiyun 		kaddr = kmap_atomic(*ppage);
135*4882a593Smuzhiyun 		if (base) {
136*4882a593Smuzhiyun 			len -= base;
137*4882a593Smuzhiyun 			if (pglen < len)
138*4882a593Smuzhiyun 				len = pglen;
139*4882a593Smuzhiyun 			ret = copy_actor(desc, kaddr + base, len);
140*4882a593Smuzhiyun 			base = 0;
141*4882a593Smuzhiyun 		} else {
142*4882a593Smuzhiyun 			if (pglen < len)
143*4882a593Smuzhiyun 				len = pglen;
144*4882a593Smuzhiyun 			ret = copy_actor(desc, kaddr, len);
145*4882a593Smuzhiyun 		}
146*4882a593Smuzhiyun 		flush_dcache_page(*ppage);
147*4882a593Smuzhiyun 		kunmap_atomic(kaddr);
148*4882a593Smuzhiyun 		copied += ret;
149*4882a593Smuzhiyun 		if (ret != len || !desc->count)
150*4882a593Smuzhiyun 			goto out;
151*4882a593Smuzhiyun 		ppage++;
152*4882a593Smuzhiyun 	} while ((pglen -= len) != 0);
153*4882a593Smuzhiyun copy_tail:
154*4882a593Smuzhiyun 	len = xdr->tail[0].iov_len;
155*4882a593Smuzhiyun 	if (base < len)
156*4882a593Smuzhiyun 		copied += copy_actor(desc, (char *)xdr->tail[0].iov_base + base, len - base);
157*4882a593Smuzhiyun out:
158*4882a593Smuzhiyun 	return copied;
159*4882a593Smuzhiyun }
160*4882a593Smuzhiyun 
161*4882a593Smuzhiyun /**
162*4882a593Smuzhiyun  * csum_partial_copy_to_xdr - checksum and copy data
163*4882a593Smuzhiyun  * @xdr: target XDR buffer
164*4882a593Smuzhiyun  * @skb: source skb
165*4882a593Smuzhiyun  *
166*4882a593Smuzhiyun  * We have set things up such that we perform the checksum of the UDP
167*4882a593Smuzhiyun  * packet in parallel with the copies into the RPC client iovec.  -DaveM
168*4882a593Smuzhiyun  */
csum_partial_copy_to_xdr(struct xdr_buf * xdr,struct sk_buff * skb)169*4882a593Smuzhiyun int csum_partial_copy_to_xdr(struct xdr_buf *xdr, struct sk_buff *skb)
170*4882a593Smuzhiyun {
171*4882a593Smuzhiyun 	struct xdr_skb_reader	desc;
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun 	desc.skb = skb;
174*4882a593Smuzhiyun 	desc.offset = 0;
175*4882a593Smuzhiyun 	desc.count = skb->len - desc.offset;
176*4882a593Smuzhiyun 
177*4882a593Smuzhiyun 	if (skb_csum_unnecessary(skb))
178*4882a593Smuzhiyun 		goto no_checksum;
179*4882a593Smuzhiyun 
180*4882a593Smuzhiyun 	desc.csum = csum_partial(skb->data, desc.offset, skb->csum);
181*4882a593Smuzhiyun 	if (xdr_partial_copy_from_skb(xdr, 0, &desc, xdr_skb_read_and_csum_bits) < 0)
182*4882a593Smuzhiyun 		return -1;
183*4882a593Smuzhiyun 	if (desc.offset != skb->len) {
184*4882a593Smuzhiyun 		__wsum csum2;
185*4882a593Smuzhiyun 		csum2 = skb_checksum(skb, desc.offset, skb->len - desc.offset, 0);
186*4882a593Smuzhiyun 		desc.csum = csum_block_add(desc.csum, csum2, desc.offset);
187*4882a593Smuzhiyun 	}
188*4882a593Smuzhiyun 	if (desc.count)
189*4882a593Smuzhiyun 		return -1;
190*4882a593Smuzhiyun 	if (csum_fold(desc.csum))
191*4882a593Smuzhiyun 		return -1;
192*4882a593Smuzhiyun 	if (unlikely(skb->ip_summed == CHECKSUM_COMPLETE) &&
193*4882a593Smuzhiyun 	    !skb->csum_complete_sw)
194*4882a593Smuzhiyun 		netdev_rx_csum_fault(skb->dev, skb);
195*4882a593Smuzhiyun 	return 0;
196*4882a593Smuzhiyun no_checksum:
197*4882a593Smuzhiyun 	if (xdr_partial_copy_from_skb(xdr, 0, &desc, xdr_skb_read_bits) < 0)
198*4882a593Smuzhiyun 		return -1;
199*4882a593Smuzhiyun 	if (desc.count)
200*4882a593Smuzhiyun 		return -1;
201*4882a593Smuzhiyun 	return 0;
202*4882a593Smuzhiyun }
203*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(csum_partial_copy_to_xdr);
204*4882a593Smuzhiyun 
xprt_sendmsg(struct socket * sock,struct msghdr * msg,size_t seek)205*4882a593Smuzhiyun static inline int xprt_sendmsg(struct socket *sock, struct msghdr *msg,
206*4882a593Smuzhiyun 			       size_t seek)
207*4882a593Smuzhiyun {
208*4882a593Smuzhiyun 	if (seek)
209*4882a593Smuzhiyun 		iov_iter_advance(&msg->msg_iter, seek);
210*4882a593Smuzhiyun 	return sock_sendmsg(sock, msg);
211*4882a593Smuzhiyun }
212*4882a593Smuzhiyun 
xprt_send_kvec(struct socket * sock,struct msghdr * msg,struct kvec * vec,size_t seek)213*4882a593Smuzhiyun static int xprt_send_kvec(struct socket *sock, struct msghdr *msg,
214*4882a593Smuzhiyun 			  struct kvec *vec, size_t seek)
215*4882a593Smuzhiyun {
216*4882a593Smuzhiyun 	iov_iter_kvec(&msg->msg_iter, WRITE, vec, 1, vec->iov_len);
217*4882a593Smuzhiyun 	return xprt_sendmsg(sock, msg, seek);
218*4882a593Smuzhiyun }
219*4882a593Smuzhiyun 
xprt_send_pagedata(struct socket * sock,struct msghdr * msg,struct xdr_buf * xdr,size_t base)220*4882a593Smuzhiyun static int xprt_send_pagedata(struct socket *sock, struct msghdr *msg,
221*4882a593Smuzhiyun 			      struct xdr_buf *xdr, size_t base)
222*4882a593Smuzhiyun {
223*4882a593Smuzhiyun 	int err;
224*4882a593Smuzhiyun 
225*4882a593Smuzhiyun 	err = xdr_alloc_bvec(xdr, GFP_KERNEL);
226*4882a593Smuzhiyun 	if (err < 0)
227*4882a593Smuzhiyun 		return err;
228*4882a593Smuzhiyun 
229*4882a593Smuzhiyun 	iov_iter_bvec(&msg->msg_iter, WRITE, xdr->bvec, xdr_buf_pagecount(xdr),
230*4882a593Smuzhiyun 		      xdr->page_len + xdr->page_base);
231*4882a593Smuzhiyun 	return xprt_sendmsg(sock, msg, base + xdr->page_base);
232*4882a593Smuzhiyun }
233*4882a593Smuzhiyun 
234*4882a593Smuzhiyun /* Common case:
235*4882a593Smuzhiyun  *  - stream transport
236*4882a593Smuzhiyun  *  - sending from byte 0 of the message
237*4882a593Smuzhiyun  *  - the message is wholly contained in @xdr's head iovec
238*4882a593Smuzhiyun  */
xprt_send_rm_and_kvec(struct socket * sock,struct msghdr * msg,rpc_fraghdr marker,struct kvec * vec,size_t base)239*4882a593Smuzhiyun static int xprt_send_rm_and_kvec(struct socket *sock, struct msghdr *msg,
240*4882a593Smuzhiyun 				 rpc_fraghdr marker, struct kvec *vec,
241*4882a593Smuzhiyun 				 size_t base)
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun 	struct kvec iov[2] = {
244*4882a593Smuzhiyun 		[0] = {
245*4882a593Smuzhiyun 			.iov_base	= &marker,
246*4882a593Smuzhiyun 			.iov_len	= sizeof(marker)
247*4882a593Smuzhiyun 		},
248*4882a593Smuzhiyun 		[1] = *vec,
249*4882a593Smuzhiyun 	};
250*4882a593Smuzhiyun 	size_t len = iov[0].iov_len + iov[1].iov_len;
251*4882a593Smuzhiyun 
252*4882a593Smuzhiyun 	iov_iter_kvec(&msg->msg_iter, WRITE, iov, 2, len);
253*4882a593Smuzhiyun 	return xprt_sendmsg(sock, msg, base);
254*4882a593Smuzhiyun }
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun /**
257*4882a593Smuzhiyun  * xprt_sock_sendmsg - write an xdr_buf directly to a socket
258*4882a593Smuzhiyun  * @sock: open socket to send on
259*4882a593Smuzhiyun  * @msg: socket message metadata
260*4882a593Smuzhiyun  * @xdr: xdr_buf containing this request
261*4882a593Smuzhiyun  * @base: starting position in the buffer
262*4882a593Smuzhiyun  * @marker: stream record marker field
263*4882a593Smuzhiyun  * @sent_p: return the total number of bytes successfully queued for sending
264*4882a593Smuzhiyun  *
265*4882a593Smuzhiyun  * Return values:
266*4882a593Smuzhiyun  *   On success, returns zero and fills in @sent_p.
267*4882a593Smuzhiyun  *   %-ENOTSOCK if  @sock is not a struct socket.
268*4882a593Smuzhiyun  */
xprt_sock_sendmsg(struct socket * sock,struct msghdr * msg,struct xdr_buf * xdr,unsigned int base,rpc_fraghdr marker,unsigned int * sent_p)269*4882a593Smuzhiyun int xprt_sock_sendmsg(struct socket *sock, struct msghdr *msg,
270*4882a593Smuzhiyun 		      struct xdr_buf *xdr, unsigned int base,
271*4882a593Smuzhiyun 		      rpc_fraghdr marker, unsigned int *sent_p)
272*4882a593Smuzhiyun {
273*4882a593Smuzhiyun 	unsigned int rmsize = marker ? sizeof(marker) : 0;
274*4882a593Smuzhiyun 	unsigned int remainder = rmsize + xdr->len - base;
275*4882a593Smuzhiyun 	unsigned int want;
276*4882a593Smuzhiyun 	int err = 0;
277*4882a593Smuzhiyun 
278*4882a593Smuzhiyun 	*sent_p = 0;
279*4882a593Smuzhiyun 
280*4882a593Smuzhiyun 	if (unlikely(!sock))
281*4882a593Smuzhiyun 		return -ENOTSOCK;
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun 	msg->msg_flags |= MSG_MORE;
284*4882a593Smuzhiyun 	want = xdr->head[0].iov_len + rmsize;
285*4882a593Smuzhiyun 	if (base < want) {
286*4882a593Smuzhiyun 		unsigned int len = want - base;
287*4882a593Smuzhiyun 
288*4882a593Smuzhiyun 		remainder -= len;
289*4882a593Smuzhiyun 		if (remainder == 0)
290*4882a593Smuzhiyun 			msg->msg_flags &= ~MSG_MORE;
291*4882a593Smuzhiyun 		if (rmsize)
292*4882a593Smuzhiyun 			err = xprt_send_rm_and_kvec(sock, msg, marker,
293*4882a593Smuzhiyun 						    &xdr->head[0], base);
294*4882a593Smuzhiyun 		else
295*4882a593Smuzhiyun 			err = xprt_send_kvec(sock, msg, &xdr->head[0], base);
296*4882a593Smuzhiyun 		if (remainder == 0 || err != len)
297*4882a593Smuzhiyun 			goto out;
298*4882a593Smuzhiyun 		*sent_p += err;
299*4882a593Smuzhiyun 		base = 0;
300*4882a593Smuzhiyun 	} else {
301*4882a593Smuzhiyun 		base -= want;
302*4882a593Smuzhiyun 	}
303*4882a593Smuzhiyun 
304*4882a593Smuzhiyun 	if (base < xdr->page_len) {
305*4882a593Smuzhiyun 		unsigned int len = xdr->page_len - base;
306*4882a593Smuzhiyun 
307*4882a593Smuzhiyun 		remainder -= len;
308*4882a593Smuzhiyun 		if (remainder == 0)
309*4882a593Smuzhiyun 			msg->msg_flags &= ~MSG_MORE;
310*4882a593Smuzhiyun 		err = xprt_send_pagedata(sock, msg, xdr, base);
311*4882a593Smuzhiyun 		if (remainder == 0 || err != len)
312*4882a593Smuzhiyun 			goto out;
313*4882a593Smuzhiyun 		*sent_p += err;
314*4882a593Smuzhiyun 		base = 0;
315*4882a593Smuzhiyun 	} else {
316*4882a593Smuzhiyun 		base -= xdr->page_len;
317*4882a593Smuzhiyun 	}
318*4882a593Smuzhiyun 
319*4882a593Smuzhiyun 	if (base >= xdr->tail[0].iov_len)
320*4882a593Smuzhiyun 		return 0;
321*4882a593Smuzhiyun 	msg->msg_flags &= ~MSG_MORE;
322*4882a593Smuzhiyun 	err = xprt_send_kvec(sock, msg, &xdr->tail[0], base);
323*4882a593Smuzhiyun out:
324*4882a593Smuzhiyun 	if (err > 0) {
325*4882a593Smuzhiyun 		*sent_p += err;
326*4882a593Smuzhiyun 		err = 0;
327*4882a593Smuzhiyun 	}
328*4882a593Smuzhiyun 	return err;
329*4882a593Smuzhiyun }
330