1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * linux/net/sunrpc/socklib.c
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Common socket helper routines for RPC client and server
6*4882a593Smuzhiyun *
7*4882a593Smuzhiyun * Copyright (C) 1995, 1996 Olaf Kirch <okir@monad.swb.de>
8*4882a593Smuzhiyun */
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #include <linux/compiler.h>
11*4882a593Smuzhiyun #include <linux/netdevice.h>
12*4882a593Smuzhiyun #include <linux/gfp.h>
13*4882a593Smuzhiyun #include <linux/skbuff.h>
14*4882a593Smuzhiyun #include <linux/types.h>
15*4882a593Smuzhiyun #include <linux/pagemap.h>
16*4882a593Smuzhiyun #include <linux/udp.h>
17*4882a593Smuzhiyun #include <linux/sunrpc/msg_prot.h>
18*4882a593Smuzhiyun #include <linux/sunrpc/xdr.h>
19*4882a593Smuzhiyun #include <linux/export.h>
20*4882a593Smuzhiyun
21*4882a593Smuzhiyun #include "socklib.h"
22*4882a593Smuzhiyun
23*4882a593Smuzhiyun /*
24*4882a593Smuzhiyun * Helper structure for copying from an sk_buff.
25*4882a593Smuzhiyun */
26*4882a593Smuzhiyun struct xdr_skb_reader {
27*4882a593Smuzhiyun struct sk_buff *skb;
28*4882a593Smuzhiyun unsigned int offset;
29*4882a593Smuzhiyun size_t count;
30*4882a593Smuzhiyun __wsum csum;
31*4882a593Smuzhiyun };
32*4882a593Smuzhiyun
33*4882a593Smuzhiyun typedef size_t (*xdr_skb_read_actor)(struct xdr_skb_reader *desc, void *to,
34*4882a593Smuzhiyun size_t len);
35*4882a593Smuzhiyun
36*4882a593Smuzhiyun /**
37*4882a593Smuzhiyun * xdr_skb_read_bits - copy some data bits from skb to internal buffer
38*4882a593Smuzhiyun * @desc: sk_buff copy helper
39*4882a593Smuzhiyun * @to: copy destination
40*4882a593Smuzhiyun * @len: number of bytes to copy
41*4882a593Smuzhiyun *
42*4882a593Smuzhiyun * Possibly called several times to iterate over an sk_buff and copy
43*4882a593Smuzhiyun * data out of it.
44*4882a593Smuzhiyun */
45*4882a593Smuzhiyun static size_t
xdr_skb_read_bits(struct xdr_skb_reader * desc,void * to,size_t len)46*4882a593Smuzhiyun xdr_skb_read_bits(struct xdr_skb_reader *desc, void *to, size_t len)
47*4882a593Smuzhiyun {
48*4882a593Smuzhiyun if (len > desc->count)
49*4882a593Smuzhiyun len = desc->count;
50*4882a593Smuzhiyun if (unlikely(skb_copy_bits(desc->skb, desc->offset, to, len)))
51*4882a593Smuzhiyun return 0;
52*4882a593Smuzhiyun desc->count -= len;
53*4882a593Smuzhiyun desc->offset += len;
54*4882a593Smuzhiyun return len;
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun
57*4882a593Smuzhiyun /**
58*4882a593Smuzhiyun * xdr_skb_read_and_csum_bits - copy and checksum from skb to buffer
59*4882a593Smuzhiyun * @desc: sk_buff copy helper
60*4882a593Smuzhiyun * @to: copy destination
61*4882a593Smuzhiyun * @len: number of bytes to copy
62*4882a593Smuzhiyun *
63*4882a593Smuzhiyun * Same as skb_read_bits, but calculate a checksum at the same time.
64*4882a593Smuzhiyun */
xdr_skb_read_and_csum_bits(struct xdr_skb_reader * desc,void * to,size_t len)65*4882a593Smuzhiyun static size_t xdr_skb_read_and_csum_bits(struct xdr_skb_reader *desc, void *to, size_t len)
66*4882a593Smuzhiyun {
67*4882a593Smuzhiyun unsigned int pos;
68*4882a593Smuzhiyun __wsum csum2;
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun if (len > desc->count)
71*4882a593Smuzhiyun len = desc->count;
72*4882a593Smuzhiyun pos = desc->offset;
73*4882a593Smuzhiyun csum2 = skb_copy_and_csum_bits(desc->skb, pos, to, len);
74*4882a593Smuzhiyun desc->csum = csum_block_add(desc->csum, csum2, pos);
75*4882a593Smuzhiyun desc->count -= len;
76*4882a593Smuzhiyun desc->offset += len;
77*4882a593Smuzhiyun return len;
78*4882a593Smuzhiyun }
79*4882a593Smuzhiyun
80*4882a593Smuzhiyun /**
81*4882a593Smuzhiyun * xdr_partial_copy_from_skb - copy data out of an skb
82*4882a593Smuzhiyun * @xdr: target XDR buffer
83*4882a593Smuzhiyun * @base: starting offset
84*4882a593Smuzhiyun * @desc: sk_buff copy helper
85*4882a593Smuzhiyun * @copy_actor: virtual method for copying data
86*4882a593Smuzhiyun *
87*4882a593Smuzhiyun */
88*4882a593Smuzhiyun static ssize_t
xdr_partial_copy_from_skb(struct xdr_buf * xdr,unsigned int base,struct xdr_skb_reader * desc,xdr_skb_read_actor copy_actor)89*4882a593Smuzhiyun xdr_partial_copy_from_skb(struct xdr_buf *xdr, unsigned int base, struct xdr_skb_reader *desc, xdr_skb_read_actor copy_actor)
90*4882a593Smuzhiyun {
91*4882a593Smuzhiyun struct page **ppage = xdr->pages;
92*4882a593Smuzhiyun unsigned int len, pglen = xdr->page_len;
93*4882a593Smuzhiyun ssize_t copied = 0;
94*4882a593Smuzhiyun size_t ret;
95*4882a593Smuzhiyun
96*4882a593Smuzhiyun len = xdr->head[0].iov_len;
97*4882a593Smuzhiyun if (base < len) {
98*4882a593Smuzhiyun len -= base;
99*4882a593Smuzhiyun ret = copy_actor(desc, (char *)xdr->head[0].iov_base + base, len);
100*4882a593Smuzhiyun copied += ret;
101*4882a593Smuzhiyun if (ret != len || !desc->count)
102*4882a593Smuzhiyun goto out;
103*4882a593Smuzhiyun base = 0;
104*4882a593Smuzhiyun } else
105*4882a593Smuzhiyun base -= len;
106*4882a593Smuzhiyun
107*4882a593Smuzhiyun if (unlikely(pglen == 0))
108*4882a593Smuzhiyun goto copy_tail;
109*4882a593Smuzhiyun if (unlikely(base >= pglen)) {
110*4882a593Smuzhiyun base -= pglen;
111*4882a593Smuzhiyun goto copy_tail;
112*4882a593Smuzhiyun }
113*4882a593Smuzhiyun if (base || xdr->page_base) {
114*4882a593Smuzhiyun pglen -= base;
115*4882a593Smuzhiyun base += xdr->page_base;
116*4882a593Smuzhiyun ppage += base >> PAGE_SHIFT;
117*4882a593Smuzhiyun base &= ~PAGE_MASK;
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun do {
120*4882a593Smuzhiyun char *kaddr;
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun /* ACL likes to be lazy in allocating pages - ACLs
123*4882a593Smuzhiyun * are small by default but can get huge. */
124*4882a593Smuzhiyun if ((xdr->flags & XDRBUF_SPARSE_PAGES) && *ppage == NULL) {
125*4882a593Smuzhiyun *ppage = alloc_page(GFP_NOWAIT | __GFP_NOWARN);
126*4882a593Smuzhiyun if (unlikely(*ppage == NULL)) {
127*4882a593Smuzhiyun if (copied == 0)
128*4882a593Smuzhiyun copied = -ENOMEM;
129*4882a593Smuzhiyun goto out;
130*4882a593Smuzhiyun }
131*4882a593Smuzhiyun }
132*4882a593Smuzhiyun
133*4882a593Smuzhiyun len = PAGE_SIZE;
134*4882a593Smuzhiyun kaddr = kmap_atomic(*ppage);
135*4882a593Smuzhiyun if (base) {
136*4882a593Smuzhiyun len -= base;
137*4882a593Smuzhiyun if (pglen < len)
138*4882a593Smuzhiyun len = pglen;
139*4882a593Smuzhiyun ret = copy_actor(desc, kaddr + base, len);
140*4882a593Smuzhiyun base = 0;
141*4882a593Smuzhiyun } else {
142*4882a593Smuzhiyun if (pglen < len)
143*4882a593Smuzhiyun len = pglen;
144*4882a593Smuzhiyun ret = copy_actor(desc, kaddr, len);
145*4882a593Smuzhiyun }
146*4882a593Smuzhiyun flush_dcache_page(*ppage);
147*4882a593Smuzhiyun kunmap_atomic(kaddr);
148*4882a593Smuzhiyun copied += ret;
149*4882a593Smuzhiyun if (ret != len || !desc->count)
150*4882a593Smuzhiyun goto out;
151*4882a593Smuzhiyun ppage++;
152*4882a593Smuzhiyun } while ((pglen -= len) != 0);
153*4882a593Smuzhiyun copy_tail:
154*4882a593Smuzhiyun len = xdr->tail[0].iov_len;
155*4882a593Smuzhiyun if (base < len)
156*4882a593Smuzhiyun copied += copy_actor(desc, (char *)xdr->tail[0].iov_base + base, len - base);
157*4882a593Smuzhiyun out:
158*4882a593Smuzhiyun return copied;
159*4882a593Smuzhiyun }
160*4882a593Smuzhiyun
161*4882a593Smuzhiyun /**
162*4882a593Smuzhiyun * csum_partial_copy_to_xdr - checksum and copy data
163*4882a593Smuzhiyun * @xdr: target XDR buffer
164*4882a593Smuzhiyun * @skb: source skb
165*4882a593Smuzhiyun *
166*4882a593Smuzhiyun * We have set things up such that we perform the checksum of the UDP
167*4882a593Smuzhiyun * packet in parallel with the copies into the RPC client iovec. -DaveM
168*4882a593Smuzhiyun */
csum_partial_copy_to_xdr(struct xdr_buf * xdr,struct sk_buff * skb)169*4882a593Smuzhiyun int csum_partial_copy_to_xdr(struct xdr_buf *xdr, struct sk_buff *skb)
170*4882a593Smuzhiyun {
171*4882a593Smuzhiyun struct xdr_skb_reader desc;
172*4882a593Smuzhiyun
173*4882a593Smuzhiyun desc.skb = skb;
174*4882a593Smuzhiyun desc.offset = 0;
175*4882a593Smuzhiyun desc.count = skb->len - desc.offset;
176*4882a593Smuzhiyun
177*4882a593Smuzhiyun if (skb_csum_unnecessary(skb))
178*4882a593Smuzhiyun goto no_checksum;
179*4882a593Smuzhiyun
180*4882a593Smuzhiyun desc.csum = csum_partial(skb->data, desc.offset, skb->csum);
181*4882a593Smuzhiyun if (xdr_partial_copy_from_skb(xdr, 0, &desc, xdr_skb_read_and_csum_bits) < 0)
182*4882a593Smuzhiyun return -1;
183*4882a593Smuzhiyun if (desc.offset != skb->len) {
184*4882a593Smuzhiyun __wsum csum2;
185*4882a593Smuzhiyun csum2 = skb_checksum(skb, desc.offset, skb->len - desc.offset, 0);
186*4882a593Smuzhiyun desc.csum = csum_block_add(desc.csum, csum2, desc.offset);
187*4882a593Smuzhiyun }
188*4882a593Smuzhiyun if (desc.count)
189*4882a593Smuzhiyun return -1;
190*4882a593Smuzhiyun if (csum_fold(desc.csum))
191*4882a593Smuzhiyun return -1;
192*4882a593Smuzhiyun if (unlikely(skb->ip_summed == CHECKSUM_COMPLETE) &&
193*4882a593Smuzhiyun !skb->csum_complete_sw)
194*4882a593Smuzhiyun netdev_rx_csum_fault(skb->dev, skb);
195*4882a593Smuzhiyun return 0;
196*4882a593Smuzhiyun no_checksum:
197*4882a593Smuzhiyun if (xdr_partial_copy_from_skb(xdr, 0, &desc, xdr_skb_read_bits) < 0)
198*4882a593Smuzhiyun return -1;
199*4882a593Smuzhiyun if (desc.count)
200*4882a593Smuzhiyun return -1;
201*4882a593Smuzhiyun return 0;
202*4882a593Smuzhiyun }
203*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(csum_partial_copy_to_xdr);
204*4882a593Smuzhiyun
xprt_sendmsg(struct socket * sock,struct msghdr * msg,size_t seek)205*4882a593Smuzhiyun static inline int xprt_sendmsg(struct socket *sock, struct msghdr *msg,
206*4882a593Smuzhiyun size_t seek)
207*4882a593Smuzhiyun {
208*4882a593Smuzhiyun if (seek)
209*4882a593Smuzhiyun iov_iter_advance(&msg->msg_iter, seek);
210*4882a593Smuzhiyun return sock_sendmsg(sock, msg);
211*4882a593Smuzhiyun }
212*4882a593Smuzhiyun
xprt_send_kvec(struct socket * sock,struct msghdr * msg,struct kvec * vec,size_t seek)213*4882a593Smuzhiyun static int xprt_send_kvec(struct socket *sock, struct msghdr *msg,
214*4882a593Smuzhiyun struct kvec *vec, size_t seek)
215*4882a593Smuzhiyun {
216*4882a593Smuzhiyun iov_iter_kvec(&msg->msg_iter, WRITE, vec, 1, vec->iov_len);
217*4882a593Smuzhiyun return xprt_sendmsg(sock, msg, seek);
218*4882a593Smuzhiyun }
219*4882a593Smuzhiyun
xprt_send_pagedata(struct socket * sock,struct msghdr * msg,struct xdr_buf * xdr,size_t base)220*4882a593Smuzhiyun static int xprt_send_pagedata(struct socket *sock, struct msghdr *msg,
221*4882a593Smuzhiyun struct xdr_buf *xdr, size_t base)
222*4882a593Smuzhiyun {
223*4882a593Smuzhiyun int err;
224*4882a593Smuzhiyun
225*4882a593Smuzhiyun err = xdr_alloc_bvec(xdr, GFP_KERNEL);
226*4882a593Smuzhiyun if (err < 0)
227*4882a593Smuzhiyun return err;
228*4882a593Smuzhiyun
229*4882a593Smuzhiyun iov_iter_bvec(&msg->msg_iter, WRITE, xdr->bvec, xdr_buf_pagecount(xdr),
230*4882a593Smuzhiyun xdr->page_len + xdr->page_base);
231*4882a593Smuzhiyun return xprt_sendmsg(sock, msg, base + xdr->page_base);
232*4882a593Smuzhiyun }
233*4882a593Smuzhiyun
234*4882a593Smuzhiyun /* Common case:
235*4882a593Smuzhiyun * - stream transport
236*4882a593Smuzhiyun * - sending from byte 0 of the message
237*4882a593Smuzhiyun * - the message is wholly contained in @xdr's head iovec
238*4882a593Smuzhiyun */
xprt_send_rm_and_kvec(struct socket * sock,struct msghdr * msg,rpc_fraghdr marker,struct kvec * vec,size_t base)239*4882a593Smuzhiyun static int xprt_send_rm_and_kvec(struct socket *sock, struct msghdr *msg,
240*4882a593Smuzhiyun rpc_fraghdr marker, struct kvec *vec,
241*4882a593Smuzhiyun size_t base)
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun struct kvec iov[2] = {
244*4882a593Smuzhiyun [0] = {
245*4882a593Smuzhiyun .iov_base = &marker,
246*4882a593Smuzhiyun .iov_len = sizeof(marker)
247*4882a593Smuzhiyun },
248*4882a593Smuzhiyun [1] = *vec,
249*4882a593Smuzhiyun };
250*4882a593Smuzhiyun size_t len = iov[0].iov_len + iov[1].iov_len;
251*4882a593Smuzhiyun
252*4882a593Smuzhiyun iov_iter_kvec(&msg->msg_iter, WRITE, iov, 2, len);
253*4882a593Smuzhiyun return xprt_sendmsg(sock, msg, base);
254*4882a593Smuzhiyun }
255*4882a593Smuzhiyun
256*4882a593Smuzhiyun /**
257*4882a593Smuzhiyun * xprt_sock_sendmsg - write an xdr_buf directly to a socket
258*4882a593Smuzhiyun * @sock: open socket to send on
259*4882a593Smuzhiyun * @msg: socket message metadata
260*4882a593Smuzhiyun * @xdr: xdr_buf containing this request
261*4882a593Smuzhiyun * @base: starting position in the buffer
262*4882a593Smuzhiyun * @marker: stream record marker field
263*4882a593Smuzhiyun * @sent_p: return the total number of bytes successfully queued for sending
264*4882a593Smuzhiyun *
265*4882a593Smuzhiyun * Return values:
266*4882a593Smuzhiyun * On success, returns zero and fills in @sent_p.
267*4882a593Smuzhiyun * %-ENOTSOCK if @sock is not a struct socket.
268*4882a593Smuzhiyun */
xprt_sock_sendmsg(struct socket * sock,struct msghdr * msg,struct xdr_buf * xdr,unsigned int base,rpc_fraghdr marker,unsigned int * sent_p)269*4882a593Smuzhiyun int xprt_sock_sendmsg(struct socket *sock, struct msghdr *msg,
270*4882a593Smuzhiyun struct xdr_buf *xdr, unsigned int base,
271*4882a593Smuzhiyun rpc_fraghdr marker, unsigned int *sent_p)
272*4882a593Smuzhiyun {
273*4882a593Smuzhiyun unsigned int rmsize = marker ? sizeof(marker) : 0;
274*4882a593Smuzhiyun unsigned int remainder = rmsize + xdr->len - base;
275*4882a593Smuzhiyun unsigned int want;
276*4882a593Smuzhiyun int err = 0;
277*4882a593Smuzhiyun
278*4882a593Smuzhiyun *sent_p = 0;
279*4882a593Smuzhiyun
280*4882a593Smuzhiyun if (unlikely(!sock))
281*4882a593Smuzhiyun return -ENOTSOCK;
282*4882a593Smuzhiyun
283*4882a593Smuzhiyun msg->msg_flags |= MSG_MORE;
284*4882a593Smuzhiyun want = xdr->head[0].iov_len + rmsize;
285*4882a593Smuzhiyun if (base < want) {
286*4882a593Smuzhiyun unsigned int len = want - base;
287*4882a593Smuzhiyun
288*4882a593Smuzhiyun remainder -= len;
289*4882a593Smuzhiyun if (remainder == 0)
290*4882a593Smuzhiyun msg->msg_flags &= ~MSG_MORE;
291*4882a593Smuzhiyun if (rmsize)
292*4882a593Smuzhiyun err = xprt_send_rm_and_kvec(sock, msg, marker,
293*4882a593Smuzhiyun &xdr->head[0], base);
294*4882a593Smuzhiyun else
295*4882a593Smuzhiyun err = xprt_send_kvec(sock, msg, &xdr->head[0], base);
296*4882a593Smuzhiyun if (remainder == 0 || err != len)
297*4882a593Smuzhiyun goto out;
298*4882a593Smuzhiyun *sent_p += err;
299*4882a593Smuzhiyun base = 0;
300*4882a593Smuzhiyun } else {
301*4882a593Smuzhiyun base -= want;
302*4882a593Smuzhiyun }
303*4882a593Smuzhiyun
304*4882a593Smuzhiyun if (base < xdr->page_len) {
305*4882a593Smuzhiyun unsigned int len = xdr->page_len - base;
306*4882a593Smuzhiyun
307*4882a593Smuzhiyun remainder -= len;
308*4882a593Smuzhiyun if (remainder == 0)
309*4882a593Smuzhiyun msg->msg_flags &= ~MSG_MORE;
310*4882a593Smuzhiyun err = xprt_send_pagedata(sock, msg, xdr, base);
311*4882a593Smuzhiyun if (remainder == 0 || err != len)
312*4882a593Smuzhiyun goto out;
313*4882a593Smuzhiyun *sent_p += err;
314*4882a593Smuzhiyun base = 0;
315*4882a593Smuzhiyun } else {
316*4882a593Smuzhiyun base -= xdr->page_len;
317*4882a593Smuzhiyun }
318*4882a593Smuzhiyun
319*4882a593Smuzhiyun if (base >= xdr->tail[0].iov_len)
320*4882a593Smuzhiyun return 0;
321*4882a593Smuzhiyun msg->msg_flags &= ~MSG_MORE;
322*4882a593Smuzhiyun err = xprt_send_kvec(sock, msg, &xdr->tail[0], base);
323*4882a593Smuzhiyun out:
324*4882a593Smuzhiyun if (err > 0) {
325*4882a593Smuzhiyun *sent_p += err;
326*4882a593Smuzhiyun err = 0;
327*4882a593Smuzhiyun }
328*4882a593Smuzhiyun return err;
329*4882a593Smuzhiyun }
330