xref: /OK3568_Linux_fs/kernel/drivers/md/dm-log-userspace-transfer.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * Copyright (C) 2006-2009 Red Hat, Inc.
3*4882a593Smuzhiyun  *
4*4882a593Smuzhiyun  * This file is released under the LGPL.
5*4882a593Smuzhiyun  */
6*4882a593Smuzhiyun 
7*4882a593Smuzhiyun #include <linux/kernel.h>
8*4882a593Smuzhiyun #include <linux/module.h>
9*4882a593Smuzhiyun #include <linux/slab.h>
10*4882a593Smuzhiyun #include <net/sock.h>
11*4882a593Smuzhiyun #include <linux/workqueue.h>
12*4882a593Smuzhiyun #include <linux/connector.h>
13*4882a593Smuzhiyun #include <linux/device-mapper.h>
14*4882a593Smuzhiyun #include <linux/dm-log-userspace.h>
15*4882a593Smuzhiyun 
16*4882a593Smuzhiyun #include "dm-log-userspace-transfer.h"
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun static uint32_t dm_ulog_seq;
19*4882a593Smuzhiyun 
20*4882a593Smuzhiyun /*
21*4882a593Smuzhiyun  * Netlink/Connector is an unreliable protocol.  How long should
22*4882a593Smuzhiyun  * we wait for a response before assuming it was lost and retrying?
23*4882a593Smuzhiyun  * (If we do receive a response after this time, it will be discarded
24*4882a593Smuzhiyun  * and the response to the resent request will be waited for.
25*4882a593Smuzhiyun  */
26*4882a593Smuzhiyun #define DM_ULOG_RETRY_TIMEOUT (15 * HZ)
27*4882a593Smuzhiyun 
28*4882a593Smuzhiyun /*
29*4882a593Smuzhiyun  * Pre-allocated space for speed
30*4882a593Smuzhiyun  */
31*4882a593Smuzhiyun #define DM_ULOG_PREALLOCED_SIZE 512
32*4882a593Smuzhiyun static struct cn_msg *prealloced_cn_msg;
33*4882a593Smuzhiyun static struct dm_ulog_request *prealloced_ulog_tfr;
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun static struct cb_id ulog_cn_id = {
36*4882a593Smuzhiyun 	.idx = CN_IDX_DM,
37*4882a593Smuzhiyun 	.val = CN_VAL_DM_USERSPACE_LOG
38*4882a593Smuzhiyun };
39*4882a593Smuzhiyun 
40*4882a593Smuzhiyun static DEFINE_MUTEX(dm_ulog_lock);
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun struct receiving_pkg {
43*4882a593Smuzhiyun 	struct list_head list;
44*4882a593Smuzhiyun 	struct completion complete;
45*4882a593Smuzhiyun 
46*4882a593Smuzhiyun 	uint32_t seq;
47*4882a593Smuzhiyun 
48*4882a593Smuzhiyun 	int error;
49*4882a593Smuzhiyun 	size_t *data_size;
50*4882a593Smuzhiyun 	char *data;
51*4882a593Smuzhiyun };
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun static DEFINE_SPINLOCK(receiving_list_lock);
54*4882a593Smuzhiyun static struct list_head receiving_list;
55*4882a593Smuzhiyun 
dm_ulog_sendto_server(struct dm_ulog_request * tfr)56*4882a593Smuzhiyun static int dm_ulog_sendto_server(struct dm_ulog_request *tfr)
57*4882a593Smuzhiyun {
58*4882a593Smuzhiyun 	int r;
59*4882a593Smuzhiyun 	struct cn_msg *msg = prealloced_cn_msg;
60*4882a593Smuzhiyun 
61*4882a593Smuzhiyun 	memset(msg, 0, sizeof(struct cn_msg));
62*4882a593Smuzhiyun 
63*4882a593Smuzhiyun 	msg->id.idx = ulog_cn_id.idx;
64*4882a593Smuzhiyun 	msg->id.val = ulog_cn_id.val;
65*4882a593Smuzhiyun 	msg->ack = 0;
66*4882a593Smuzhiyun 	msg->seq = tfr->seq;
67*4882a593Smuzhiyun 	msg->len = sizeof(struct dm_ulog_request) + tfr->data_size;
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun 	r = cn_netlink_send(msg, 0, 0, gfp_any());
70*4882a593Smuzhiyun 
71*4882a593Smuzhiyun 	return r;
72*4882a593Smuzhiyun }
73*4882a593Smuzhiyun 
74*4882a593Smuzhiyun /*
75*4882a593Smuzhiyun  * Parameters for this function can be either msg or tfr, but not
76*4882a593Smuzhiyun  * both.  This function fills in the reply for a waiting request.
77*4882a593Smuzhiyun  * If just msg is given, then the reply is simply an ACK from userspace
78*4882a593Smuzhiyun  * that the request was received.
79*4882a593Smuzhiyun  *
80*4882a593Smuzhiyun  * Returns: 0 on success, -ENOENT on failure
81*4882a593Smuzhiyun  */
fill_pkg(struct cn_msg * msg,struct dm_ulog_request * tfr)82*4882a593Smuzhiyun static int fill_pkg(struct cn_msg *msg, struct dm_ulog_request *tfr)
83*4882a593Smuzhiyun {
84*4882a593Smuzhiyun 	uint32_t rtn_seq = (msg) ? msg->seq : (tfr) ? tfr->seq : 0;
85*4882a593Smuzhiyun 	struct receiving_pkg *pkg;
86*4882a593Smuzhiyun 
87*4882a593Smuzhiyun 	/*
88*4882a593Smuzhiyun 	 * The 'receiving_pkg' entries in this list are statically
89*4882a593Smuzhiyun 	 * allocated on the stack in 'dm_consult_userspace'.
90*4882a593Smuzhiyun 	 * Each process that is waiting for a reply from the user
91*4882a593Smuzhiyun 	 * space server will have an entry in this list.
92*4882a593Smuzhiyun 	 *
93*4882a593Smuzhiyun 	 * We are safe to do it this way because the stack space
94*4882a593Smuzhiyun 	 * is unique to each process, but still addressable by
95*4882a593Smuzhiyun 	 * other processes.
96*4882a593Smuzhiyun 	 */
97*4882a593Smuzhiyun 	list_for_each_entry(pkg, &receiving_list, list) {
98*4882a593Smuzhiyun 		if (rtn_seq != pkg->seq)
99*4882a593Smuzhiyun 			continue;
100*4882a593Smuzhiyun 
101*4882a593Smuzhiyun 		if (msg) {
102*4882a593Smuzhiyun 			pkg->error = -msg->ack;
103*4882a593Smuzhiyun 			/*
104*4882a593Smuzhiyun 			 * If we are trying again, we will need to know our
105*4882a593Smuzhiyun 			 * storage capacity.  Otherwise, along with the
106*4882a593Smuzhiyun 			 * error code, we make explicit that we have no data.
107*4882a593Smuzhiyun 			 */
108*4882a593Smuzhiyun 			if (pkg->error != -EAGAIN)
109*4882a593Smuzhiyun 				*(pkg->data_size) = 0;
110*4882a593Smuzhiyun 		} else if (tfr->data_size > *(pkg->data_size)) {
111*4882a593Smuzhiyun 			DMERR("Insufficient space to receive package [%u] "
112*4882a593Smuzhiyun 			      "(%u vs %zu)", tfr->request_type,
113*4882a593Smuzhiyun 			      tfr->data_size, *(pkg->data_size));
114*4882a593Smuzhiyun 
115*4882a593Smuzhiyun 			*(pkg->data_size) = 0;
116*4882a593Smuzhiyun 			pkg->error = -ENOSPC;
117*4882a593Smuzhiyun 		} else {
118*4882a593Smuzhiyun 			pkg->error = tfr->error;
119*4882a593Smuzhiyun 			memcpy(pkg->data, tfr->data, tfr->data_size);
120*4882a593Smuzhiyun 			*(pkg->data_size) = tfr->data_size;
121*4882a593Smuzhiyun 		}
122*4882a593Smuzhiyun 		complete(&pkg->complete);
123*4882a593Smuzhiyun 		return 0;
124*4882a593Smuzhiyun 	}
125*4882a593Smuzhiyun 
126*4882a593Smuzhiyun 	return -ENOENT;
127*4882a593Smuzhiyun }
128*4882a593Smuzhiyun 
129*4882a593Smuzhiyun /*
130*4882a593Smuzhiyun  * This is the connector callback that delivers data
131*4882a593Smuzhiyun  * that was sent from userspace.
132*4882a593Smuzhiyun  */
cn_ulog_callback(struct cn_msg * msg,struct netlink_skb_parms * nsp)133*4882a593Smuzhiyun static void cn_ulog_callback(struct cn_msg *msg, struct netlink_skb_parms *nsp)
134*4882a593Smuzhiyun {
135*4882a593Smuzhiyun 	struct dm_ulog_request *tfr = (struct dm_ulog_request *)(msg + 1);
136*4882a593Smuzhiyun 
137*4882a593Smuzhiyun 	if (!capable(CAP_SYS_ADMIN))
138*4882a593Smuzhiyun 		return;
139*4882a593Smuzhiyun 
140*4882a593Smuzhiyun 	spin_lock(&receiving_list_lock);
141*4882a593Smuzhiyun 	if (msg->len == 0)
142*4882a593Smuzhiyun 		fill_pkg(msg, NULL);
143*4882a593Smuzhiyun 	else if (msg->len < sizeof(*tfr))
144*4882a593Smuzhiyun 		DMERR("Incomplete message received (expected %u, got %u): [%u]",
145*4882a593Smuzhiyun 		      (unsigned)sizeof(*tfr), msg->len, msg->seq);
146*4882a593Smuzhiyun 	else
147*4882a593Smuzhiyun 		fill_pkg(NULL, tfr);
148*4882a593Smuzhiyun 	spin_unlock(&receiving_list_lock);
149*4882a593Smuzhiyun }
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun /**
152*4882a593Smuzhiyun  * dm_consult_userspace
153*4882a593Smuzhiyun  * @uuid: log's universal unique identifier (must be DM_UUID_LEN in size)
154*4882a593Smuzhiyun  * @luid: log's local unique identifier
155*4882a593Smuzhiyun  * @request_type:  found in include/linux/dm-log-userspace.h
156*4882a593Smuzhiyun  * @data: data to tx to the server
157*4882a593Smuzhiyun  * @data_size: size of data in bytes
158*4882a593Smuzhiyun  * @rdata: place to put return data from server
159*4882a593Smuzhiyun  * @rdata_size: value-result (amount of space given/amount of space used)
160*4882a593Smuzhiyun  *
161*4882a593Smuzhiyun  * rdata_size is undefined on failure.
162*4882a593Smuzhiyun  *
163*4882a593Smuzhiyun  * Memory used to communicate with userspace is zero'ed
164*4882a593Smuzhiyun  * before populating to ensure that no unwanted bits leak
165*4882a593Smuzhiyun  * from kernel space to user-space.  All userspace log communications
166*4882a593Smuzhiyun  * between kernel and user space go through this function.
167*4882a593Smuzhiyun  *
168*4882a593Smuzhiyun  * Returns: 0 on success, -EXXX on failure
169*4882a593Smuzhiyun  **/
dm_consult_userspace(const char * uuid,uint64_t luid,int request_type,char * data,size_t data_size,char * rdata,size_t * rdata_size)170*4882a593Smuzhiyun int dm_consult_userspace(const char *uuid, uint64_t luid, int request_type,
171*4882a593Smuzhiyun 			 char *data, size_t data_size,
172*4882a593Smuzhiyun 			 char *rdata, size_t *rdata_size)
173*4882a593Smuzhiyun {
174*4882a593Smuzhiyun 	int r = 0;
175*4882a593Smuzhiyun 	unsigned long tmo;
176*4882a593Smuzhiyun 	size_t dummy = 0;
177*4882a593Smuzhiyun 	int overhead_size = sizeof(struct dm_ulog_request) + sizeof(struct cn_msg);
178*4882a593Smuzhiyun 	struct dm_ulog_request *tfr = prealloced_ulog_tfr;
179*4882a593Smuzhiyun 	struct receiving_pkg pkg;
180*4882a593Smuzhiyun 
181*4882a593Smuzhiyun 	/*
182*4882a593Smuzhiyun 	 * Given the space needed to hold the 'struct cn_msg' and
183*4882a593Smuzhiyun 	 * 'struct dm_ulog_request' - do we have enough payload
184*4882a593Smuzhiyun 	 * space remaining?
185*4882a593Smuzhiyun 	 */
186*4882a593Smuzhiyun 	if (data_size > (DM_ULOG_PREALLOCED_SIZE - overhead_size)) {
187*4882a593Smuzhiyun 		DMINFO("Size of tfr exceeds preallocated size");
188*4882a593Smuzhiyun 		return -EINVAL;
189*4882a593Smuzhiyun 	}
190*4882a593Smuzhiyun 
191*4882a593Smuzhiyun 	if (!rdata_size)
192*4882a593Smuzhiyun 		rdata_size = &dummy;
193*4882a593Smuzhiyun resend:
194*4882a593Smuzhiyun 	/*
195*4882a593Smuzhiyun 	 * We serialize the sending of requests so we can
196*4882a593Smuzhiyun 	 * use the preallocated space.
197*4882a593Smuzhiyun 	 */
198*4882a593Smuzhiyun 	mutex_lock(&dm_ulog_lock);
199*4882a593Smuzhiyun 
200*4882a593Smuzhiyun 	memset(tfr, 0, DM_ULOG_PREALLOCED_SIZE - sizeof(struct cn_msg));
201*4882a593Smuzhiyun 	memcpy(tfr->uuid, uuid, DM_UUID_LEN);
202*4882a593Smuzhiyun 	tfr->version = DM_ULOG_REQUEST_VERSION;
203*4882a593Smuzhiyun 	tfr->luid = luid;
204*4882a593Smuzhiyun 	tfr->seq = dm_ulog_seq++;
205*4882a593Smuzhiyun 
206*4882a593Smuzhiyun 	/*
207*4882a593Smuzhiyun 	 * Must be valid request type (all other bits set to
208*4882a593Smuzhiyun 	 * zero).  This reserves other bits for possible future
209*4882a593Smuzhiyun 	 * use.
210*4882a593Smuzhiyun 	 */
211*4882a593Smuzhiyun 	tfr->request_type = request_type & DM_ULOG_REQUEST_MASK;
212*4882a593Smuzhiyun 
213*4882a593Smuzhiyun 	tfr->data_size = data_size;
214*4882a593Smuzhiyun 	if (data && data_size)
215*4882a593Smuzhiyun 		memcpy(tfr->data, data, data_size);
216*4882a593Smuzhiyun 
217*4882a593Smuzhiyun 	memset(&pkg, 0, sizeof(pkg));
218*4882a593Smuzhiyun 	init_completion(&pkg.complete);
219*4882a593Smuzhiyun 	pkg.seq = tfr->seq;
220*4882a593Smuzhiyun 	pkg.data_size = rdata_size;
221*4882a593Smuzhiyun 	pkg.data = rdata;
222*4882a593Smuzhiyun 	spin_lock(&receiving_list_lock);
223*4882a593Smuzhiyun 	list_add(&(pkg.list), &receiving_list);
224*4882a593Smuzhiyun 	spin_unlock(&receiving_list_lock);
225*4882a593Smuzhiyun 
226*4882a593Smuzhiyun 	r = dm_ulog_sendto_server(tfr);
227*4882a593Smuzhiyun 
228*4882a593Smuzhiyun 	mutex_unlock(&dm_ulog_lock);
229*4882a593Smuzhiyun 
230*4882a593Smuzhiyun 	if (r) {
231*4882a593Smuzhiyun 		DMERR("Unable to send log request [%u] to userspace: %d",
232*4882a593Smuzhiyun 		      request_type, r);
233*4882a593Smuzhiyun 		spin_lock(&receiving_list_lock);
234*4882a593Smuzhiyun 		list_del_init(&(pkg.list));
235*4882a593Smuzhiyun 		spin_unlock(&receiving_list_lock);
236*4882a593Smuzhiyun 
237*4882a593Smuzhiyun 		goto out;
238*4882a593Smuzhiyun 	}
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun 	tmo = wait_for_completion_timeout(&(pkg.complete), DM_ULOG_RETRY_TIMEOUT);
241*4882a593Smuzhiyun 	spin_lock(&receiving_list_lock);
242*4882a593Smuzhiyun 	list_del_init(&(pkg.list));
243*4882a593Smuzhiyun 	spin_unlock(&receiving_list_lock);
244*4882a593Smuzhiyun 	if (!tmo) {
245*4882a593Smuzhiyun 		DMWARN("[%s] Request timed out: [%u/%u] - retrying",
246*4882a593Smuzhiyun 		       (strlen(uuid) > 8) ?
247*4882a593Smuzhiyun 		       (uuid + (strlen(uuid) - 8)) : (uuid),
248*4882a593Smuzhiyun 		       request_type, pkg.seq);
249*4882a593Smuzhiyun 		goto resend;
250*4882a593Smuzhiyun 	}
251*4882a593Smuzhiyun 
252*4882a593Smuzhiyun 	r = pkg.error;
253*4882a593Smuzhiyun 	if (r == -EAGAIN)
254*4882a593Smuzhiyun 		goto resend;
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun out:
257*4882a593Smuzhiyun 	return r;
258*4882a593Smuzhiyun }
259*4882a593Smuzhiyun 
dm_ulog_tfr_init(void)260*4882a593Smuzhiyun int dm_ulog_tfr_init(void)
261*4882a593Smuzhiyun {
262*4882a593Smuzhiyun 	int r;
263*4882a593Smuzhiyun 	void *prealloced;
264*4882a593Smuzhiyun 
265*4882a593Smuzhiyun 	INIT_LIST_HEAD(&receiving_list);
266*4882a593Smuzhiyun 
267*4882a593Smuzhiyun 	prealloced = kmalloc(DM_ULOG_PREALLOCED_SIZE, GFP_KERNEL);
268*4882a593Smuzhiyun 	if (!prealloced)
269*4882a593Smuzhiyun 		return -ENOMEM;
270*4882a593Smuzhiyun 
271*4882a593Smuzhiyun 	prealloced_cn_msg = prealloced;
272*4882a593Smuzhiyun 	prealloced_ulog_tfr = prealloced + sizeof(struct cn_msg);
273*4882a593Smuzhiyun 
274*4882a593Smuzhiyun 	r = cn_add_callback(&ulog_cn_id, "dmlogusr", cn_ulog_callback);
275*4882a593Smuzhiyun 	if (r) {
276*4882a593Smuzhiyun 		kfree(prealloced_cn_msg);
277*4882a593Smuzhiyun 		return r;
278*4882a593Smuzhiyun 	}
279*4882a593Smuzhiyun 
280*4882a593Smuzhiyun 	return 0;
281*4882a593Smuzhiyun }
282*4882a593Smuzhiyun 
dm_ulog_tfr_exit(void)283*4882a593Smuzhiyun void dm_ulog_tfr_exit(void)
284*4882a593Smuzhiyun {
285*4882a593Smuzhiyun 	cn_del_callback(&ulog_cn_id);
286*4882a593Smuzhiyun 	kfree(prealloced_cn_msg);
287*4882a593Smuzhiyun }
288