1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * Copyright (C) 2006-2009 Red Hat, Inc.
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * This file is released under the LGPL.
5*4882a593Smuzhiyun */
6*4882a593Smuzhiyun
7*4882a593Smuzhiyun #include <linux/kernel.h>
8*4882a593Smuzhiyun #include <linux/module.h>
9*4882a593Smuzhiyun #include <linux/slab.h>
10*4882a593Smuzhiyun #include <net/sock.h>
11*4882a593Smuzhiyun #include <linux/workqueue.h>
12*4882a593Smuzhiyun #include <linux/connector.h>
13*4882a593Smuzhiyun #include <linux/device-mapper.h>
14*4882a593Smuzhiyun #include <linux/dm-log-userspace.h>
15*4882a593Smuzhiyun
16*4882a593Smuzhiyun #include "dm-log-userspace-transfer.h"
17*4882a593Smuzhiyun
18*4882a593Smuzhiyun static uint32_t dm_ulog_seq;
19*4882a593Smuzhiyun
20*4882a593Smuzhiyun /*
21*4882a593Smuzhiyun * Netlink/Connector is an unreliable protocol. How long should
22*4882a593Smuzhiyun * we wait for a response before assuming it was lost and retrying?
23*4882a593Smuzhiyun * (If we do receive a response after this time, it will be discarded
24*4882a593Smuzhiyun * and the response to the resent request will be waited for.
25*4882a593Smuzhiyun */
26*4882a593Smuzhiyun #define DM_ULOG_RETRY_TIMEOUT (15 * HZ)
27*4882a593Smuzhiyun
28*4882a593Smuzhiyun /*
29*4882a593Smuzhiyun * Pre-allocated space for speed
30*4882a593Smuzhiyun */
31*4882a593Smuzhiyun #define DM_ULOG_PREALLOCED_SIZE 512
32*4882a593Smuzhiyun static struct cn_msg *prealloced_cn_msg;
33*4882a593Smuzhiyun static struct dm_ulog_request *prealloced_ulog_tfr;
34*4882a593Smuzhiyun
35*4882a593Smuzhiyun static struct cb_id ulog_cn_id = {
36*4882a593Smuzhiyun .idx = CN_IDX_DM,
37*4882a593Smuzhiyun .val = CN_VAL_DM_USERSPACE_LOG
38*4882a593Smuzhiyun };
39*4882a593Smuzhiyun
40*4882a593Smuzhiyun static DEFINE_MUTEX(dm_ulog_lock);
41*4882a593Smuzhiyun
42*4882a593Smuzhiyun struct receiving_pkg {
43*4882a593Smuzhiyun struct list_head list;
44*4882a593Smuzhiyun struct completion complete;
45*4882a593Smuzhiyun
46*4882a593Smuzhiyun uint32_t seq;
47*4882a593Smuzhiyun
48*4882a593Smuzhiyun int error;
49*4882a593Smuzhiyun size_t *data_size;
50*4882a593Smuzhiyun char *data;
51*4882a593Smuzhiyun };
52*4882a593Smuzhiyun
53*4882a593Smuzhiyun static DEFINE_SPINLOCK(receiving_list_lock);
54*4882a593Smuzhiyun static struct list_head receiving_list;
55*4882a593Smuzhiyun
dm_ulog_sendto_server(struct dm_ulog_request * tfr)56*4882a593Smuzhiyun static int dm_ulog_sendto_server(struct dm_ulog_request *tfr)
57*4882a593Smuzhiyun {
58*4882a593Smuzhiyun int r;
59*4882a593Smuzhiyun struct cn_msg *msg = prealloced_cn_msg;
60*4882a593Smuzhiyun
61*4882a593Smuzhiyun memset(msg, 0, sizeof(struct cn_msg));
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun msg->id.idx = ulog_cn_id.idx;
64*4882a593Smuzhiyun msg->id.val = ulog_cn_id.val;
65*4882a593Smuzhiyun msg->ack = 0;
66*4882a593Smuzhiyun msg->seq = tfr->seq;
67*4882a593Smuzhiyun msg->len = sizeof(struct dm_ulog_request) + tfr->data_size;
68*4882a593Smuzhiyun
69*4882a593Smuzhiyun r = cn_netlink_send(msg, 0, 0, gfp_any());
70*4882a593Smuzhiyun
71*4882a593Smuzhiyun return r;
72*4882a593Smuzhiyun }
73*4882a593Smuzhiyun
74*4882a593Smuzhiyun /*
75*4882a593Smuzhiyun * Parameters for this function can be either msg or tfr, but not
76*4882a593Smuzhiyun * both. This function fills in the reply for a waiting request.
77*4882a593Smuzhiyun * If just msg is given, then the reply is simply an ACK from userspace
78*4882a593Smuzhiyun * that the request was received.
79*4882a593Smuzhiyun *
80*4882a593Smuzhiyun * Returns: 0 on success, -ENOENT on failure
81*4882a593Smuzhiyun */
fill_pkg(struct cn_msg * msg,struct dm_ulog_request * tfr)82*4882a593Smuzhiyun static int fill_pkg(struct cn_msg *msg, struct dm_ulog_request *tfr)
83*4882a593Smuzhiyun {
84*4882a593Smuzhiyun uint32_t rtn_seq = (msg) ? msg->seq : (tfr) ? tfr->seq : 0;
85*4882a593Smuzhiyun struct receiving_pkg *pkg;
86*4882a593Smuzhiyun
87*4882a593Smuzhiyun /*
88*4882a593Smuzhiyun * The 'receiving_pkg' entries in this list are statically
89*4882a593Smuzhiyun * allocated on the stack in 'dm_consult_userspace'.
90*4882a593Smuzhiyun * Each process that is waiting for a reply from the user
91*4882a593Smuzhiyun * space server will have an entry in this list.
92*4882a593Smuzhiyun *
93*4882a593Smuzhiyun * We are safe to do it this way because the stack space
94*4882a593Smuzhiyun * is unique to each process, but still addressable by
95*4882a593Smuzhiyun * other processes.
96*4882a593Smuzhiyun */
97*4882a593Smuzhiyun list_for_each_entry(pkg, &receiving_list, list) {
98*4882a593Smuzhiyun if (rtn_seq != pkg->seq)
99*4882a593Smuzhiyun continue;
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun if (msg) {
102*4882a593Smuzhiyun pkg->error = -msg->ack;
103*4882a593Smuzhiyun /*
104*4882a593Smuzhiyun * If we are trying again, we will need to know our
105*4882a593Smuzhiyun * storage capacity. Otherwise, along with the
106*4882a593Smuzhiyun * error code, we make explicit that we have no data.
107*4882a593Smuzhiyun */
108*4882a593Smuzhiyun if (pkg->error != -EAGAIN)
109*4882a593Smuzhiyun *(pkg->data_size) = 0;
110*4882a593Smuzhiyun } else if (tfr->data_size > *(pkg->data_size)) {
111*4882a593Smuzhiyun DMERR("Insufficient space to receive package [%u] "
112*4882a593Smuzhiyun "(%u vs %zu)", tfr->request_type,
113*4882a593Smuzhiyun tfr->data_size, *(pkg->data_size));
114*4882a593Smuzhiyun
115*4882a593Smuzhiyun *(pkg->data_size) = 0;
116*4882a593Smuzhiyun pkg->error = -ENOSPC;
117*4882a593Smuzhiyun } else {
118*4882a593Smuzhiyun pkg->error = tfr->error;
119*4882a593Smuzhiyun memcpy(pkg->data, tfr->data, tfr->data_size);
120*4882a593Smuzhiyun *(pkg->data_size) = tfr->data_size;
121*4882a593Smuzhiyun }
122*4882a593Smuzhiyun complete(&pkg->complete);
123*4882a593Smuzhiyun return 0;
124*4882a593Smuzhiyun }
125*4882a593Smuzhiyun
126*4882a593Smuzhiyun return -ENOENT;
127*4882a593Smuzhiyun }
128*4882a593Smuzhiyun
129*4882a593Smuzhiyun /*
130*4882a593Smuzhiyun * This is the connector callback that delivers data
131*4882a593Smuzhiyun * that was sent from userspace.
132*4882a593Smuzhiyun */
cn_ulog_callback(struct cn_msg * msg,struct netlink_skb_parms * nsp)133*4882a593Smuzhiyun static void cn_ulog_callback(struct cn_msg *msg, struct netlink_skb_parms *nsp)
134*4882a593Smuzhiyun {
135*4882a593Smuzhiyun struct dm_ulog_request *tfr = (struct dm_ulog_request *)(msg + 1);
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun if (!capable(CAP_SYS_ADMIN))
138*4882a593Smuzhiyun return;
139*4882a593Smuzhiyun
140*4882a593Smuzhiyun spin_lock(&receiving_list_lock);
141*4882a593Smuzhiyun if (msg->len == 0)
142*4882a593Smuzhiyun fill_pkg(msg, NULL);
143*4882a593Smuzhiyun else if (msg->len < sizeof(*tfr))
144*4882a593Smuzhiyun DMERR("Incomplete message received (expected %u, got %u): [%u]",
145*4882a593Smuzhiyun (unsigned)sizeof(*tfr), msg->len, msg->seq);
146*4882a593Smuzhiyun else
147*4882a593Smuzhiyun fill_pkg(NULL, tfr);
148*4882a593Smuzhiyun spin_unlock(&receiving_list_lock);
149*4882a593Smuzhiyun }
150*4882a593Smuzhiyun
151*4882a593Smuzhiyun /**
152*4882a593Smuzhiyun * dm_consult_userspace
153*4882a593Smuzhiyun * @uuid: log's universal unique identifier (must be DM_UUID_LEN in size)
154*4882a593Smuzhiyun * @luid: log's local unique identifier
155*4882a593Smuzhiyun * @request_type: found in include/linux/dm-log-userspace.h
156*4882a593Smuzhiyun * @data: data to tx to the server
157*4882a593Smuzhiyun * @data_size: size of data in bytes
158*4882a593Smuzhiyun * @rdata: place to put return data from server
159*4882a593Smuzhiyun * @rdata_size: value-result (amount of space given/amount of space used)
160*4882a593Smuzhiyun *
161*4882a593Smuzhiyun * rdata_size is undefined on failure.
162*4882a593Smuzhiyun *
163*4882a593Smuzhiyun * Memory used to communicate with userspace is zero'ed
164*4882a593Smuzhiyun * before populating to ensure that no unwanted bits leak
165*4882a593Smuzhiyun * from kernel space to user-space. All userspace log communications
166*4882a593Smuzhiyun * between kernel and user space go through this function.
167*4882a593Smuzhiyun *
168*4882a593Smuzhiyun * Returns: 0 on success, -EXXX on failure
169*4882a593Smuzhiyun **/
dm_consult_userspace(const char * uuid,uint64_t luid,int request_type,char * data,size_t data_size,char * rdata,size_t * rdata_size)170*4882a593Smuzhiyun int dm_consult_userspace(const char *uuid, uint64_t luid, int request_type,
171*4882a593Smuzhiyun char *data, size_t data_size,
172*4882a593Smuzhiyun char *rdata, size_t *rdata_size)
173*4882a593Smuzhiyun {
174*4882a593Smuzhiyun int r = 0;
175*4882a593Smuzhiyun unsigned long tmo;
176*4882a593Smuzhiyun size_t dummy = 0;
177*4882a593Smuzhiyun int overhead_size = sizeof(struct dm_ulog_request) + sizeof(struct cn_msg);
178*4882a593Smuzhiyun struct dm_ulog_request *tfr = prealloced_ulog_tfr;
179*4882a593Smuzhiyun struct receiving_pkg pkg;
180*4882a593Smuzhiyun
181*4882a593Smuzhiyun /*
182*4882a593Smuzhiyun * Given the space needed to hold the 'struct cn_msg' and
183*4882a593Smuzhiyun * 'struct dm_ulog_request' - do we have enough payload
184*4882a593Smuzhiyun * space remaining?
185*4882a593Smuzhiyun */
186*4882a593Smuzhiyun if (data_size > (DM_ULOG_PREALLOCED_SIZE - overhead_size)) {
187*4882a593Smuzhiyun DMINFO("Size of tfr exceeds preallocated size");
188*4882a593Smuzhiyun return -EINVAL;
189*4882a593Smuzhiyun }
190*4882a593Smuzhiyun
191*4882a593Smuzhiyun if (!rdata_size)
192*4882a593Smuzhiyun rdata_size = &dummy;
193*4882a593Smuzhiyun resend:
194*4882a593Smuzhiyun /*
195*4882a593Smuzhiyun * We serialize the sending of requests so we can
196*4882a593Smuzhiyun * use the preallocated space.
197*4882a593Smuzhiyun */
198*4882a593Smuzhiyun mutex_lock(&dm_ulog_lock);
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun memset(tfr, 0, DM_ULOG_PREALLOCED_SIZE - sizeof(struct cn_msg));
201*4882a593Smuzhiyun memcpy(tfr->uuid, uuid, DM_UUID_LEN);
202*4882a593Smuzhiyun tfr->version = DM_ULOG_REQUEST_VERSION;
203*4882a593Smuzhiyun tfr->luid = luid;
204*4882a593Smuzhiyun tfr->seq = dm_ulog_seq++;
205*4882a593Smuzhiyun
206*4882a593Smuzhiyun /*
207*4882a593Smuzhiyun * Must be valid request type (all other bits set to
208*4882a593Smuzhiyun * zero). This reserves other bits for possible future
209*4882a593Smuzhiyun * use.
210*4882a593Smuzhiyun */
211*4882a593Smuzhiyun tfr->request_type = request_type & DM_ULOG_REQUEST_MASK;
212*4882a593Smuzhiyun
213*4882a593Smuzhiyun tfr->data_size = data_size;
214*4882a593Smuzhiyun if (data && data_size)
215*4882a593Smuzhiyun memcpy(tfr->data, data, data_size);
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun memset(&pkg, 0, sizeof(pkg));
218*4882a593Smuzhiyun init_completion(&pkg.complete);
219*4882a593Smuzhiyun pkg.seq = tfr->seq;
220*4882a593Smuzhiyun pkg.data_size = rdata_size;
221*4882a593Smuzhiyun pkg.data = rdata;
222*4882a593Smuzhiyun spin_lock(&receiving_list_lock);
223*4882a593Smuzhiyun list_add(&(pkg.list), &receiving_list);
224*4882a593Smuzhiyun spin_unlock(&receiving_list_lock);
225*4882a593Smuzhiyun
226*4882a593Smuzhiyun r = dm_ulog_sendto_server(tfr);
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun mutex_unlock(&dm_ulog_lock);
229*4882a593Smuzhiyun
230*4882a593Smuzhiyun if (r) {
231*4882a593Smuzhiyun DMERR("Unable to send log request [%u] to userspace: %d",
232*4882a593Smuzhiyun request_type, r);
233*4882a593Smuzhiyun spin_lock(&receiving_list_lock);
234*4882a593Smuzhiyun list_del_init(&(pkg.list));
235*4882a593Smuzhiyun spin_unlock(&receiving_list_lock);
236*4882a593Smuzhiyun
237*4882a593Smuzhiyun goto out;
238*4882a593Smuzhiyun }
239*4882a593Smuzhiyun
240*4882a593Smuzhiyun tmo = wait_for_completion_timeout(&(pkg.complete), DM_ULOG_RETRY_TIMEOUT);
241*4882a593Smuzhiyun spin_lock(&receiving_list_lock);
242*4882a593Smuzhiyun list_del_init(&(pkg.list));
243*4882a593Smuzhiyun spin_unlock(&receiving_list_lock);
244*4882a593Smuzhiyun if (!tmo) {
245*4882a593Smuzhiyun DMWARN("[%s] Request timed out: [%u/%u] - retrying",
246*4882a593Smuzhiyun (strlen(uuid) > 8) ?
247*4882a593Smuzhiyun (uuid + (strlen(uuid) - 8)) : (uuid),
248*4882a593Smuzhiyun request_type, pkg.seq);
249*4882a593Smuzhiyun goto resend;
250*4882a593Smuzhiyun }
251*4882a593Smuzhiyun
252*4882a593Smuzhiyun r = pkg.error;
253*4882a593Smuzhiyun if (r == -EAGAIN)
254*4882a593Smuzhiyun goto resend;
255*4882a593Smuzhiyun
256*4882a593Smuzhiyun out:
257*4882a593Smuzhiyun return r;
258*4882a593Smuzhiyun }
259*4882a593Smuzhiyun
dm_ulog_tfr_init(void)260*4882a593Smuzhiyun int dm_ulog_tfr_init(void)
261*4882a593Smuzhiyun {
262*4882a593Smuzhiyun int r;
263*4882a593Smuzhiyun void *prealloced;
264*4882a593Smuzhiyun
265*4882a593Smuzhiyun INIT_LIST_HEAD(&receiving_list);
266*4882a593Smuzhiyun
267*4882a593Smuzhiyun prealloced = kmalloc(DM_ULOG_PREALLOCED_SIZE, GFP_KERNEL);
268*4882a593Smuzhiyun if (!prealloced)
269*4882a593Smuzhiyun return -ENOMEM;
270*4882a593Smuzhiyun
271*4882a593Smuzhiyun prealloced_cn_msg = prealloced;
272*4882a593Smuzhiyun prealloced_ulog_tfr = prealloced + sizeof(struct cn_msg);
273*4882a593Smuzhiyun
274*4882a593Smuzhiyun r = cn_add_callback(&ulog_cn_id, "dmlogusr", cn_ulog_callback);
275*4882a593Smuzhiyun if (r) {
276*4882a593Smuzhiyun kfree(prealloced_cn_msg);
277*4882a593Smuzhiyun return r;
278*4882a593Smuzhiyun }
279*4882a593Smuzhiyun
280*4882a593Smuzhiyun return 0;
281*4882a593Smuzhiyun }
282*4882a593Smuzhiyun
dm_ulog_tfr_exit(void)283*4882a593Smuzhiyun void dm_ulog_tfr_exit(void)
284*4882a593Smuzhiyun {
285*4882a593Smuzhiyun cn_del_callback(&ulog_cn_id);
286*4882a593Smuzhiyun kfree(prealloced_cn_msg);
287*4882a593Smuzhiyun }
288