1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * Copyright (C) 2011 Red Hat, Inc.
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * This file is released under the GPL.
5*4882a593Smuzhiyun */
6*4882a593Smuzhiyun #include "dm-block-manager.h"
7*4882a593Smuzhiyun #include "dm-persistent-data-internal.h"
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun #include <linux/dm-bufio.h>
10*4882a593Smuzhiyun #include <linux/crc32c.h>
11*4882a593Smuzhiyun #include <linux/module.h>
12*4882a593Smuzhiyun #include <linux/slab.h>
13*4882a593Smuzhiyun #include <linux/rwsem.h>
14*4882a593Smuzhiyun #include <linux/device-mapper.h>
15*4882a593Smuzhiyun #include <linux/stacktrace.h>
16*4882a593Smuzhiyun #include <linux/sched/task.h>
17*4882a593Smuzhiyun
18*4882a593Smuzhiyun #define DM_MSG_PREFIX "block manager"
19*4882a593Smuzhiyun
20*4882a593Smuzhiyun /*----------------------------------------------------------------*/
21*4882a593Smuzhiyun
22*4882a593Smuzhiyun #ifdef CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING
23*4882a593Smuzhiyun
24*4882a593Smuzhiyun /*
25*4882a593Smuzhiyun * This is a read/write semaphore with a couple of differences.
26*4882a593Smuzhiyun *
27*4882a593Smuzhiyun * i) There is a restriction on the number of concurrent read locks that
28*4882a593Smuzhiyun * may be held at once. This is just an implementation detail.
29*4882a593Smuzhiyun *
30*4882a593Smuzhiyun * ii) Recursive locking attempts are detected and return EINVAL. A stack
31*4882a593Smuzhiyun * trace is also emitted for the previous lock acquisition.
32*4882a593Smuzhiyun *
33*4882a593Smuzhiyun * iii) Priority is given to write locks.
34*4882a593Smuzhiyun */
35*4882a593Smuzhiyun #define MAX_HOLDERS 4
36*4882a593Smuzhiyun #define MAX_STACK 10
37*4882a593Smuzhiyun
38*4882a593Smuzhiyun struct stack_store {
39*4882a593Smuzhiyun unsigned int nr_entries;
40*4882a593Smuzhiyun unsigned long entries[MAX_STACK];
41*4882a593Smuzhiyun };
42*4882a593Smuzhiyun
43*4882a593Smuzhiyun struct block_lock {
44*4882a593Smuzhiyun spinlock_t lock;
45*4882a593Smuzhiyun __s32 count;
46*4882a593Smuzhiyun struct list_head waiters;
47*4882a593Smuzhiyun struct task_struct *holders[MAX_HOLDERS];
48*4882a593Smuzhiyun
49*4882a593Smuzhiyun #ifdef CONFIG_DM_DEBUG_BLOCK_STACK_TRACING
50*4882a593Smuzhiyun struct stack_store traces[MAX_HOLDERS];
51*4882a593Smuzhiyun #endif
52*4882a593Smuzhiyun };
53*4882a593Smuzhiyun
54*4882a593Smuzhiyun struct waiter {
55*4882a593Smuzhiyun struct list_head list;
56*4882a593Smuzhiyun struct task_struct *task;
57*4882a593Smuzhiyun int wants_write;
58*4882a593Smuzhiyun };
59*4882a593Smuzhiyun
__find_holder(struct block_lock * lock,struct task_struct * task)60*4882a593Smuzhiyun static unsigned __find_holder(struct block_lock *lock,
61*4882a593Smuzhiyun struct task_struct *task)
62*4882a593Smuzhiyun {
63*4882a593Smuzhiyun unsigned i;
64*4882a593Smuzhiyun
65*4882a593Smuzhiyun for (i = 0; i < MAX_HOLDERS; i++)
66*4882a593Smuzhiyun if (lock->holders[i] == task)
67*4882a593Smuzhiyun break;
68*4882a593Smuzhiyun
69*4882a593Smuzhiyun BUG_ON(i == MAX_HOLDERS);
70*4882a593Smuzhiyun return i;
71*4882a593Smuzhiyun }
72*4882a593Smuzhiyun
73*4882a593Smuzhiyun /* call this *after* you increment lock->count */
__add_holder(struct block_lock * lock,struct task_struct * task)74*4882a593Smuzhiyun static void __add_holder(struct block_lock *lock, struct task_struct *task)
75*4882a593Smuzhiyun {
76*4882a593Smuzhiyun unsigned h = __find_holder(lock, NULL);
77*4882a593Smuzhiyun #ifdef CONFIG_DM_DEBUG_BLOCK_STACK_TRACING
78*4882a593Smuzhiyun struct stack_store *t;
79*4882a593Smuzhiyun #endif
80*4882a593Smuzhiyun
81*4882a593Smuzhiyun get_task_struct(task);
82*4882a593Smuzhiyun lock->holders[h] = task;
83*4882a593Smuzhiyun
84*4882a593Smuzhiyun #ifdef CONFIG_DM_DEBUG_BLOCK_STACK_TRACING
85*4882a593Smuzhiyun t = lock->traces + h;
86*4882a593Smuzhiyun t->nr_entries = stack_trace_save(t->entries, MAX_STACK, 2);
87*4882a593Smuzhiyun #endif
88*4882a593Smuzhiyun }
89*4882a593Smuzhiyun
90*4882a593Smuzhiyun /* call this *before* you decrement lock->count */
__del_holder(struct block_lock * lock,struct task_struct * task)91*4882a593Smuzhiyun static void __del_holder(struct block_lock *lock, struct task_struct *task)
92*4882a593Smuzhiyun {
93*4882a593Smuzhiyun unsigned h = __find_holder(lock, task);
94*4882a593Smuzhiyun lock->holders[h] = NULL;
95*4882a593Smuzhiyun put_task_struct(task);
96*4882a593Smuzhiyun }
97*4882a593Smuzhiyun
__check_holder(struct block_lock * lock)98*4882a593Smuzhiyun static int __check_holder(struct block_lock *lock)
99*4882a593Smuzhiyun {
100*4882a593Smuzhiyun unsigned i;
101*4882a593Smuzhiyun
102*4882a593Smuzhiyun for (i = 0; i < MAX_HOLDERS; i++) {
103*4882a593Smuzhiyun if (lock->holders[i] == current) {
104*4882a593Smuzhiyun DMERR("recursive lock detected in metadata");
105*4882a593Smuzhiyun #ifdef CONFIG_DM_DEBUG_BLOCK_STACK_TRACING
106*4882a593Smuzhiyun DMERR("previously held here:");
107*4882a593Smuzhiyun stack_trace_print(lock->traces[i].entries,
108*4882a593Smuzhiyun lock->traces[i].nr_entries, 4);
109*4882a593Smuzhiyun
110*4882a593Smuzhiyun DMERR("subsequent acquisition attempted here:");
111*4882a593Smuzhiyun dump_stack();
112*4882a593Smuzhiyun #endif
113*4882a593Smuzhiyun return -EINVAL;
114*4882a593Smuzhiyun }
115*4882a593Smuzhiyun }
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun return 0;
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun
__wait(struct waiter * w)120*4882a593Smuzhiyun static void __wait(struct waiter *w)
121*4882a593Smuzhiyun {
122*4882a593Smuzhiyun for (;;) {
123*4882a593Smuzhiyun set_current_state(TASK_UNINTERRUPTIBLE);
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun if (!w->task)
126*4882a593Smuzhiyun break;
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun schedule();
129*4882a593Smuzhiyun }
130*4882a593Smuzhiyun
131*4882a593Smuzhiyun set_current_state(TASK_RUNNING);
132*4882a593Smuzhiyun }
133*4882a593Smuzhiyun
__wake_waiter(struct waiter * w)134*4882a593Smuzhiyun static void __wake_waiter(struct waiter *w)
135*4882a593Smuzhiyun {
136*4882a593Smuzhiyun struct task_struct *task;
137*4882a593Smuzhiyun
138*4882a593Smuzhiyun list_del(&w->list);
139*4882a593Smuzhiyun task = w->task;
140*4882a593Smuzhiyun smp_mb();
141*4882a593Smuzhiyun w->task = NULL;
142*4882a593Smuzhiyun wake_up_process(task);
143*4882a593Smuzhiyun }
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun /*
146*4882a593Smuzhiyun * We either wake a few readers or a single writer.
147*4882a593Smuzhiyun */
__wake_many(struct block_lock * lock)148*4882a593Smuzhiyun static void __wake_many(struct block_lock *lock)
149*4882a593Smuzhiyun {
150*4882a593Smuzhiyun struct waiter *w, *tmp;
151*4882a593Smuzhiyun
152*4882a593Smuzhiyun BUG_ON(lock->count < 0);
153*4882a593Smuzhiyun list_for_each_entry_safe(w, tmp, &lock->waiters, list) {
154*4882a593Smuzhiyun if (lock->count >= MAX_HOLDERS)
155*4882a593Smuzhiyun return;
156*4882a593Smuzhiyun
157*4882a593Smuzhiyun if (w->wants_write) {
158*4882a593Smuzhiyun if (lock->count > 0)
159*4882a593Smuzhiyun return; /* still read locked */
160*4882a593Smuzhiyun
161*4882a593Smuzhiyun lock->count = -1;
162*4882a593Smuzhiyun __add_holder(lock, w->task);
163*4882a593Smuzhiyun __wake_waiter(w);
164*4882a593Smuzhiyun return;
165*4882a593Smuzhiyun }
166*4882a593Smuzhiyun
167*4882a593Smuzhiyun lock->count++;
168*4882a593Smuzhiyun __add_holder(lock, w->task);
169*4882a593Smuzhiyun __wake_waiter(w);
170*4882a593Smuzhiyun }
171*4882a593Smuzhiyun }
172*4882a593Smuzhiyun
bl_init(struct block_lock * lock)173*4882a593Smuzhiyun static void bl_init(struct block_lock *lock)
174*4882a593Smuzhiyun {
175*4882a593Smuzhiyun int i;
176*4882a593Smuzhiyun
177*4882a593Smuzhiyun spin_lock_init(&lock->lock);
178*4882a593Smuzhiyun lock->count = 0;
179*4882a593Smuzhiyun INIT_LIST_HEAD(&lock->waiters);
180*4882a593Smuzhiyun for (i = 0; i < MAX_HOLDERS; i++)
181*4882a593Smuzhiyun lock->holders[i] = NULL;
182*4882a593Smuzhiyun }
183*4882a593Smuzhiyun
__available_for_read(struct block_lock * lock)184*4882a593Smuzhiyun static int __available_for_read(struct block_lock *lock)
185*4882a593Smuzhiyun {
186*4882a593Smuzhiyun return lock->count >= 0 &&
187*4882a593Smuzhiyun lock->count < MAX_HOLDERS &&
188*4882a593Smuzhiyun list_empty(&lock->waiters);
189*4882a593Smuzhiyun }
190*4882a593Smuzhiyun
bl_down_read(struct block_lock * lock)191*4882a593Smuzhiyun static int bl_down_read(struct block_lock *lock)
192*4882a593Smuzhiyun {
193*4882a593Smuzhiyun int r;
194*4882a593Smuzhiyun struct waiter w;
195*4882a593Smuzhiyun
196*4882a593Smuzhiyun spin_lock(&lock->lock);
197*4882a593Smuzhiyun r = __check_holder(lock);
198*4882a593Smuzhiyun if (r) {
199*4882a593Smuzhiyun spin_unlock(&lock->lock);
200*4882a593Smuzhiyun return r;
201*4882a593Smuzhiyun }
202*4882a593Smuzhiyun
203*4882a593Smuzhiyun if (__available_for_read(lock)) {
204*4882a593Smuzhiyun lock->count++;
205*4882a593Smuzhiyun __add_holder(lock, current);
206*4882a593Smuzhiyun spin_unlock(&lock->lock);
207*4882a593Smuzhiyun return 0;
208*4882a593Smuzhiyun }
209*4882a593Smuzhiyun
210*4882a593Smuzhiyun get_task_struct(current);
211*4882a593Smuzhiyun
212*4882a593Smuzhiyun w.task = current;
213*4882a593Smuzhiyun w.wants_write = 0;
214*4882a593Smuzhiyun list_add_tail(&w.list, &lock->waiters);
215*4882a593Smuzhiyun spin_unlock(&lock->lock);
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun __wait(&w);
218*4882a593Smuzhiyun put_task_struct(current);
219*4882a593Smuzhiyun return 0;
220*4882a593Smuzhiyun }
221*4882a593Smuzhiyun
bl_down_read_nonblock(struct block_lock * lock)222*4882a593Smuzhiyun static int bl_down_read_nonblock(struct block_lock *lock)
223*4882a593Smuzhiyun {
224*4882a593Smuzhiyun int r;
225*4882a593Smuzhiyun
226*4882a593Smuzhiyun spin_lock(&lock->lock);
227*4882a593Smuzhiyun r = __check_holder(lock);
228*4882a593Smuzhiyun if (r)
229*4882a593Smuzhiyun goto out;
230*4882a593Smuzhiyun
231*4882a593Smuzhiyun if (__available_for_read(lock)) {
232*4882a593Smuzhiyun lock->count++;
233*4882a593Smuzhiyun __add_holder(lock, current);
234*4882a593Smuzhiyun r = 0;
235*4882a593Smuzhiyun } else
236*4882a593Smuzhiyun r = -EWOULDBLOCK;
237*4882a593Smuzhiyun
238*4882a593Smuzhiyun out:
239*4882a593Smuzhiyun spin_unlock(&lock->lock);
240*4882a593Smuzhiyun return r;
241*4882a593Smuzhiyun }
242*4882a593Smuzhiyun
bl_up_read(struct block_lock * lock)243*4882a593Smuzhiyun static void bl_up_read(struct block_lock *lock)
244*4882a593Smuzhiyun {
245*4882a593Smuzhiyun spin_lock(&lock->lock);
246*4882a593Smuzhiyun BUG_ON(lock->count <= 0);
247*4882a593Smuzhiyun __del_holder(lock, current);
248*4882a593Smuzhiyun --lock->count;
249*4882a593Smuzhiyun if (!list_empty(&lock->waiters))
250*4882a593Smuzhiyun __wake_many(lock);
251*4882a593Smuzhiyun spin_unlock(&lock->lock);
252*4882a593Smuzhiyun }
253*4882a593Smuzhiyun
bl_down_write(struct block_lock * lock)254*4882a593Smuzhiyun static int bl_down_write(struct block_lock *lock)
255*4882a593Smuzhiyun {
256*4882a593Smuzhiyun int r;
257*4882a593Smuzhiyun struct waiter w;
258*4882a593Smuzhiyun
259*4882a593Smuzhiyun spin_lock(&lock->lock);
260*4882a593Smuzhiyun r = __check_holder(lock);
261*4882a593Smuzhiyun if (r) {
262*4882a593Smuzhiyun spin_unlock(&lock->lock);
263*4882a593Smuzhiyun return r;
264*4882a593Smuzhiyun }
265*4882a593Smuzhiyun
266*4882a593Smuzhiyun if (lock->count == 0 && list_empty(&lock->waiters)) {
267*4882a593Smuzhiyun lock->count = -1;
268*4882a593Smuzhiyun __add_holder(lock, current);
269*4882a593Smuzhiyun spin_unlock(&lock->lock);
270*4882a593Smuzhiyun return 0;
271*4882a593Smuzhiyun }
272*4882a593Smuzhiyun
273*4882a593Smuzhiyun get_task_struct(current);
274*4882a593Smuzhiyun w.task = current;
275*4882a593Smuzhiyun w.wants_write = 1;
276*4882a593Smuzhiyun
277*4882a593Smuzhiyun /*
278*4882a593Smuzhiyun * Writers given priority. We know there's only one mutator in the
279*4882a593Smuzhiyun * system, so ignoring the ordering reversal.
280*4882a593Smuzhiyun */
281*4882a593Smuzhiyun list_add(&w.list, &lock->waiters);
282*4882a593Smuzhiyun spin_unlock(&lock->lock);
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun __wait(&w);
285*4882a593Smuzhiyun put_task_struct(current);
286*4882a593Smuzhiyun
287*4882a593Smuzhiyun return 0;
288*4882a593Smuzhiyun }
289*4882a593Smuzhiyun
bl_up_write(struct block_lock * lock)290*4882a593Smuzhiyun static void bl_up_write(struct block_lock *lock)
291*4882a593Smuzhiyun {
292*4882a593Smuzhiyun spin_lock(&lock->lock);
293*4882a593Smuzhiyun __del_holder(lock, current);
294*4882a593Smuzhiyun lock->count = 0;
295*4882a593Smuzhiyun if (!list_empty(&lock->waiters))
296*4882a593Smuzhiyun __wake_many(lock);
297*4882a593Smuzhiyun spin_unlock(&lock->lock);
298*4882a593Smuzhiyun }
299*4882a593Smuzhiyun
report_recursive_bug(dm_block_t b,int r)300*4882a593Smuzhiyun static void report_recursive_bug(dm_block_t b, int r)
301*4882a593Smuzhiyun {
302*4882a593Smuzhiyun if (r == -EINVAL)
303*4882a593Smuzhiyun DMERR("recursive acquisition of block %llu requested.",
304*4882a593Smuzhiyun (unsigned long long) b);
305*4882a593Smuzhiyun }
306*4882a593Smuzhiyun
307*4882a593Smuzhiyun #else /* !CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING */
308*4882a593Smuzhiyun
309*4882a593Smuzhiyun #define bl_init(x) do { } while (0)
310*4882a593Smuzhiyun #define bl_down_read(x) 0
311*4882a593Smuzhiyun #define bl_down_read_nonblock(x) 0
312*4882a593Smuzhiyun #define bl_up_read(x) do { } while (0)
313*4882a593Smuzhiyun #define bl_down_write(x) 0
314*4882a593Smuzhiyun #define bl_up_write(x) do { } while (0)
315*4882a593Smuzhiyun #define report_recursive_bug(x, y) do { } while (0)
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun #endif /* CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING */
318*4882a593Smuzhiyun
319*4882a593Smuzhiyun /*----------------------------------------------------------------*/
320*4882a593Smuzhiyun
321*4882a593Smuzhiyun /*
322*4882a593Smuzhiyun * Block manager is currently implemented using dm-bufio. struct
323*4882a593Smuzhiyun * dm_block_manager and struct dm_block map directly onto a couple of
324*4882a593Smuzhiyun * structs in the bufio interface. I want to retain the freedom to move
325*4882a593Smuzhiyun * away from bufio in the future. So these structs are just cast within
326*4882a593Smuzhiyun * this .c file, rather than making it through to the public interface.
327*4882a593Smuzhiyun */
to_buffer(struct dm_block * b)328*4882a593Smuzhiyun static struct dm_buffer *to_buffer(struct dm_block *b)
329*4882a593Smuzhiyun {
330*4882a593Smuzhiyun return (struct dm_buffer *) b;
331*4882a593Smuzhiyun }
332*4882a593Smuzhiyun
dm_block_location(struct dm_block * b)333*4882a593Smuzhiyun dm_block_t dm_block_location(struct dm_block *b)
334*4882a593Smuzhiyun {
335*4882a593Smuzhiyun return dm_bufio_get_block_number(to_buffer(b));
336*4882a593Smuzhiyun }
337*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_block_location);
338*4882a593Smuzhiyun
dm_block_data(struct dm_block * b)339*4882a593Smuzhiyun void *dm_block_data(struct dm_block *b)
340*4882a593Smuzhiyun {
341*4882a593Smuzhiyun return dm_bufio_get_block_data(to_buffer(b));
342*4882a593Smuzhiyun }
343*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_block_data);
344*4882a593Smuzhiyun
345*4882a593Smuzhiyun struct buffer_aux {
346*4882a593Smuzhiyun struct dm_block_validator *validator;
347*4882a593Smuzhiyun int write_locked;
348*4882a593Smuzhiyun
349*4882a593Smuzhiyun #ifdef CONFIG_DM_DEBUG_BLOCK_MANAGER_LOCKING
350*4882a593Smuzhiyun struct block_lock lock;
351*4882a593Smuzhiyun #endif
352*4882a593Smuzhiyun };
353*4882a593Smuzhiyun
dm_block_manager_alloc_callback(struct dm_buffer * buf)354*4882a593Smuzhiyun static void dm_block_manager_alloc_callback(struct dm_buffer *buf)
355*4882a593Smuzhiyun {
356*4882a593Smuzhiyun struct buffer_aux *aux = dm_bufio_get_aux_data(buf);
357*4882a593Smuzhiyun aux->validator = NULL;
358*4882a593Smuzhiyun bl_init(&aux->lock);
359*4882a593Smuzhiyun }
360*4882a593Smuzhiyun
dm_block_manager_write_callback(struct dm_buffer * buf)361*4882a593Smuzhiyun static void dm_block_manager_write_callback(struct dm_buffer *buf)
362*4882a593Smuzhiyun {
363*4882a593Smuzhiyun struct buffer_aux *aux = dm_bufio_get_aux_data(buf);
364*4882a593Smuzhiyun if (aux->validator) {
365*4882a593Smuzhiyun aux->validator->prepare_for_write(aux->validator, (struct dm_block *) buf,
366*4882a593Smuzhiyun dm_bufio_get_block_size(dm_bufio_get_client(buf)));
367*4882a593Smuzhiyun }
368*4882a593Smuzhiyun }
369*4882a593Smuzhiyun
370*4882a593Smuzhiyun /*----------------------------------------------------------------
371*4882a593Smuzhiyun * Public interface
372*4882a593Smuzhiyun *--------------------------------------------------------------*/
373*4882a593Smuzhiyun struct dm_block_manager {
374*4882a593Smuzhiyun struct dm_bufio_client *bufio;
375*4882a593Smuzhiyun bool read_only:1;
376*4882a593Smuzhiyun };
377*4882a593Smuzhiyun
dm_block_manager_create(struct block_device * bdev,unsigned block_size,unsigned max_held_per_thread)378*4882a593Smuzhiyun struct dm_block_manager *dm_block_manager_create(struct block_device *bdev,
379*4882a593Smuzhiyun unsigned block_size,
380*4882a593Smuzhiyun unsigned max_held_per_thread)
381*4882a593Smuzhiyun {
382*4882a593Smuzhiyun int r;
383*4882a593Smuzhiyun struct dm_block_manager *bm;
384*4882a593Smuzhiyun
385*4882a593Smuzhiyun bm = kmalloc(sizeof(*bm), GFP_KERNEL);
386*4882a593Smuzhiyun if (!bm) {
387*4882a593Smuzhiyun r = -ENOMEM;
388*4882a593Smuzhiyun goto bad;
389*4882a593Smuzhiyun }
390*4882a593Smuzhiyun
391*4882a593Smuzhiyun bm->bufio = dm_bufio_client_create(bdev, block_size, max_held_per_thread,
392*4882a593Smuzhiyun sizeof(struct buffer_aux),
393*4882a593Smuzhiyun dm_block_manager_alloc_callback,
394*4882a593Smuzhiyun dm_block_manager_write_callback);
395*4882a593Smuzhiyun if (IS_ERR(bm->bufio)) {
396*4882a593Smuzhiyun r = PTR_ERR(bm->bufio);
397*4882a593Smuzhiyun kfree(bm);
398*4882a593Smuzhiyun goto bad;
399*4882a593Smuzhiyun }
400*4882a593Smuzhiyun
401*4882a593Smuzhiyun bm->read_only = false;
402*4882a593Smuzhiyun
403*4882a593Smuzhiyun return bm;
404*4882a593Smuzhiyun
405*4882a593Smuzhiyun bad:
406*4882a593Smuzhiyun return ERR_PTR(r);
407*4882a593Smuzhiyun }
408*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_block_manager_create);
409*4882a593Smuzhiyun
dm_block_manager_destroy(struct dm_block_manager * bm)410*4882a593Smuzhiyun void dm_block_manager_destroy(struct dm_block_manager *bm)
411*4882a593Smuzhiyun {
412*4882a593Smuzhiyun dm_bufio_client_destroy(bm->bufio);
413*4882a593Smuzhiyun kfree(bm);
414*4882a593Smuzhiyun }
415*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_block_manager_destroy);
416*4882a593Smuzhiyun
dm_bm_block_size(struct dm_block_manager * bm)417*4882a593Smuzhiyun unsigned dm_bm_block_size(struct dm_block_manager *bm)
418*4882a593Smuzhiyun {
419*4882a593Smuzhiyun return dm_bufio_get_block_size(bm->bufio);
420*4882a593Smuzhiyun }
421*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_block_size);
422*4882a593Smuzhiyun
dm_bm_nr_blocks(struct dm_block_manager * bm)423*4882a593Smuzhiyun dm_block_t dm_bm_nr_blocks(struct dm_block_manager *bm)
424*4882a593Smuzhiyun {
425*4882a593Smuzhiyun return dm_bufio_get_device_size(bm->bufio);
426*4882a593Smuzhiyun }
427*4882a593Smuzhiyun
dm_bm_validate_buffer(struct dm_block_manager * bm,struct dm_buffer * buf,struct buffer_aux * aux,struct dm_block_validator * v)428*4882a593Smuzhiyun static int dm_bm_validate_buffer(struct dm_block_manager *bm,
429*4882a593Smuzhiyun struct dm_buffer *buf,
430*4882a593Smuzhiyun struct buffer_aux *aux,
431*4882a593Smuzhiyun struct dm_block_validator *v)
432*4882a593Smuzhiyun {
433*4882a593Smuzhiyun if (unlikely(!aux->validator)) {
434*4882a593Smuzhiyun int r;
435*4882a593Smuzhiyun if (!v)
436*4882a593Smuzhiyun return 0;
437*4882a593Smuzhiyun r = v->check(v, (struct dm_block *) buf, dm_bufio_get_block_size(bm->bufio));
438*4882a593Smuzhiyun if (unlikely(r)) {
439*4882a593Smuzhiyun DMERR_LIMIT("%s validator check failed for block %llu", v->name,
440*4882a593Smuzhiyun (unsigned long long) dm_bufio_get_block_number(buf));
441*4882a593Smuzhiyun return r;
442*4882a593Smuzhiyun }
443*4882a593Smuzhiyun aux->validator = v;
444*4882a593Smuzhiyun } else {
445*4882a593Smuzhiyun if (unlikely(aux->validator != v)) {
446*4882a593Smuzhiyun DMERR_LIMIT("validator mismatch (old=%s vs new=%s) for block %llu",
447*4882a593Smuzhiyun aux->validator->name, v ? v->name : "NULL",
448*4882a593Smuzhiyun (unsigned long long) dm_bufio_get_block_number(buf));
449*4882a593Smuzhiyun return -EINVAL;
450*4882a593Smuzhiyun }
451*4882a593Smuzhiyun }
452*4882a593Smuzhiyun
453*4882a593Smuzhiyun return 0;
454*4882a593Smuzhiyun }
dm_bm_read_lock(struct dm_block_manager * bm,dm_block_t b,struct dm_block_validator * v,struct dm_block ** result)455*4882a593Smuzhiyun int dm_bm_read_lock(struct dm_block_manager *bm, dm_block_t b,
456*4882a593Smuzhiyun struct dm_block_validator *v,
457*4882a593Smuzhiyun struct dm_block **result)
458*4882a593Smuzhiyun {
459*4882a593Smuzhiyun struct buffer_aux *aux;
460*4882a593Smuzhiyun void *p;
461*4882a593Smuzhiyun int r;
462*4882a593Smuzhiyun
463*4882a593Smuzhiyun p = dm_bufio_read(bm->bufio, b, (struct dm_buffer **) result);
464*4882a593Smuzhiyun if (IS_ERR(p))
465*4882a593Smuzhiyun return PTR_ERR(p);
466*4882a593Smuzhiyun
467*4882a593Smuzhiyun aux = dm_bufio_get_aux_data(to_buffer(*result));
468*4882a593Smuzhiyun r = bl_down_read(&aux->lock);
469*4882a593Smuzhiyun if (unlikely(r)) {
470*4882a593Smuzhiyun dm_bufio_release(to_buffer(*result));
471*4882a593Smuzhiyun report_recursive_bug(b, r);
472*4882a593Smuzhiyun return r;
473*4882a593Smuzhiyun }
474*4882a593Smuzhiyun
475*4882a593Smuzhiyun aux->write_locked = 0;
476*4882a593Smuzhiyun
477*4882a593Smuzhiyun r = dm_bm_validate_buffer(bm, to_buffer(*result), aux, v);
478*4882a593Smuzhiyun if (unlikely(r)) {
479*4882a593Smuzhiyun bl_up_read(&aux->lock);
480*4882a593Smuzhiyun dm_bufio_release(to_buffer(*result));
481*4882a593Smuzhiyun return r;
482*4882a593Smuzhiyun }
483*4882a593Smuzhiyun
484*4882a593Smuzhiyun return 0;
485*4882a593Smuzhiyun }
486*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_read_lock);
487*4882a593Smuzhiyun
dm_bm_write_lock(struct dm_block_manager * bm,dm_block_t b,struct dm_block_validator * v,struct dm_block ** result)488*4882a593Smuzhiyun int dm_bm_write_lock(struct dm_block_manager *bm,
489*4882a593Smuzhiyun dm_block_t b, struct dm_block_validator *v,
490*4882a593Smuzhiyun struct dm_block **result)
491*4882a593Smuzhiyun {
492*4882a593Smuzhiyun struct buffer_aux *aux;
493*4882a593Smuzhiyun void *p;
494*4882a593Smuzhiyun int r;
495*4882a593Smuzhiyun
496*4882a593Smuzhiyun if (dm_bm_is_read_only(bm))
497*4882a593Smuzhiyun return -EPERM;
498*4882a593Smuzhiyun
499*4882a593Smuzhiyun p = dm_bufio_read(bm->bufio, b, (struct dm_buffer **) result);
500*4882a593Smuzhiyun if (IS_ERR(p))
501*4882a593Smuzhiyun return PTR_ERR(p);
502*4882a593Smuzhiyun
503*4882a593Smuzhiyun aux = dm_bufio_get_aux_data(to_buffer(*result));
504*4882a593Smuzhiyun r = bl_down_write(&aux->lock);
505*4882a593Smuzhiyun if (r) {
506*4882a593Smuzhiyun dm_bufio_release(to_buffer(*result));
507*4882a593Smuzhiyun report_recursive_bug(b, r);
508*4882a593Smuzhiyun return r;
509*4882a593Smuzhiyun }
510*4882a593Smuzhiyun
511*4882a593Smuzhiyun aux->write_locked = 1;
512*4882a593Smuzhiyun
513*4882a593Smuzhiyun r = dm_bm_validate_buffer(bm, to_buffer(*result), aux, v);
514*4882a593Smuzhiyun if (unlikely(r)) {
515*4882a593Smuzhiyun bl_up_write(&aux->lock);
516*4882a593Smuzhiyun dm_bufio_release(to_buffer(*result));
517*4882a593Smuzhiyun return r;
518*4882a593Smuzhiyun }
519*4882a593Smuzhiyun
520*4882a593Smuzhiyun return 0;
521*4882a593Smuzhiyun }
522*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_write_lock);
523*4882a593Smuzhiyun
dm_bm_read_try_lock(struct dm_block_manager * bm,dm_block_t b,struct dm_block_validator * v,struct dm_block ** result)524*4882a593Smuzhiyun int dm_bm_read_try_lock(struct dm_block_manager *bm,
525*4882a593Smuzhiyun dm_block_t b, struct dm_block_validator *v,
526*4882a593Smuzhiyun struct dm_block **result)
527*4882a593Smuzhiyun {
528*4882a593Smuzhiyun struct buffer_aux *aux;
529*4882a593Smuzhiyun void *p;
530*4882a593Smuzhiyun int r;
531*4882a593Smuzhiyun
532*4882a593Smuzhiyun p = dm_bufio_get(bm->bufio, b, (struct dm_buffer **) result);
533*4882a593Smuzhiyun if (IS_ERR(p))
534*4882a593Smuzhiyun return PTR_ERR(p);
535*4882a593Smuzhiyun if (unlikely(!p))
536*4882a593Smuzhiyun return -EWOULDBLOCK;
537*4882a593Smuzhiyun
538*4882a593Smuzhiyun aux = dm_bufio_get_aux_data(to_buffer(*result));
539*4882a593Smuzhiyun r = bl_down_read_nonblock(&aux->lock);
540*4882a593Smuzhiyun if (r < 0) {
541*4882a593Smuzhiyun dm_bufio_release(to_buffer(*result));
542*4882a593Smuzhiyun report_recursive_bug(b, r);
543*4882a593Smuzhiyun return r;
544*4882a593Smuzhiyun }
545*4882a593Smuzhiyun aux->write_locked = 0;
546*4882a593Smuzhiyun
547*4882a593Smuzhiyun r = dm_bm_validate_buffer(bm, to_buffer(*result), aux, v);
548*4882a593Smuzhiyun if (unlikely(r)) {
549*4882a593Smuzhiyun bl_up_read(&aux->lock);
550*4882a593Smuzhiyun dm_bufio_release(to_buffer(*result));
551*4882a593Smuzhiyun return r;
552*4882a593Smuzhiyun }
553*4882a593Smuzhiyun
554*4882a593Smuzhiyun return 0;
555*4882a593Smuzhiyun }
556*4882a593Smuzhiyun
dm_bm_write_lock_zero(struct dm_block_manager * bm,dm_block_t b,struct dm_block_validator * v,struct dm_block ** result)557*4882a593Smuzhiyun int dm_bm_write_lock_zero(struct dm_block_manager *bm,
558*4882a593Smuzhiyun dm_block_t b, struct dm_block_validator *v,
559*4882a593Smuzhiyun struct dm_block **result)
560*4882a593Smuzhiyun {
561*4882a593Smuzhiyun int r;
562*4882a593Smuzhiyun struct buffer_aux *aux;
563*4882a593Smuzhiyun void *p;
564*4882a593Smuzhiyun
565*4882a593Smuzhiyun if (dm_bm_is_read_only(bm))
566*4882a593Smuzhiyun return -EPERM;
567*4882a593Smuzhiyun
568*4882a593Smuzhiyun p = dm_bufio_new(bm->bufio, b, (struct dm_buffer **) result);
569*4882a593Smuzhiyun if (IS_ERR(p))
570*4882a593Smuzhiyun return PTR_ERR(p);
571*4882a593Smuzhiyun
572*4882a593Smuzhiyun memset(p, 0, dm_bm_block_size(bm));
573*4882a593Smuzhiyun
574*4882a593Smuzhiyun aux = dm_bufio_get_aux_data(to_buffer(*result));
575*4882a593Smuzhiyun r = bl_down_write(&aux->lock);
576*4882a593Smuzhiyun if (r) {
577*4882a593Smuzhiyun dm_bufio_release(to_buffer(*result));
578*4882a593Smuzhiyun return r;
579*4882a593Smuzhiyun }
580*4882a593Smuzhiyun
581*4882a593Smuzhiyun aux->write_locked = 1;
582*4882a593Smuzhiyun aux->validator = v;
583*4882a593Smuzhiyun
584*4882a593Smuzhiyun return 0;
585*4882a593Smuzhiyun }
586*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_write_lock_zero);
587*4882a593Smuzhiyun
dm_bm_unlock(struct dm_block * b)588*4882a593Smuzhiyun void dm_bm_unlock(struct dm_block *b)
589*4882a593Smuzhiyun {
590*4882a593Smuzhiyun struct buffer_aux *aux;
591*4882a593Smuzhiyun aux = dm_bufio_get_aux_data(to_buffer(b));
592*4882a593Smuzhiyun
593*4882a593Smuzhiyun if (aux->write_locked) {
594*4882a593Smuzhiyun dm_bufio_mark_buffer_dirty(to_buffer(b));
595*4882a593Smuzhiyun bl_up_write(&aux->lock);
596*4882a593Smuzhiyun } else
597*4882a593Smuzhiyun bl_up_read(&aux->lock);
598*4882a593Smuzhiyun
599*4882a593Smuzhiyun dm_bufio_release(to_buffer(b));
600*4882a593Smuzhiyun }
601*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_unlock);
602*4882a593Smuzhiyun
dm_bm_flush(struct dm_block_manager * bm)603*4882a593Smuzhiyun int dm_bm_flush(struct dm_block_manager *bm)
604*4882a593Smuzhiyun {
605*4882a593Smuzhiyun if (dm_bm_is_read_only(bm))
606*4882a593Smuzhiyun return -EPERM;
607*4882a593Smuzhiyun
608*4882a593Smuzhiyun return dm_bufio_write_dirty_buffers(bm->bufio);
609*4882a593Smuzhiyun }
610*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_flush);
611*4882a593Smuzhiyun
dm_bm_prefetch(struct dm_block_manager * bm,dm_block_t b)612*4882a593Smuzhiyun void dm_bm_prefetch(struct dm_block_manager *bm, dm_block_t b)
613*4882a593Smuzhiyun {
614*4882a593Smuzhiyun dm_bufio_prefetch(bm->bufio, b, 1);
615*4882a593Smuzhiyun }
616*4882a593Smuzhiyun
dm_bm_is_read_only(struct dm_block_manager * bm)617*4882a593Smuzhiyun bool dm_bm_is_read_only(struct dm_block_manager *bm)
618*4882a593Smuzhiyun {
619*4882a593Smuzhiyun return (bm ? bm->read_only : true);
620*4882a593Smuzhiyun }
621*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_is_read_only);
622*4882a593Smuzhiyun
dm_bm_set_read_only(struct dm_block_manager * bm)623*4882a593Smuzhiyun void dm_bm_set_read_only(struct dm_block_manager *bm)
624*4882a593Smuzhiyun {
625*4882a593Smuzhiyun if (bm)
626*4882a593Smuzhiyun bm->read_only = true;
627*4882a593Smuzhiyun }
628*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_set_read_only);
629*4882a593Smuzhiyun
dm_bm_set_read_write(struct dm_block_manager * bm)630*4882a593Smuzhiyun void dm_bm_set_read_write(struct dm_block_manager *bm)
631*4882a593Smuzhiyun {
632*4882a593Smuzhiyun if (bm)
633*4882a593Smuzhiyun bm->read_only = false;
634*4882a593Smuzhiyun }
635*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_set_read_write);
636*4882a593Smuzhiyun
dm_bm_checksum(const void * data,size_t len,u32 init_xor)637*4882a593Smuzhiyun u32 dm_bm_checksum(const void *data, size_t len, u32 init_xor)
638*4882a593Smuzhiyun {
639*4882a593Smuzhiyun return crc32c(~(u32) 0, data, len) ^ init_xor;
640*4882a593Smuzhiyun }
641*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(dm_bm_checksum);
642*4882a593Smuzhiyun
643*4882a593Smuzhiyun /*----------------------------------------------------------------*/
644*4882a593Smuzhiyun
645*4882a593Smuzhiyun MODULE_LICENSE("GPL");
646*4882a593Smuzhiyun MODULE_AUTHOR("Joe Thornber <dm-devel@redhat.com>");
647*4882a593Smuzhiyun MODULE_DESCRIPTION("Immutable metadata library for dm");
648*4882a593Smuzhiyun
649*4882a593Smuzhiyun /*----------------------------------------------------------------*/
650