Lines Matching +full:- +full:c
4 * Copyright (C) 2006-2008 Nokia Corporation.
5 * Copyright (C) 2006, 2007 University of Szeged, Hungary
7 * SPDX-License-Identifier: GPL-2.0+
15 * This file implements UBIFS I/O subsystem which provides various I/O-related
17 * write-buffering support. Write buffers help to save space which otherwise
19 * Instead, data first goes to the write-buffer and is flushed when the
23 * UBIFS distinguishes between minimum write size (@c->min_io_size) and maximum
24 * write size (@c->max_write_size). The latter is the maximum amount of bytes
26 * @c->max_write_size units should presumably be faster. Obviously,
27 * @c->min_io_size <= @c->max_write_size. Write-buffers are of
28 * @c->max_write_size bytes in size for maximum performance. However, when a
29 * write-buffer is flushed, only the portion of it (aligned to @c->min_io_size
30 * boundary) which contains data is written, not the whole write-buffer,
31 * because this is more space-efficient.
34 * hand, we want to write in optimal @c->max_write_size bytes chunks, which
35 * also means aligning writes at the @c->max_write_size bytes offsets. On the
38 * the next write offset to be not aligned to @c->max_write_size bytes. So the
39 * have to make sure that the write-buffer offset (@wbuf->offs) becomes aligned
40 * to @c->max_write_size bytes again. We do this by temporarily shrinking
41 * write-buffer size (@wbuf->size).
43 * Write-buffers are defined by 'struct ubifs_wbuf' objects and protected by
44 * mutexes defined inside these objects. Since sometimes upper-level code
45 * has to lock the write-buffer (e.g. journal space reservation code), many
46 * functions related to write-buffers have "nolock" suffix which means that the
47 * caller has to lock the write-buffer before calling this function.
49 * UBIFS stores nodes at 64 bit-aligned addresses. If the node length is not
72 * ubifs_ro_mode - switch UBIFS to read read-only mode.
73 * @c: UBIFS file-system description object
76 void ubifs_ro_mode(struct ubifs_info *c, int err) in ubifs_ro_mode() argument
78 if (!c->ro_error) { in ubifs_ro_mode()
79 c->ro_error = 1; in ubifs_ro_mode()
80 c->no_chk_data_crc = 0; in ubifs_ro_mode()
81 c->vfs_sb->s_flags |= MS_RDONLY; in ubifs_ro_mode()
82 ubifs_warn(c, "switched to read-only mode, error %d", err); in ubifs_ro_mode()
93 int ubifs_leb_read(const struct ubifs_info *c, int lnum, void *buf, int offs, in ubifs_leb_read() argument
98 err = ubi_read(c->ubi, lnum, buf, offs, len); in ubifs_leb_read()
100 * In case of %-EBADMSG print the error message only if the in ubifs_leb_read()
103 if (err && (err != -EBADMSG || even_ebadmsg)) { in ubifs_leb_read()
104 ubifs_err(c, "reading %d bytes from LEB %d:%d failed, error %d", in ubifs_leb_read()
111 int ubifs_leb_write(struct ubifs_info *c, int lnum, const void *buf, int offs, in ubifs_leb_write() argument
116 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_leb_write()
117 if (c->ro_error) in ubifs_leb_write()
118 return -EROFS; in ubifs_leb_write()
119 if (!dbg_is_tst_rcvry(c)) in ubifs_leb_write()
120 err = ubi_leb_write(c->ubi, lnum, buf, offs, len); in ubifs_leb_write()
123 err = dbg_leb_write(c, lnum, buf, offs, len); in ubifs_leb_write()
126 ubifs_err(c, "writing %d bytes to LEB %d:%d failed, error %d", in ubifs_leb_write()
128 ubifs_ro_mode(c, err); in ubifs_leb_write()
134 int ubifs_leb_change(struct ubifs_info *c, int lnum, const void *buf, int len) in ubifs_leb_change() argument
138 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_leb_change()
139 if (c->ro_error) in ubifs_leb_change()
140 return -EROFS; in ubifs_leb_change()
141 if (!dbg_is_tst_rcvry(c)) in ubifs_leb_change()
142 err = ubi_leb_change(c->ubi, lnum, buf, len); in ubifs_leb_change()
145 err = dbg_leb_change(c, lnum, buf, len); in ubifs_leb_change()
148 ubifs_err(c, "changing %d bytes in LEB %d failed, error %d", in ubifs_leb_change()
150 ubifs_ro_mode(c, err); in ubifs_leb_change()
156 int ubifs_leb_unmap(struct ubifs_info *c, int lnum) in ubifs_leb_unmap() argument
160 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_leb_unmap()
161 if (c->ro_error) in ubifs_leb_unmap()
162 return -EROFS; in ubifs_leb_unmap()
163 if (!dbg_is_tst_rcvry(c)) in ubifs_leb_unmap()
164 err = ubi_leb_unmap(c->ubi, lnum); in ubifs_leb_unmap()
167 err = dbg_leb_unmap(c, lnum); in ubifs_leb_unmap()
170 ubifs_err(c, "unmap LEB %d failed, error %d", lnum, err); in ubifs_leb_unmap()
171 ubifs_ro_mode(c, err); in ubifs_leb_unmap()
177 int ubifs_leb_map(struct ubifs_info *c, int lnum) in ubifs_leb_map() argument
181 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_leb_map()
182 if (c->ro_error) in ubifs_leb_map()
183 return -EROFS; in ubifs_leb_map()
184 if (!dbg_is_tst_rcvry(c)) in ubifs_leb_map()
185 err = ubi_leb_map(c->ubi, lnum); in ubifs_leb_map()
188 err = dbg_leb_map(c, lnum); in ubifs_leb_map()
191 ubifs_err(c, "mapping LEB %d failed, error %d", lnum, err); in ubifs_leb_map()
192 ubifs_ro_mode(c, err); in ubifs_leb_map()
198 int ubifs_is_mapped(const struct ubifs_info *c, int lnum) in ubifs_is_mapped() argument
202 err = ubi_is_mapped(c->ubi, lnum); in ubifs_is_mapped()
204 ubifs_err(c, "ubi_is_mapped failed for LEB %d, error %d", in ubifs_is_mapped()
212 * ubifs_check_node - check node.
213 * @c: UBIFS file-system description object
222 * feeds it a file-system image with incorrect nodes. For example, too large
226 * This function may skip data nodes CRC checking if @c->no_chk_data_crc is
228 * @must_chk_crc is true, then @c->no_chk_data_crc is ignored and CRC is
229 * checked. Similarly, if @c->mounting or @c->remounting_rw is true (we are
230 * mounting or re-mounting to R/W mode), @c->no_chk_data_crc is ignored and CRC
231 * is checked. This is because during mounting or re-mounting from R/O mode to
236 * This function returns zero in case of success and %-EUCLEAN in case of bad
239 int ubifs_check_node(const struct ubifs_info *c, const void *buf, int lnum, in ubifs_check_node() argument
242 int err = -EINVAL, type, node_len; in ubifs_check_node()
246 ubifs_assert(lnum >= 0 && lnum < c->leb_cnt && offs >= 0); in ubifs_check_node()
247 ubifs_assert(!(offs & 7) && offs < c->leb_size); in ubifs_check_node()
249 magic = le32_to_cpu(ch->magic); in ubifs_check_node()
252 ubifs_err(c, "bad magic %#08x, expected %#08x", in ubifs_check_node()
254 err = -EUCLEAN; in ubifs_check_node()
258 type = ch->node_type; in ubifs_check_node()
261 ubifs_err(c, "bad node type %d", type); in ubifs_check_node()
265 node_len = le32_to_cpu(ch->len); in ubifs_check_node()
266 if (node_len + offs > c->leb_size) in ubifs_check_node()
269 if (c->ranges[type].max_len == 0) { in ubifs_check_node()
270 if (node_len != c->ranges[type].len) in ubifs_check_node()
272 } else if (node_len < c->ranges[type].min_len || in ubifs_check_node()
273 node_len > c->ranges[type].max_len) in ubifs_check_node()
276 if (!must_chk_crc && type == UBIFS_DATA_NODE && !c->mounting && in ubifs_check_node()
277 !c->remounting_rw && c->no_chk_data_crc) in ubifs_check_node()
280 crc = crc32(UBIFS_CRC32_INIT, buf + 8, node_len - 8); in ubifs_check_node()
281 node_crc = le32_to_cpu(ch->crc); in ubifs_check_node()
284 ubifs_err(c, "bad CRC: calculated %#08x, read %#08x", in ubifs_check_node()
286 err = -EUCLEAN; in ubifs_check_node()
294 ubifs_err(c, "bad node length %d", node_len); in ubifs_check_node()
297 ubifs_err(c, "bad node at LEB %d:%d", lnum, offs); in ubifs_check_node()
298 ubifs_dump_node(c, buf); in ubifs_check_node()
305 * ubifs_pad - pad flash space.
306 * @c: UBIFS file-system description object
310 * The flash media obliges us to write only in chunks of %c->min_io_size and
311 * when we have to write less data we add padding node to the write-buffer and
317 * Padding nodes are also used to fill gaps when the "commit-in-gaps" method is
320 void ubifs_pad(const struct ubifs_info *c, void *buf, int pad) in ubifs_pad() argument
330 ch->magic = cpu_to_le32(UBIFS_NODE_MAGIC); in ubifs_pad()
331 ch->node_type = UBIFS_PAD_NODE; in ubifs_pad()
332 ch->group_type = UBIFS_NO_NODE_GROUP; in ubifs_pad()
333 ch->padding[0] = ch->padding[1] = 0; in ubifs_pad()
334 ch->sqnum = 0; in ubifs_pad()
335 ch->len = cpu_to_le32(UBIFS_PAD_NODE_SZ); in ubifs_pad()
336 pad -= UBIFS_PAD_NODE_SZ; in ubifs_pad()
337 pad_node->pad_len = cpu_to_le32(pad); in ubifs_pad()
338 crc = crc32(UBIFS_CRC32_INIT, buf + 8, UBIFS_PAD_NODE_SZ - 8); in ubifs_pad()
339 ch->crc = cpu_to_le32(crc); in ubifs_pad()
347 * next_sqnum - get next sequence number.
348 * @c: UBIFS file-system description object
350 static unsigned long long next_sqnum(struct ubifs_info *c) in next_sqnum() argument
354 spin_lock(&c->cnt_lock); in next_sqnum()
355 sqnum = ++c->max_sqnum; in next_sqnum()
356 spin_unlock(&c->cnt_lock); in next_sqnum()
360 ubifs_err(c, "sequence number overflow %llu, end of life", in next_sqnum()
362 ubifs_ro_mode(c, -EINVAL); in next_sqnum()
364 ubifs_warn(c, "running out of sequence numbers, end of life soon"); in next_sqnum()
371 * ubifs_prepare_node - prepare node to be written to flash.
372 * @c: UBIFS file-system description object
377 * This function prepares node at @node to be written to the media - it
381 void ubifs_prepare_node(struct ubifs_info *c, void *node, int len, int pad) in ubifs_prepare_node() argument
385 unsigned long long sqnum = next_sqnum(c); in ubifs_prepare_node()
389 ch->magic = cpu_to_le32(UBIFS_NODE_MAGIC); in ubifs_prepare_node()
390 ch->len = cpu_to_le32(len); in ubifs_prepare_node()
391 ch->group_type = UBIFS_NO_NODE_GROUP; in ubifs_prepare_node()
392 ch->sqnum = cpu_to_le64(sqnum); in ubifs_prepare_node()
393 ch->padding[0] = ch->padding[1] = 0; in ubifs_prepare_node()
394 crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); in ubifs_prepare_node()
395 ch->crc = cpu_to_le32(crc); in ubifs_prepare_node()
399 pad = ALIGN(len, c->min_io_size) - len; in ubifs_prepare_node()
400 ubifs_pad(c, node + len, pad); in ubifs_prepare_node()
405 * ubifs_prep_grp_node - prepare node of a group to be written to flash.
406 * @c: UBIFS file-system description object
411 * This function prepares node at @node to be written to the media - it
414 void ubifs_prep_grp_node(struct ubifs_info *c, void *node, int len, int last) in ubifs_prep_grp_node() argument
418 unsigned long long sqnum = next_sqnum(c); in ubifs_prep_grp_node()
422 ch->magic = cpu_to_le32(UBIFS_NODE_MAGIC); in ubifs_prep_grp_node()
423 ch->len = cpu_to_le32(len); in ubifs_prep_grp_node()
425 ch->group_type = UBIFS_LAST_OF_NODE_GROUP; in ubifs_prep_grp_node()
427 ch->group_type = UBIFS_IN_NODE_GROUP; in ubifs_prep_grp_node()
428 ch->sqnum = cpu_to_le64(sqnum); in ubifs_prep_grp_node()
429 ch->padding[0] = ch->padding[1] = 0; in ubifs_prep_grp_node()
430 crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); in ubifs_prep_grp_node()
431 ch->crc = cpu_to_le32(crc); in ubifs_prep_grp_node()
436 * wbuf_timer_callback - write-buffer timer callback function.
437 * @timer: timer data (write-buffer descriptor)
439 * This function is called when the write-buffer timer expires.
445 dbg_io("jhead %s", dbg_jhead(wbuf->jhead)); in wbuf_timer_callback_nolock()
446 wbuf->need_sync = 1; in wbuf_timer_callback_nolock()
447 wbuf->c->need_wbuf_sync = 1; in wbuf_timer_callback_nolock()
448 ubifs_wake_up_bgt(wbuf->c); in wbuf_timer_callback_nolock()
453 * new_wbuf_timer - start new write-buffer timer.
454 * @wbuf: write-buffer descriptor
458 ubifs_assert(!hrtimer_active(&wbuf->timer)); in new_wbuf_timer_nolock()
460 if (wbuf->no_timer) in new_wbuf_timer_nolock()
462 dbg_io("set timer for jhead %s, %llu-%llu millisecs", in new_wbuf_timer_nolock()
463 dbg_jhead(wbuf->jhead), in new_wbuf_timer_nolock()
464 div_u64(ktime_to_ns(wbuf->softlimit), USEC_PER_SEC), in new_wbuf_timer_nolock()
465 div_u64(ktime_to_ns(wbuf->softlimit) + wbuf->delta, in new_wbuf_timer_nolock()
467 hrtimer_start_range_ns(&wbuf->timer, wbuf->softlimit, wbuf->delta, in new_wbuf_timer_nolock()
473 * cancel_wbuf_timer - cancel write-buffer timer.
474 * @wbuf: write-buffer descriptor
478 if (wbuf->no_timer) in cancel_wbuf_timer_nolock()
480 wbuf->need_sync = 0; in cancel_wbuf_timer_nolock()
482 hrtimer_cancel(&wbuf->timer); in cancel_wbuf_timer_nolock()
487 * ubifs_wbuf_sync_nolock - synchronize write-buffer.
488 * @wbuf: write-buffer to synchronize
490 * This function synchronizes write-buffer @buf and returns zero in case of
493 * Note, although write-buffers are of @c->max_write_size, this function does
494 * not necessarily writes all @c->max_write_size bytes to the flash. Instead,
495 * if the write-buffer is only partially filled with data, only the used part
496 * of the write-buffer (aligned on @c->min_io_size boundary) is synchronized.
501 struct ubifs_info *c = wbuf->c; in ubifs_wbuf_sync_nolock() local
505 if (!wbuf->used || wbuf->lnum == -1) in ubifs_wbuf_sync_nolock()
506 /* Write-buffer is empty or not seeked */ in ubifs_wbuf_sync_nolock()
510 wbuf->lnum, wbuf->offs, wbuf->used, dbg_jhead(wbuf->jhead)); in ubifs_wbuf_sync_nolock()
511 ubifs_assert(!(wbuf->avail & 7)); in ubifs_wbuf_sync_nolock()
512 ubifs_assert(wbuf->offs + wbuf->size <= c->leb_size); in ubifs_wbuf_sync_nolock()
513 ubifs_assert(wbuf->size >= c->min_io_size); in ubifs_wbuf_sync_nolock()
514 ubifs_assert(wbuf->size <= c->max_write_size); in ubifs_wbuf_sync_nolock()
515 ubifs_assert(wbuf->size % c->min_io_size == 0); in ubifs_wbuf_sync_nolock()
516 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_wbuf_sync_nolock()
517 if (c->leb_size - wbuf->offs >= c->max_write_size) in ubifs_wbuf_sync_nolock()
518 ubifs_assert(!((wbuf->offs + wbuf->size) % c->max_write_size)); in ubifs_wbuf_sync_nolock()
520 if (c->ro_error) in ubifs_wbuf_sync_nolock()
521 return -EROFS; in ubifs_wbuf_sync_nolock()
527 sync_len = ALIGN(wbuf->used, c->min_io_size); in ubifs_wbuf_sync_nolock()
528 dirt = sync_len - wbuf->used; in ubifs_wbuf_sync_nolock()
530 ubifs_pad(c, wbuf->buf + wbuf->used, dirt); in ubifs_wbuf_sync_nolock()
531 err = ubifs_leb_write(c, wbuf->lnum, wbuf->buf, wbuf->offs, sync_len); in ubifs_wbuf_sync_nolock()
535 spin_lock(&wbuf->lock); in ubifs_wbuf_sync_nolock()
536 wbuf->offs += sync_len; in ubifs_wbuf_sync_nolock()
538 * Now @wbuf->offs is not necessarily aligned to @c->max_write_size. in ubifs_wbuf_sync_nolock()
540 * @c->max_write_size chunks and to @c->max_write_size-aligned offset. in ubifs_wbuf_sync_nolock()
541 * Thus, if @wbuf->offs is not aligned to @c->max_write_size now, make in ubifs_wbuf_sync_nolock()
542 * sure that @wbuf->offs + @wbuf->size is aligned to in ubifs_wbuf_sync_nolock()
543 * @c->max_write_size. This way we make sure that after next in ubifs_wbuf_sync_nolock()
544 * write-buffer flush we are again at the optimal offset (aligned to in ubifs_wbuf_sync_nolock()
545 * @c->max_write_size). in ubifs_wbuf_sync_nolock()
547 if (c->leb_size - wbuf->offs < c->max_write_size) in ubifs_wbuf_sync_nolock()
548 wbuf->size = c->leb_size - wbuf->offs; in ubifs_wbuf_sync_nolock()
549 else if (wbuf->offs & (c->max_write_size - 1)) in ubifs_wbuf_sync_nolock()
550 wbuf->size = ALIGN(wbuf->offs, c->max_write_size) - wbuf->offs; in ubifs_wbuf_sync_nolock()
552 wbuf->size = c->max_write_size; in ubifs_wbuf_sync_nolock()
553 wbuf->avail = wbuf->size; in ubifs_wbuf_sync_nolock()
554 wbuf->used = 0; in ubifs_wbuf_sync_nolock()
555 wbuf->next_ino = 0; in ubifs_wbuf_sync_nolock()
556 spin_unlock(&wbuf->lock); in ubifs_wbuf_sync_nolock()
558 if (wbuf->sync_callback) in ubifs_wbuf_sync_nolock()
559 err = wbuf->sync_callback(c, wbuf->lnum, in ubifs_wbuf_sync_nolock()
560 c->leb_size - wbuf->offs, dirt); in ubifs_wbuf_sync_nolock()
565 * ubifs_wbuf_seek_nolock - seek write-buffer.
566 * @wbuf: write-buffer
570 * This function targets the write-buffer to logical eraseblock @lnum:@offs.
571 * The write-buffer has to be empty. Returns zero in case of success and a
576 const struct ubifs_info *c = wbuf->c; in ubifs_wbuf_seek_nolock() local
578 dbg_io("LEB %d:%d, jhead %s", lnum, offs, dbg_jhead(wbuf->jhead)); in ubifs_wbuf_seek_nolock()
579 ubifs_assert(lnum >= 0 && lnum < c->leb_cnt); in ubifs_wbuf_seek_nolock()
580 ubifs_assert(offs >= 0 && offs <= c->leb_size); in ubifs_wbuf_seek_nolock()
581 ubifs_assert(offs % c->min_io_size == 0 && !(offs & 7)); in ubifs_wbuf_seek_nolock()
582 ubifs_assert(lnum != wbuf->lnum); in ubifs_wbuf_seek_nolock()
583 ubifs_assert(wbuf->used == 0); in ubifs_wbuf_seek_nolock()
585 spin_lock(&wbuf->lock); in ubifs_wbuf_seek_nolock()
586 wbuf->lnum = lnum; in ubifs_wbuf_seek_nolock()
587 wbuf->offs = offs; in ubifs_wbuf_seek_nolock()
588 if (c->leb_size - wbuf->offs < c->max_write_size) in ubifs_wbuf_seek_nolock()
589 wbuf->size = c->leb_size - wbuf->offs; in ubifs_wbuf_seek_nolock()
590 else if (wbuf->offs & (c->max_write_size - 1)) in ubifs_wbuf_seek_nolock()
591 wbuf->size = ALIGN(wbuf->offs, c->max_write_size) - wbuf->offs; in ubifs_wbuf_seek_nolock()
593 wbuf->size = c->max_write_size; in ubifs_wbuf_seek_nolock()
594 wbuf->avail = wbuf->size; in ubifs_wbuf_seek_nolock()
595 wbuf->used = 0; in ubifs_wbuf_seek_nolock()
596 spin_unlock(&wbuf->lock); in ubifs_wbuf_seek_nolock()
603 * ubifs_bg_wbufs_sync - synchronize write-buffers.
604 * @c: UBIFS file-system description object
606 * This function is called by background thread to synchronize write-buffers.
610 int ubifs_bg_wbufs_sync(struct ubifs_info *c) in ubifs_bg_wbufs_sync() argument
614 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_bg_wbufs_sync()
615 if (!c->need_wbuf_sync) in ubifs_bg_wbufs_sync()
617 c->need_wbuf_sync = 0; in ubifs_bg_wbufs_sync()
619 if (c->ro_error) { in ubifs_bg_wbufs_sync()
620 err = -EROFS; in ubifs_bg_wbufs_sync()
625 for (i = 0; i < c->jhead_cnt; i++) { in ubifs_bg_wbufs_sync()
626 struct ubifs_wbuf *wbuf = &c->jheads[i].wbuf; in ubifs_bg_wbufs_sync()
634 if (mutex_is_locked(&wbuf->io_mutex)) in ubifs_bg_wbufs_sync()
637 mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); in ubifs_bg_wbufs_sync()
638 if (!wbuf->need_sync) { in ubifs_bg_wbufs_sync()
639 mutex_unlock(&wbuf->io_mutex); in ubifs_bg_wbufs_sync()
644 mutex_unlock(&wbuf->io_mutex); in ubifs_bg_wbufs_sync()
646 ubifs_err(c, "cannot sync write-buffer, error %d", err); in ubifs_bg_wbufs_sync()
647 ubifs_ro_mode(c, err); in ubifs_bg_wbufs_sync()
656 for (i = 0; i < c->jhead_cnt; i++) { in ubifs_bg_wbufs_sync()
657 struct ubifs_wbuf *wbuf = &c->jheads[i].wbuf; in ubifs_bg_wbufs_sync()
659 mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); in ubifs_bg_wbufs_sync()
661 mutex_unlock(&wbuf->io_mutex); in ubifs_bg_wbufs_sync()
667 * ubifs_wbuf_write_nolock - write data to flash via write-buffer.
668 * @wbuf: write-buffer
672 * This function writes data to flash via write-buffer @wbuf. This means that
674 * does not take whole max. write unit (@c->max_write_size). Instead, the node
675 * will sit in RAM until the write-buffer is synchronized (e.g., by timer, or
676 * because more data are appended to the write-buffer).
680 * space in this logical eraseblock, %-ENOSPC is returned.
684 struct ubifs_info *c = wbuf->c; in ubifs_wbuf_write_nolock() local
688 dbg_ntype(((struct ubifs_ch *)buf)->node_type), in ubifs_wbuf_write_nolock()
689 dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs + wbuf->used); in ubifs_wbuf_write_nolock()
690 ubifs_assert(len > 0 && wbuf->lnum >= 0 && wbuf->lnum < c->leb_cnt); in ubifs_wbuf_write_nolock()
691 ubifs_assert(wbuf->offs >= 0 && wbuf->offs % c->min_io_size == 0); in ubifs_wbuf_write_nolock()
692 ubifs_assert(!(wbuf->offs & 7) && wbuf->offs <= c->leb_size); in ubifs_wbuf_write_nolock()
693 ubifs_assert(wbuf->avail > 0 && wbuf->avail <= wbuf->size); in ubifs_wbuf_write_nolock()
694 ubifs_assert(wbuf->size >= c->min_io_size); in ubifs_wbuf_write_nolock()
695 ubifs_assert(wbuf->size <= c->max_write_size); in ubifs_wbuf_write_nolock()
696 ubifs_assert(wbuf->size % c->min_io_size == 0); in ubifs_wbuf_write_nolock()
697 ubifs_assert(mutex_is_locked(&wbuf->io_mutex)); in ubifs_wbuf_write_nolock()
698 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_wbuf_write_nolock()
699 ubifs_assert(!c->space_fixup); in ubifs_wbuf_write_nolock()
700 if (c->leb_size - wbuf->offs >= c->max_write_size) in ubifs_wbuf_write_nolock()
701 ubifs_assert(!((wbuf->offs + wbuf->size) % c->max_write_size)); in ubifs_wbuf_write_nolock()
703 if (c->leb_size - wbuf->offs - wbuf->used < aligned_len) { in ubifs_wbuf_write_nolock()
704 err = -ENOSPC; in ubifs_wbuf_write_nolock()
710 if (c->ro_error) in ubifs_wbuf_write_nolock()
711 return -EROFS; in ubifs_wbuf_write_nolock()
713 if (aligned_len <= wbuf->avail) { in ubifs_wbuf_write_nolock()
716 * write-buffer. in ubifs_wbuf_write_nolock()
718 memcpy(wbuf->buf + wbuf->used, buf, len); in ubifs_wbuf_write_nolock()
720 if (aligned_len == wbuf->avail) { in ubifs_wbuf_write_nolock()
722 dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs); in ubifs_wbuf_write_nolock()
723 err = ubifs_leb_write(c, wbuf->lnum, wbuf->buf, in ubifs_wbuf_write_nolock()
724 wbuf->offs, wbuf->size); in ubifs_wbuf_write_nolock()
728 spin_lock(&wbuf->lock); in ubifs_wbuf_write_nolock()
729 wbuf->offs += wbuf->size; in ubifs_wbuf_write_nolock()
730 if (c->leb_size - wbuf->offs >= c->max_write_size) in ubifs_wbuf_write_nolock()
731 wbuf->size = c->max_write_size; in ubifs_wbuf_write_nolock()
733 wbuf->size = c->leb_size - wbuf->offs; in ubifs_wbuf_write_nolock()
734 wbuf->avail = wbuf->size; in ubifs_wbuf_write_nolock()
735 wbuf->used = 0; in ubifs_wbuf_write_nolock()
736 wbuf->next_ino = 0; in ubifs_wbuf_write_nolock()
737 spin_unlock(&wbuf->lock); in ubifs_wbuf_write_nolock()
739 spin_lock(&wbuf->lock); in ubifs_wbuf_write_nolock()
740 wbuf->avail -= aligned_len; in ubifs_wbuf_write_nolock()
741 wbuf->used += aligned_len; in ubifs_wbuf_write_nolock()
742 spin_unlock(&wbuf->lock); in ubifs_wbuf_write_nolock()
750 if (wbuf->used) { in ubifs_wbuf_write_nolock()
754 * write-buffer and switch to the next max. write unit. in ubifs_wbuf_write_nolock()
757 dbg_jhead(wbuf->jhead), wbuf->lnum, wbuf->offs); in ubifs_wbuf_write_nolock()
758 memcpy(wbuf->buf + wbuf->used, buf, wbuf->avail); in ubifs_wbuf_write_nolock()
759 err = ubifs_leb_write(c, wbuf->lnum, wbuf->buf, wbuf->offs, in ubifs_wbuf_write_nolock()
760 wbuf->size); in ubifs_wbuf_write_nolock()
764 wbuf->offs += wbuf->size; in ubifs_wbuf_write_nolock()
765 len -= wbuf->avail; in ubifs_wbuf_write_nolock()
766 aligned_len -= wbuf->avail; in ubifs_wbuf_write_nolock()
767 written += wbuf->avail; in ubifs_wbuf_write_nolock()
768 } else if (wbuf->offs & (c->max_write_size - 1)) { in ubifs_wbuf_write_nolock()
770 * The write-buffer offset is not aligned to in ubifs_wbuf_write_nolock()
771 * @c->max_write_size and @wbuf->size is less than in ubifs_wbuf_write_nolock()
772 * @c->max_write_size. Write @wbuf->size bytes to make sure the in ubifs_wbuf_write_nolock()
773 * following writes are done in optimal @c->max_write_size in ubifs_wbuf_write_nolock()
777 wbuf->size, wbuf->lnum, wbuf->offs); in ubifs_wbuf_write_nolock()
778 err = ubifs_leb_write(c, wbuf->lnum, buf, wbuf->offs, in ubifs_wbuf_write_nolock()
779 wbuf->size); in ubifs_wbuf_write_nolock()
783 wbuf->offs += wbuf->size; in ubifs_wbuf_write_nolock()
784 len -= wbuf->size; in ubifs_wbuf_write_nolock()
785 aligned_len -= wbuf->size; in ubifs_wbuf_write_nolock()
786 written += wbuf->size; in ubifs_wbuf_write_nolock()
792 * We align node length to 8-byte boundary because we anyway flash wbuf in ubifs_wbuf_write_nolock()
795 n = aligned_len >> c->max_write_shift; in ubifs_wbuf_write_nolock()
797 n <<= c->max_write_shift; in ubifs_wbuf_write_nolock()
798 dbg_io("write %d bytes to LEB %d:%d", n, wbuf->lnum, in ubifs_wbuf_write_nolock()
799 wbuf->offs); in ubifs_wbuf_write_nolock()
800 err = ubifs_leb_write(c, wbuf->lnum, buf + written, in ubifs_wbuf_write_nolock()
801 wbuf->offs, n); in ubifs_wbuf_write_nolock()
804 wbuf->offs += n; in ubifs_wbuf_write_nolock()
805 aligned_len -= n; in ubifs_wbuf_write_nolock()
806 len -= n; in ubifs_wbuf_write_nolock()
810 spin_lock(&wbuf->lock); in ubifs_wbuf_write_nolock()
814 * max. write unit, so write it to the write-buffer and we are in ubifs_wbuf_write_nolock()
817 memcpy(wbuf->buf, buf + written, len); in ubifs_wbuf_write_nolock()
819 if (c->leb_size - wbuf->offs >= c->max_write_size) in ubifs_wbuf_write_nolock()
820 wbuf->size = c->max_write_size; in ubifs_wbuf_write_nolock()
822 wbuf->size = c->leb_size - wbuf->offs; in ubifs_wbuf_write_nolock()
823 wbuf->avail = wbuf->size - aligned_len; in ubifs_wbuf_write_nolock()
824 wbuf->used = aligned_len; in ubifs_wbuf_write_nolock()
825 wbuf->next_ino = 0; in ubifs_wbuf_write_nolock()
826 spin_unlock(&wbuf->lock); in ubifs_wbuf_write_nolock()
829 if (wbuf->sync_callback) { in ubifs_wbuf_write_nolock()
830 int free = c->leb_size - wbuf->offs - wbuf->used; in ubifs_wbuf_write_nolock()
832 err = wbuf->sync_callback(c, wbuf->lnum, free, 0); in ubifs_wbuf_write_nolock()
837 if (wbuf->used) in ubifs_wbuf_write_nolock()
843 ubifs_err(c, "cannot write %d bytes to LEB %d:%d, error %d", in ubifs_wbuf_write_nolock()
844 len, wbuf->lnum, wbuf->offs, err); in ubifs_wbuf_write_nolock()
845 ubifs_dump_node(c, buf); in ubifs_wbuf_write_nolock()
847 ubifs_dump_leb(c, wbuf->lnum); in ubifs_wbuf_write_nolock()
852 * ubifs_write_node - write node to the media.
853 * @c: UBIFS file-system description object
865 int ubifs_write_node(struct ubifs_info *c, void *buf, int len, int lnum, in ubifs_write_node() argument
868 int err, buf_len = ALIGN(len, c->min_io_size); in ubifs_write_node()
871 lnum, offs, dbg_ntype(((struct ubifs_ch *)buf)->node_type), len, in ubifs_write_node()
873 ubifs_assert(lnum >= 0 && lnum < c->leb_cnt && offs >= 0); in ubifs_write_node()
874 ubifs_assert(offs % c->min_io_size == 0 && offs < c->leb_size); in ubifs_write_node()
875 ubifs_assert(!c->ro_media && !c->ro_mount); in ubifs_write_node()
876 ubifs_assert(!c->space_fixup); in ubifs_write_node()
878 if (c->ro_error) in ubifs_write_node()
879 return -EROFS; in ubifs_write_node()
881 ubifs_prepare_node(c, buf, len, 1); in ubifs_write_node()
882 err = ubifs_leb_write(c, lnum, buf, offs, buf_len); in ubifs_write_node()
884 ubifs_dump_node(c, buf); in ubifs_write_node()
891 * ubifs_read_node_wbuf - read node from the media or write-buffer.
892 * @wbuf: wbuf to check for un-written data
900 * in @buf. If the node partially or fully sits in the write-buffer, this
902 * Returns zero in case of success, %-EUCLEAN if CRC mismatched and a negative
908 const struct ubifs_info *c = wbuf->c; in ubifs_read_node_wbuf() local
913 dbg_ntype(type), len, dbg_jhead(wbuf->jhead)); in ubifs_read_node_wbuf()
914 ubifs_assert(wbuf && lnum >= 0 && lnum < c->leb_cnt && offs >= 0); in ubifs_read_node_wbuf()
915 ubifs_assert(!(offs & 7) && offs < c->leb_size); in ubifs_read_node_wbuf()
918 spin_lock(&wbuf->lock); in ubifs_read_node_wbuf()
919 overlap = (lnum == wbuf->lnum && offs + len > wbuf->offs); in ubifs_read_node_wbuf()
921 /* We may safely unlock the write-buffer and read the data */ in ubifs_read_node_wbuf()
922 spin_unlock(&wbuf->lock); in ubifs_read_node_wbuf()
923 return ubifs_read_node(c, buf, type, len, lnum, offs); in ubifs_read_node_wbuf()
927 rlen = wbuf->offs - offs; in ubifs_read_node_wbuf()
931 /* Copy the rest from the write-buffer */ in ubifs_read_node_wbuf()
932 memcpy(buf + rlen, wbuf->buf + offs + rlen - wbuf->offs, len - rlen); in ubifs_read_node_wbuf()
933 spin_unlock(&wbuf->lock); in ubifs_read_node_wbuf()
936 /* Read everything that goes before write-buffer */ in ubifs_read_node_wbuf()
937 err = ubifs_leb_read(c, lnum, buf, offs, rlen, 0); in ubifs_read_node_wbuf()
938 if (err && err != -EBADMSG) in ubifs_read_node_wbuf()
942 if (type != ch->node_type) { in ubifs_read_node_wbuf()
943 ubifs_err(c, "bad node type (%d but expected %d)", in ubifs_read_node_wbuf()
944 ch->node_type, type); in ubifs_read_node_wbuf()
948 err = ubifs_check_node(c, buf, lnum, offs, 0, 0); in ubifs_read_node_wbuf()
950 ubifs_err(c, "expected node type %d", type); in ubifs_read_node_wbuf()
954 rlen = le32_to_cpu(ch->len); in ubifs_read_node_wbuf()
956 ubifs_err(c, "bad node length %d, expected %d", rlen, len); in ubifs_read_node_wbuf()
963 ubifs_err(c, "bad node at LEB %d:%d", lnum, offs); in ubifs_read_node_wbuf()
964 ubifs_dump_node(c, buf); in ubifs_read_node_wbuf()
966 return -EINVAL; in ubifs_read_node_wbuf()
970 * ubifs_read_node - read node.
971 * @c: UBIFS file-system description object
979 * stores in @buf. Returns zero in case of success, %-EUCLEAN if CRC mismatched
982 int ubifs_read_node(const struct ubifs_info *c, void *buf, int type, int len, in ubifs_read_node() argument
989 ubifs_assert(lnum >= 0 && lnum < c->leb_cnt && offs >= 0); in ubifs_read_node()
990 ubifs_assert(len >= UBIFS_CH_SZ && offs + len <= c->leb_size); in ubifs_read_node()
991 ubifs_assert(!(offs & 7) && offs < c->leb_size); in ubifs_read_node()
994 err = ubifs_leb_read(c, lnum, buf, offs, len, 0); in ubifs_read_node()
995 if (err && err != -EBADMSG) in ubifs_read_node()
998 if (type != ch->node_type) { in ubifs_read_node()
999 ubifs_errc(c, "bad node type (%d but expected %d)", in ubifs_read_node()
1000 ch->node_type, type); in ubifs_read_node()
1004 err = ubifs_check_node(c, buf, lnum, offs, 0, 0); in ubifs_read_node()
1006 ubifs_errc(c, "expected node type %d", type); in ubifs_read_node()
1010 l = le32_to_cpu(ch->len); in ubifs_read_node()
1012 ubifs_errc(c, "bad node length %d, expected %d", l, len); in ubifs_read_node()
1019 ubifs_errc(c, "bad node at LEB %d:%d, LEB mapping status %d", lnum, in ubifs_read_node()
1020 offs, ubi_is_mapped(c->ubi, lnum)); in ubifs_read_node()
1021 if (!c->probing) { in ubifs_read_node()
1022 ubifs_dump_node(c, buf); in ubifs_read_node()
1025 return -EINVAL; in ubifs_read_node()
1029 * ubifs_wbuf_init - initialize write-buffer.
1030 * @c: UBIFS file-system description object
1031 * @wbuf: write-buffer to initialize
1033 * This function initializes write-buffer. Returns zero in case of success
1034 * %-ENOMEM in case of failure.
1036 int ubifs_wbuf_init(struct ubifs_info *c, struct ubifs_wbuf *wbuf) in ubifs_wbuf_init() argument
1040 wbuf->buf = kmalloc(c->max_write_size, GFP_KERNEL); in ubifs_wbuf_init()
1041 if (!wbuf->buf) in ubifs_wbuf_init()
1042 return -ENOMEM; in ubifs_wbuf_init()
1044 size = (c->max_write_size / UBIFS_CH_SZ + 1) * sizeof(ino_t); in ubifs_wbuf_init()
1045 wbuf->inodes = kmalloc(size, GFP_KERNEL); in ubifs_wbuf_init()
1046 if (!wbuf->inodes) { in ubifs_wbuf_init()
1047 kfree(wbuf->buf); in ubifs_wbuf_init()
1048 wbuf->buf = NULL; in ubifs_wbuf_init()
1049 return -ENOMEM; in ubifs_wbuf_init()
1052 wbuf->used = 0; in ubifs_wbuf_init()
1053 wbuf->lnum = wbuf->offs = -1; in ubifs_wbuf_init()
1056 * write-buffer size has to be set to @c->max_write_size. Otherwise, in ubifs_wbuf_init()
1060 size = c->max_write_size - (c->leb_start % c->max_write_size); in ubifs_wbuf_init()
1061 wbuf->avail = wbuf->size = size; in ubifs_wbuf_init()
1062 wbuf->sync_callback = NULL; in ubifs_wbuf_init()
1063 mutex_init(&wbuf->io_mutex); in ubifs_wbuf_init()
1064 spin_lock_init(&wbuf->lock); in ubifs_wbuf_init()
1065 wbuf->c = c; in ubifs_wbuf_init()
1066 wbuf->next_ino = 0; in ubifs_wbuf_init()
1069 hrtimer_init(&wbuf->timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL); in ubifs_wbuf_init()
1070 wbuf->timer.function = wbuf_timer_callback_nolock; in ubifs_wbuf_init()
1071 wbuf->softlimit = ktime_set(WBUF_TIMEOUT_SOFTLIMIT, 0); in ubifs_wbuf_init()
1072 wbuf->delta = WBUF_TIMEOUT_HARDLIMIT - WBUF_TIMEOUT_SOFTLIMIT; in ubifs_wbuf_init()
1073 wbuf->delta *= 1000000000ULL; in ubifs_wbuf_init()
1074 ubifs_assert(wbuf->delta <= ULONG_MAX); in ubifs_wbuf_init()
1080 * ubifs_wbuf_add_ino_nolock - add an inode number into the wbuf inode array.
1081 * @wbuf: the write-buffer where to add
1084 * This function adds an inode number to the inode array of the write-buffer.
1088 if (!wbuf->buf) in ubifs_wbuf_add_ino_nolock()
1092 spin_lock(&wbuf->lock); in ubifs_wbuf_add_ino_nolock()
1093 if (wbuf->used) in ubifs_wbuf_add_ino_nolock()
1094 wbuf->inodes[wbuf->next_ino++] = inum; in ubifs_wbuf_add_ino_nolock()
1095 spin_unlock(&wbuf->lock); in ubifs_wbuf_add_ino_nolock()
1099 * wbuf_has_ino - returns if the wbuf contains data from the inode.
1100 * @wbuf: the write-buffer
1103 * This function returns with %1 if the write-buffer contains some data from the
1110 spin_lock(&wbuf->lock); in wbuf_has_ino()
1111 for (i = 0; i < wbuf->next_ino; i++) in wbuf_has_ino()
1112 if (inum == wbuf->inodes[i]) { in wbuf_has_ino()
1116 spin_unlock(&wbuf->lock); in wbuf_has_ino()
1122 * ubifs_sync_wbufs_by_inode - synchronize write-buffers for an inode.
1123 * @c: UBIFS file-system description object
1126 * This function synchronizes write-buffers which contain nodes belonging to
1130 int ubifs_sync_wbufs_by_inode(struct ubifs_info *c, struct inode *inode) in ubifs_sync_wbufs_by_inode() argument
1134 for (i = 0; i < c->jhead_cnt; i++) { in ubifs_sync_wbufs_by_inode()
1135 struct ubifs_wbuf *wbuf = &c->jheads[i].wbuf; in ubifs_sync_wbufs_by_inode()
1141 * a _copy_ of corresponding on-flash node which sits in ubifs_sync_wbufs_by_inode()
1146 if (!wbuf_has_ino(wbuf, inode->i_ino)) in ubifs_sync_wbufs_by_inode()
1149 mutex_lock_nested(&wbuf->io_mutex, wbuf->jhead); in ubifs_sync_wbufs_by_inode()
1150 if (wbuf_has_ino(wbuf, inode->i_ino)) in ubifs_sync_wbufs_by_inode()
1152 mutex_unlock(&wbuf->io_mutex); in ubifs_sync_wbufs_by_inode()
1155 ubifs_ro_mode(c, err); in ubifs_sync_wbufs_by_inode()