Lines Matching refs:sectors

318 		r10_bio->devs[slot].addr + (r10_bio->sectors);  in update_head_pos()
406 r10_bio->sectors, in close_write()
512 r10_bio->sectors, in raid10_end_write_request()
709 int sectors = r10_bio->sectors; in read_balance() local
736 && (this_sector + sectors >= conf->next_resync)) || in read_balance()
739 this_sector + sectors))) in read_balance()
754 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
760 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
764 if (is_badblock(rdev, dev_sector, sectors, in read_balance()
775 if (!do_balance && sectors > bad_sectors) in read_balance()
776 sectors = bad_sectors; in read_balance()
777 if (best_good_sectors > sectors) in read_balance()
778 best_good_sectors = sectors; in read_balance()
793 best_good_sectors = sectors; in read_balance()
1101 struct bio *bio, sector_t sectors) in regular_request_wait() argument
1106 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait()
1112 sectors); in regular_request_wait()
1159 regular_request_wait(mddev, conf, bio, r10_bio->sectors); in raid10_read_request()
1184 r10_bio->sectors = max_sectors; in raid10_read_request()
1285 sector_t sectors; in raid10_write_request() local
1304 sectors = r10_bio->sectors; in raid10_write_request()
1305 regular_request_wait(mddev, conf, bio, sectors); in raid10_write_request()
1309 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request()
1310 : (bio->bi_iter.bi_sector + sectors > conf->reshape_safe && in raid10_write_request()
1345 max_sectors = r10_bio->sectors; in raid10_write_request()
1457 if (max_sectors < r10_bio->sectors) in raid10_write_request()
1458 r10_bio->sectors = max_sectors; in raid10_write_request()
1460 if (r10_bio->sectors < bio_sectors(bio)) { in raid10_write_request()
1461 struct bio *split = bio_split(bio, r10_bio->sectors, in raid10_write_request()
1472 md_bitmap_startwrite(mddev->bitmap, r10_bio->sector, r10_bio->sectors, 0); in raid10_write_request()
1483 static void __make_request(struct mddev *mddev, struct bio *bio, int sectors) in __make_request() argument
1491 r10_bio->sectors = sectors; in __make_request()
1510 int sectors = bio_sectors(bio); in raid10_make_request() local
1524 sectors > chunk_sects in raid10_make_request()
1528 sectors = chunk_sects - in raid10_make_request()
1531 __make_request(mddev, bio, sectors); in raid10_make_request()
1880 atomic_add(r10_bio->sectors, in __end_sync_read()
1920 sector_t s = r10_bio->sectors; in end_sync_request()
1970 r10_bio->sectors, in end_sync_write()
2015 fbio->bi_iter.bi_size = r10_bio->sectors << 9; in sync_request_write()
2019 vcnt = (r10_bio->sectors + (PAGE_SIZE >> 9) - 1) >> (PAGE_SHIFT - 9); in sync_request_write()
2041 int sectors = r10_bio->sectors; in sync_request_write() local
2044 if (sectors < (len / 512)) in sync_request_write()
2045 len = sectors * 512; in sync_request_write()
2050 sectors -= len/512; in sync_request_write()
2054 atomic64_add(r10_bio->sectors, &mddev->resync_mismatches); in sync_request_write()
2113 md_done_sync(mddev, r10_bio->sectors, 1); in sync_request_write()
2141 int sectors = r10_bio->sectors; in fix_recovery_read_error() local
2147 while (sectors) { in fix_recovery_read_error()
2148 int s = sectors; in fix_recovery_read_error()
2205 sectors -= s; in fix_recovery_read_error()
2286 int sectors, struct page *page, int rw) in r10_sync_page_io() argument
2291 if (is_badblock(rdev, sector, sectors, &first_bad, &bad_sectors) in r10_sync_page_io()
2294 if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false)) in r10_sync_page_io()
2304 if (!rdev_set_badblocks(rdev, sector, sectors, 0)) in r10_sync_page_io()
2320 int sectors = r10_bio->sectors; in fix_read_error() local
2351 while(sectors) { in fix_read_error()
2352 int s = sectors; in fix_read_error()
2497 sectors -= s; in fix_read_error()
2521 int sectors; in narrow_write_error() local
2522 int sect_to_write = r10_bio->sectors; in narrow_write_error()
2531 sectors = ((r10_bio->sector + block_sectors) in narrow_write_error()
2538 if (sectors > sect_to_write) in narrow_write_error()
2539 sectors = sect_to_write; in narrow_write_error()
2542 bio_trim(wbio, sector - bio->bi_iter.bi_sector, sectors); in narrow_write_error()
2552 sectors, 0) in narrow_write_error()
2556 sect_to_write -= sectors; in narrow_write_error()
2557 sector += sectors; in narrow_write_error()
2558 sectors = block_sectors; in narrow_write_error()
2620 r10_bio->sectors, 0); in handle_write_completed()
2625 r10_bio->sectors, 0)) in handle_write_completed()
2637 r10_bio->sectors, 0); in handle_write_completed()
2642 r10_bio->sectors, 0)) in handle_write_completed()
2657 r10_bio->sectors, 0); in handle_write_completed()
2674 r10_bio->sectors, 0); in handle_write_completed()
3319 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()
3427 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3481 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3512 raid10_size(struct mddev *mddev, sector_t sectors, int raid_disks) in raid10_size() argument
3520 if (!sectors) in raid10_size()
3521 sectors = conf->dev_sectors; in raid10_size()
3523 size = sectors >> conf->geo.chunk_shift; in raid10_size()
3944 static int raid10_resize(struct mddev *mddev, sector_t sectors) in raid10_resize() argument
3968 size = raid10_size(mddev, sectors, 0); in raid10_resize()
3978 if (sectors > mddev->dev_sectors && in raid10_resize()
3983 calc_sectors(conf, sectors); in raid10_resize()
4017 rdev->sectors = size; in raid10_takeover_raid0()
4525 r10_bio->sectors = last - sector_nr + 1; in reshape_request()
4630 r10_bio->sectors = nr_sectors; in reshape_request()
4633 md_sync_acct_bio(read_bio, r10_bio->sectors); in reshape_request()
4670 md_done_sync(mddev, r10_bio->sectors, 0); in reshape_request_write()
4696 md_sync_acct_bio(b, r10_bio->sectors); in reshape_request_write()
4739 int sectors = r10_bio->sectors; in handle_reshape_read_error() local
4758 while (sectors) { in handle_reshape_read_error()
4759 int s = sectors; in handle_reshape_read_error()
4803 sectors -= s; in handle_reshape_read_error()
4841 md_done_sync(r10_bio->mddev, r10_bio->sectors, 1); in end_reshape_request()