Lines Matching refs:btt
255 static void btt_debugfs_init(struct btt *btt) in btt_debugfs_init() argument
260 btt->debugfs_dir = debugfs_create_dir(dev_name(&btt->nd_btt->dev), in btt_debugfs_init()
262 if (IS_ERR_OR_NULL(btt->debugfs_dir)) in btt_debugfs_init()
265 list_for_each_entry(arena, &btt->arena_list, list) { in btt_debugfs_init()
266 arena_debugfs_init(arena, btt->debugfs_dir, i); in btt_debugfs_init()
745 static struct arena_info *alloc_arena(struct btt *btt, size_t size, in alloc_arena() argument
755 arena->nd_btt = btt->nd_btt; in alloc_arena()
756 arena->sector_size = btt->sector_size; in alloc_arena()
764 arena->external_lbasize = btt->lbasize; in alloc_arena()
768 arena->version_major = btt->nd_btt->version_major; in alloc_arena()
769 arena->version_minor = btt->nd_btt->version_minor; in alloc_arena()
802 static void free_arenas(struct btt *btt) in free_arenas() argument
806 list_for_each_entry_safe(arena, next, &btt->arena_list, list) { in free_arenas()
846 static int discover_arenas(struct btt *btt) in discover_arenas() argument
851 size_t remaining = btt->rawsize; in discover_arenas()
862 arena = alloc_arena(btt, 0, 0, 0); in discover_arenas()
873 if (!nd_btt_arena_is_valid(btt->nd_btt, super)) { in discover_arenas()
874 if (remaining == btt->rawsize) { in discover_arenas()
875 btt->init_state = INIT_NOTFOUND; in discover_arenas()
908 list_add_tail(&arena->list, &btt->arena_list); in discover_arenas()
918 btt->num_arenas = num_arenas; in discover_arenas()
919 btt->nlba = cur_nlba; in discover_arenas()
920 btt->init_state = INIT_READY; in discover_arenas()
927 free_arenas(btt); in discover_arenas()
933 static int create_arenas(struct btt *btt) in create_arenas() argument
935 size_t remaining = btt->rawsize; in create_arenas()
946 arena = alloc_arena(btt, arena_size, btt->nlba, cur_off); in create_arenas()
948 free_arenas(btt); in create_arenas()
951 btt->nlba += arena->external_nlba; in create_arenas()
957 list_add_tail(&arena->list, &btt->arena_list); in create_arenas()
1025 static int btt_meta_init(struct btt *btt) in btt_meta_init() argument
1030 mutex_lock(&btt->init_lock); in btt_meta_init()
1031 list_for_each_entry(arena, &btt->arena_list, list) { in btt_meta_init()
1049 btt->init_state = INIT_READY; in btt_meta_init()
1052 mutex_unlock(&btt->init_lock); in btt_meta_init()
1056 static u32 btt_meta_size(struct btt *btt) in btt_meta_size() argument
1058 return btt->lbasize - btt->sector_size; in btt_meta_size()
1068 static int lba_to_arena(struct btt *btt, sector_t sector, __u32 *premap, in lba_to_arena() argument
1072 __u64 lba = div_u64(sector << SECTOR_SHIFT, btt->sector_size); in lba_to_arena()
1074 list_for_each_entry(arena_list, &btt->arena_list, list) { in lba_to_arena()
1141 static int btt_rw_integrity(struct btt *btt, struct bio_integrity_payload *bip, in btt_rw_integrity() argument
1144 unsigned int len = btt_meta_size(btt); in btt_rw_integrity()
1151 meta_nsoff = to_namespace_offset(arena, postmap) + btt->sector_size; in btt_rw_integrity()
1190 static int btt_rw_integrity(struct btt *btt, struct bio_integrity_payload *bip, in btt_rw_integrity() argument
1197 static int btt_read_pg(struct btt *btt, struct bio_integrity_payload *bip, in btt_read_pg() argument
1209 lane = nd_region_acquire_lane(btt->nd_region); in btt_read_pg()
1211 ret = lba_to_arena(btt, sector, &premap, &arena); in btt_read_pg()
1215 cur_len = min(btt->sector_size, len); in btt_read_pg()
1273 ret = btt_rw_integrity(btt, bip, arena, postmap, READ); in btt_read_pg()
1279 nd_region_release_lane(btt->nd_region, lane); in btt_read_pg()
1283 sector += btt->sector_size >> SECTOR_SHIFT; in btt_read_pg()
1291 nd_region_release_lane(btt->nd_region, lane); in btt_read_pg()
1300 static bool btt_is_badblock(struct btt *btt, struct arena_info *arena, in btt_is_badblock() argument
1307 return is_bad_pmem(btt->phys_bb, phys_sector, arena->internal_lbasize); in btt_is_badblock()
1310 static int btt_write_pg(struct btt *btt, struct bio_integrity_payload *bip, in btt_write_pg() argument
1325 lane = nd_region_acquire_lane(btt->nd_region); in btt_write_pg()
1327 ret = lba_to_arena(btt, sector, &premap, &arena); in btt_write_pg()
1330 cur_len = min(btt->sector_size, len); in btt_write_pg()
1337 if (btt_is_badblock(btt, arena, arena->freelist[lane].block)) in btt_write_pg()
1342 nd_region_release_lane(btt->nd_region, lane); in btt_write_pg()
1370 ret = btt_rw_integrity(btt, bip, arena, new_postmap, in btt_write_pg()
1403 nd_region_release_lane(btt->nd_region, lane); in btt_write_pg()
1413 sector += btt->sector_size >> SECTOR_SHIFT; in btt_write_pg()
1421 nd_region_release_lane(btt->nd_region, lane); in btt_write_pg()
1425 static int btt_do_bvec(struct btt *btt, struct bio_integrity_payload *bip, in btt_do_bvec() argument
1432 ret = btt_read_pg(btt, bip, page, off, sector, len); in btt_do_bvec()
1436 ret = btt_write_pg(btt, bip, sector, page, off, len); in btt_do_bvec()
1445 struct btt *btt = bio->bi_disk->private_data; in btt_submit_bio() local
1461 if (len > PAGE_SIZE || len < btt->sector_size || in btt_submit_bio()
1462 len % btt->sector_size) { in btt_submit_bio()
1463 dev_err_ratelimited(&btt->nd_btt->dev, in btt_submit_bio()
1469 err = btt_do_bvec(btt, bip, bvec.bv_page, len, bvec.bv_offset, in btt_submit_bio()
1472 dev_err(&btt->nd_btt->dev, in btt_submit_bio()
1491 struct btt *btt = bdev->bd_disk->private_data; in btt_rw_page() local
1494 rc = btt_do_bvec(btt, NULL, page, thp_size(page), 0, op, sector); in btt_rw_page()
1518 static int btt_blk_init(struct btt *btt) in btt_blk_init() argument
1520 struct nd_btt *nd_btt = btt->nd_btt; in btt_blk_init()
1524 btt->btt_queue = blk_alloc_queue(NUMA_NO_NODE); in btt_blk_init()
1525 if (!btt->btt_queue) in btt_blk_init()
1528 btt->btt_disk = alloc_disk(0); in btt_blk_init()
1529 if (!btt->btt_disk) { in btt_blk_init()
1530 blk_cleanup_queue(btt->btt_queue); in btt_blk_init()
1534 nvdimm_namespace_disk_name(ndns, btt->btt_disk->disk_name); in btt_blk_init()
1535 btt->btt_disk->first_minor = 0; in btt_blk_init()
1536 btt->btt_disk->fops = &btt_fops; in btt_blk_init()
1537 btt->btt_disk->private_data = btt; in btt_blk_init()
1538 btt->btt_disk->queue = btt->btt_queue; in btt_blk_init()
1539 btt->btt_disk->flags = GENHD_FL_EXT_DEVT; in btt_blk_init()
1541 blk_queue_logical_block_size(btt->btt_queue, btt->sector_size); in btt_blk_init()
1542 blk_queue_max_hw_sectors(btt->btt_queue, UINT_MAX); in btt_blk_init()
1543 blk_queue_flag_set(QUEUE_FLAG_NONROT, btt->btt_queue); in btt_blk_init()
1545 if (btt_meta_size(btt)) { in btt_blk_init()
1546 int rc = nd_integrity_init(btt->btt_disk, btt_meta_size(btt)); in btt_blk_init()
1549 del_gendisk(btt->btt_disk); in btt_blk_init()
1550 put_disk(btt->btt_disk); in btt_blk_init()
1551 blk_cleanup_queue(btt->btt_queue); in btt_blk_init()
1555 set_capacity(btt->btt_disk, btt->nlba * btt->sector_size >> 9); in btt_blk_init()
1556 device_add_disk(&btt->nd_btt->dev, btt->btt_disk, NULL); in btt_blk_init()
1557 btt->nd_btt->size = btt->nlba * (u64)btt->sector_size; in btt_blk_init()
1558 nvdimm_check_and_set_ro(btt->btt_disk); in btt_blk_init()
1563 static void btt_blk_cleanup(struct btt *btt) in btt_blk_cleanup() argument
1565 del_gendisk(btt->btt_disk); in btt_blk_cleanup()
1566 put_disk(btt->btt_disk); in btt_blk_cleanup()
1567 blk_cleanup_queue(btt->btt_queue); in btt_blk_cleanup()
1587 static struct btt *btt_init(struct nd_btt *nd_btt, unsigned long long rawsize, in btt_init()
1591 struct btt *btt; in btt_init() local
1595 btt = devm_kzalloc(dev, sizeof(struct btt), GFP_KERNEL); in btt_init()
1596 if (!btt) in btt_init()
1599 btt->nd_btt = nd_btt; in btt_init()
1600 btt->rawsize = rawsize; in btt_init()
1601 btt->lbasize = lbasize; in btt_init()
1602 btt->sector_size = ((lbasize >= 4096) ? 4096 : 512); in btt_init()
1603 INIT_LIST_HEAD(&btt->arena_list); in btt_init()
1604 mutex_init(&btt->init_lock); in btt_init()
1605 btt->nd_region = nd_region; in btt_init()
1607 btt->phys_bb = &nsio->bb; in btt_init()
1609 ret = discover_arenas(btt); in btt_init()
1615 if (btt->init_state != INIT_READY && nd_region->ro) { in btt_init()
1619 } else if (btt->init_state != INIT_READY) { in btt_init()
1620 btt->num_arenas = (rawsize / ARENA_MAX_SIZE) + in btt_init()
1623 btt->num_arenas, rawsize); in btt_init()
1625 ret = create_arenas(btt); in btt_init()
1631 ret = btt_meta_init(btt); in btt_init()
1638 ret = btt_blk_init(btt); in btt_init()
1644 btt_debugfs_init(btt); in btt_init()
1646 return btt; in btt_init()
1658 static void btt_fini(struct btt *btt) in btt_fini() argument
1660 if (btt) { in btt_fini()
1661 btt_blk_cleanup(btt); in btt_fini()
1662 free_arenas(btt); in btt_fini()
1663 debugfs_remove_recursive(btt->debugfs_dir); in btt_fini()
1672 struct btt *btt; in nvdimm_namespace_attach_btt() local
1706 btt = btt_init(nd_btt, rawsize, nd_btt->lbasize, nd_btt->uuid, in nvdimm_namespace_attach_btt()
1708 if (!btt) in nvdimm_namespace_attach_btt()
1710 nd_btt->btt = btt; in nvdimm_namespace_attach_btt()
1718 struct btt *btt = nd_btt->btt; in nvdimm_namespace_detach_btt() local
1720 btt_fini(btt); in nvdimm_namespace_detach_btt()
1721 nd_btt->btt = NULL; in nvdimm_namespace_detach_btt()