Lines Matching refs:aeb
118 struct ubi_ainf_peb *aeb; in add_to_list() local
130 aeb = kmem_cache_alloc(ai->aeb_slab_cache, GFP_KERNEL); in add_to_list()
131 if (!aeb) in add_to_list()
134 aeb->pnum = pnum; in add_to_list()
135 aeb->vol_id = vol_id; in add_to_list()
136 aeb->lnum = lnum; in add_to_list()
137 aeb->ec = ec; in add_to_list()
139 list_add(&aeb->u.list, list); in add_to_list()
141 list_add_tail(&aeb->u.list, list); in add_to_list()
158 struct ubi_ainf_peb *aeb; in add_corrupted() local
162 aeb = kmem_cache_alloc(ai->aeb_slab_cache, GFP_KERNEL); in add_corrupted()
163 if (!aeb) in add_corrupted()
167 aeb->pnum = pnum; in add_corrupted()
168 aeb->ec = ec; in add_corrupted()
169 list_add(&aeb->u.list, &ai->corr); in add_corrupted()
320 int ubi_compare_lebs(struct ubi_device *ubi, const struct ubi_ainf_peb *aeb, in ubi_compare_lebs() argument
328 if (sqnum2 == aeb->sqnum) { in ubi_compare_lebs()
342 second_is_newer = (sqnum2 > aeb->sqnum); in ubi_compare_lebs()
361 if (!aeb->copy_flag) { in ubi_compare_lebs()
372 pnum = aeb->pnum; in ubi_compare_lebs()
451 struct ubi_ainf_peb *aeb; in ubi_add_to_av() local
477 aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); in ubi_add_to_av()
478 if (lnum != aeb->lnum) { in ubi_add_to_av()
479 if (lnum < aeb->lnum) in ubi_add_to_av()
492 aeb->pnum, aeb->sqnum, aeb->ec); in ubi_add_to_av()
507 if (aeb->sqnum == sqnum && sqnum != 0) { in ubi_add_to_av()
510 ubi_dump_aeb(aeb, 0); in ubi_add_to_av()
519 cmp_res = ubi_compare_lebs(ubi, aeb, pnum, vid_hdr); in ubi_add_to_av()
532 err = add_to_list(ai, aeb->pnum, aeb->vol_id, in ubi_add_to_av()
533 aeb->lnum, aeb->ec, cmp_res & 4, in ubi_add_to_av()
538 aeb->ec = ec; in ubi_add_to_av()
539 aeb->pnum = pnum; in ubi_add_to_av()
540 aeb->vol_id = vol_id; in ubi_add_to_av()
541 aeb->lnum = lnum; in ubi_add_to_av()
542 aeb->scrub = ((cmp_res & 2) || bitflips); in ubi_add_to_av()
543 aeb->copy_flag = vid_hdr->copy_flag; in ubi_add_to_av()
544 aeb->sqnum = sqnum; in ubi_add_to_av()
570 aeb = kmem_cache_alloc(ai->aeb_slab_cache, GFP_KERNEL); in ubi_add_to_av()
571 if (!aeb) in ubi_add_to_av()
574 aeb->ec = ec; in ubi_add_to_av()
575 aeb->pnum = pnum; in ubi_add_to_av()
576 aeb->vol_id = vol_id; in ubi_add_to_av()
577 aeb->lnum = lnum; in ubi_add_to_av()
578 aeb->scrub = bitflips; in ubi_add_to_av()
579 aeb->copy_flag = vid_hdr->copy_flag; in ubi_add_to_av()
580 aeb->sqnum = sqnum; in ubi_add_to_av()
588 rb_link_node(&aeb->u.rb, parent, p); in ubi_add_to_av()
589 rb_insert_color(&aeb->u.rb, &av->root); in ubi_add_to_av()
630 struct ubi_ainf_peb *aeb; in ubi_remove_av() local
635 aeb = rb_entry(rb, struct ubi_ainf_peb, u.rb); in ubi_remove_av()
636 rb_erase(&aeb->u.rb, &av->root); in ubi_remove_av()
637 list_add_tail(&aeb->u.list, &ai->erase); in ubi_remove_av()
709 struct ubi_ainf_peb *aeb, *tmp_aeb; in ubi_early_get_peb() local
712 aeb = list_entry(ai->free.next, struct ubi_ainf_peb, u.list); in ubi_early_get_peb()
713 list_del(&aeb->u.list); in ubi_early_get_peb()
714 dbg_bld("return free PEB %d, EC %d", aeb->pnum, aeb->ec); in ubi_early_get_peb()
715 return aeb; in ubi_early_get_peb()
724 list_for_each_entry_safe(aeb, tmp_aeb, &ai->erase, u.list) { in ubi_early_get_peb()
725 if (aeb->ec == UBI_UNKNOWN) in ubi_early_get_peb()
726 aeb->ec = ai->mean_ec; in ubi_early_get_peb()
728 err = early_erase_peb(ubi, ai, aeb->pnum, aeb->ec+1); in ubi_early_get_peb()
732 aeb->ec += 1; in ubi_early_get_peb()
733 list_del(&aeb->u.list); in ubi_early_get_peb()
734 dbg_bld("return PEB %d, EC %d", aeb->pnum, aeb->ec); in ubi_early_get_peb()
735 return aeb; in ubi_early_get_peb()
1066 struct ubi_ainf_peb *aeb; in late_analysis() local
1081 list_for_each_entry(aeb, &ai->corr, u.list) in late_analysis()
1082 pr_cont(" %d", aeb->pnum); in late_analysis()
1135 struct ubi_ainf_peb *aeb; in destroy_av() local
1144 aeb = rb_entry(this, struct ubi_ainf_peb, u.rb); in destroy_av()
1147 if (this->rb_left == &aeb->u.rb) in destroy_av()
1153 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_av()
1165 struct ubi_ainf_peb *aeb, *aeb_tmp; in destroy_ai() local
1169 list_for_each_entry_safe(aeb, aeb_tmp, &ai->alien, u.list) { in destroy_ai()
1170 list_del(&aeb->u.list); in destroy_ai()
1171 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1173 list_for_each_entry_safe(aeb, aeb_tmp, &ai->erase, u.list) { in destroy_ai()
1174 list_del(&aeb->u.list); in destroy_ai()
1175 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1177 list_for_each_entry_safe(aeb, aeb_tmp, &ai->corr, u.list) { in destroy_ai()
1178 list_del(&aeb->u.list); in destroy_ai()
1179 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1181 list_for_each_entry_safe(aeb, aeb_tmp, &ai->free, u.list) { in destroy_ai()
1182 list_del(&aeb->u.list); in destroy_ai()
1183 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1229 struct ubi_ainf_peb *aeb; in scan_all() local
1265 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) in scan_all()
1266 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1267 aeb->ec = ai->mean_ec; in scan_all()
1270 list_for_each_entry(aeb, &ai->free, u.list) { in scan_all()
1271 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1272 aeb->ec = ai->mean_ec; in scan_all()
1275 list_for_each_entry(aeb, &ai->corr, u.list) in scan_all()
1276 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1277 aeb->ec = ai->mean_ec; in scan_all()
1279 list_for_each_entry(aeb, &ai->erase, u.list) in scan_all()
1280 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1281 aeb->ec = ai->mean_ec; in scan_all()
1505 struct ubi_ainf_peb *aeb, *last_aeb; in self_check_ai() local
1557 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) { in self_check_ai()
1560 last_aeb = aeb; in self_check_ai()
1563 if (aeb->pnum < 0 || aeb->ec < 0) { in self_check_ai()
1568 if (aeb->ec < ai->min_ec) { in self_check_ai()
1570 ai->min_ec, aeb->ec); in self_check_ai()
1574 if (aeb->ec > ai->max_ec) { in self_check_ai()
1576 ai->max_ec, aeb->ec); in self_check_ai()
1580 if (aeb->pnum >= ubi->peb_count) { in self_check_ai()
1582 aeb->pnum, ubi->peb_count); in self_check_ai()
1587 if (aeb->lnum >= av->used_ebs) { in self_check_ai()
1598 if (aeb->lnum > av->highest_lnum) { in self_check_ai()
1613 aeb = last_aeb; in self_check_ai()
1615 if (aeb->lnum != av->highest_lnum) { in self_check_ai()
1630 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) { in self_check_ai()
1635 last_aeb = aeb; in self_check_ai()
1637 err = ubi_io_read_vid_hdr(ubi, aeb->pnum, vidh, 1); in self_check_ai()
1653 if (aeb->sqnum != be64_to_cpu(vidh->sqnum)) { in self_check_ai()
1654 ubi_err(ubi, "bad sqnum %llu", aeb->sqnum); in self_check_ai()
1668 if (aeb->lnum != be32_to_cpu(vidh->lnum)) { in self_check_ai()
1669 ubi_err(ubi, "bad lnum %d", aeb->lnum); in self_check_ai()
1717 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) in self_check_ai()
1718 buf[aeb->pnum] = 1; in self_check_ai()
1720 list_for_each_entry(aeb, &ai->free, u.list) in self_check_ai()
1721 buf[aeb->pnum] = 1; in self_check_ai()
1723 list_for_each_entry(aeb, &ai->corr, u.list) in self_check_ai()
1724 buf[aeb->pnum] = 1; in self_check_ai()
1726 list_for_each_entry(aeb, &ai->erase, u.list) in self_check_ai()
1727 buf[aeb->pnum] = 1; in self_check_ai()
1729 list_for_each_entry(aeb, &ai->alien, u.list) in self_check_ai()
1730 buf[aeb->pnum] = 1; in self_check_ai()
1745 ubi_err(ubi, "bad attaching information about LEB %d", aeb->lnum); in self_check_ai()
1746 ubi_dump_aeb(aeb, 0); in self_check_ai()