| /OK3568_Linux_fs/kernel/drivers/lightnvm/ |
| H A D | pblk-read.c | 40 static int pblk_read_ppalist_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_ppalist_rq() argument 44 void *meta_list = rqd->meta_list; in pblk_read_ppalist_rq() 48 nr_secs = pblk_lookup_l2p_seq(pblk, rqd->ppa_list, blba, rqd->nr_ppas, in pblk_read_ppalist_rq() 58 if (pblk_ppa_empty(rqd->ppa_list[i])) { in pblk_read_ppalist_rq() 62 } else if (pblk_addr_in_cache(rqd->ppa_list[i])) { in pblk_read_ppalist_rq() 69 rqd->ppa_list[i])) { in pblk_read_ppalist_rq() 99 rqd->is_seq = 1; in pblk_read_ppalist_rq() 109 static void pblk_read_check_seq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_check_seq() argument 112 void *meta_list = rqd->meta_list; in pblk_read_check_seq() 113 int nr_lbas = rqd->nr_ppas; in pblk_read_check_seq() [all …]
|
| H A D | pblk-write.c | 22 static unsigned long pblk_end_w_bio(struct pblk *pblk, struct nvm_rq *rqd, in pblk_end_w_bio() argument 53 pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid, in pblk_end_w_bio() 57 atomic_long_add(rqd->nr_ppas, &pblk->sync_writes); in pblk_end_w_bio() 62 bio_put(rqd->bio); in pblk_end_w_bio() 63 pblk_free_rqd(pblk, rqd, PBLK_WRITE); in pblk_end_w_bio() 69 struct nvm_rq *rqd, in pblk_end_queued_w_bio() argument 73 return pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_end_queued_w_bio() 76 static void pblk_complete_write(struct pblk *pblk, struct nvm_rq *rqd, in pblk_complete_write() argument 90 pos = pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_complete_write() 94 rqd = nvm_rq_from_c_ctx(c); in pblk_complete_write() [all …]
|
| H A D | pblk-core.c | 79 static void __pblk_end_io_erase(struct pblk *pblk, struct nvm_rq *rqd) in __pblk_end_io_erase() argument 87 line = pblk_ppa_to_line(pblk, rqd->ppa_addr); in __pblk_end_io_erase() 88 pos = pblk_ppa_to_pos(geo, rqd->ppa_addr); in __pblk_end_io_erase() 93 if (rqd->error) { in __pblk_end_io_erase() 95 &rqd->ppa_addr, PBLK_CHUNK_RESET_FAILED); in __pblk_end_io_erase() 98 pblk_mark_bb(pblk, line, rqd->ppa_addr); in __pblk_end_io_erase() 101 &rqd->ppa_addr, PBLK_CHUNK_RESET_DONE); in __pblk_end_io_erase() 106 trace_pblk_chunk_state(pblk_disk_name(pblk), &rqd->ppa_addr, in __pblk_end_io_erase() 113 static void pblk_end_io_erase(struct nvm_rq *rqd) in pblk_end_io_erase() argument 115 struct pblk *pblk = rqd->private; in pblk_end_io_erase() [all …]
|
| H A D | core.c | 606 static void nvm_rq_tgt_to_dev(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_tgt_to_dev() argument 608 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in nvm_rq_tgt_to_dev() 610 nvm_ppa_tgt_to_dev(tgt_dev, ppa_list, rqd->nr_ppas); in nvm_rq_tgt_to_dev() 613 static void nvm_rq_dev_to_tgt(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_dev_to_tgt() argument 615 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in nvm_rq_dev_to_tgt() 617 nvm_ppa_dev_to_tgt(tgt_dev, ppa_list, rqd->nr_ppas); in nvm_rq_dev_to_tgt() 671 static int nvm_set_rqd_ppalist(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, in nvm_set_rqd_ppalist() argument 680 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist() 681 rqd->ppa_addr = ppas[0]; in nvm_set_rqd_ppalist() 686 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist() [all …]
|
| H A D | pblk-recovery.c | 145 struct nvm_rq *rqd; member 158 static void pblk_end_io_recov(struct nvm_rq *rqd) in pblk_end_io_recov() argument 160 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_end_io_recov() 161 struct pblk_pad_rq *pad_rq = rqd->private; in pblk_end_io_recov() 166 pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT); in pblk_end_io_recov() 180 struct nvm_rq *rqd; in pblk_recov_pad_line() local 214 rqd = pblk_alloc_rqd(pblk, PBLK_WRITE_INT); in pblk_recov_pad_line() 216 ret = pblk_alloc_rqd_meta(pblk, rqd); in pblk_recov_pad_line() 218 pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT); in pblk_recov_pad_line() 222 rqd->bio = NULL; in pblk_recov_pad_line() [all …]
|
| H A D | pblk-map.c | 96 int pblk_map_rq(struct pblk *pblk, struct nvm_rq *rqd, unsigned int sentry, in pblk_map_rq() argument 100 void *meta_list = pblk_get_meta_for_writes(pblk, rqd); in pblk_map_rq() 102 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_map_rq() 108 for (i = off; i < rqd->nr_ppas; i += min) { in pblk_map_rq() 122 int pblk_map_erase_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_map_erase_rq() argument 129 void *meta_list = pblk_get_meta_for_writes(pblk, rqd); in pblk_map_erase_rq() 131 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_map_erase_rq() 139 for (i = 0; i < rqd->nr_ppas; i += min) { in pblk_map_erase_rq() 155 return pblk_map_rq(pblk, rqd, sentry, lun_bitmap, in pblk_map_erase_rq() 170 return pblk_map_rq(pblk, rqd, sentry, lun_bitmap, in pblk_map_erase_rq()
|
| H A D | pblk.h | 134 struct nvm_rq *rqd; member 738 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, 765 void pblk_free_rqd(struct pblk *pblk, struct nvm_rq *rqd, int type); 766 int pblk_alloc_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd); 767 void pblk_free_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd); 769 int pblk_setup_w_rec_rq(struct pblk *pblk, struct nvm_rq *rqd, 776 void pblk_log_write_err(struct pblk *pblk, struct nvm_rq *rqd); 777 void pblk_log_read_err(struct pblk *pblk, struct nvm_rq *rqd); 778 int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf); 779 int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf); [all …]
|
| H A D | pblk-rb.c | 550 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, in pblk_rb_read_to_bio() argument 556 struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); in pblk_rb_read_to_bio() 557 struct bio *bio = rqd->bio; in pblk_rb_read_to_bio()
|
| /OK3568_Linux_fs/kernel/block/ |
| H A D | blk-rq-qos.c | 116 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() argument 128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 129 if (rqd->scale_step > 0) in rq_depth_calc_max_depth() 130 rqd->max_depth = 1; in rq_depth_calc_max_depth() 132 rqd->max_depth = 2; in rq_depth_calc_max_depth() 143 depth = min_t(unsigned int, rqd->default_depth, in rq_depth_calc_max_depth() 144 rqd->queue_depth); in rq_depth_calc_max_depth() 145 if (rqd->scale_step > 0) in rq_depth_calc_max_depth() 146 depth = 1 + ((depth - 1) >> min(31, rqd->scale_step)); in rq_depth_calc_max_depth() 147 else if (rqd->scale_step < 0) { in rq_depth_calc_max_depth() [all …]
|
| H A D | blk-wbt.c | 238 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() local 282 if (rqd->scale_step) in latency_exceeded() 291 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() local 293 trace_wbt_step(bdi, msg, rqd->scale_step, rwb->cur_win_nsec, in rwb_trace_step() 294 rwb->wb_background, rwb->wb_normal, rqd->max_depth); in rwb_trace_step() 331 struct rq_depth *rqd = &rwb->rq_depth; in rwb_arm_timer() local 333 if (rqd->scale_step > 0) { in rwb_arm_timer() 341 int_sqrt((rqd->scale_step + 1) << 8)); in rwb_arm_timer() 356 struct rq_depth *rqd = &rwb->rq_depth; in wb_timer_fn() local 362 trace_wbt_timer(rwb->rqos.q->backing_dev_info, status, rqd->scale_step, in wb_timer_fn() [all …]
|
| H A D | blk-rq-qos.h | 142 bool rq_depth_scale_up(struct rq_depth *rqd); 143 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle); 144 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
|
| /OK3568_Linux_fs/kernel/drivers/nvme/host/ |
| H A D | lightnvm.c | 624 static inline void nvme_nvm_rqtocmd(struct nvm_rq *rqd, struct nvme_ns *ns, in nvme_nvm_rqtocmd() argument 627 c->ph_rw.opcode = rqd->opcode; in nvme_nvm_rqtocmd() 629 c->ph_rw.spba = cpu_to_le64(rqd->ppa_addr.ppa); in nvme_nvm_rqtocmd() 630 c->ph_rw.metadata = cpu_to_le64(rqd->dma_meta_list); in nvme_nvm_rqtocmd() 631 c->ph_rw.control = cpu_to_le16(rqd->flags); in nvme_nvm_rqtocmd() 632 c->ph_rw.length = cpu_to_le16(rqd->nr_ppas - 1); in nvme_nvm_rqtocmd() 637 struct nvm_rq *rqd = rq->end_io_data; in nvme_nvm_end_io() local 639 rqd->ppa_status = le64_to_cpu(nvme_req(rq)->result.u64); in nvme_nvm_end_io() 640 rqd->error = nvme_req(rq)->status; in nvme_nvm_end_io() 641 nvm_end_io(rqd); in nvme_nvm_end_io() [all …]
|
| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | lightnvm.h | 321 static inline struct ppa_addr *nvm_rq_to_ppa_list(struct nvm_rq *rqd) in nvm_rq_to_ppa_list() argument 323 return (rqd->nr_ppas > 1) ? rqd->ppa_list : &rqd->ppa_addr; in nvm_rq_to_ppa_list()
|
| /OK3568_Linux_fs/kernel/drivers/dma/ |
| H A D | pl330.c | 539 struct list_head rqd; member 1914 list_add_tail(&descdone->rqd, &pl330->req_done); in pl330_update() 1922 struct dma_pl330_desc, rqd); in pl330_update() 1923 list_del(&descdone->rqd); in pl330_update()
|
| /OK3568_Linux_fs/kernel/arch/m68k/ifpsp060/src/ |
| H A D | ilsp.S | 299 swap %d5 # same as r*b if previous step rqd
|
| /OK3568_Linux_fs/prebuilts/gcc/linux-x86/aarch64/gcc-arm-10.3-2021.07-x86_64-aarch64-none-linux-gnu/share/info/ |
| H A D | as.info | 25275 rqd 64 bit destination register 25297 and rd,addr cp rd,imm16 divl rqd,@rs 25298 and rd,addr(rs) cp rd,rs divl rqd,addr 25299 and rd,imm16 cpb @rd,imm8 divl rqd,addr(rs) 25300 and rd,rs cpb addr(rd),imm8 divl rqd,imm32 25301 andb rbd,@rs cpb addr,imm8 divl rqd,rrs 25318 call addr cpl rrd,imm32 extsl rqd 25329 incb addr(rd),imm4m1 ldctl ctrl,rs multl rqd,@rs 25330 incb addr,imm4m1 ldctl rd,ctrl multl rqd,addr 25331 incb rbd,imm4m1 ldd @rs,@rd,rr multl rqd,addr(rs) [all …]
|
| /OK3568_Linux_fs/prebuilts/gcc/linux-x86/arm/gcc-arm-10.3-2021.07-x86_64-arm-none-linux-gnueabihf/share/info/ |
| H A D | as.info | 25275 rqd 64 bit destination register 25297 and rd,addr cp rd,imm16 divl rqd,@rs 25298 and rd,addr(rs) cp rd,rs divl rqd,addr 25299 and rd,imm16 cpb @rd,imm8 divl rqd,addr(rs) 25300 and rd,rs cpb addr(rd),imm8 divl rqd,imm32 25301 andb rbd,@rs cpb addr,imm8 divl rqd,rrs 25318 call addr cpl rrd,imm32 extsl rqd 25329 incb addr(rd),imm4m1 ldctl ctrl,rs multl rqd,@rs 25330 incb addr,imm4m1 ldctl rd,ctrl multl rqd,addr 25331 incb rbd,imm4m1 ldd @rs,@rd,rr multl rqd,addr(rs) [all …]
|