Searched refs:flush_rq (Results 1 – 5 of 5) sorted by relevance
| /OK3568_Linux_fs/kernel/block/ |
| H A D | blk-flush.c | 215 static void flush_end_io(struct request *flush_rq, blk_status_t error) in flush_end_io() argument 217 struct request_queue *q = flush_rq->q; in flush_end_io() 221 struct blk_flush_queue *fq = blk_get_flush_queue(q, flush_rq->mq_ctx); in flush_end_io() 226 if (!refcount_dec_and_test(&flush_rq->ref)) { in flush_end_io() 232 blk_account_io_flush(flush_rq); in flush_end_io() 238 WRITE_ONCE(flush_rq->state, MQ_RQ_IDLE); in flush_end_io() 245 flush_rq->tag = BLK_MQ_NO_TAG; in flush_end_io() 247 blk_mq_put_driver_tag(flush_rq); in flush_end_io() 248 flush_rq->internal_tag = BLK_MQ_NO_TAG; in flush_end_io() 292 struct request *flush_rq = fq->flush_rq; in blk_kick_flush() local [all …]
|
| H A D | blk-mq.c | 2674 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument 2683 WARN_ON_ONCE(refcount_read(&flush_rq->ref) != 0); in blk_mq_clear_flush_rq_mapping() 2686 cmpxchg(&tags->rqs[i], flush_rq, NULL); in blk_mq_clear_flush_rq_mapping() 2703 struct request *flush_rq = hctx->fq->flush_rq; in blk_mq_exit_hctx() local 2709 set->queue_depth, flush_rq); in blk_mq_exit_hctx() 2711 set->ops->exit_request(set, flush_rq, hctx_idx); in blk_mq_exit_hctx() 2768 if (blk_mq_init_request(set, hctx->fq->flush_rq, hctx_idx, in blk_mq_init_hctx()
|
| H A D | blk.h | 26 struct request *flush_rq; member
|
| /OK3568_Linux_fs/kernel/drivers/infiniband/hw/i40iw/ |
| H A D | i40iw_ctrl.c | 2695 bool flush_sq = false, flush_rq = false; in i40iw_sc_qp_flush_wqes() local 2697 if (info->rq && !qp->flush_rq) in i40iw_sc_qp_flush_wqes() 2698 flush_rq = true; in i40iw_sc_qp_flush_wqes() 2704 qp->flush_rq |= flush_rq; in i40iw_sc_qp_flush_wqes() 2705 if (!flush_sq && !flush_rq) in i40iw_sc_qp_flush_wqes() 2713 if (flush_rq) { in i40iw_sc_qp_flush_wqes() 2734 LS_64(flush_rq, I40IW_CQPSQ_FWQE_FLUSHRQ) | in i40iw_sc_qp_flush_wqes()
|
| H A D | i40iw_type.h | 404 bool flush_rq; member
|