Lines Matching refs:queue_depth

2350 		for (i = 0; i < set->queue_depth; i++) {  in blk_mq_clear_rq_mapping()
2674 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument
2685 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
2709 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
2885 set->queue_depth, set->reserved_tags, flags); in __blk_mq_alloc_map_and_request()
2890 set->queue_depth); in __blk_mq_alloc_map_and_request()
3172 unsigned int queue_depth, in blk_mq_init_sq_queue() argument
3182 set->queue_depth = queue_depth; in blk_mq_init_sq_queue()
3353 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
3422 depth = set->queue_depth; in blk_mq_alloc_map_and_requests()
3428 set->queue_depth >>= 1; in blk_mq_alloc_map_and_requests()
3429 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) { in blk_mq_alloc_map_and_requests()
3433 } while (set->queue_depth); in blk_mq_alloc_map_and_requests()
3435 if (!set->queue_depth || err) { in blk_mq_alloc_map_and_requests()
3440 if (depth != set->queue_depth) in blk_mq_alloc_map_and_requests()
3442 depth, set->queue_depth); in blk_mq_alloc_map_and_requests()
3521 if (!set->queue_depth) in blk_mq_alloc_tag_set()
3523 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) in blk_mq_alloc_tag_set()
3532 if (set->queue_depth > BLK_MQ_MAX_DEPTH) { in blk_mq_alloc_tag_set()
3535 set->queue_depth = BLK_MQ_MAX_DEPTH; in blk_mq_alloc_tag_set()
3551 set->queue_depth = min(64U, set->queue_depth); in blk_mq_alloc_tag_set()