Lines Matching refs:request_queue
35 struct request_queue;
138 struct request_queue *q;
410 struct request_queue { struct
649 void blk_queue_flag_set(unsigned int flag, struct request_queue *q);
650 void blk_queue_flag_clear(unsigned int flag, struct request_queue *q);
651 bool blk_queue_flag_test_and_set(unsigned int flag, struct request_queue *q);
691 extern void blk_set_pm_only(struct request_queue *q);
692 extern void blk_clear_pm_only(struct request_queue *q);
710 static inline bool queue_is_mq(struct request_queue *q) in queue_is_mq()
716 blk_queue_zoned_model(struct request_queue *q) in blk_queue_zoned_model()
723 static inline bool blk_queue_is_zoned(struct request_queue *q) in blk_queue_is_zoned()
734 static inline sector_t blk_queue_zone_sectors(struct request_queue *q) in blk_queue_zone_sectors()
740 static inline unsigned int blk_queue_nr_zones(struct request_queue *q) in blk_queue_nr_zones()
745 static inline unsigned int blk_queue_zone_no(struct request_queue *q, in blk_queue_zone_no()
753 static inline bool blk_queue_zone_is_seq(struct request_queue *q, in blk_queue_zone_is_seq()
763 static inline void blk_queue_max_open_zones(struct request_queue *q, in blk_queue_max_open_zones()
769 static inline unsigned int queue_max_open_zones(const struct request_queue *q) in queue_max_open_zones()
774 static inline void blk_queue_max_active_zones(struct request_queue *q, in blk_queue_max_active_zones()
780 static inline unsigned int queue_max_active_zones(const struct request_queue *q) in queue_max_active_zones()
785 static inline unsigned int blk_queue_nr_zones(struct request_queue *q) in blk_queue_nr_zones()
789 static inline bool blk_queue_zone_is_seq(struct request_queue *q, in blk_queue_zone_is_seq()
794 static inline unsigned int blk_queue_zone_no(struct request_queue *q, in blk_queue_zone_no()
799 static inline unsigned int queue_max_open_zones(const struct request_queue *q) in queue_max_open_zones()
803 static inline unsigned int queue_max_active_zones(const struct request_queue *q) in queue_max_active_zones()
845 static inline unsigned int blk_queue_depth(struct request_queue *q) in blk_queue_depth()
924 extern void blk_rq_init(struct request_queue *q, struct request *rq);
926 extern struct request *blk_get_request(struct request_queue *, unsigned int op,
928 extern int blk_lld_busy(struct request_queue *q);
934 extern blk_status_t blk_insert_cloned_request(struct request_queue *q,
941 extern int scsi_cmd_ioctl(struct request_queue *, struct gendisk *, fmode_t,
943 extern int sg_scsi_ioctl(struct request_queue *, struct gendisk *, fmode_t,
948 extern int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags);
949 extern void blk_queue_exit(struct request_queue *q);
950 extern void blk_sync_queue(struct request_queue *q);
951 extern int blk_rq_map_user(struct request_queue *, struct request *,
955 extern int blk_rq_map_kern(struct request_queue *, struct request *, void *, unsigned int, gfp_t);
956 extern int blk_rq_map_user_iov(struct request_queue *, struct request *,
959 extern void blk_execute_rq(struct request_queue *, struct gendisk *,
961 extern void blk_execute_rq_nowait(struct request_queue *, struct gendisk *,
970 int blk_poll(struct request_queue *q, blk_qc_t cookie, bool spin);
972 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue()
1071 static inline unsigned int blk_queue_get_max_sectors(struct request_queue *q, in blk_queue_get_max_sectors()
1091 static inline unsigned int blk_max_size_offset(struct request_queue *q, in blk_max_size_offset()
1113 struct request_queue *q = rq->q; in blk_rq_get_max_sectors()
1154 extern void blk_cleanup_queue(struct request_queue *);
1155 extern void blk_queue_bounce_limit(struct request_queue *, u64);
1156 extern void blk_queue_max_hw_sectors(struct request_queue *, unsigned int);
1157 extern void blk_queue_chunk_sectors(struct request_queue *, unsigned int);
1158 extern void blk_queue_max_segments(struct request_queue *, unsigned short);
1159 extern void blk_queue_max_discard_segments(struct request_queue *,
1161 extern void blk_queue_max_segment_size(struct request_queue *, unsigned int);
1162 extern void blk_queue_max_discard_sectors(struct request_queue *q,
1164 extern void blk_queue_max_write_same_sectors(struct request_queue *q,
1166 extern void blk_queue_max_write_zeroes_sectors(struct request_queue *q,
1168 extern void blk_queue_logical_block_size(struct request_queue *, unsigned int);
1169 extern void blk_queue_max_zone_append_sectors(struct request_queue *q,
1171 extern void blk_queue_physical_block_size(struct request_queue *, unsigned int);
1172 extern void blk_queue_alignment_offset(struct request_queue *q,
1174 void blk_queue_update_readahead(struct request_queue *q);
1176 extern void blk_queue_io_min(struct request_queue *q, unsigned int min);
1178 extern void blk_queue_io_opt(struct request_queue *q, unsigned int opt);
1179 extern void blk_set_queue_depth(struct request_queue *q, unsigned int depth);
1186 extern void blk_queue_update_dma_pad(struct request_queue *, unsigned int);
1187 extern void blk_queue_segment_boundary(struct request_queue *, unsigned long);
1188 extern void blk_queue_virt_boundary(struct request_queue *, unsigned long);
1189 extern void blk_queue_dma_alignment(struct request_queue *, int);
1190 extern void blk_queue_update_dma_alignment(struct request_queue *, int);
1191 extern void blk_queue_rq_timeout(struct request_queue *, unsigned int);
1192 extern void blk_queue_write_cache(struct request_queue *q, bool enabled, bool fua);
1193 extern void blk_queue_required_elevator_features(struct request_queue *q,
1195 extern bool blk_queue_can_use_dma_map_merging(struct request_queue *q,
1223 int __blk_rq_map_sg(struct request_queue *q, struct request *rq,
1225 static inline int blk_rq_map_sg(struct request_queue *q, struct request *rq, in blk_rq_map_sg()
1234 bool __must_check blk_get_queue(struct request_queue *);
1235 struct request_queue *blk_alloc_queue(int node_id);
1236 extern void blk_put_queue(struct request_queue *);
1237 extern void blk_set_queue_dying(struct request_queue *);
1395 static inline unsigned long queue_segment_boundary(const struct request_queue *q) in queue_segment_boundary()
1400 static inline unsigned long queue_virt_boundary(const struct request_queue *q) in queue_virt_boundary()
1405 static inline unsigned int queue_max_sectors(const struct request_queue *q) in queue_max_sectors()
1410 static inline unsigned int queue_max_hw_sectors(const struct request_queue *q) in queue_max_hw_sectors()
1415 static inline unsigned short queue_max_segments(const struct request_queue *q) in queue_max_segments()
1420 static inline unsigned short queue_max_discard_segments(const struct request_queue *q) in queue_max_discard_segments()
1425 static inline unsigned int queue_max_segment_size(const struct request_queue *q) in queue_max_segment_size()
1430 static inline unsigned int queue_max_zone_append_sectors(const struct request_queue *q) in queue_max_zone_append_sectors()
1438 static inline unsigned queue_logical_block_size(const struct request_queue *q) in queue_logical_block_size()
1453 static inline unsigned int queue_physical_block_size(const struct request_queue *q) in queue_physical_block_size()
1463 static inline unsigned int queue_io_min(const struct request_queue *q) in queue_io_min()
1473 static inline unsigned int queue_io_opt(const struct request_queue *q) in queue_io_opt()
1483 static inline int queue_alignment_offset(const struct request_queue *q) in queue_alignment_offset()
1502 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset()
1512 static inline int queue_discard_alignment(const struct request_queue *q) in queue_discard_alignment()
1561 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
1571 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_same()
1581 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors()
1591 struct request_queue *q = bdev_get_queue(bdev); in bdev_zoned_model()
1601 struct request_queue *q = bdev_get_queue(bdev); in bdev_is_zoned()
1611 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors()
1620 struct request_queue *q = bdev_get_queue(bdev); in bdev_max_open_zones()
1629 struct request_queue *q = bdev_get_queue(bdev); in bdev_max_active_zones()
1636 static inline int queue_dma_alignment(const struct request_queue *q) in queue_dma_alignment()
1641 static inline int blk_rq_aligned(struct request_queue *q, unsigned long addr, in blk_rq_aligned()
1705 extern int blk_rq_map_integrity_sg(struct request_queue *, struct bio *,
1707 extern int blk_rq_count_integrity_sg(struct request_queue *, struct bio *);
1726 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity()
1736 static inline void blk_queue_max_integrity_segments(struct request_queue *q, in blk_queue_max_integrity_segments()
1743 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments()
1792 static inline int blk_rq_count_integrity_sg(struct request_queue *q, in blk_rq_count_integrity_sg()
1797 static inline int blk_rq_map_integrity_sg(struct request_queue *q, in blk_rq_map_integrity_sg()
1812 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity()
1827 static inline void blk_queue_max_integrity_segments(struct request_queue *q, in blk_queue_max_integrity_segments()
1831 static inline unsigned short queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments()
1857 bool blk_ksm_register(struct blk_keyslot_manager *ksm, struct request_queue *q);
1859 void blk_ksm_unregister(struct request_queue *q);
1864 struct request_queue *q) in blk_ksm_register()
1869 static inline void blk_ksm_unregister(struct request_queue *q) { } in blk_ksm_unregister()