diff options
Diffstat (limited to 'include/linux/blkdev.h')
-rw-r--r-- | include/linux/blkdev.h | 140 |
1 files changed, 70 insertions, 70 deletions
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index 695e34964cb7..4be37de02052 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h | |||
@@ -37,7 +37,6 @@ | |||
37 | struct scsi_ioctl_command; | 37 | struct scsi_ioctl_command; |
38 | 38 | ||
39 | struct request_queue; | 39 | struct request_queue; |
40 | typedef struct request_queue request_queue_t; | ||
41 | struct elevator_queue; | 40 | struct elevator_queue; |
42 | typedef struct elevator_queue elevator_t; | 41 | typedef struct elevator_queue elevator_t; |
43 | struct request_pm_state; | 42 | struct request_pm_state; |
@@ -233,7 +232,7 @@ struct request { | |||
233 | struct list_head queuelist; | 232 | struct list_head queuelist; |
234 | struct list_head donelist; | 233 | struct list_head donelist; |
235 | 234 | ||
236 | request_queue_t *q; | 235 | struct request_queue *q; |
237 | 236 | ||
238 | unsigned int cmd_flags; | 237 | unsigned int cmd_flags; |
239 | enum rq_cmd_type_bits cmd_type; | 238 | enum rq_cmd_type_bits cmd_type; |
@@ -337,15 +336,15 @@ struct request_pm_state | |||
337 | 336 | ||
338 | #include <linux/elevator.h> | 337 | #include <linux/elevator.h> |
339 | 338 | ||
340 | typedef void (request_fn_proc) (request_queue_t *q); | 339 | typedef void (request_fn_proc) (struct request_queue *q); |
341 | typedef int (make_request_fn) (request_queue_t *q, struct bio *bio); | 340 | typedef int (make_request_fn) (struct request_queue *q, struct bio *bio); |
342 | typedef int (prep_rq_fn) (request_queue_t *, struct request *); | 341 | typedef int (prep_rq_fn) (struct request_queue *, struct request *); |
343 | typedef void (unplug_fn) (request_queue_t *); | 342 | typedef void (unplug_fn) (struct request_queue *); |
344 | 343 | ||
345 | struct bio_vec; | 344 | struct bio_vec; |
346 | typedef int (merge_bvec_fn) (request_queue_t *, struct bio *, struct bio_vec *); | 345 | typedef int (merge_bvec_fn) (struct request_queue *, struct bio *, struct bio_vec *); |
347 | typedef int (issue_flush_fn) (request_queue_t *, struct gendisk *, sector_t *); | 346 | typedef int (issue_flush_fn) (struct request_queue *, struct gendisk *, sector_t *); |
348 | typedef void (prepare_flush_fn) (request_queue_t *, struct request *); | 347 | typedef void (prepare_flush_fn) (struct request_queue *, struct request *); |
349 | typedef void (softirq_done_fn)(struct request *); | 348 | typedef void (softirq_done_fn)(struct request *); |
350 | 349 | ||
351 | enum blk_queue_state { | 350 | enum blk_queue_state { |
@@ -626,13 +625,13 @@ extern unsigned long blk_max_low_pfn, blk_max_pfn; | |||
626 | 625 | ||
627 | #ifdef CONFIG_BOUNCE | 626 | #ifdef CONFIG_BOUNCE |
628 | extern int init_emergency_isa_pool(void); | 627 | extern int init_emergency_isa_pool(void); |
629 | extern void blk_queue_bounce(request_queue_t *q, struct bio **bio); | 628 | extern void blk_queue_bounce(struct request_queue *q, struct bio **bio); |
630 | #else | 629 | #else |
631 | static inline int init_emergency_isa_pool(void) | 630 | static inline int init_emergency_isa_pool(void) |
632 | { | 631 | { |
633 | return 0; | 632 | return 0; |
634 | } | 633 | } |
635 | static inline void blk_queue_bounce(request_queue_t *q, struct bio **bio) | 634 | static inline void blk_queue_bounce(struct request_queue *q, struct bio **bio) |
636 | { | 635 | { |
637 | } | 636 | } |
638 | #endif /* CONFIG_MMU */ | 637 | #endif /* CONFIG_MMU */ |
@@ -646,14 +645,14 @@ extern void blk_unregister_queue(struct gendisk *disk); | |||
646 | extern void register_disk(struct gendisk *dev); | 645 | extern void register_disk(struct gendisk *dev); |
647 | extern void generic_make_request(struct bio *bio); | 646 | extern void generic_make_request(struct bio *bio); |
648 | extern void blk_put_request(struct request *); | 647 | extern void blk_put_request(struct request *); |
649 | extern void __blk_put_request(request_queue_t *, struct request *); | 648 | extern void __blk_put_request(struct request_queue *, struct request *); |
650 | extern void blk_end_sync_rq(struct request *rq, int error); | 649 | extern void blk_end_sync_rq(struct request *rq, int error); |
651 | extern struct request *blk_get_request(request_queue_t *, int, gfp_t); | 650 | extern struct request *blk_get_request(struct request_queue *, int, gfp_t); |
652 | extern void blk_insert_request(request_queue_t *, struct request *, int, void *); | 651 | extern void blk_insert_request(struct request_queue *, struct request *, int, void *); |
653 | extern void blk_requeue_request(request_queue_t *, struct request *); | 652 | extern void blk_requeue_request(struct request_queue *, struct request *); |
654 | extern void blk_plug_device(request_queue_t *); | 653 | extern void blk_plug_device(struct request_queue *); |
655 | extern int blk_remove_plug(request_queue_t *); | 654 | extern int blk_remove_plug(struct request_queue *); |
656 | extern void blk_recount_segments(request_queue_t *, struct bio *); | 655 | extern void blk_recount_segments(struct request_queue *, struct bio *); |
657 | extern int scsi_cmd_ioctl(struct file *, struct request_queue *, | 656 | extern int scsi_cmd_ioctl(struct file *, struct request_queue *, |
658 | struct gendisk *, unsigned int, void __user *); | 657 | struct gendisk *, unsigned int, void __user *); |
659 | extern int sg_scsi_ioctl(struct file *, struct request_queue *, | 658 | extern int sg_scsi_ioctl(struct file *, struct request_queue *, |
@@ -662,14 +661,15 @@ extern int sg_scsi_ioctl(struct file *, struct request_queue *, | |||
662 | /* | 661 | /* |
663 | * Temporary export, until SCSI gets fixed up. | 662 | * Temporary export, until SCSI gets fixed up. |
664 | */ | 663 | */ |
665 | extern int ll_back_merge_fn(request_queue_t *, struct request *, struct bio *); | 664 | extern int ll_back_merge_fn(struct request_queue *, struct request *, |
665 | struct bio *); | ||
666 | 666 | ||
667 | /* | 667 | /* |
668 | * A queue has just exitted congestion. Note this in the global counter of | 668 | * A queue has just exitted congestion. Note this in the global counter of |
669 | * congested queues, and wake up anyone who was waiting for requests to be | 669 | * congested queues, and wake up anyone who was waiting for requests to be |
670 | * put back. | 670 | * put back. |
671 | */ | 671 | */ |
672 | static inline void blk_clear_queue_congested(request_queue_t *q, int rw) | 672 | static inline void blk_clear_queue_congested(struct request_queue *q, int rw) |
673 | { | 673 | { |
674 | clear_bdi_congested(&q->backing_dev_info, rw); | 674 | clear_bdi_congested(&q->backing_dev_info, rw); |
675 | } | 675 | } |
@@ -678,29 +678,29 @@ static inline void blk_clear_queue_congested(request_queue_t *q, int rw) | |||
678 | * A queue has just entered congestion. Flag that in the queue's VM-visible | 678 | * A queue has just entered congestion. Flag that in the queue's VM-visible |
679 | * state flags and increment the global gounter of congested queues. | 679 | * state flags and increment the global gounter of congested queues. |
680 | */ | 680 | */ |
681 | static inline void blk_set_queue_congested(request_queue_t *q, int rw) | 681 | static inline void blk_set_queue_congested(struct request_queue *q, int rw) |
682 | { | 682 | { |
683 | set_bdi_congested(&q->backing_dev_info, rw); | 683 | set_bdi_congested(&q->backing_dev_info, rw); |
684 | } | 684 | } |
685 | 685 | ||
686 | extern void blk_start_queue(request_queue_t *q); | 686 | extern void blk_start_queue(struct request_queue *q); |
687 | extern void blk_stop_queue(request_queue_t *q); | 687 | extern void blk_stop_queue(struct request_queue *q); |
688 | extern void blk_sync_queue(struct request_queue *q); | 688 | extern void blk_sync_queue(struct request_queue *q); |
689 | extern void __blk_stop_queue(request_queue_t *q); | 689 | extern void __blk_stop_queue(struct request_queue *q); |
690 | extern void blk_run_queue(request_queue_t *); | 690 | extern void blk_run_queue(struct request_queue *); |
691 | extern void blk_start_queueing(request_queue_t *); | 691 | extern void blk_start_queueing(struct request_queue *); |
692 | extern int blk_rq_map_user(request_queue_t *, struct request *, void __user *, unsigned long); | 692 | extern int blk_rq_map_user(struct request_queue *, struct request *, void __user *, unsigned long); |
693 | extern int blk_rq_unmap_user(struct bio *); | 693 | extern int blk_rq_unmap_user(struct bio *); |
694 | extern int blk_rq_map_kern(request_queue_t *, struct request *, void *, unsigned int, gfp_t); | 694 | extern int blk_rq_map_kern(struct request_queue *, struct request *, void *, unsigned int, gfp_t); |
695 | extern int blk_rq_map_user_iov(request_queue_t *, struct request *, | 695 | extern int blk_rq_map_user_iov(struct request_queue *, struct request *, |
696 | struct sg_iovec *, int, unsigned int); | 696 | struct sg_iovec *, int, unsigned int); |
697 | extern int blk_execute_rq(request_queue_t *, struct gendisk *, | 697 | extern int blk_execute_rq(struct request_queue *, struct gendisk *, |
698 | struct request *, int); | 698 | struct request *, int); |
699 | extern void blk_execute_rq_nowait(request_queue_t *, struct gendisk *, | 699 | extern void blk_execute_rq_nowait(struct request_queue *, struct gendisk *, |
700 | struct request *, int, rq_end_io_fn *); | 700 | struct request *, int, rq_end_io_fn *); |
701 | extern int blk_verify_command(unsigned char *, int); | 701 | extern int blk_verify_command(unsigned char *, int); |
702 | 702 | ||
703 | static inline request_queue_t *bdev_get_queue(struct block_device *bdev) | 703 | static inline struct request_queue *bdev_get_queue(struct block_device *bdev) |
704 | { | 704 | { |
705 | return bdev->bd_disk->queue; | 705 | return bdev->bd_disk->queue; |
706 | } | 706 | } |
@@ -749,41 +749,41 @@ static inline void blkdev_dequeue_request(struct request *req) | |||
749 | /* | 749 | /* |
750 | * Access functions for manipulating queue properties | 750 | * Access functions for manipulating queue properties |
751 | */ | 751 | */ |
752 | extern request_queue_t *blk_init_queue_node(request_fn_proc *rfn, | 752 | extern struct request_queue *blk_init_queue_node(request_fn_proc *rfn, |
753 | spinlock_t *lock, int node_id); | 753 | spinlock_t *lock, int node_id); |
754 | extern request_queue_t *blk_init_queue(request_fn_proc *, spinlock_t *); | 754 | extern struct request_queue *blk_init_queue(request_fn_proc *, spinlock_t *); |
755 | extern void blk_cleanup_queue(request_queue_t *); | 755 | extern void blk_cleanup_queue(struct request_queue *); |
756 | extern void blk_queue_make_request(request_queue_t *, make_request_fn *); | 756 | extern void blk_queue_make_request(struct request_queue *, make_request_fn *); |
757 | extern void blk_queue_bounce_limit(request_queue_t *, u64); | 757 | extern void blk_queue_bounce_limit(struct request_queue *, u64); |
758 | extern void blk_queue_max_sectors(request_queue_t *, unsigned int); | 758 | extern void blk_queue_max_sectors(struct request_queue *, unsigned int); |
759 | extern void blk_queue_max_phys_segments(request_queue_t *, unsigned short); | 759 | extern void blk_queue_max_phys_segments(struct request_queue *, unsigned short); |
760 | extern void blk_queue_max_hw_segments(request_queue_t *, unsigned short); | 760 | extern void blk_queue_max_hw_segments(struct request_queue *, unsigned short); |
761 | extern void blk_queue_max_segment_size(request_queue_t *, unsigned int); | 761 | extern void blk_queue_max_segment_size(struct request_queue *, unsigned int); |
762 | extern void blk_queue_hardsect_size(request_queue_t *, unsigned short); | 762 | extern void blk_queue_hardsect_size(struct request_queue *, unsigned short); |
763 | extern void blk_queue_stack_limits(request_queue_t *t, request_queue_t *b); | 763 | extern void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b); |
764 | extern void blk_queue_segment_boundary(request_queue_t *, unsigned long); | 764 | extern void blk_queue_segment_boundary(struct request_queue *, unsigned long); |
765 | extern void blk_queue_prep_rq(request_queue_t *, prep_rq_fn *pfn); | 765 | extern void blk_queue_prep_rq(struct request_queue *, prep_rq_fn *pfn); |
766 | extern void blk_queue_merge_bvec(request_queue_t *, merge_bvec_fn *); | 766 | extern void blk_queue_merge_bvec(struct request_queue *, merge_bvec_fn *); |
767 | extern void blk_queue_dma_alignment(request_queue_t *, int); | 767 | extern void blk_queue_dma_alignment(struct request_queue *, int); |
768 | extern void blk_queue_softirq_done(request_queue_t *, softirq_done_fn *); | 768 | extern void blk_queue_softirq_done(struct request_queue *, softirq_done_fn *); |
769 | extern struct backing_dev_info *blk_get_backing_dev_info(struct block_device *bdev); | 769 | extern struct backing_dev_info *blk_get_backing_dev_info(struct block_device *bdev); |
770 | extern int blk_queue_ordered(request_queue_t *, unsigned, prepare_flush_fn *); | 770 | extern int blk_queue_ordered(struct request_queue *, unsigned, prepare_flush_fn *); |
771 | extern void blk_queue_issue_flush_fn(request_queue_t *, issue_flush_fn *); | 771 | extern void blk_queue_issue_flush_fn(struct request_queue *, issue_flush_fn *); |
772 | extern int blk_do_ordered(request_queue_t *, struct request **); | 772 | extern int blk_do_ordered(struct request_queue *, struct request **); |
773 | extern unsigned blk_ordered_cur_seq(request_queue_t *); | 773 | extern unsigned blk_ordered_cur_seq(struct request_queue *); |
774 | extern unsigned blk_ordered_req_seq(struct request *); | 774 | extern unsigned blk_ordered_req_seq(struct request *); |
775 | extern void blk_ordered_complete_seq(request_queue_t *, unsigned, int); | 775 | extern void blk_ordered_complete_seq(struct request_queue *, unsigned, int); |
776 | 776 | ||
777 | extern int blk_rq_map_sg(request_queue_t *, struct request *, struct scatterlist *); | 777 | extern int blk_rq_map_sg(struct request_queue *, struct request *, struct scatterlist *); |
778 | extern void blk_dump_rq_flags(struct request *, char *); | 778 | extern void blk_dump_rq_flags(struct request *, char *); |
779 | extern void generic_unplug_device(request_queue_t *); | 779 | extern void generic_unplug_device(struct request_queue *); |
780 | extern void __generic_unplug_device(request_queue_t *); | 780 | extern void __generic_unplug_device(struct request_queue *); |
781 | extern long nr_blockdev_pages(void); | 781 | extern long nr_blockdev_pages(void); |
782 | 782 | ||
783 | int blk_get_queue(request_queue_t *); | 783 | int blk_get_queue(struct request_queue *); |
784 | request_queue_t *blk_alloc_queue(gfp_t); | 784 | struct request_queue *blk_alloc_queue(gfp_t); |
785 | request_queue_t *blk_alloc_queue_node(gfp_t, int); | 785 | struct request_queue *blk_alloc_queue_node(gfp_t, int); |
786 | extern void blk_put_queue(request_queue_t *); | 786 | extern void blk_put_queue(struct request_queue *); |
787 | 787 | ||
788 | /* | 788 | /* |
789 | * tag stuff | 789 | * tag stuff |
@@ -791,13 +791,13 @@ extern void blk_put_queue(request_queue_t *); | |||
791 | #define blk_queue_tag_depth(q) ((q)->queue_tags->busy) | 791 | #define blk_queue_tag_depth(q) ((q)->queue_tags->busy) |
792 | #define blk_queue_tag_queue(q) ((q)->queue_tags->busy < (q)->queue_tags->max_depth) | 792 | #define blk_queue_tag_queue(q) ((q)->queue_tags->busy < (q)->queue_tags->max_depth) |
793 | #define blk_rq_tagged(rq) ((rq)->cmd_flags & REQ_QUEUED) | 793 | #define blk_rq_tagged(rq) ((rq)->cmd_flags & REQ_QUEUED) |
794 | extern int blk_queue_start_tag(request_queue_t *, struct request *); | 794 | extern int blk_queue_start_tag(struct request_queue *, struct request *); |
795 | extern struct request *blk_queue_find_tag(request_queue_t *, int); | 795 | extern struct request *blk_queue_find_tag(struct request_queue *, int); |
796 | extern void blk_queue_end_tag(request_queue_t *, struct request *); | 796 | extern void blk_queue_end_tag(struct request_queue *, struct request *); |
797 | extern int blk_queue_init_tags(request_queue_t *, int, struct blk_queue_tag *); | 797 | extern int blk_queue_init_tags(struct request_queue *, int, struct blk_queue_tag *); |
798 | extern void blk_queue_free_tags(request_queue_t *); | 798 | extern void blk_queue_free_tags(struct request_queue *); |
799 | extern int blk_queue_resize_tags(request_queue_t *, int); | 799 | extern int blk_queue_resize_tags(struct request_queue *, int); |
800 | extern void blk_queue_invalidate_tags(request_queue_t *); | 800 | extern void blk_queue_invalidate_tags(struct request_queue *); |
801 | extern struct blk_queue_tag *blk_init_tags(int); | 801 | extern struct blk_queue_tag *blk_init_tags(int); |
802 | extern void blk_free_tags(struct blk_queue_tag *); | 802 | extern void blk_free_tags(struct blk_queue_tag *); |
803 | 803 | ||
@@ -809,7 +809,7 @@ static inline struct request *blk_map_queue_find_tag(struct blk_queue_tag *bqt, | |||
809 | return bqt->tag_index[tag]; | 809 | return bqt->tag_index[tag]; |
810 | } | 810 | } |
811 | 811 | ||
812 | extern void blk_rq_bio_prep(request_queue_t *, struct request *, struct bio *); | 812 | extern void blk_rq_bio_prep(struct request_queue *, struct request *, struct bio *); |
813 | extern int blkdev_issue_flush(struct block_device *, sector_t *); | 813 | extern int blkdev_issue_flush(struct block_device *, sector_t *); |
814 | 814 | ||
815 | #define MAX_PHYS_SEGMENTS 128 | 815 | #define MAX_PHYS_SEGMENTS 128 |
@@ -821,7 +821,7 @@ extern int blkdev_issue_flush(struct block_device *, sector_t *); | |||
821 | 821 | ||
822 | #define blkdev_entry_to_request(entry) list_entry((entry), struct request, queuelist) | 822 | #define blkdev_entry_to_request(entry) list_entry((entry), struct request, queuelist) |
823 | 823 | ||
824 | static inline int queue_hardsect_size(request_queue_t *q) | 824 | static inline int queue_hardsect_size(struct request_queue *q) |
825 | { | 825 | { |
826 | int retval = 512; | 826 | int retval = 512; |
827 | 827 | ||
@@ -836,7 +836,7 @@ static inline int bdev_hardsect_size(struct block_device *bdev) | |||
836 | return queue_hardsect_size(bdev_get_queue(bdev)); | 836 | return queue_hardsect_size(bdev_get_queue(bdev)); |
837 | } | 837 | } |
838 | 838 | ||
839 | static inline int queue_dma_alignment(request_queue_t *q) | 839 | static inline int queue_dma_alignment(struct request_queue *q) |
840 | { | 840 | { |
841 | int retval = 511; | 841 | int retval = 511; |
842 | 842 | ||