diff options
author | Jens Axboe <axboe@suse.de> | 2006-07-28 03:36:46 -0400 |
---|---|---|
committer | Jens Axboe <axboe@nelson.home.kernel.dk> | 2006-09-30 14:29:38 -0400 |
commit | a3b05e8f58c95dfccbf2c824d0c68e5990571f24 (patch) | |
tree | 768aad0f032790df38b277eff062c47b6ecff533 /include/linux/blkdev.h | |
parent | 1ea25ecb7256978947c258f08a30c878eebe9edb (diff) |
[PATCH] Kill various deprecated/unused block layer defines/functions
Signed-off-by: Jens Axboe <axboe@suse.de>
Diffstat (limited to 'include/linux/blkdev.h')
-rw-r--r-- | include/linux/blkdev.h | 29 |
1 files changed, 0 insertions, 29 deletions
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index 79cb9fa8034a..593386162f47 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h | |||
@@ -579,12 +579,6 @@ static inline void blk_clear_queue_full(struct request_queue *q, int rw) | |||
579 | (!((rq)->cmd_flags & RQ_NOMERGE_FLAGS) && blk_fs_request((rq))) | 579 | (!((rq)->cmd_flags & RQ_NOMERGE_FLAGS) && blk_fs_request((rq))) |
580 | 580 | ||
581 | /* | 581 | /* |
582 | * noop, requests are automagically marked as active/inactive by I/O | ||
583 | * scheduler -- see elv_next_request | ||
584 | */ | ||
585 | #define blk_queue_headactive(q, head_active) | ||
586 | |||
587 | /* | ||
588 | * q->prep_rq_fn return values | 582 | * q->prep_rq_fn return values |
589 | */ | 583 | */ |
590 | #define BLKPREP_OK 0 /* serve it */ | 584 | #define BLKPREP_OK 0 /* serve it */ |
@@ -621,11 +615,6 @@ static inline void blk_queue_bounce(request_queue_t *q, struct bio **bio) | |||
621 | if ((rq->bio)) \ | 615 | if ((rq->bio)) \ |
622 | for (_bio = (rq)->bio; _bio; _bio = _bio->bi_next) | 616 | for (_bio = (rq)->bio; _bio; _bio = _bio->bi_next) |
623 | 617 | ||
624 | struct sec_size { | ||
625 | unsigned block_size; | ||
626 | unsigned block_size_bits; | ||
627 | }; | ||
628 | |||
629 | extern int blk_register_queue(struct gendisk *disk); | 618 | extern int blk_register_queue(struct gendisk *disk); |
630 | extern void blk_unregister_queue(struct gendisk *disk); | 619 | extern void blk_unregister_queue(struct gendisk *disk); |
631 | extern void register_disk(struct gendisk *dev); | 620 | extern void register_disk(struct gendisk *dev); |
@@ -690,16 +679,6 @@ extern void end_that_request_last(struct request *, int); | |||
690 | extern void end_request(struct request *req, int uptodate); | 679 | extern void end_request(struct request *req, int uptodate); |
691 | extern void blk_complete_request(struct request *); | 680 | extern void blk_complete_request(struct request *); |
692 | 681 | ||
693 | static inline int rq_all_done(struct request *rq, unsigned int nr_bytes) | ||
694 | { | ||
695 | if (blk_fs_request(rq)) | ||
696 | return (nr_bytes >= (rq->hard_nr_sectors << 9)); | ||
697 | else if (blk_pc_request(rq)) | ||
698 | return nr_bytes >= rq->data_len; | ||
699 | |||
700 | return 0; | ||
701 | } | ||
702 | |||
703 | /* | 682 | /* |
704 | * end_that_request_first/chunk() takes an uptodate argument. we account | 683 | * end_that_request_first/chunk() takes an uptodate argument. we account |
705 | * any value <= as an io error. 0 means -EIO for compatability reasons, | 684 | * any value <= as an io error. 0 means -EIO for compatability reasons, |
@@ -807,14 +786,6 @@ static inline int queue_dma_alignment(request_queue_t *q) | |||
807 | return retval; | 786 | return retval; |
808 | } | 787 | } |
809 | 788 | ||
810 | static inline int bdev_dma_aligment(struct block_device *bdev) | ||
811 | { | ||
812 | return queue_dma_alignment(bdev_get_queue(bdev)); | ||
813 | } | ||
814 | |||
815 | #define blk_finished_io(nsects) do { } while (0) | ||
816 | #define blk_started_io(nsects) do { } while (0) | ||
817 | |||
818 | /* assumes size > 256 */ | 789 | /* assumes size > 256 */ |
819 | static inline unsigned int blksize_bits(unsigned int size) | 790 | static inline unsigned int blksize_bits(unsigned int size) |
820 | { | 791 | { |