diff options
Diffstat (limited to 'include/linux/blkdev.h')
-rw-r--r-- | include/linux/blkdev.h | 18 |
1 files changed, 6 insertions, 12 deletions
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index ef1afc178c0a..b54a0348a890 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h | |||
@@ -285,16 +285,12 @@ enum blk_queue_state { | |||
285 | Queue_up, | 285 | Queue_up, |
286 | }; | 286 | }; |
287 | 287 | ||
288 | #define BLK_TAGS_PER_LONG (sizeof(unsigned long) * 8) | ||
289 | #define BLK_TAGS_MASK (BLK_TAGS_PER_LONG - 1) | ||
290 | |||
291 | struct blk_queue_tag { | 288 | struct blk_queue_tag { |
292 | struct request **tag_index; /* map of busy tags */ | 289 | struct request **tag_index; /* map of busy tags */ |
293 | unsigned long *tag_map; /* bit map of free/busy tags */ | 290 | unsigned long *tag_map; /* bit map of free/busy tags */ |
294 | struct list_head busy_list; /* fifo list of busy tags */ | 291 | struct list_head busy_list; /* fifo list of busy tags */ |
295 | int busy; /* current depth */ | 292 | int busy; /* current depth */ |
296 | int max_depth; /* what we will send to device */ | 293 | int max_depth; /* what we will send to device */ |
297 | int real_max_depth; /* what the array can hold */ | ||
298 | atomic_t refcnt; /* map can be shared */ | 294 | atomic_t refcnt; /* map can be shared */ |
299 | }; | 295 | }; |
300 | 296 | ||
@@ -396,6 +392,7 @@ struct request_queue | |||
396 | */ | 392 | */ |
397 | unsigned int sg_timeout; | 393 | unsigned int sg_timeout; |
398 | unsigned int sg_reserved_size; | 394 | unsigned int sg_reserved_size; |
395 | int node; | ||
399 | 396 | ||
400 | struct list_head drain_list; | 397 | struct list_head drain_list; |
401 | 398 | ||
@@ -542,15 +539,12 @@ extern void generic_make_request(struct bio *bio); | |||
542 | extern void blk_put_request(struct request *); | 539 | extern void blk_put_request(struct request *); |
543 | extern void blk_end_sync_rq(struct request *rq); | 540 | extern void blk_end_sync_rq(struct request *rq); |
544 | extern void blk_attempt_remerge(request_queue_t *, struct request *); | 541 | extern void blk_attempt_remerge(request_queue_t *, struct request *); |
545 | extern void __blk_attempt_remerge(request_queue_t *, struct request *); | ||
546 | extern struct request *blk_get_request(request_queue_t *, int, int); | 542 | extern struct request *blk_get_request(request_queue_t *, int, int); |
547 | extern void blk_insert_request(request_queue_t *, struct request *, int, void *, int); | 543 | extern void blk_insert_request(request_queue_t *, struct request *, int, void *); |
548 | extern void blk_requeue_request(request_queue_t *, struct request *); | 544 | extern void blk_requeue_request(request_queue_t *, struct request *); |
549 | extern void blk_plug_device(request_queue_t *); | 545 | extern void blk_plug_device(request_queue_t *); |
550 | extern int blk_remove_plug(request_queue_t *); | 546 | extern int blk_remove_plug(request_queue_t *); |
551 | extern void blk_recount_segments(request_queue_t *, struct bio *); | 547 | extern void blk_recount_segments(request_queue_t *, struct bio *); |
552 | extern int blk_phys_contig_segment(request_queue_t *q, struct bio *, struct bio *); | ||
553 | extern int blk_hw_contig_segment(request_queue_t *q, struct bio *, struct bio *); | ||
554 | extern int scsi_cmd_ioctl(struct file *, struct gendisk *, unsigned int, void __user *); | 548 | extern int scsi_cmd_ioctl(struct file *, struct gendisk *, unsigned int, void __user *); |
555 | extern void blk_start_queue(request_queue_t *q); | 549 | extern void blk_start_queue(request_queue_t *q); |
556 | extern void blk_stop_queue(request_queue_t *q); | 550 | extern void blk_stop_queue(request_queue_t *q); |
@@ -615,6 +609,8 @@ static inline void blkdev_dequeue_request(struct request *req) | |||
615 | /* | 609 | /* |
616 | * Access functions for manipulating queue properties | 610 | * Access functions for manipulating queue properties |
617 | */ | 611 | */ |
612 | extern request_queue_t *blk_init_queue_node(request_fn_proc *rfn, | ||
613 | spinlock_t *lock, int node_id); | ||
618 | extern request_queue_t *blk_init_queue(request_fn_proc *, spinlock_t *); | 614 | extern request_queue_t *blk_init_queue(request_fn_proc *, spinlock_t *); |
619 | extern void blk_cleanup_queue(request_queue_t *); | 615 | extern void blk_cleanup_queue(request_queue_t *); |
620 | extern void blk_queue_make_request(request_queue_t *, make_request_fn *); | 616 | extern void blk_queue_make_request(request_queue_t *, make_request_fn *); |
@@ -632,7 +628,6 @@ extern void blk_queue_dma_alignment(request_queue_t *, int); | |||
632 | extern struct backing_dev_info *blk_get_backing_dev_info(struct block_device *bdev); | 628 | extern struct backing_dev_info *blk_get_backing_dev_info(struct block_device *bdev); |
633 | extern void blk_queue_ordered(request_queue_t *, int); | 629 | extern void blk_queue_ordered(request_queue_t *, int); |
634 | extern void blk_queue_issue_flush_fn(request_queue_t *, issue_flush_fn *); | 630 | extern void blk_queue_issue_flush_fn(request_queue_t *, issue_flush_fn *); |
635 | extern int blkdev_scsi_issue_flush_fn(request_queue_t *, struct gendisk *, sector_t *); | ||
636 | extern struct request *blk_start_pre_flush(request_queue_t *,struct request *); | 631 | extern struct request *blk_start_pre_flush(request_queue_t *,struct request *); |
637 | extern int blk_complete_barrier_rq(request_queue_t *, struct request *, int); | 632 | extern int blk_complete_barrier_rq(request_queue_t *, struct request *, int); |
638 | extern int blk_complete_barrier_rq_locked(request_queue_t *, struct request *, int); | 633 | extern int blk_complete_barrier_rq_locked(request_queue_t *, struct request *, int); |
@@ -646,7 +641,8 @@ extern void blk_wait_queue_drained(request_queue_t *, int); | |||
646 | extern void blk_finish_queue_drain(request_queue_t *); | 641 | extern void blk_finish_queue_drain(request_queue_t *); |
647 | 642 | ||
648 | int blk_get_queue(request_queue_t *); | 643 | int blk_get_queue(request_queue_t *); |
649 | request_queue_t *blk_alloc_queue(int); | 644 | request_queue_t *blk_alloc_queue(int gfp_mask); |
645 | request_queue_t *blk_alloc_queue_node(int,int); | ||
650 | #define blk_put_queue(q) blk_cleanup_queue((q)) | 646 | #define blk_put_queue(q) blk_cleanup_queue((q)) |
651 | 647 | ||
652 | /* | 648 | /* |
@@ -675,8 +671,6 @@ extern int blkdev_issue_flush(struct block_device *, sector_t *); | |||
675 | 671 | ||
676 | #define blkdev_entry_to_request(entry) list_entry((entry), struct request, queuelist) | 672 | #define blkdev_entry_to_request(entry) list_entry((entry), struct request, queuelist) |
677 | 673 | ||
678 | extern void drive_stat_acct(struct request *, int, int); | ||
679 | |||
680 | static inline int queue_hardsect_size(request_queue_t *q) | 674 | static inline int queue_hardsect_size(request_queue_t *q) |
681 | { | 675 | { |
682 | int retval = 512; | 676 | int retval = 512; |