diff options
| author | Ingo Molnar <mingo@elte.hu> | 2008-10-12 09:17:14 -0400 |
|---|---|---|
| committer | Ingo Molnar <mingo@elte.hu> | 2008-10-12 09:17:14 -0400 |
| commit | 620f2efcdc5c7a2db68da41bc3df3cf9a718024e (patch) | |
| tree | b1a0411e2588953777d0b10245b12044c33cef54 /include/linux/blkdev.h | |
| parent | 04944b793e18ece23f63c0252646b310c1845940 (diff) | |
| parent | fd048088306656824958e7783ffcee27e241b361 (diff) | |
Merge branch 'linus' into x86/xsave
Diffstat (limited to 'include/linux/blkdev.h')
| -rw-r--r-- | include/linux/blkdev.h | 171 |
1 files changed, 107 insertions, 64 deletions
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index e61f22be4d0e..a92d9e4ea96e 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h | |||
| @@ -16,7 +16,9 @@ | |||
| 16 | #include <linux/bio.h> | 16 | #include <linux/bio.h> |
| 17 | #include <linux/module.h> | 17 | #include <linux/module.h> |
| 18 | #include <linux/stringify.h> | 18 | #include <linux/stringify.h> |
| 19 | #include <linux/gfp.h> | ||
| 19 | #include <linux/bsg.h> | 20 | #include <linux/bsg.h> |
| 21 | #include <linux/smp.h> | ||
| 20 | 22 | ||
| 21 | #include <asm/scatterlist.h> | 23 | #include <asm/scatterlist.h> |
| 22 | 24 | ||
| @@ -54,7 +56,6 @@ enum rq_cmd_type_bits { | |||
| 54 | REQ_TYPE_PM_SUSPEND, /* suspend request */ | 56 | REQ_TYPE_PM_SUSPEND, /* suspend request */ |
| 55 | REQ_TYPE_PM_RESUME, /* resume request */ | 57 | REQ_TYPE_PM_RESUME, /* resume request */ |
| 56 | REQ_TYPE_PM_SHUTDOWN, /* shutdown request */ | 58 | REQ_TYPE_PM_SHUTDOWN, /* shutdown request */ |
| 57 | REQ_TYPE_FLUSH, /* flush request */ | ||
| 58 | REQ_TYPE_SPECIAL, /* driver defined type */ | 59 | REQ_TYPE_SPECIAL, /* driver defined type */ |
| 59 | REQ_TYPE_LINUX_BLOCK, /* generic block layer message */ | 60 | REQ_TYPE_LINUX_BLOCK, /* generic block layer message */ |
| 60 | /* | 61 | /* |
| @@ -76,19 +77,18 @@ enum rq_cmd_type_bits { | |||
| 76 | * | 77 | * |
| 77 | */ | 78 | */ |
| 78 | enum { | 79 | enum { |
| 79 | /* | ||
| 80 | * just examples for now | ||
| 81 | */ | ||
| 82 | REQ_LB_OP_EJECT = 0x40, /* eject request */ | 80 | REQ_LB_OP_EJECT = 0x40, /* eject request */ |
| 83 | REQ_LB_OP_FLUSH = 0x41, /* flush device */ | 81 | REQ_LB_OP_FLUSH = 0x41, /* flush request */ |
| 82 | REQ_LB_OP_DISCARD = 0x42, /* discard sectors */ | ||
| 84 | }; | 83 | }; |
| 85 | 84 | ||
| 86 | /* | 85 | /* |
| 87 | * request type modified bits. first three bits match BIO_RW* bits, important | 86 | * request type modified bits. first two bits match BIO_RW* bits, important |
| 88 | */ | 87 | */ |
| 89 | enum rq_flag_bits { | 88 | enum rq_flag_bits { |
| 90 | __REQ_RW, /* not set, read. set, write */ | 89 | __REQ_RW, /* not set, read. set, write */ |
| 91 | __REQ_FAILFAST, /* no low level driver retries */ | 90 | __REQ_FAILFAST, /* no low level driver retries */ |
| 91 | __REQ_DISCARD, /* request to discard sectors */ | ||
| 92 | __REQ_SORTED, /* elevator knows about this request */ | 92 | __REQ_SORTED, /* elevator knows about this request */ |
| 93 | __REQ_SOFTBARRIER, /* may not be passed by ioscheduler */ | 93 | __REQ_SOFTBARRIER, /* may not be passed by ioscheduler */ |
| 94 | __REQ_HARDBARRIER, /* may not be passed by drive either */ | 94 | __REQ_HARDBARRIER, /* may not be passed by drive either */ |
| @@ -111,6 +111,7 @@ enum rq_flag_bits { | |||
| 111 | }; | 111 | }; |
| 112 | 112 | ||
| 113 | #define REQ_RW (1 << __REQ_RW) | 113 | #define REQ_RW (1 << __REQ_RW) |
| 114 | #define REQ_DISCARD (1 << __REQ_DISCARD) | ||
| 114 | #define REQ_FAILFAST (1 << __REQ_FAILFAST) | 115 | #define REQ_FAILFAST (1 << __REQ_FAILFAST) |
| 115 | #define REQ_SORTED (1 << __REQ_SORTED) | 116 | #define REQ_SORTED (1 << __REQ_SORTED) |
| 116 | #define REQ_SOFTBARRIER (1 << __REQ_SOFTBARRIER) | 117 | #define REQ_SOFTBARRIER (1 << __REQ_SOFTBARRIER) |
| @@ -140,12 +141,14 @@ enum rq_flag_bits { | |||
| 140 | */ | 141 | */ |
| 141 | struct request { | 142 | struct request { |
| 142 | struct list_head queuelist; | 143 | struct list_head queuelist; |
| 143 | struct list_head donelist; | 144 | struct call_single_data csd; |
| 145 | int cpu; | ||
| 144 | 146 | ||
| 145 | struct request_queue *q; | 147 | struct request_queue *q; |
| 146 | 148 | ||
| 147 | unsigned int cmd_flags; | 149 | unsigned int cmd_flags; |
| 148 | enum rq_cmd_type_bits cmd_type; | 150 | enum rq_cmd_type_bits cmd_type; |
| 151 | unsigned long atomic_flags; | ||
| 149 | 152 | ||
| 150 | /* Maintain bio traversal state for part by part I/O submission. | 153 | /* Maintain bio traversal state for part by part I/O submission. |
| 151 | * hard_* are block layer internals, no driver should touch them! | 154 | * hard_* are block layer internals, no driver should touch them! |
| @@ -190,13 +193,6 @@ struct request { | |||
| 190 | */ | 193 | */ |
| 191 | unsigned short nr_phys_segments; | 194 | unsigned short nr_phys_segments; |
| 192 | 195 | ||
| 193 | /* Number of scatter-gather addr+len pairs after | ||
| 194 | * physical and DMA remapping hardware coalescing is performed. | ||
| 195 | * This is the number of scatter-gather entries the driver | ||
| 196 | * will actually have to deal with after DMA mapping is done. | ||
| 197 | */ | ||
| 198 | unsigned short nr_hw_segments; | ||
| 199 | |||
| 200 | unsigned short ioprio; | 196 | unsigned short ioprio; |
| 201 | 197 | ||
| 202 | void *special; | 198 | void *special; |
| @@ -220,6 +216,8 @@ struct request { | |||
| 220 | void *data; | 216 | void *data; |
| 221 | void *sense; | 217 | void *sense; |
| 222 | 218 | ||
| 219 | unsigned long deadline; | ||
| 220 | struct list_head timeout_list; | ||
| 223 | unsigned int timeout; | 221 | unsigned int timeout; |
| 224 | int retries; | 222 | int retries; |
| 225 | 223 | ||
| @@ -233,6 +231,11 @@ struct request { | |||
| 233 | struct request *next_rq; | 231 | struct request *next_rq; |
| 234 | }; | 232 | }; |
| 235 | 233 | ||
| 234 | static inline unsigned short req_get_ioprio(struct request *req) | ||
| 235 | { | ||
| 236 | return req->ioprio; | ||
| 237 | } | ||
| 238 | |||
| 236 | /* | 239 | /* |
| 237 | * State information carried for REQ_TYPE_PM_SUSPEND and REQ_TYPE_PM_RESUME | 240 | * State information carried for REQ_TYPE_PM_SUSPEND and REQ_TYPE_PM_RESUME |
| 238 | * requests. Some step values could eventually be made generic. | 241 | * requests. Some step values could eventually be made generic. |
| @@ -252,6 +255,7 @@ typedef void (request_fn_proc) (struct request_queue *q); | |||
| 252 | typedef int (make_request_fn) (struct request_queue *q, struct bio *bio); | 255 | typedef int (make_request_fn) (struct request_queue *q, struct bio *bio); |
| 253 | typedef int (prep_rq_fn) (struct request_queue *, struct request *); | 256 | typedef int (prep_rq_fn) (struct request_queue *, struct request *); |
| 254 | typedef void (unplug_fn) (struct request_queue *); | 257 | typedef void (unplug_fn) (struct request_queue *); |
| 258 | typedef int (prepare_discard_fn) (struct request_queue *, struct request *); | ||
| 255 | 259 | ||
| 256 | struct bio_vec; | 260 | struct bio_vec; |
| 257 | struct bvec_merge_data { | 261 | struct bvec_merge_data { |
| @@ -265,6 +269,15 @@ typedef int (merge_bvec_fn) (struct request_queue *, struct bvec_merge_data *, | |||
| 265 | typedef void (prepare_flush_fn) (struct request_queue *, struct request *); | 269 | typedef void (prepare_flush_fn) (struct request_queue *, struct request *); |
| 266 | typedef void (softirq_done_fn)(struct request *); | 270 | typedef void (softirq_done_fn)(struct request *); |
| 267 | typedef int (dma_drain_needed_fn)(struct request *); | 271 | typedef int (dma_drain_needed_fn)(struct request *); |
| 272 | typedef int (lld_busy_fn) (struct request_queue *q); | ||
| 273 | |||
| 274 | enum blk_eh_timer_return { | ||
| 275 | BLK_EH_NOT_HANDLED, | ||
| 276 | BLK_EH_HANDLED, | ||
| 277 | BLK_EH_RESET_TIMER, | ||
| 278 | }; | ||
| 279 | |||
| 280 | typedef enum blk_eh_timer_return (rq_timed_out_fn)(struct request *); | ||
| 268 | 281 | ||
| 269 | enum blk_queue_state { | 282 | enum blk_queue_state { |
| 270 | Queue_down, | 283 | Queue_down, |
| @@ -280,6 +293,15 @@ struct blk_queue_tag { | |||
| 280 | atomic_t refcnt; /* map can be shared */ | 293 | atomic_t refcnt; /* map can be shared */ |
| 281 | }; | 294 | }; |
| 282 | 295 | ||
| 296 | #define BLK_SCSI_MAX_CMDS (256) | ||
| 297 | #define BLK_SCSI_CMD_PER_LONG (BLK_SCSI_MAX_CMDS / (sizeof(long) * 8)) | ||
| 298 | |||
| 299 | struct blk_cmd_filter { | ||
| 300 | unsigned long read_ok[BLK_SCSI_CMD_PER_LONG]; | ||
| 301 | unsigned long write_ok[BLK_SCSI_CMD_PER_LONG]; | ||
| 302 | struct kobject kobj; | ||
| 303 | }; | ||
| 304 | |||
| 283 | struct request_queue | 305 | struct request_queue |
| 284 | { | 306 | { |
| 285 | /* | 307 | /* |
| @@ -298,10 +320,13 @@ struct request_queue | |||
| 298 | make_request_fn *make_request_fn; | 320 | make_request_fn *make_request_fn; |
| 299 | prep_rq_fn *prep_rq_fn; | 321 | prep_rq_fn *prep_rq_fn; |
| 300 | unplug_fn *unplug_fn; | 322 | unplug_fn *unplug_fn; |
| 323 | prepare_discard_fn *prepare_discard_fn; | ||
| 301 | merge_bvec_fn *merge_bvec_fn; | 324 | merge_bvec_fn *merge_bvec_fn; |
| 302 | prepare_flush_fn *prepare_flush_fn; | 325 | prepare_flush_fn *prepare_flush_fn; |
| 303 | softirq_done_fn *softirq_done_fn; | 326 | softirq_done_fn *softirq_done_fn; |
| 327 | rq_timed_out_fn *rq_timed_out_fn; | ||
| 304 | dma_drain_needed_fn *dma_drain_needed; | 328 | dma_drain_needed_fn *dma_drain_needed; |
| 329 | lld_busy_fn *lld_busy_fn; | ||
| 305 | 330 | ||
| 306 | /* | 331 | /* |
| 307 | * Dispatch queue sorting | 332 | * Dispatch queue sorting |
| @@ -376,6 +401,10 @@ struct request_queue | |||
| 376 | unsigned int nr_sorted; | 401 | unsigned int nr_sorted; |
| 377 | unsigned int in_flight; | 402 | unsigned int in_flight; |
| 378 | 403 | ||
| 404 | unsigned int rq_timeout; | ||
| 405 | struct timer_list timeout; | ||
| 406 | struct list_head timeout_list; | ||
| 407 | |||
| 379 | /* | 408 | /* |
| 380 | * sg stuff | 409 | * sg stuff |
| 381 | */ | 410 | */ |
| @@ -398,6 +427,7 @@ struct request_queue | |||
| 398 | #if defined(CONFIG_BLK_DEV_BSG) | 427 | #if defined(CONFIG_BLK_DEV_BSG) |
| 399 | struct bsg_class_device bsg_dev; | 428 | struct bsg_class_device bsg_dev; |
| 400 | #endif | 429 | #endif |
| 430 | struct blk_cmd_filter cmd_filter; | ||
| 401 | }; | 431 | }; |
| 402 | 432 | ||
| 403 | #define QUEUE_FLAG_CLUSTER 0 /* cluster several segments into 1 */ | 433 | #define QUEUE_FLAG_CLUSTER 0 /* cluster several segments into 1 */ |
| @@ -411,6 +441,10 @@ struct request_queue | |||
| 411 | #define QUEUE_FLAG_ELVSWITCH 8 /* don't use elevator, just do FIFO */ | 441 | #define QUEUE_FLAG_ELVSWITCH 8 /* don't use elevator, just do FIFO */ |
| 412 | #define QUEUE_FLAG_BIDI 9 /* queue supports bidi requests */ | 442 | #define QUEUE_FLAG_BIDI 9 /* queue supports bidi requests */ |
| 413 | #define QUEUE_FLAG_NOMERGES 10 /* disable merge attempts */ | 443 | #define QUEUE_FLAG_NOMERGES 10 /* disable merge attempts */ |
| 444 | #define QUEUE_FLAG_SAME_COMP 11 /* force complete on same CPU */ | ||
| 445 | #define QUEUE_FLAG_FAIL_IO 12 /* fake timeout */ | ||
| 446 | #define QUEUE_FLAG_STACKABLE 13 /* supports request stacking */ | ||
| 447 | #define QUEUE_FLAG_NONROT 14 /* non-rotational device (SSD) */ | ||
| 414 | 448 | ||
| 415 | static inline int queue_is_locked(struct request_queue *q) | 449 | static inline int queue_is_locked(struct request_queue *q) |
| 416 | { | 450 | { |
| @@ -516,7 +550,10 @@ enum { | |||
| 516 | #define blk_queue_tagged(q) test_bit(QUEUE_FLAG_QUEUED, &(q)->queue_flags) | 550 | #define blk_queue_tagged(q) test_bit(QUEUE_FLAG_QUEUED, &(q)->queue_flags) |
| 517 | #define blk_queue_stopped(q) test_bit(QUEUE_FLAG_STOPPED, &(q)->queue_flags) | 551 | #define blk_queue_stopped(q) test_bit(QUEUE_FLAG_STOPPED, &(q)->queue_flags) |
| 518 | #define blk_queue_nomerges(q) test_bit(QUEUE_FLAG_NOMERGES, &(q)->queue_flags) | 552 | #define blk_queue_nomerges(q) test_bit(QUEUE_FLAG_NOMERGES, &(q)->queue_flags) |
| 553 | #define blk_queue_nonrot(q) test_bit(QUEUE_FLAG_NONROT, &(q)->queue_flags) | ||
| 519 | #define blk_queue_flushing(q) ((q)->ordseq) | 554 | #define blk_queue_flushing(q) ((q)->ordseq) |
| 555 | #define blk_queue_stackable(q) \ | ||
| 556 | test_bit(QUEUE_FLAG_STACKABLE, &(q)->queue_flags) | ||
| 520 | 557 | ||
| 521 | #define blk_fs_request(rq) ((rq)->cmd_type == REQ_TYPE_FS) | 558 | #define blk_fs_request(rq) ((rq)->cmd_type == REQ_TYPE_FS) |
| 522 | #define blk_pc_request(rq) ((rq)->cmd_type == REQ_TYPE_BLOCK_PC) | 559 | #define blk_pc_request(rq) ((rq)->cmd_type == REQ_TYPE_BLOCK_PC) |
| @@ -526,16 +563,18 @@ enum { | |||
| 526 | #define blk_noretry_request(rq) ((rq)->cmd_flags & REQ_FAILFAST) | 563 | #define blk_noretry_request(rq) ((rq)->cmd_flags & REQ_FAILFAST) |
| 527 | #define blk_rq_started(rq) ((rq)->cmd_flags & REQ_STARTED) | 564 | #define blk_rq_started(rq) ((rq)->cmd_flags & REQ_STARTED) |
| 528 | 565 | ||
| 529 | #define blk_account_rq(rq) (blk_rq_started(rq) && blk_fs_request(rq)) | 566 | #define blk_account_rq(rq) (blk_rq_started(rq) && (blk_fs_request(rq) || blk_discard_rq(rq))) |
| 530 | 567 | ||
| 531 | #define blk_pm_suspend_request(rq) ((rq)->cmd_type == REQ_TYPE_PM_SUSPEND) | 568 | #define blk_pm_suspend_request(rq) ((rq)->cmd_type == REQ_TYPE_PM_SUSPEND) |
| 532 | #define blk_pm_resume_request(rq) ((rq)->cmd_type == REQ_TYPE_PM_RESUME) | 569 | #define blk_pm_resume_request(rq) ((rq)->cmd_type == REQ_TYPE_PM_RESUME) |
| 533 | #define blk_pm_request(rq) \ | 570 | #define blk_pm_request(rq) \ |
| 534 | (blk_pm_suspend_request(rq) || blk_pm_resume_request(rq)) | 571 | (blk_pm_suspend_request(rq) || blk_pm_resume_request(rq)) |
| 535 | 572 | ||
| 573 | #define blk_rq_cpu_valid(rq) ((rq)->cpu != -1) | ||
| 536 | #define blk_sorted_rq(rq) ((rq)->cmd_flags & REQ_SORTED) | 574 | #define blk_sorted_rq(rq) ((rq)->cmd_flags & REQ_SORTED) |
| 537 | #define blk_barrier_rq(rq) ((rq)->cmd_flags & REQ_HARDBARRIER) | 575 | #define blk_barrier_rq(rq) ((rq)->cmd_flags & REQ_HARDBARRIER) |
| 538 | #define blk_fua_rq(rq) ((rq)->cmd_flags & REQ_FUA) | 576 | #define blk_fua_rq(rq) ((rq)->cmd_flags & REQ_FUA) |
| 577 | #define blk_discard_rq(rq) ((rq)->cmd_flags & REQ_DISCARD) | ||
| 539 | #define blk_bidi_rq(rq) ((rq)->next_rq != NULL) | 578 | #define blk_bidi_rq(rq) ((rq)->next_rq != NULL) |
| 540 | #define blk_empty_barrier(rq) (blk_barrier_rq(rq) && blk_fs_request(rq) && !(rq)->hard_nr_sectors) | 579 | #define blk_empty_barrier(rq) (blk_barrier_rq(rq) && blk_fs_request(rq) && !(rq)->hard_nr_sectors) |
| 541 | /* rq->queuelist of dequeued request must be list_empty() */ | 580 | /* rq->queuelist of dequeued request must be list_empty() */ |
| @@ -582,7 +621,8 @@ static inline void blk_clear_queue_full(struct request_queue *q, int rw) | |||
| 582 | #define RQ_NOMERGE_FLAGS \ | 621 | #define RQ_NOMERGE_FLAGS \ |
| 583 | (REQ_NOMERGE | REQ_STARTED | REQ_HARDBARRIER | REQ_SOFTBARRIER) | 622 | (REQ_NOMERGE | REQ_STARTED | REQ_HARDBARRIER | REQ_SOFTBARRIER) |
| 584 | #define rq_mergeable(rq) \ | 623 | #define rq_mergeable(rq) \ |
| 585 | (!((rq)->cmd_flags & RQ_NOMERGE_FLAGS) && blk_fs_request((rq))) | 624 | (!((rq)->cmd_flags & RQ_NOMERGE_FLAGS) && \ |
| 625 | (blk_discard_rq(rq) || blk_fs_request((rq)))) | ||
| 586 | 626 | ||
| 587 | /* | 627 | /* |
| 588 | * q->prep_rq_fn return values | 628 | * q->prep_rq_fn return values |
| @@ -627,6 +667,12 @@ static inline void blk_queue_bounce(struct request_queue *q, struct bio **bio) | |||
| 627 | } | 667 | } |
| 628 | #endif /* CONFIG_MMU */ | 668 | #endif /* CONFIG_MMU */ |
| 629 | 669 | ||
| 670 | struct rq_map_data { | ||
| 671 | struct page **pages; | ||
| 672 | int page_order; | ||
| 673 | int nr_entries; | ||
| 674 | }; | ||
| 675 | |||
| 630 | struct req_iterator { | 676 | struct req_iterator { |
| 631 | int i; | 677 | int i; |
| 632 | struct bio *bio; | 678 | struct bio *bio; |
| @@ -654,6 +700,10 @@ extern void __blk_put_request(struct request_queue *, struct request *); | |||
| 654 | extern struct request *blk_get_request(struct request_queue *, int, gfp_t); | 700 | extern struct request *blk_get_request(struct request_queue *, int, gfp_t); |
| 655 | extern void blk_insert_request(struct request_queue *, struct request *, int, void *); | 701 | extern void blk_insert_request(struct request_queue *, struct request *, int, void *); |
| 656 | extern void blk_requeue_request(struct request_queue *, struct request *); | 702 | extern void blk_requeue_request(struct request_queue *, struct request *); |
| 703 | extern int blk_rq_check_limits(struct request_queue *q, struct request *rq); | ||
| 704 | extern int blk_lld_busy(struct request_queue *q); | ||
| 705 | extern int blk_insert_cloned_request(struct request_queue *q, | ||
| 706 | struct request *rq); | ||
| 657 | extern void blk_plug_device(struct request_queue *); | 707 | extern void blk_plug_device(struct request_queue *); |
| 658 | extern void blk_plug_device_unlocked(struct request_queue *); | 708 | extern void blk_plug_device_unlocked(struct request_queue *); |
| 659 | extern int blk_remove_plug(struct request_queue *); | 709 | extern int blk_remove_plug(struct request_queue *); |
| @@ -695,11 +745,14 @@ extern void __blk_stop_queue(struct request_queue *q); | |||
| 695 | extern void __blk_run_queue(struct request_queue *); | 745 | extern void __blk_run_queue(struct request_queue *); |
| 696 | extern void blk_run_queue(struct request_queue *); | 746 | extern void blk_run_queue(struct request_queue *); |
| 697 | extern void blk_start_queueing(struct request_queue *); | 747 | extern void blk_start_queueing(struct request_queue *); |
| 698 | extern int blk_rq_map_user(struct request_queue *, struct request *, void __user *, unsigned long); | 748 | extern int blk_rq_map_user(struct request_queue *, struct request *, |
| 749 | struct rq_map_data *, void __user *, unsigned long, | ||
| 750 | gfp_t); | ||
| 699 | extern int blk_rq_unmap_user(struct bio *); | 751 | extern int blk_rq_unmap_user(struct bio *); |
| 700 | extern int blk_rq_map_kern(struct request_queue *, struct request *, void *, unsigned int, gfp_t); | 752 | extern int blk_rq_map_kern(struct request_queue *, struct request *, void *, unsigned int, gfp_t); |
| 701 | extern int blk_rq_map_user_iov(struct request_queue *, struct request *, | 753 | extern int blk_rq_map_user_iov(struct request_queue *, struct request *, |
| 702 | struct sg_iovec *, int, unsigned int); | 754 | struct rq_map_data *, struct sg_iovec *, int, |
| 755 | unsigned int, gfp_t); | ||
| 703 | extern int blk_execute_rq(struct request_queue *, struct gendisk *, | 756 | extern int blk_execute_rq(struct request_queue *, struct gendisk *, |
| 704 | struct request *, int); | 757 | struct request *, int); |
| 705 | extern void blk_execute_rq_nowait(struct request_queue *, struct gendisk *, | 758 | extern void blk_execute_rq_nowait(struct request_queue *, struct gendisk *, |
| @@ -740,12 +793,15 @@ extern int __blk_end_request(struct request *rq, int error, | |||
| 740 | extern int blk_end_bidi_request(struct request *rq, int error, | 793 | extern int blk_end_bidi_request(struct request *rq, int error, |
| 741 | unsigned int nr_bytes, unsigned int bidi_bytes); | 794 | unsigned int nr_bytes, unsigned int bidi_bytes); |
| 742 | extern void end_request(struct request *, int); | 795 | extern void end_request(struct request *, int); |
| 743 | extern void end_queued_request(struct request *, int); | ||
| 744 | extern void end_dequeued_request(struct request *, int); | ||
| 745 | extern int blk_end_request_callback(struct request *rq, int error, | 796 | extern int blk_end_request_callback(struct request *rq, int error, |
| 746 | unsigned int nr_bytes, | 797 | unsigned int nr_bytes, |
| 747 | int (drv_callback)(struct request *)); | 798 | int (drv_callback)(struct request *)); |
| 748 | extern void blk_complete_request(struct request *); | 799 | extern void blk_complete_request(struct request *); |
| 800 | extern void __blk_complete_request(struct request *); | ||
| 801 | extern void blk_abort_request(struct request *); | ||
| 802 | extern void blk_abort_queue(struct request_queue *); | ||
| 803 | extern void blk_update_request(struct request *rq, int error, | ||
| 804 | unsigned int nr_bytes); | ||
| 749 | 805 | ||
| 750 | /* | 806 | /* |
| 751 | * blk_end_request() takes bytes instead of sectors as a complete size. | 807 | * blk_end_request() takes bytes instead of sectors as a complete size. |
| @@ -780,12 +836,16 @@ extern void blk_queue_update_dma_pad(struct request_queue *, unsigned int); | |||
| 780 | extern int blk_queue_dma_drain(struct request_queue *q, | 836 | extern int blk_queue_dma_drain(struct request_queue *q, |
| 781 | dma_drain_needed_fn *dma_drain_needed, | 837 | dma_drain_needed_fn *dma_drain_needed, |
| 782 | void *buf, unsigned int size); | 838 | void *buf, unsigned int size); |
| 839 | extern void blk_queue_lld_busy(struct request_queue *q, lld_busy_fn *fn); | ||
| 783 | extern void blk_queue_segment_boundary(struct request_queue *, unsigned long); | 840 | extern void blk_queue_segment_boundary(struct request_queue *, unsigned long); |
| 784 | extern void blk_queue_prep_rq(struct request_queue *, prep_rq_fn *pfn); | 841 | extern void blk_queue_prep_rq(struct request_queue *, prep_rq_fn *pfn); |
| 785 | extern void blk_queue_merge_bvec(struct request_queue *, merge_bvec_fn *); | 842 | extern void blk_queue_merge_bvec(struct request_queue *, merge_bvec_fn *); |
| 786 | extern void blk_queue_dma_alignment(struct request_queue *, int); | 843 | extern void blk_queue_dma_alignment(struct request_queue *, int); |
| 787 | extern void blk_queue_update_dma_alignment(struct request_queue *, int); | 844 | extern void blk_queue_update_dma_alignment(struct request_queue *, int); |
| 788 | extern void blk_queue_softirq_done(struct request_queue *, softirq_done_fn *); | 845 | extern void blk_queue_softirq_done(struct request_queue *, softirq_done_fn *); |
| 846 | extern void blk_queue_set_discard(struct request_queue *, prepare_discard_fn *); | ||
| 847 | extern void blk_queue_rq_timed_out(struct request_queue *, rq_timed_out_fn *); | ||
| 848 | extern void blk_queue_rq_timeout(struct request_queue *, unsigned int); | ||
| 789 | extern struct backing_dev_info *blk_get_backing_dev_info(struct block_device *bdev); | 849 | extern struct backing_dev_info *blk_get_backing_dev_info(struct block_device *bdev); |
| 790 | extern int blk_queue_ordered(struct request_queue *, unsigned, prepare_flush_fn *); | 850 | extern int blk_queue_ordered(struct request_queue *, unsigned, prepare_flush_fn *); |
| 791 | extern int blk_do_ordered(struct request_queue *, struct request **); | 851 | extern int blk_do_ordered(struct request_queue *, struct request **); |
| @@ -807,8 +867,6 @@ extern void blk_put_queue(struct request_queue *); | |||
| 807 | /* | 867 | /* |
| 808 | * tag stuff | 868 | * tag stuff |
| 809 | */ | 869 | */ |
| 810 | #define blk_queue_tag_depth(q) ((q)->queue_tags->busy) | ||
| 811 | #define blk_queue_tag_queue(q) ((q)->queue_tags->busy < (q)->queue_tags->max_depth) | ||
| 812 | #define blk_rq_tagged(rq) ((rq)->cmd_flags & REQ_QUEUED) | 870 | #define blk_rq_tagged(rq) ((rq)->cmd_flags & REQ_QUEUED) |
| 813 | extern int blk_queue_start_tag(struct request_queue *, struct request *); | 871 | extern int blk_queue_start_tag(struct request_queue *, struct request *); |
| 814 | extern struct request *blk_queue_find_tag(struct request_queue *, int); | 872 | extern struct request *blk_queue_find_tag(struct request_queue *, int); |
| @@ -829,15 +887,23 @@ static inline struct request *blk_map_queue_find_tag(struct blk_queue_tag *bqt, | |||
| 829 | } | 887 | } |
| 830 | 888 | ||
| 831 | extern int blkdev_issue_flush(struct block_device *, sector_t *); | 889 | extern int blkdev_issue_flush(struct block_device *, sector_t *); |
| 890 | extern int blkdev_issue_discard(struct block_device *, | ||
| 891 | sector_t sector, sector_t nr_sects, gfp_t); | ||
| 892 | |||
| 893 | static inline int sb_issue_discard(struct super_block *sb, | ||
| 894 | sector_t block, sector_t nr_blocks) | ||
| 895 | { | ||
| 896 | block <<= (sb->s_blocksize_bits - 9); | ||
| 897 | nr_blocks <<= (sb->s_blocksize_bits - 9); | ||
| 898 | return blkdev_issue_discard(sb->s_bdev, block, nr_blocks, GFP_KERNEL); | ||
| 899 | } | ||
| 832 | 900 | ||
| 833 | /* | 901 | /* |
| 834 | * command filter functions | 902 | * command filter functions |
| 835 | */ | 903 | */ |
| 836 | extern int blk_verify_command(struct file *file, unsigned char *cmd); | 904 | extern int blk_verify_command(struct blk_cmd_filter *filter, |
| 837 | extern int blk_cmd_filter_verify_command(struct blk_scsi_cmd_filter *filter, | 905 | unsigned char *cmd, int has_write_perm); |
| 838 | unsigned char *cmd, mode_t *f_mode); | 906 | extern void blk_set_cmd_filter_defaults(struct blk_cmd_filter *filter); |
| 839 | extern int blk_register_filter(struct gendisk *disk); | ||
| 840 | extern void blk_unregister_filter(struct gendisk *disk); | ||
| 841 | 907 | ||
| 842 | #define MAX_PHYS_SEGMENTS 128 | 908 | #define MAX_PHYS_SEGMENTS 128 |
| 843 | #define MAX_HW_SEGMENTS 128 | 909 | #define MAX_HW_SEGMENTS 128 |
| @@ -868,6 +934,13 @@ static inline int queue_dma_alignment(struct request_queue *q) | |||
| 868 | return q ? q->dma_alignment : 511; | 934 | return q ? q->dma_alignment : 511; |
| 869 | } | 935 | } |
| 870 | 936 | ||
| 937 | static inline int blk_rq_aligned(struct request_queue *q, void *addr, | ||
| 938 | unsigned int len) | ||
| 939 | { | ||
| 940 | unsigned int alignment = queue_dma_alignment(q) | q->dma_pad_mask; | ||
| 941 | return !((unsigned long)addr & alignment) && !(len & alignment); | ||
| 942 | } | ||
| 943 | |||
| 871 | /* assumes size > 256 */ | 944 | /* assumes size > 256 */ |
| 872 | static inline unsigned int blksize_bits(unsigned int size) | 945 | static inline unsigned int blksize_bits(unsigned int size) |
| 873 | { | 946 | { |
| @@ -894,7 +967,7 @@ static inline void put_dev_sector(Sector p) | |||
| 894 | } | 967 | } |
| 895 | 968 | ||
| 896 | struct work_struct; | 969 | struct work_struct; |
| 897 | int kblockd_schedule_work(struct work_struct *work); | 970 | int kblockd_schedule_work(struct request_queue *q, struct work_struct *work); |
| 898 | void kblockd_flush_work(struct work_struct *work); | 971 | void kblockd_flush_work(struct work_struct *work); |
| 899 | 972 | ||
| 900 | #define MODULE_ALIAS_BLOCKDEV(major,minor) \ | 973 | #define MODULE_ALIAS_BLOCKDEV(major,minor) \ |
| @@ -939,49 +1012,19 @@ struct blk_integrity { | |||
| 939 | 1012 | ||
| 940 | extern int blk_integrity_register(struct gendisk *, struct blk_integrity *); | 1013 | extern int blk_integrity_register(struct gendisk *, struct blk_integrity *); |
| 941 | extern void blk_integrity_unregister(struct gendisk *); | 1014 | extern void blk_integrity_unregister(struct gendisk *); |
| 942 | extern int blk_integrity_compare(struct block_device *, struct block_device *); | 1015 | extern int blk_integrity_compare(struct gendisk *, struct gendisk *); |
| 943 | extern int blk_rq_map_integrity_sg(struct request *, struct scatterlist *); | 1016 | extern int blk_rq_map_integrity_sg(struct request *, struct scatterlist *); |
| 944 | extern int blk_rq_count_integrity_sg(struct request *); | 1017 | extern int blk_rq_count_integrity_sg(struct request *); |
| 945 | 1018 | ||
| 946 | static inline unsigned short blk_integrity_tuple_size(struct blk_integrity *bi) | 1019 | static inline |
| 947 | { | 1020 | struct blk_integrity *bdev_get_integrity(struct block_device *bdev) |
| 948 | if (bi) | ||
| 949 | return bi->tuple_size; | ||
| 950 | |||
| 951 | return 0; | ||
| 952 | } | ||
| 953 | |||
| 954 | static inline struct blk_integrity *bdev_get_integrity(struct block_device *bdev) | ||
| 955 | { | 1021 | { |
| 956 | return bdev->bd_disk->integrity; | 1022 | return bdev->bd_disk->integrity; |
| 957 | } | 1023 | } |
| 958 | 1024 | ||
| 959 | static inline unsigned int bdev_get_tag_size(struct block_device *bdev) | 1025 | static inline struct blk_integrity *blk_get_integrity(struct gendisk *disk) |
| 960 | { | 1026 | { |
| 961 | struct blk_integrity *bi = bdev_get_integrity(bdev); | 1027 | return disk->integrity; |
| 962 | |||
| 963 | if (bi) | ||
| 964 | return bi->tag_size; | ||
| 965 | |||
| 966 | return 0; | ||
| 967 | } | ||
| 968 | |||
| 969 | static inline int bdev_integrity_enabled(struct block_device *bdev, int rw) | ||
| 970 | { | ||
| 971 | struct blk_integrity *bi = bdev_get_integrity(bdev); | ||
| 972 | |||
| 973 | if (bi == NULL) | ||
| 974 | return 0; | ||
| 975 | |||
| 976 | if (rw == READ && bi->verify_fn != NULL && | ||
| 977 | (bi->flags & INTEGRITY_FLAG_READ)) | ||
| 978 | return 1; | ||
| 979 | |||
| 980 | if (rw == WRITE && bi->generate_fn != NULL && | ||
| 981 | (bi->flags & INTEGRITY_FLAG_WRITE)) | ||
| 982 | return 1; | ||
| 983 | |||
| 984 | return 0; | ||
| 985 | } | 1028 | } |
| 986 | 1029 | ||
| 987 | static inline int blk_integrity_rq(struct request *rq) | 1030 | static inline int blk_integrity_rq(struct request *rq) |
| @@ -998,7 +1041,7 @@ static inline int blk_integrity_rq(struct request *rq) | |||
| 998 | #define blk_rq_count_integrity_sg(a) (0) | 1041 | #define blk_rq_count_integrity_sg(a) (0) |
| 999 | #define blk_rq_map_integrity_sg(a, b) (0) | 1042 | #define blk_rq_map_integrity_sg(a, b) (0) |
| 1000 | #define bdev_get_integrity(a) (0) | 1043 | #define bdev_get_integrity(a) (0) |
| 1001 | #define bdev_get_tag_size(a) (0) | 1044 | #define blk_get_integrity(a) (0) |
| 1002 | #define blk_integrity_compare(a, b) (0) | 1045 | #define blk_integrity_compare(a, b) (0) |
| 1003 | #define blk_integrity_register(a, b) (0) | 1046 | #define blk_integrity_register(a, b) (0) |
| 1004 | #define blk_integrity_unregister(a) do { } while (0); | 1047 | #define blk_integrity_unregister(a) do { } while (0); |
