diff options
author | Jens Axboe <jaxboe@fusionio.com> | 2011-05-20 14:33:15 -0400 |
---|---|---|
committer | Jens Axboe <jaxboe@fusionio.com> | 2011-05-20 14:33:15 -0400 |
commit | 698567f3fa790fea37509a54dea855302dd88331 (patch) | |
tree | 7a1df976a0eb12cab03e82c18809a30d5482fee4 /include/linux/blkdev.h | |
parent | d70d0711edd8076ec2ce0ed109106e2df950681b (diff) | |
parent | 61c4f2c81c61f73549928dfd9f3e8f26aa36a8cf (diff) |
Merge commit 'v2.6.39' into for-2.6.40/core
Since for-2.6.40/core was forked off the 2.6.39 devel tree, we've
had churn in the core area that makes it difficult to handle
patches for eg cfq or blk-throttle. Instead of requiring that they
be based in older versions with bugs that have been fixed later
in the rc cycle, merge in 2.6.39 final.
Also fixes up conflicts in the below files.
Conflicts:
drivers/block/paride/pcd.c
drivers/cdrom/viocd.c
drivers/ide/ide-cd.c
Signed-off-by: Jens Axboe <jaxboe@fusionio.com>
Diffstat (limited to 'include/linux/blkdev.h')
-rw-r--r-- | include/linux/blkdev.h | 26 |
1 files changed, 13 insertions, 13 deletions
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index 520d8618ed76..ae9091a68480 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h | |||
@@ -390,20 +390,19 @@ struct request_queue | |||
390 | #define QUEUE_FLAG_SYNCFULL 3 /* read queue has been filled */ | 390 | #define QUEUE_FLAG_SYNCFULL 3 /* read queue has been filled */ |
391 | #define QUEUE_FLAG_ASYNCFULL 4 /* write queue has been filled */ | 391 | #define QUEUE_FLAG_ASYNCFULL 4 /* write queue has been filled */ |
392 | #define QUEUE_FLAG_DEAD 5 /* queue being torn down */ | 392 | #define QUEUE_FLAG_DEAD 5 /* queue being torn down */ |
393 | #define QUEUE_FLAG_REENTER 6 /* Re-entrancy avoidance */ | 393 | #define QUEUE_FLAG_ELVSWITCH 6 /* don't use elevator, just do FIFO */ |
394 | #define QUEUE_FLAG_ELVSWITCH 7 /* don't use elevator, just do FIFO */ | 394 | #define QUEUE_FLAG_BIDI 7 /* queue supports bidi requests */ |
395 | #define QUEUE_FLAG_BIDI 8 /* queue supports bidi requests */ | 395 | #define QUEUE_FLAG_NOMERGES 8 /* disable merge attempts */ |
396 | #define QUEUE_FLAG_NOMERGES 9 /* disable merge attempts */ | 396 | #define QUEUE_FLAG_SAME_COMP 9 /* force complete on same CPU */ |
397 | #define QUEUE_FLAG_SAME_COMP 10 /* force complete on same CPU */ | 397 | #define QUEUE_FLAG_FAIL_IO 10 /* fake timeout */ |
398 | #define QUEUE_FLAG_FAIL_IO 11 /* fake timeout */ | 398 | #define QUEUE_FLAG_STACKABLE 11 /* supports request stacking */ |
399 | #define QUEUE_FLAG_STACKABLE 12 /* supports request stacking */ | 399 | #define QUEUE_FLAG_NONROT 12 /* non-rotational device (SSD) */ |
400 | #define QUEUE_FLAG_NONROT 13 /* non-rotational device (SSD) */ | ||
401 | #define QUEUE_FLAG_VIRT QUEUE_FLAG_NONROT /* paravirt device */ | 400 | #define QUEUE_FLAG_VIRT QUEUE_FLAG_NONROT /* paravirt device */ |
402 | #define QUEUE_FLAG_IO_STAT 15 /* do IO stats */ | 401 | #define QUEUE_FLAG_IO_STAT 13 /* do IO stats */ |
403 | #define QUEUE_FLAG_DISCARD 16 /* supports DISCARD */ | 402 | #define QUEUE_FLAG_DISCARD 14 /* supports DISCARD */ |
404 | #define QUEUE_FLAG_NOXMERGES 17 /* No extended merges */ | 403 | #define QUEUE_FLAG_NOXMERGES 15 /* No extended merges */ |
405 | #define QUEUE_FLAG_ADD_RANDOM 18 /* Contributes to random pool */ | 404 | #define QUEUE_FLAG_ADD_RANDOM 16 /* Contributes to random pool */ |
406 | #define QUEUE_FLAG_SECDISCARD 19 /* supports SECDISCARD */ | 405 | #define QUEUE_FLAG_SECDISCARD 17 /* supports SECDISCARD */ |
407 | 406 | ||
408 | #define QUEUE_FLAG_DEFAULT ((1 << QUEUE_FLAG_IO_STAT) | \ | 407 | #define QUEUE_FLAG_DEFAULT ((1 << QUEUE_FLAG_IO_STAT) | \ |
409 | (1 << QUEUE_FLAG_STACKABLE) | \ | 408 | (1 << QUEUE_FLAG_STACKABLE) | \ |
@@ -701,6 +700,7 @@ extern void blk_sync_queue(struct request_queue *q); | |||
701 | extern void __blk_stop_queue(struct request_queue *q); | 700 | extern void __blk_stop_queue(struct request_queue *q); |
702 | extern void __blk_run_queue(struct request_queue *q); | 701 | extern void __blk_run_queue(struct request_queue *q); |
703 | extern void blk_run_queue(struct request_queue *); | 702 | extern void blk_run_queue(struct request_queue *); |
703 | extern void blk_run_queue_async(struct request_queue *q); | ||
704 | extern int blk_rq_map_user(struct request_queue *, struct request *, | 704 | extern int blk_rq_map_user(struct request_queue *, struct request *, |
705 | struct rq_map_data *, void __user *, unsigned long, | 705 | struct rq_map_data *, void __user *, unsigned long, |
706 | gfp_t); | 706 | gfp_t); |