diff options
| -rw-r--r-- | block/blk-throttle.c | 2 | ||||
| -rw-r--r-- | drivers/block/amiflop.c | 2 | ||||
| -rw-r--r-- | drivers/block/ataflop.c | 2 | ||||
| -rw-r--r-- | drivers/block/cciss.c | 3 | ||||
| -rw-r--r-- | drivers/block/xen-blkfront.c | 55 | ||||
| -rw-r--r-- | fs/ioprio.c | 31 |
6 files changed, 43 insertions, 52 deletions
diff --git a/block/blk-throttle.c b/block/blk-throttle.c index 56ad4531b412..004be80fd894 100644 --- a/block/blk-throttle.c +++ b/block/blk-throttle.c | |||
| @@ -645,7 +645,7 @@ static int throtl_dispatch_tg(struct throtl_data *td, struct throtl_grp *tg, | |||
| 645 | { | 645 | { |
| 646 | unsigned int nr_reads = 0, nr_writes = 0; | 646 | unsigned int nr_reads = 0, nr_writes = 0; |
| 647 | unsigned int max_nr_reads = throtl_grp_quantum*3/4; | 647 | unsigned int max_nr_reads = throtl_grp_quantum*3/4; |
| 648 | unsigned int max_nr_writes = throtl_grp_quantum - nr_reads; | 648 | unsigned int max_nr_writes = throtl_grp_quantum - max_nr_reads; |
| 649 | struct bio *bio; | 649 | struct bio *bio; |
| 650 | 650 | ||
| 651 | /* Try to dispatch 75% READS and 25% WRITES */ | 651 | /* Try to dispatch 75% READS and 25% WRITES */ |
diff --git a/drivers/block/amiflop.c b/drivers/block/amiflop.c index a1725e6488d3..7888501ad9ee 100644 --- a/drivers/block/amiflop.c +++ b/drivers/block/amiflop.c | |||
| @@ -1341,7 +1341,7 @@ static struct request *set_next_request(void) | |||
| 1341 | { | 1341 | { |
| 1342 | struct request_queue *q; | 1342 | struct request_queue *q; |
| 1343 | int cnt = FD_MAX_UNITS; | 1343 | int cnt = FD_MAX_UNITS; |
| 1344 | struct request *rq; | 1344 | struct request *rq = NULL; |
| 1345 | 1345 | ||
| 1346 | /* Find next queue we can dispatch from */ | 1346 | /* Find next queue we can dispatch from */ |
| 1347 | fdc_queue = fdc_queue + 1; | 1347 | fdc_queue = fdc_queue + 1; |
diff --git a/drivers/block/ataflop.c b/drivers/block/ataflop.c index 4e4cc6c828cb..605a67e40bbf 100644 --- a/drivers/block/ataflop.c +++ b/drivers/block/ataflop.c | |||
| @@ -1399,7 +1399,7 @@ static struct request *set_next_request(void) | |||
| 1399 | { | 1399 | { |
| 1400 | struct request_queue *q; | 1400 | struct request_queue *q; |
| 1401 | int old_pos = fdc_queue; | 1401 | int old_pos = fdc_queue; |
| 1402 | struct request *rq; | 1402 | struct request *rq = NULL; |
| 1403 | 1403 | ||
| 1404 | do { | 1404 | do { |
| 1405 | q = unit[fdc_queue].disk->queue; | 1405 | q = unit[fdc_queue].disk->queue; |
diff --git a/drivers/block/cciss.c b/drivers/block/cciss.c index a67d0a611a8a..f291587d753e 100644 --- a/drivers/block/cciss.c +++ b/drivers/block/cciss.c | |||
| @@ -66,6 +66,7 @@ MODULE_VERSION("3.6.26"); | |||
| 66 | MODULE_LICENSE("GPL"); | 66 | MODULE_LICENSE("GPL"); |
| 67 | 67 | ||
| 68 | static DEFINE_MUTEX(cciss_mutex); | 68 | static DEFINE_MUTEX(cciss_mutex); |
| 69 | static struct proc_dir_entry *proc_cciss; | ||
| 69 | 70 | ||
| 70 | #include "cciss_cmd.h" | 71 | #include "cciss_cmd.h" |
| 71 | #include "cciss.h" | 72 | #include "cciss.h" |
| @@ -363,8 +364,6 @@ static const char *raid_label[] = { "0", "4", "1(1+0)", "5", "5+1", "ADG", | |||
| 363 | #define ENG_GIG_FACTOR (ENG_GIG/512) | 364 | #define ENG_GIG_FACTOR (ENG_GIG/512) |
| 364 | #define ENGAGE_SCSI "engage scsi" | 365 | #define ENGAGE_SCSI "engage scsi" |
| 365 | 366 | ||
| 366 | static struct proc_dir_entry *proc_cciss; | ||
| 367 | |||
| 368 | static void cciss_seq_show_header(struct seq_file *seq) | 367 | static void cciss_seq_show_header(struct seq_file *seq) |
| 369 | { | 368 | { |
| 370 | ctlr_info_t *h = seq->private; | 369 | ctlr_info_t *h = seq->private; |
diff --git a/drivers/block/xen-blkfront.c b/drivers/block/xen-blkfront.c index 255035cfc88a..4f9e22f29138 100644 --- a/drivers/block/xen-blkfront.c +++ b/drivers/block/xen-blkfront.c | |||
| @@ -65,7 +65,7 @@ enum blkif_state { | |||
| 65 | 65 | ||
| 66 | struct blk_shadow { | 66 | struct blk_shadow { |
| 67 | struct blkif_request req; | 67 | struct blkif_request req; |
| 68 | unsigned long request; | 68 | struct request *request; |
| 69 | unsigned long frame[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 69 | unsigned long frame[BLKIF_MAX_SEGMENTS_PER_REQUEST]; |
| 70 | }; | 70 | }; |
| 71 | 71 | ||
| @@ -136,7 +136,7 @@ static void add_id_to_freelist(struct blkfront_info *info, | |||
| 136 | unsigned long id) | 136 | unsigned long id) |
| 137 | { | 137 | { |
| 138 | info->shadow[id].req.id = info->shadow_free; | 138 | info->shadow[id].req.id = info->shadow_free; |
| 139 | info->shadow[id].request = 0; | 139 | info->shadow[id].request = NULL; |
| 140 | info->shadow_free = id; | 140 | info->shadow_free = id; |
| 141 | } | 141 | } |
| 142 | 142 | ||
| @@ -245,14 +245,11 @@ static int blkif_ioctl(struct block_device *bdev, fmode_t mode, | |||
| 245 | } | 245 | } |
| 246 | 246 | ||
| 247 | /* | 247 | /* |
| 248 | * blkif_queue_request | 248 | * Generate a Xen blkfront IO request from a blk layer request. Reads |
| 249 | * and writes are handled as expected. Since we lack a loose flush | ||
| 250 | * request, we map flushes into a full ordered barrier. | ||
| 249 | * | 251 | * |
| 250 | * request block io | 252 | * @req: a request struct |
| 251 | * | ||
| 252 | * id: for guest use only. | ||
| 253 | * operation: BLKIF_OP_{READ,WRITE,PROBE} | ||
| 254 | * buffer: buffer to read/write into. this should be a | ||
| 255 | * virtual address in the guest os. | ||
| 256 | */ | 253 | */ |
| 257 | static int blkif_queue_request(struct request *req) | 254 | static int blkif_queue_request(struct request *req) |
| 258 | { | 255 | { |
| @@ -281,7 +278,7 @@ static int blkif_queue_request(struct request *req) | |||
| 281 | /* Fill out a communications ring structure. */ | 278 | /* Fill out a communications ring structure. */ |
| 282 | ring_req = RING_GET_REQUEST(&info->ring, info->ring.req_prod_pvt); | 279 | ring_req = RING_GET_REQUEST(&info->ring, info->ring.req_prod_pvt); |
| 283 | id = get_id_from_freelist(info); | 280 | id = get_id_from_freelist(info); |
| 284 | info->shadow[id].request = (unsigned long)req; | 281 | info->shadow[id].request = req; |
| 285 | 282 | ||
| 286 | ring_req->id = id; | 283 | ring_req->id = id; |
| 287 | ring_req->sector_number = (blkif_sector_t)blk_rq_pos(req); | 284 | ring_req->sector_number = (blkif_sector_t)blk_rq_pos(req); |
| @@ -290,6 +287,18 @@ static int blkif_queue_request(struct request *req) | |||
| 290 | ring_req->operation = rq_data_dir(req) ? | 287 | ring_req->operation = rq_data_dir(req) ? |
| 291 | BLKIF_OP_WRITE : BLKIF_OP_READ; | 288 | BLKIF_OP_WRITE : BLKIF_OP_READ; |
| 292 | 289 | ||
| 290 | if (req->cmd_flags & (REQ_FLUSH | REQ_FUA)) { | ||
| 291 | /* | ||
| 292 | * Ideally we could just do an unordered | ||
| 293 | * flush-to-disk, but all we have is a full write | ||
| 294 | * barrier at the moment. However, a barrier write is | ||
| 295 | * a superset of FUA, so we can implement it the same | ||
| 296 | * way. (It's also a FLUSH+FUA, since it is | ||
| 297 | * guaranteed ordered WRT previous writes.) | ||
| 298 | */ | ||
| 299 | ring_req->operation = BLKIF_OP_WRITE_BARRIER; | ||
| 300 | } | ||
| 301 | |||
| 293 | ring_req->nr_segments = blk_rq_map_sg(req->q, req, info->sg); | 302 | ring_req->nr_segments = blk_rq_map_sg(req->q, req, info->sg); |
| 294 | BUG_ON(ring_req->nr_segments > BLKIF_MAX_SEGMENTS_PER_REQUEST); | 303 | BUG_ON(ring_req->nr_segments > BLKIF_MAX_SEGMENTS_PER_REQUEST); |
| 295 | 304 | ||
| @@ -634,7 +643,7 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id) | |||
| 634 | 643 | ||
| 635 | bret = RING_GET_RESPONSE(&info->ring, i); | 644 | bret = RING_GET_RESPONSE(&info->ring, i); |
| 636 | id = bret->id; | 645 | id = bret->id; |
| 637 | req = (struct request *)info->shadow[id].request; | 646 | req = info->shadow[id].request; |
| 638 | 647 | ||
| 639 | blkif_completion(&info->shadow[id]); | 648 | blkif_completion(&info->shadow[id]); |
| 640 | 649 | ||
| @@ -647,6 +656,16 @@ static irqreturn_t blkif_interrupt(int irq, void *dev_id) | |||
| 647 | printk(KERN_WARNING "blkfront: %s: write barrier op failed\n", | 656 | printk(KERN_WARNING "blkfront: %s: write barrier op failed\n", |
| 648 | info->gd->disk_name); | 657 | info->gd->disk_name); |
| 649 | error = -EOPNOTSUPP; | 658 | error = -EOPNOTSUPP; |
| 659 | } | ||
| 660 | if (unlikely(bret->status == BLKIF_RSP_ERROR && | ||
| 661 | info->shadow[id].req.nr_segments == 0)) { | ||
| 662 | printk(KERN_WARNING "blkfront: %s: empty write barrier op failed\n", | ||
| 663 | info->gd->disk_name); | ||
| 664 | error = -EOPNOTSUPP; | ||
| 665 | } | ||
| 666 | if (unlikely(error)) { | ||
| 667 | if (error == -EOPNOTSUPP) | ||
| 668 | error = 0; | ||
| 650 | info->feature_flush = 0; | 669 | info->feature_flush = 0; |
| 651 | xlvbd_flush(info); | 670 | xlvbd_flush(info); |
| 652 | } | 671 | } |
| @@ -899,7 +918,7 @@ static int blkif_recover(struct blkfront_info *info) | |||
| 899 | /* Stage 3: Find pending requests and requeue them. */ | 918 | /* Stage 3: Find pending requests and requeue them. */ |
| 900 | for (i = 0; i < BLK_RING_SIZE; i++) { | 919 | for (i = 0; i < BLK_RING_SIZE; i++) { |
| 901 | /* Not in use? */ | 920 | /* Not in use? */ |
| 902 | if (copy[i].request == 0) | 921 | if (!copy[i].request) |
| 903 | continue; | 922 | continue; |
| 904 | 923 | ||
| 905 | /* Grab a request slot and copy shadow state into it. */ | 924 | /* Grab a request slot and copy shadow state into it. */ |
| @@ -916,9 +935,7 @@ static int blkif_recover(struct blkfront_info *info) | |||
| 916 | req->seg[j].gref, | 935 | req->seg[j].gref, |
| 917 | info->xbdev->otherend_id, | 936 | info->xbdev->otherend_id, |
| 918 | pfn_to_mfn(info->shadow[req->id].frame[j]), | 937 | pfn_to_mfn(info->shadow[req->id].frame[j]), |
| 919 | rq_data_dir( | 938 | rq_data_dir(info->shadow[req->id].request)); |
| 920 | (struct request *) | ||
| 921 | info->shadow[req->id].request)); | ||
| 922 | info->shadow[req->id].req = *req; | 939 | info->shadow[req->id].req = *req; |
| 923 | 940 | ||
| 924 | info->ring.req_prod_pvt++; | 941 | info->ring.req_prod_pvt++; |
| @@ -1067,14 +1084,8 @@ static void blkfront_connect(struct blkfront_info *info) | |||
| 1067 | */ | 1084 | */ |
| 1068 | info->feature_flush = 0; | 1085 | info->feature_flush = 0; |
| 1069 | 1086 | ||
| 1070 | /* | ||
| 1071 | * The driver doesn't properly handled empty flushes, so | ||
| 1072 | * lets disable barrier support for now. | ||
| 1073 | */ | ||
| 1074 | #if 0 | ||
| 1075 | if (!err && barrier) | 1087 | if (!err && barrier) |
| 1076 | info->feature_flush = REQ_FLUSH; | 1088 | info->feature_flush = REQ_FLUSH | REQ_FUA; |
| 1077 | #endif | ||
| 1078 | 1089 | ||
| 1079 | err = xlvbd_alloc_gendisk(sectors, info, binfo, sector_size); | 1090 | err = xlvbd_alloc_gendisk(sectors, info, binfo, sector_size); |
| 1080 | if (err) { | 1091 | if (err) { |
diff --git a/fs/ioprio.c b/fs/ioprio.c index 2f7d05c89922..7da2a06508e5 100644 --- a/fs/ioprio.c +++ b/fs/ioprio.c | |||
| @@ -103,22 +103,15 @@ SYSCALL_DEFINE3(ioprio_set, int, which, int, who, int, ioprio) | |||
| 103 | } | 103 | } |
| 104 | 104 | ||
| 105 | ret = -ESRCH; | 105 | ret = -ESRCH; |
| 106 | /* | 106 | rcu_read_lock(); |
| 107 | * We want IOPRIO_WHO_PGRP/IOPRIO_WHO_USER to be "atomic", | ||
| 108 | * so we can't use rcu_read_lock(). See re-copy of ->ioprio | ||
| 109 | * in copy_process(). | ||
| 110 | */ | ||
| 111 | read_lock(&tasklist_lock); | ||
| 112 | switch (which) { | 107 | switch (which) { |
| 113 | case IOPRIO_WHO_PROCESS: | 108 | case IOPRIO_WHO_PROCESS: |
| 114 | rcu_read_lock(); | ||
| 115 | if (!who) | 109 | if (!who) |
| 116 | p = current; | 110 | p = current; |
| 117 | else | 111 | else |
| 118 | p = find_task_by_vpid(who); | 112 | p = find_task_by_vpid(who); |
| 119 | if (p) | 113 | if (p) |
| 120 | ret = set_task_ioprio(p, ioprio); | 114 | ret = set_task_ioprio(p, ioprio); |
| 121 | rcu_read_unlock(); | ||
| 122 | break; | 115 | break; |
| 123 | case IOPRIO_WHO_PGRP: | 116 | case IOPRIO_WHO_PGRP: |
| 124 | if (!who) | 117 | if (!who) |
| @@ -141,12 +134,7 @@ SYSCALL_DEFINE3(ioprio_set, int, which, int, who, int, ioprio) | |||
| 141 | break; | 134 | break; |
| 142 | 135 | ||
| 143 | do_each_thread(g, p) { | 136 | do_each_thread(g, p) { |
| 144 | int match; | 137 | if (__task_cred(p)->uid != who) |
| 145 | |||
| 146 | rcu_read_lock(); | ||
| 147 | match = __task_cred(p)->uid == who; | ||
| 148 | rcu_read_unlock(); | ||
| 149 | if (!match) | ||
| 150 | continue; | 138 | continue; |
| 151 | ret = set_task_ioprio(p, ioprio); | 139 | ret = set_task_ioprio(p, ioprio); |
| 152 | if (ret) | 140 | if (ret) |
| @@ -160,7 +148,7 @@ free_uid: | |||
| 160 | ret = -EINVAL; | 148 | ret = -EINVAL; |
| 161 | } | 149 | } |
| 162 | 150 | ||
| 163 | read_unlock(&tasklist_lock); | 151 | rcu_read_unlock(); |
| 164 | return ret; | 152 | return ret; |
| 165 | } | 153 | } |
| 166 | 154 | ||
| @@ -204,17 +192,15 @@ SYSCALL_DEFINE2(ioprio_get, int, which, int, who) | |||
| 204 | int ret = -ESRCH; | 192 | int ret = -ESRCH; |
| 205 | int tmpio; | 193 | int tmpio; |
| 206 | 194 | ||
| 207 | read_lock(&tasklist_lock); | 195 | rcu_read_lock(); |
| 208 | switch (which) { | 196 | switch (which) { |
| 209 | case IOPRIO_WHO_PROCESS: | 197 | case IOPRIO_WHO_PROCESS: |
| 210 | rcu_read_lock(); | ||
| 211 | if (!who) | 198 | if (!who) |
| 212 | p = current; | 199 | p = current; |
| 213 | else | 200 | else |
| 214 | p = find_task_by_vpid(who); | 201 | p = find_task_by_vpid(who); |
| 215 | if (p) | 202 | if (p) |
| 216 | ret = get_task_ioprio(p); | 203 | ret = get_task_ioprio(p); |
| 217 | rcu_read_unlock(); | ||
| 218 | break; | 204 | break; |
| 219 | case IOPRIO_WHO_PGRP: | 205 | case IOPRIO_WHO_PGRP: |
| 220 | if (!who) | 206 | if (!who) |
| @@ -241,12 +227,7 @@ SYSCALL_DEFINE2(ioprio_get, int, which, int, who) | |||
| 241 | break; | 227 | break; |
| 242 | 228 | ||
| 243 | do_each_thread(g, p) { | 229 | do_each_thread(g, p) { |
| 244 | int match; | 230 | if (__task_cred(p)->uid != user->uid) |
| 245 | |||
| 246 | rcu_read_lock(); | ||
| 247 | match = __task_cred(p)->uid == user->uid; | ||
| 248 | rcu_read_unlock(); | ||
| 249 | if (!match) | ||
| 250 | continue; | 231 | continue; |
| 251 | tmpio = get_task_ioprio(p); | 232 | tmpio = get_task_ioprio(p); |
| 252 | if (tmpio < 0) | 233 | if (tmpio < 0) |
| @@ -264,6 +245,6 @@ SYSCALL_DEFINE2(ioprio_get, int, which, int, who) | |||
| 264 | ret = -EINVAL; | 245 | ret = -EINVAL; |
| 265 | } | 246 | } |
| 266 | 247 | ||
| 267 | read_unlock(&tasklist_lock); | 248 | rcu_read_unlock(); |
| 268 | return ret; | 249 | return ret; |
| 269 | } | 250 | } |
