aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMing Lei <ming.lei@canonical.com>2015-09-17 11:58:38 -0400
committerJens Axboe <axboe@fb.com>2015-09-17 11:58:38 -0400
commit52cc6eead9095e2faf2ec7afc013aa3af1f01ac5 (patch)
treeae5daeeaa270e5fb28b4c0a92f82a3dfe8f6e4fc
parent6fe810bda0bd9a5d7674fc671fac27b8aa8ec243 (diff)
block: blk-merge: fast-clone bio when splitting rw bios
biovecs has become immutable since v3.13, so it isn't necessary to allocate biovecs for the new cloned bios, then we can save one extra biovecs allocation/copy, and the allocation is often not fixed-length and a bit more expensive. For example, if the 'max_sectors_kb' of null blk's queue is set as 16(32 sectors) via sysfs just for making more splits, this patch can increase throught about ~70% in the sequential read test over null_blk(direct io, bs: 1M). Cc: Christoph Hellwig <hch@infradead.org> Cc: Kent Overstreet <kent.overstreet@gmail.com> Cc: Ming Lin <ming.l@ssi.samsung.com> Cc: Dongsu Park <dpark@posteo.net> Signed-off-by: Ming Lei <ming.lei@canonical.com> This fixes a performance regression introduced by commit 54efd50bfd, and allows us to take full advantage of the fact that we have immutable bio_vecs. Hand applied, as it rejected violently with commit 5014c311baa2. Signed-off-by: Jens Axboe <axboe@fb.com>
-rw-r--r--block/blk-merge.c19
1 files changed, 4 insertions, 15 deletions
diff --git a/block/blk-merge.c b/block/blk-merge.c
index 574ea7c0468f..c4e9c37f3e38 100644
--- a/block/blk-merge.c
+++ b/block/blk-merge.c
@@ -66,15 +66,12 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
66 struct bio *bio, 66 struct bio *bio,
67 struct bio_set *bs) 67 struct bio_set *bs)
68{ 68{
69 struct bio *split;
70 struct bio_vec bv, bvprv, *bvprvp = NULL; 69 struct bio_vec bv, bvprv, *bvprvp = NULL;
71 struct bvec_iter iter; 70 struct bvec_iter iter;
72 unsigned seg_size = 0, nsegs = 0, sectors = 0; 71 unsigned seg_size = 0, nsegs = 0, sectors = 0;
73 72
74 bio_for_each_segment(bv, bio, iter) { 73 bio_for_each_segment(bv, bio, iter) {
75 sectors += bv.bv_len >> 9; 74 if (sectors + (bv.bv_len >> 9) > queue_max_sectors(q))
76
77 if (sectors > queue_max_sectors(q))
78 goto split; 75 goto split;
79 76
80 /* 77 /*
@@ -95,6 +92,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
95 seg_size += bv.bv_len; 92 seg_size += bv.bv_len;
96 bvprv = bv; 93 bvprv = bv;
97 bvprvp = &bv; 94 bvprvp = &bv;
95 sectors += bv.bv_len >> 9;
98 continue; 96 continue;
99 } 97 }
100new_segment: 98new_segment:
@@ -105,21 +103,12 @@ new_segment:
105 bvprv = bv; 103 bvprv = bv;
106 bvprvp = &bv; 104 bvprvp = &bv;
107 seg_size = bv.bv_len; 105 seg_size = bv.bv_len;
106 sectors += bv.bv_len >> 9;
108 } 107 }
109 108
110 return NULL; 109 return NULL;
111split: 110split:
112 split = bio_clone_bioset(bio, GFP_NOIO, bs); 111 return bio_split(bio, sectors, GFP_NOIO, bs);
113
114 split->bi_iter.bi_size -= iter.bi_size;
115 bio->bi_iter = iter;
116
117 if (bio_integrity(bio)) {
118 bio_integrity_advance(bio, split->bi_iter.bi_size);
119 bio_integrity_trim(split, 0, bio_sectors(split));
120 }
121
122 return split;
123} 112}
124 113
125void blk_queue_split(struct request_queue *q, struct bio **bio, 114void blk_queue_split(struct request_queue *q, struct bio **bio,