aboutsummaryrefslogtreecommitdiffstats
path: root/fs/ext4
diff options
context:
space:
mode:
authorJiri Kosina <jkosina@suse.cz>2009-12-07 12:36:35 -0500
committerJiri Kosina <jkosina@suse.cz>2009-12-07 12:36:35 -0500
commitd014d043869cdc591f3a33243d3481fa4479c2d0 (patch)
tree63626829498e647ba058a1ce06419fe7e4d5f97d /fs/ext4
parent6ec22f9b037fc0c2e00ddb7023fad279c365324d (diff)
parent6070d81eb5f2d4943223c96e7609a53cdc984364 (diff)
Merge branch 'for-next' into for-linus
Conflicts: kernel/irq/chip.c
Diffstat (limited to 'fs/ext4')
-rw-r--r--fs/ext4/inode.c8
-rw-r--r--fs/ext4/mballoc.c2
2 files changed, 5 insertions, 5 deletions
diff --git a/fs/ext4/inode.c b/fs/ext4/inode.c
index 2c8caa51add..4e8e2f15b8b 100644
--- a/fs/ext4/inode.c
+++ b/fs/ext4/inode.c
@@ -2933,7 +2933,7 @@ retry:
2933 ret = write_cache_pages(mapping, wbc, __mpage_da_writepage, 2933 ret = write_cache_pages(mapping, wbc, __mpage_da_writepage,
2934 &mpd); 2934 &mpd);
2935 /* 2935 /*
2936 * If we have a contigous extent of pages and we 2936 * If we have a contiguous extent of pages and we
2937 * haven't done the I/O yet, map the blocks and submit 2937 * haven't done the I/O yet, map the blocks and submit
2938 * them for I/O. 2938 * them for I/O.
2939 */ 2939 */
@@ -4064,7 +4064,7 @@ static Indirect *ext4_find_shared(struct inode *inode, int depth,
4064 int k, err; 4064 int k, err;
4065 4065
4066 *top = 0; 4066 *top = 0;
4067 /* Make k index the deepest non-null offest + 1 */ 4067 /* Make k index the deepest non-null offset + 1 */
4068 for (k = depth; k > 1 && !offsets[k-1]; k--) 4068 for (k = depth; k > 1 && !offsets[k-1]; k--)
4069 ; 4069 ;
4070 partial = ext4_get_branch(inode, k, offsets, chain, &err); 4070 partial = ext4_get_branch(inode, k, offsets, chain, &err);
@@ -5376,7 +5376,7 @@ static int ext4_index_trans_blocks(struct inode *inode, int nrblocks, int chunk)
5376 * worse case, the indexs blocks spread over different block groups 5376 * worse case, the indexs blocks spread over different block groups
5377 * 5377 *
5378 * If datablocks are discontiguous, they are possible to spread over 5378 * If datablocks are discontiguous, they are possible to spread over
5379 * different block groups too. If they are contiugous, with flexbg, 5379 * different block groups too. If they are contiuguous, with flexbg,
5380 * they could still across block group boundary. 5380 * they could still across block group boundary.
5381 * 5381 *
5382 * Also account for superblock, inode, quota and xattr blocks 5382 * Also account for superblock, inode, quota and xattr blocks
@@ -5452,7 +5452,7 @@ int ext4_writepage_trans_blocks(struct inode *inode)
5452 * Calculate the journal credits for a chunk of data modification. 5452 * Calculate the journal credits for a chunk of data modification.
5453 * 5453 *
5454 * This is called from DIO, fallocate or whoever calling 5454 * This is called from DIO, fallocate or whoever calling
5455 * ext4_get_blocks() to map/allocate a chunk of contigous disk blocks. 5455 * ext4_get_blocks() to map/allocate a chunk of contiguous disk blocks.
5456 * 5456 *
5457 * journal buffers for data blocks are not included here, as DIO 5457 * journal buffers for data blocks are not included here, as DIO
5458 * and fallocate do no need to journal data buffers. 5458 * and fallocate do no need to journal data buffers.
diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c
index bba12824def..74e495dabe0 100644
--- a/fs/ext4/mballoc.c
+++ b/fs/ext4/mballoc.c
@@ -142,7 +142,7 @@
142 * 2 blocks and the order of allocation is >= sbi->s_mb_order2_reqs. The 142 * 2 blocks and the order of allocation is >= sbi->s_mb_order2_reqs. The
143 * value of s_mb_order2_reqs can be tuned via 143 * value of s_mb_order2_reqs can be tuned via
144 * /sys/fs/ext4/<partition>/mb_order2_req. If the request len is equal to 144 * /sys/fs/ext4/<partition>/mb_order2_req. If the request len is equal to
145 * stripe size (sbi->s_stripe), we try to search for contigous block in 145 * stripe size (sbi->s_stripe), we try to search for contiguous block in
146 * stripe size. This should result in better allocation on RAID setups. If 146 * stripe size. This should result in better allocation on RAID setups. If
147 * not, we search in the specific group using bitmap for best extents. The 147 * not, we search in the specific group using bitmap for best extents. The
148 * tunable min_to_scan and max_to_scan control the behaviour here. 148 * tunable min_to_scan and max_to_scan control the behaviour here.