diff options
author | Theodore Ts'o <tytso@mit.edu> | 2010-05-16 20:00:00 -0400 |
---|---|---|
committer | Theodore Ts'o <tytso@mit.edu> | 2010-05-16 20:00:00 -0400 |
commit | 2ed886852adfcb070bf350e66a0da0d98b2f3ab5 (patch) | |
tree | bcec0a1004f413b70087e2c43097892f87f21cc3 /fs/ext4/inode.c | |
parent | e35fd6609b2fee54484d520deccb8f18bf7d38f3 (diff) |
ext4: Convert callers of ext4_get_blocks() to use ext4_map_blocks()
This saves a huge amount of stack space by avoiding unnecesary struct
buffer_head's from being allocated on the stack.
In addition, to make the code easier to understand, collapse and
refactor ext4_get_block(), ext4_get_block_write(),
noalloc_get_block_write(), into a single function.
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
Diffstat (limited to 'fs/ext4/inode.c')
-rw-r--r-- | fs/ext4/inode.c | 327 |
1 files changed, 127 insertions, 200 deletions
diff --git a/fs/ext4/inode.c b/fs/ext4/inode.c index ff2f5fd681b5..0b1d7c89f93f 100644 --- a/fs/ext4/inode.c +++ b/fs/ext4/inode.c | |||
@@ -1336,133 +1336,112 @@ int ext4_map_blocks(handle_t *handle, struct inode *inode, | |||
1336 | return retval; | 1336 | return retval; |
1337 | } | 1337 | } |
1338 | 1338 | ||
1339 | int ext4_get_blocks(handle_t *handle, struct inode *inode, sector_t block, | ||
1340 | unsigned int max_blocks, struct buffer_head *bh, | ||
1341 | int flags) | ||
1342 | { | ||
1343 | struct ext4_map_blocks map; | ||
1344 | int ret; | ||
1345 | |||
1346 | map.m_lblk = block; | ||
1347 | map.m_len = max_blocks; | ||
1348 | |||
1349 | ret = ext4_map_blocks(handle, inode, &map, flags); | ||
1350 | if (ret < 0) | ||
1351 | return ret; | ||
1352 | |||
1353 | bh->b_blocknr = map.m_pblk; | ||
1354 | bh->b_size = inode->i_sb->s_blocksize * map.m_len; | ||
1355 | bh->b_bdev = inode->i_sb->s_bdev; | ||
1356 | bh->b_state = (bh->b_state & ~EXT4_MAP_FLAGS) | map.m_flags; | ||
1357 | return ret; | ||
1358 | } | ||
1359 | |||
1360 | /* Maximum number of blocks we map for direct IO at once. */ | 1339 | /* Maximum number of blocks we map for direct IO at once. */ |
1361 | #define DIO_MAX_BLOCKS 4096 | 1340 | #define DIO_MAX_BLOCKS 4096 |
1362 | 1341 | ||
1363 | int ext4_get_block(struct inode *inode, sector_t iblock, | 1342 | static int _ext4_get_block(struct inode *inode, sector_t iblock, |
1364 | struct buffer_head *bh_result, int create) | 1343 | struct buffer_head *bh, int flags) |
1365 | { | 1344 | { |
1366 | handle_t *handle = ext4_journal_current_handle(); | 1345 | handle_t *handle = ext4_journal_current_handle(); |
1346 | struct ext4_map_blocks map; | ||
1367 | int ret = 0, started = 0; | 1347 | int ret = 0, started = 0; |
1368 | unsigned max_blocks = bh_result->b_size >> inode->i_blkbits; | ||
1369 | int dio_credits; | 1348 | int dio_credits; |
1370 | 1349 | ||
1371 | if (create && !handle) { | 1350 | map.m_lblk = iblock; |
1351 | map.m_len = bh->b_size >> inode->i_blkbits; | ||
1352 | |||
1353 | if (flags && !handle) { | ||
1372 | /* Direct IO write... */ | 1354 | /* Direct IO write... */ |
1373 | if (max_blocks > DIO_MAX_BLOCKS) | 1355 | if (map.m_len > DIO_MAX_BLOCKS) |
1374 | max_blocks = DIO_MAX_BLOCKS; | 1356 | map.m_len = DIO_MAX_BLOCKS; |
1375 | dio_credits = ext4_chunk_trans_blocks(inode, max_blocks); | 1357 | dio_credits = ext4_chunk_trans_blocks(inode, map.m_len); |
1376 | handle = ext4_journal_start(inode, dio_credits); | 1358 | handle = ext4_journal_start(inode, dio_credits); |
1377 | if (IS_ERR(handle)) { | 1359 | if (IS_ERR(handle)) { |
1378 | ret = PTR_ERR(handle); | 1360 | ret = PTR_ERR(handle); |
1379 | goto out; | 1361 | return ret; |
1380 | } | 1362 | } |
1381 | started = 1; | 1363 | started = 1; |
1382 | } | 1364 | } |
1383 | 1365 | ||
1384 | ret = ext4_get_blocks(handle, inode, iblock, max_blocks, bh_result, | 1366 | ret = ext4_map_blocks(handle, inode, &map, flags); |
1385 | create ? EXT4_GET_BLOCKS_CREATE : 0); | ||
1386 | if (ret > 0) { | 1367 | if (ret > 0) { |
1387 | bh_result->b_size = (ret << inode->i_blkbits); | 1368 | map_bh(bh, inode->i_sb, map.m_pblk); |
1369 | bh->b_state = (bh->b_state & ~EXT4_MAP_FLAGS) | map.m_flags; | ||
1370 | bh->b_size = inode->i_sb->s_blocksize * map.m_len; | ||
1388 | ret = 0; | 1371 | ret = 0; |
1389 | } | 1372 | } |
1390 | if (started) | 1373 | if (started) |
1391 | ext4_journal_stop(handle); | 1374 | ext4_journal_stop(handle); |
1392 | out: | ||
1393 | return ret; | 1375 | return ret; |
1394 | } | 1376 | } |
1395 | 1377 | ||
1378 | int ext4_get_block(struct inode *inode, sector_t iblock, | ||
1379 | struct buffer_head *bh, int create) | ||
1380 | { | ||
1381 | return _ext4_get_block(inode, iblock, bh, | ||
1382 | create ? EXT4_GET_BLOCKS_CREATE : 0); | ||
1383 | } | ||
1384 | |||
1396 | /* | 1385 | /* |
1397 | * `handle' can be NULL if create is zero | 1386 | * `handle' can be NULL if create is zero |
1398 | */ | 1387 | */ |
1399 | struct buffer_head *ext4_getblk(handle_t *handle, struct inode *inode, | 1388 | struct buffer_head *ext4_getblk(handle_t *handle, struct inode *inode, |
1400 | ext4_lblk_t block, int create, int *errp) | 1389 | ext4_lblk_t block, int create, int *errp) |
1401 | { | 1390 | { |
1402 | struct buffer_head dummy; | 1391 | struct ext4_map_blocks map; |
1392 | struct buffer_head *bh; | ||
1403 | int fatal = 0, err; | 1393 | int fatal = 0, err; |
1404 | int flags = 0; | ||
1405 | 1394 | ||
1406 | J_ASSERT(handle != NULL || create == 0); | 1395 | J_ASSERT(handle != NULL || create == 0); |
1407 | 1396 | ||
1408 | dummy.b_state = 0; | 1397 | map.m_lblk = block; |
1409 | dummy.b_blocknr = -1000; | 1398 | map.m_len = 1; |
1410 | buffer_trace_init(&dummy.b_history); | 1399 | err = ext4_map_blocks(handle, inode, &map, |
1411 | if (create) | 1400 | create ? EXT4_GET_BLOCKS_CREATE : 0); |
1412 | flags |= EXT4_GET_BLOCKS_CREATE; | 1401 | |
1413 | err = ext4_get_blocks(handle, inode, block, 1, &dummy, flags); | 1402 | if (err < 0) |
1414 | /* | 1403 | *errp = err; |
1415 | * ext4_get_blocks() returns number of blocks mapped. 0 in | 1404 | if (err <= 0) |
1416 | * case of a HOLE. | 1405 | return NULL; |
1417 | */ | 1406 | *errp = 0; |
1418 | if (err > 0) { | 1407 | |
1419 | if (err > 1) | 1408 | bh = sb_getblk(inode->i_sb, map.m_pblk); |
1420 | WARN_ON(1); | 1409 | if (!bh) { |
1421 | err = 0; | 1410 | *errp = -EIO; |
1411 | return NULL; | ||
1422 | } | 1412 | } |
1423 | *errp = err; | 1413 | if (map.m_flags & EXT4_MAP_NEW) { |
1424 | if (!err && buffer_mapped(&dummy)) { | 1414 | J_ASSERT(create != 0); |
1425 | struct buffer_head *bh; | 1415 | J_ASSERT(handle != NULL); |
1426 | bh = sb_getblk(inode->i_sb, dummy.b_blocknr); | ||
1427 | if (!bh) { | ||
1428 | *errp = -EIO; | ||
1429 | goto err; | ||
1430 | } | ||
1431 | if (buffer_new(&dummy)) { | ||
1432 | J_ASSERT(create != 0); | ||
1433 | J_ASSERT(handle != NULL); | ||
1434 | 1416 | ||
1435 | /* | 1417 | /* |
1436 | * Now that we do not always journal data, we should | 1418 | * Now that we do not always journal data, we should |
1437 | * keep in mind whether this should always journal the | 1419 | * keep in mind whether this should always journal the |
1438 | * new buffer as metadata. For now, regular file | 1420 | * new buffer as metadata. For now, regular file |
1439 | * writes use ext4_get_block instead, so it's not a | 1421 | * writes use ext4_get_block instead, so it's not a |
1440 | * problem. | 1422 | * problem. |
1441 | */ | 1423 | */ |
1442 | lock_buffer(bh); | 1424 | lock_buffer(bh); |
1443 | BUFFER_TRACE(bh, "call get_create_access"); | 1425 | BUFFER_TRACE(bh, "call get_create_access"); |
1444 | fatal = ext4_journal_get_create_access(handle, bh); | 1426 | fatal = ext4_journal_get_create_access(handle, bh); |
1445 | if (!fatal && !buffer_uptodate(bh)) { | 1427 | if (!fatal && !buffer_uptodate(bh)) { |
1446 | memset(bh->b_data, 0, inode->i_sb->s_blocksize); | 1428 | memset(bh->b_data, 0, inode->i_sb->s_blocksize); |
1447 | set_buffer_uptodate(bh); | 1429 | set_buffer_uptodate(bh); |
1448 | } | ||
1449 | unlock_buffer(bh); | ||
1450 | BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); | ||
1451 | err = ext4_handle_dirty_metadata(handle, inode, bh); | ||
1452 | if (!fatal) | ||
1453 | fatal = err; | ||
1454 | } else { | ||
1455 | BUFFER_TRACE(bh, "not a new buffer"); | ||
1456 | } | ||
1457 | if (fatal) { | ||
1458 | *errp = fatal; | ||
1459 | brelse(bh); | ||
1460 | bh = NULL; | ||
1461 | } | 1430 | } |
1462 | return bh; | 1431 | unlock_buffer(bh); |
1432 | BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata"); | ||
1433 | err = ext4_handle_dirty_metadata(handle, inode, bh); | ||
1434 | if (!fatal) | ||
1435 | fatal = err; | ||
1436 | } else { | ||
1437 | BUFFER_TRACE(bh, "not a new buffer"); | ||
1463 | } | 1438 | } |
1464 | err: | 1439 | if (fatal) { |
1465 | return NULL; | 1440 | *errp = fatal; |
1441 | brelse(bh); | ||
1442 | bh = NULL; | ||
1443 | } | ||
1444 | return bh; | ||
1466 | } | 1445 | } |
1467 | 1446 | ||
1468 | struct buffer_head *ext4_bread(handle_t *handle, struct inode *inode, | 1447 | struct buffer_head *ext4_bread(handle_t *handle, struct inode *inode, |
@@ -2050,28 +2029,23 @@ static int mpage_da_submit_io(struct mpage_da_data *mpd) | |||
2050 | /* | 2029 | /* |
2051 | * mpage_put_bnr_to_bhs - walk blocks and assign them actual numbers | 2030 | * mpage_put_bnr_to_bhs - walk blocks and assign them actual numbers |
2052 | * | 2031 | * |
2053 | * @mpd->inode - inode to walk through | ||
2054 | * @exbh->b_blocknr - first block on a disk | ||
2055 | * @exbh->b_size - amount of space in bytes | ||
2056 | * @logical - first logical block to start assignment with | ||
2057 | * | ||
2058 | * the function goes through all passed space and put actual disk | 2032 | * the function goes through all passed space and put actual disk |
2059 | * block numbers into buffer heads, dropping BH_Delay and BH_Unwritten | 2033 | * block numbers into buffer heads, dropping BH_Delay and BH_Unwritten |
2060 | */ | 2034 | */ |
2061 | static void mpage_put_bnr_to_bhs(struct mpage_da_data *mpd, sector_t logical, | 2035 | static void mpage_put_bnr_to_bhs(struct mpage_da_data *mpd, |
2062 | struct buffer_head *exbh) | 2036 | struct ext4_map_blocks *map) |
2063 | { | 2037 | { |
2064 | struct inode *inode = mpd->inode; | 2038 | struct inode *inode = mpd->inode; |
2065 | struct address_space *mapping = inode->i_mapping; | 2039 | struct address_space *mapping = inode->i_mapping; |
2066 | int blocks = exbh->b_size >> inode->i_blkbits; | 2040 | int blocks = map->m_len; |
2067 | sector_t pblock = exbh->b_blocknr, cur_logical; | 2041 | sector_t pblock = map->m_pblk, cur_logical; |
2068 | struct buffer_head *head, *bh; | 2042 | struct buffer_head *head, *bh; |
2069 | pgoff_t index, end; | 2043 | pgoff_t index, end; |
2070 | struct pagevec pvec; | 2044 | struct pagevec pvec; |
2071 | int nr_pages, i; | 2045 | int nr_pages, i; |
2072 | 2046 | ||
2073 | index = logical >> (PAGE_CACHE_SHIFT - inode->i_blkbits); | 2047 | index = map->m_lblk >> (PAGE_CACHE_SHIFT - inode->i_blkbits); |
2074 | end = (logical + blocks - 1) >> (PAGE_CACHE_SHIFT - inode->i_blkbits); | 2048 | end = (map->m_lblk + blocks - 1) >> (PAGE_CACHE_SHIFT - inode->i_blkbits); |
2075 | cur_logical = index << (PAGE_CACHE_SHIFT - inode->i_blkbits); | 2049 | cur_logical = index << (PAGE_CACHE_SHIFT - inode->i_blkbits); |
2076 | 2050 | ||
2077 | pagevec_init(&pvec, 0); | 2051 | pagevec_init(&pvec, 0); |
@@ -2098,17 +2072,16 @@ static void mpage_put_bnr_to_bhs(struct mpage_da_data *mpd, sector_t logical, | |||
2098 | 2072 | ||
2099 | /* skip blocks out of the range */ | 2073 | /* skip blocks out of the range */ |
2100 | do { | 2074 | do { |
2101 | if (cur_logical >= logical) | 2075 | if (cur_logical >= map->m_lblk) |
2102 | break; | 2076 | break; |
2103 | cur_logical++; | 2077 | cur_logical++; |
2104 | } while ((bh = bh->b_this_page) != head); | 2078 | } while ((bh = bh->b_this_page) != head); |
2105 | 2079 | ||
2106 | do { | 2080 | do { |
2107 | if (cur_logical >= logical + blocks) | 2081 | if (cur_logical >= map->m_lblk + blocks) |
2108 | break; | 2082 | break; |
2109 | 2083 | ||
2110 | if (buffer_delay(bh) || | 2084 | if (buffer_delay(bh) || buffer_unwritten(bh)) { |
2111 | buffer_unwritten(bh)) { | ||
2112 | 2085 | ||
2113 | BUG_ON(bh->b_bdev != inode->i_sb->s_bdev); | 2086 | BUG_ON(bh->b_bdev != inode->i_sb->s_bdev); |
2114 | 2087 | ||
@@ -2127,7 +2100,7 @@ static void mpage_put_bnr_to_bhs(struct mpage_da_data *mpd, sector_t logical, | |||
2127 | } else if (buffer_mapped(bh)) | 2100 | } else if (buffer_mapped(bh)) |
2128 | BUG_ON(bh->b_blocknr != pblock); | 2101 | BUG_ON(bh->b_blocknr != pblock); |
2129 | 2102 | ||
2130 | if (buffer_uninit(exbh)) | 2103 | if (map->m_flags & EXT4_MAP_UNINIT) |
2131 | set_buffer_uninit(bh); | 2104 | set_buffer_uninit(bh); |
2132 | cur_logical++; | 2105 | cur_logical++; |
2133 | pblock++; | 2106 | pblock++; |
@@ -2138,21 +2111,6 @@ static void mpage_put_bnr_to_bhs(struct mpage_da_data *mpd, sector_t logical, | |||
2138 | } | 2111 | } |
2139 | 2112 | ||
2140 | 2113 | ||
2141 | /* | ||
2142 | * __unmap_underlying_blocks - just a helper function to unmap | ||
2143 | * set of blocks described by @bh | ||
2144 | */ | ||
2145 | static inline void __unmap_underlying_blocks(struct inode *inode, | ||
2146 | struct buffer_head *bh) | ||
2147 | { | ||
2148 | struct block_device *bdev = inode->i_sb->s_bdev; | ||
2149 | int blocks, i; | ||
2150 | |||
2151 | blocks = bh->b_size >> inode->i_blkbits; | ||
2152 | for (i = 0; i < blocks; i++) | ||
2153 | unmap_underlying_metadata(bdev, bh->b_blocknr + i); | ||
2154 | } | ||
2155 | |||
2156 | static void ext4_da_block_invalidatepages(struct mpage_da_data *mpd, | 2114 | static void ext4_da_block_invalidatepages(struct mpage_da_data *mpd, |
2157 | sector_t logical, long blk_cnt) | 2115 | sector_t logical, long blk_cnt) |
2158 | { | 2116 | { |
@@ -2214,7 +2172,7 @@ static void ext4_print_free_blocks(struct inode *inode) | |||
2214 | static int mpage_da_map_blocks(struct mpage_da_data *mpd) | 2172 | static int mpage_da_map_blocks(struct mpage_da_data *mpd) |
2215 | { | 2173 | { |
2216 | int err, blks, get_blocks_flags; | 2174 | int err, blks, get_blocks_flags; |
2217 | struct buffer_head new; | 2175 | struct ext4_map_blocks map; |
2218 | sector_t next = mpd->b_blocknr; | 2176 | sector_t next = mpd->b_blocknr; |
2219 | unsigned max_blocks = mpd->b_size >> mpd->inode->i_blkbits; | 2177 | unsigned max_blocks = mpd->b_size >> mpd->inode->i_blkbits; |
2220 | loff_t disksize = EXT4_I(mpd->inode)->i_disksize; | 2178 | loff_t disksize = EXT4_I(mpd->inode)->i_disksize; |
@@ -2255,15 +2213,15 @@ static int mpage_da_map_blocks(struct mpage_da_data *mpd) | |||
2255 | * EXT4_GET_BLOCKS_DELALLOC_RESERVE so the delalloc accounting | 2213 | * EXT4_GET_BLOCKS_DELALLOC_RESERVE so the delalloc accounting |
2256 | * variables are updated after the blocks have been allocated. | 2214 | * variables are updated after the blocks have been allocated. |
2257 | */ | 2215 | */ |
2258 | new.b_state = 0; | 2216 | map.m_lblk = next; |
2217 | map.m_len = max_blocks; | ||
2259 | get_blocks_flags = EXT4_GET_BLOCKS_CREATE; | 2218 | get_blocks_flags = EXT4_GET_BLOCKS_CREATE; |
2260 | if (ext4_should_dioread_nolock(mpd->inode)) | 2219 | if (ext4_should_dioread_nolock(mpd->inode)) |
2261 | get_blocks_flags |= EXT4_GET_BLOCKS_IO_CREATE_EXT; | 2220 | get_blocks_flags |= EXT4_GET_BLOCKS_IO_CREATE_EXT; |
2262 | if (mpd->b_state & (1 << BH_Delay)) | 2221 | if (mpd->b_state & (1 << BH_Delay)) |
2263 | get_blocks_flags |= EXT4_GET_BLOCKS_DELALLOC_RESERVE; | 2222 | get_blocks_flags |= EXT4_GET_BLOCKS_DELALLOC_RESERVE; |
2264 | 2223 | ||
2265 | blks = ext4_get_blocks(handle, mpd->inode, next, max_blocks, | 2224 | blks = ext4_map_blocks(handle, mpd->inode, &map, get_blocks_flags); |
2266 | &new, get_blocks_flags); | ||
2267 | if (blks < 0) { | 2225 | if (blks < 0) { |
2268 | err = blks; | 2226 | err = blks; |
2269 | /* | 2227 | /* |
@@ -2305,10 +2263,13 @@ static int mpage_da_map_blocks(struct mpage_da_data *mpd) | |||
2305 | } | 2263 | } |
2306 | BUG_ON(blks == 0); | 2264 | BUG_ON(blks == 0); |
2307 | 2265 | ||
2308 | new.b_size = (blks << mpd->inode->i_blkbits); | 2266 | if (map.m_flags & EXT4_MAP_NEW) { |
2267 | struct block_device *bdev = mpd->inode->i_sb->s_bdev; | ||
2268 | int i; | ||
2309 | 2269 | ||
2310 | if (buffer_new(&new)) | 2270 | for (i = 0; i < map.m_len; i++) |
2311 | __unmap_underlying_blocks(mpd->inode, &new); | 2271 | unmap_underlying_metadata(bdev, map.m_pblk + i); |
2272 | } | ||
2312 | 2273 | ||
2313 | /* | 2274 | /* |
2314 | * If blocks are delayed marked, we need to | 2275 | * If blocks are delayed marked, we need to |
@@ -2316,7 +2277,7 @@ static int mpage_da_map_blocks(struct mpage_da_data *mpd) | |||
2316 | */ | 2277 | */ |
2317 | if ((mpd->b_state & (1 << BH_Delay)) || | 2278 | if ((mpd->b_state & (1 << BH_Delay)) || |
2318 | (mpd->b_state & (1 << BH_Unwritten))) | 2279 | (mpd->b_state & (1 << BH_Unwritten))) |
2319 | mpage_put_bnr_to_bhs(mpd, next, &new); | 2280 | mpage_put_bnr_to_bhs(mpd, &map); |
2320 | 2281 | ||
2321 | if (ext4_should_order_data(mpd->inode)) { | 2282 | if (ext4_should_order_data(mpd->inode)) { |
2322 | err = ext4_jbd2_file_inode(handle, mpd->inode); | 2283 | err = ext4_jbd2_file_inode(handle, mpd->inode); |
@@ -2534,8 +2495,9 @@ static int __mpage_da_writepage(struct page *page, | |||
2534 | * initialized properly. | 2495 | * initialized properly. |
2535 | */ | 2496 | */ |
2536 | static int ext4_da_get_block_prep(struct inode *inode, sector_t iblock, | 2497 | static int ext4_da_get_block_prep(struct inode *inode, sector_t iblock, |
2537 | struct buffer_head *bh_result, int create) | 2498 | struct buffer_head *bh, int create) |
2538 | { | 2499 | { |
2500 | struct ext4_map_blocks map; | ||
2539 | int ret = 0; | 2501 | int ret = 0; |
2540 | sector_t invalid_block = ~((sector_t) 0xffff); | 2502 | sector_t invalid_block = ~((sector_t) 0xffff); |
2541 | 2503 | ||
@@ -2543,16 +2505,22 @@ static int ext4_da_get_block_prep(struct inode *inode, sector_t iblock, | |||
2543 | invalid_block = ~0; | 2505 | invalid_block = ~0; |
2544 | 2506 | ||
2545 | BUG_ON(create == 0); | 2507 | BUG_ON(create == 0); |
2546 | BUG_ON(bh_result->b_size != inode->i_sb->s_blocksize); | 2508 | BUG_ON(bh->b_size != inode->i_sb->s_blocksize); |
2509 | |||
2510 | map.m_lblk = iblock; | ||
2511 | map.m_len = 1; | ||
2547 | 2512 | ||
2548 | /* | 2513 | /* |
2549 | * first, we need to know whether the block is allocated already | 2514 | * first, we need to know whether the block is allocated already |
2550 | * preallocated blocks are unmapped but should treated | 2515 | * preallocated blocks are unmapped but should treated |
2551 | * the same as allocated blocks. | 2516 | * the same as allocated blocks. |
2552 | */ | 2517 | */ |
2553 | ret = ext4_get_blocks(NULL, inode, iblock, 1, bh_result, 0); | 2518 | ret = ext4_map_blocks(NULL, inode, &map, 0); |
2554 | if ((ret == 0) && !buffer_delay(bh_result)) { | 2519 | if (ret < 0) |
2555 | /* the block isn't (pre)allocated yet, let's reserve space */ | 2520 | return ret; |
2521 | if (ret == 0) { | ||
2522 | if (buffer_delay(bh)) | ||
2523 | return 0; /* Not sure this could or should happen */ | ||
2556 | /* | 2524 | /* |
2557 | * XXX: __block_prepare_write() unmaps passed block, | 2525 | * XXX: __block_prepare_write() unmaps passed block, |
2558 | * is it OK? | 2526 | * is it OK? |
@@ -2562,26 +2530,26 @@ static int ext4_da_get_block_prep(struct inode *inode, sector_t iblock, | |||
2562 | /* not enough space to reserve */ | 2530 | /* not enough space to reserve */ |
2563 | return ret; | 2531 | return ret; |
2564 | 2532 | ||
2565 | map_bh(bh_result, inode->i_sb, invalid_block); | 2533 | map_bh(bh, inode->i_sb, invalid_block); |
2566 | set_buffer_new(bh_result); | 2534 | set_buffer_new(bh); |
2567 | set_buffer_delay(bh_result); | 2535 | set_buffer_delay(bh); |
2568 | } else if (ret > 0) { | 2536 | return 0; |
2569 | bh_result->b_size = (ret << inode->i_blkbits); | ||
2570 | if (buffer_unwritten(bh_result)) { | ||
2571 | /* A delayed write to unwritten bh should | ||
2572 | * be marked new and mapped. Mapped ensures | ||
2573 | * that we don't do get_block multiple times | ||
2574 | * when we write to the same offset and new | ||
2575 | * ensures that we do proper zero out for | ||
2576 | * partial write. | ||
2577 | */ | ||
2578 | set_buffer_new(bh_result); | ||
2579 | set_buffer_mapped(bh_result); | ||
2580 | } | ||
2581 | ret = 0; | ||
2582 | } | 2537 | } |
2583 | 2538 | ||
2584 | return ret; | 2539 | map_bh(bh, inode->i_sb, map.m_pblk); |
2540 | bh->b_state = (bh->b_state & ~EXT4_MAP_FLAGS) | map.m_flags; | ||
2541 | |||
2542 | if (buffer_unwritten(bh)) { | ||
2543 | /* A delayed write to unwritten bh should be marked | ||
2544 | * new and mapped. Mapped ensures that we don't do | ||
2545 | * get_block multiple times when we write to the same | ||
2546 | * offset and new ensures that we do proper zero out | ||
2547 | * for partial write. | ||
2548 | */ | ||
2549 | set_buffer_new(bh); | ||
2550 | set_buffer_mapped(bh); | ||
2551 | } | ||
2552 | return 0; | ||
2585 | } | 2553 | } |
2586 | 2554 | ||
2587 | /* | 2555 | /* |
@@ -2603,21 +2571,8 @@ static int ext4_da_get_block_prep(struct inode *inode, sector_t iblock, | |||
2603 | static int noalloc_get_block_write(struct inode *inode, sector_t iblock, | 2571 | static int noalloc_get_block_write(struct inode *inode, sector_t iblock, |
2604 | struct buffer_head *bh_result, int create) | 2572 | struct buffer_head *bh_result, int create) |
2605 | { | 2573 | { |
2606 | int ret = 0; | ||
2607 | unsigned max_blocks = bh_result->b_size >> inode->i_blkbits; | ||
2608 | |||
2609 | BUG_ON(bh_result->b_size != inode->i_sb->s_blocksize); | 2574 | BUG_ON(bh_result->b_size != inode->i_sb->s_blocksize); |
2610 | 2575 | return _ext4_get_block(inode, iblock, bh_result, 0); | |
2611 | /* | ||
2612 | * we don't want to do block allocation in writepage | ||
2613 | * so call get_block_wrap with create = 0 | ||
2614 | */ | ||
2615 | ret = ext4_get_blocks(NULL, inode, iblock, max_blocks, bh_result, 0); | ||
2616 | if (ret > 0) { | ||
2617 | bh_result->b_size = (ret << inode->i_blkbits); | ||
2618 | ret = 0; | ||
2619 | } | ||
2620 | return ret; | ||
2621 | } | 2576 | } |
2622 | 2577 | ||
2623 | static int bget_one(handle_t *handle, struct buffer_head *bh) | 2578 | static int bget_one(handle_t *handle, struct buffer_head *bh) |
@@ -3644,46 +3599,18 @@ out: | |||
3644 | return ret; | 3599 | return ret; |
3645 | } | 3600 | } |
3646 | 3601 | ||
3602 | /* | ||
3603 | * ext4_get_block used when preparing for a DIO write or buffer write. | ||
3604 | * We allocate an uinitialized extent if blocks haven't been allocated. | ||
3605 | * The extent will be converted to initialized after the IO is complete. | ||
3606 | */ | ||
3647 | static int ext4_get_block_write(struct inode *inode, sector_t iblock, | 3607 | static int ext4_get_block_write(struct inode *inode, sector_t iblock, |
3648 | struct buffer_head *bh_result, int create) | 3608 | struct buffer_head *bh_result, int create) |
3649 | { | 3609 | { |
3650 | handle_t *handle = ext4_journal_current_handle(); | ||
3651 | int ret = 0; | ||
3652 | unsigned max_blocks = bh_result->b_size >> inode->i_blkbits; | ||
3653 | int dio_credits; | ||
3654 | int started = 0; | ||
3655 | |||
3656 | ext4_debug("ext4_get_block_write: inode %lu, create flag %d\n", | 3610 | ext4_debug("ext4_get_block_write: inode %lu, create flag %d\n", |
3657 | inode->i_ino, create); | 3611 | inode->i_ino, create); |
3658 | /* | 3612 | return _ext4_get_block(inode, iblock, bh_result, |
3659 | * ext4_get_block in prepare for a DIO write or buffer write. | 3613 | EXT4_GET_BLOCKS_IO_CREATE_EXT); |
3660 | * We allocate an uinitialized extent if blocks haven't been allocated. | ||
3661 | * The extent will be converted to initialized after IO complete. | ||
3662 | */ | ||
3663 | create = EXT4_GET_BLOCKS_IO_CREATE_EXT; | ||
3664 | |||
3665 | if (!handle) { | ||
3666 | if (max_blocks > DIO_MAX_BLOCKS) | ||
3667 | max_blocks = DIO_MAX_BLOCKS; | ||
3668 | dio_credits = ext4_chunk_trans_blocks(inode, max_blocks); | ||
3669 | handle = ext4_journal_start(inode, dio_credits); | ||
3670 | if (IS_ERR(handle)) { | ||
3671 | ret = PTR_ERR(handle); | ||
3672 | goto out; | ||
3673 | } | ||
3674 | started = 1; | ||
3675 | } | ||
3676 | |||
3677 | ret = ext4_get_blocks(handle, inode, iblock, max_blocks, bh_result, | ||
3678 | create); | ||
3679 | if (ret > 0) { | ||
3680 | bh_result->b_size = (ret << inode->i_blkbits); | ||
3681 | ret = 0; | ||
3682 | } | ||
3683 | if (started) | ||
3684 | ext4_journal_stop(handle); | ||
3685 | out: | ||
3686 | return ret; | ||
3687 | } | 3614 | } |
3688 | 3615 | ||
3689 | static void dump_completed_IO(struct inode * inode) | 3616 | static void dump_completed_IO(struct inode * inode) |