diff options
Diffstat (limited to 'fs/ext4/extents.c')
| -rw-r--r-- | fs/ext4/extents.c | 260 |
1 files changed, 196 insertions, 64 deletions
diff --git a/fs/ext4/extents.c b/fs/ext4/extents.c index 765a4826b118..94c8ee81f5e1 100644 --- a/fs/ext4/extents.c +++ b/fs/ext4/extents.c | |||
| @@ -195,8 +195,7 @@ static ext4_fsblk_t ext4_ext_find_goal(struct inode *inode, | |||
| 195 | if (S_ISREG(inode->i_mode)) | 195 | if (S_ISREG(inode->i_mode)) |
| 196 | block_group++; | 196 | block_group++; |
| 197 | } | 197 | } |
| 198 | bg_start = (block_group * EXT4_BLOCKS_PER_GROUP(inode->i_sb)) + | 198 | bg_start = ext4_group_first_block_no(inode->i_sb, block_group); |
| 199 | le32_to_cpu(EXT4_SB(inode->i_sb)->s_es->s_first_data_block); | ||
| 200 | last_block = ext4_blocks_count(EXT4_SB(inode->i_sb)->s_es) - 1; | 199 | last_block = ext4_blocks_count(EXT4_SB(inode->i_sb)->s_es) - 1; |
| 201 | 200 | ||
| 202 | /* | 201 | /* |
| @@ -440,7 +439,7 @@ static int __ext4_ext_check(const char *function, struct inode *inode, | |||
| 440 | return 0; | 439 | return 0; |
| 441 | 440 | ||
| 442 | corrupted: | 441 | corrupted: |
| 443 | ext4_error(inode->i_sb, function, | 442 | __ext4_error(inode->i_sb, function, |
| 444 | "bad header/extent in inode #%lu: %s - magic %x, " | 443 | "bad header/extent in inode #%lu: %s - magic %x, " |
| 445 | "entries %u, max %u(%u), depth %u(%u)", | 444 | "entries %u, max %u(%u), depth %u(%u)", |
| 446 | inode->i_ino, error_msg, le16_to_cpu(eh->eh_magic), | 445 | inode->i_ino, error_msg, le16_to_cpu(eh->eh_magic), |
| @@ -703,7 +702,12 @@ ext4_ext_find_extent(struct inode *inode, ext4_lblk_t block, | |||
| 703 | } | 702 | } |
| 704 | eh = ext_block_hdr(bh); | 703 | eh = ext_block_hdr(bh); |
| 705 | ppos++; | 704 | ppos++; |
| 706 | BUG_ON(ppos > depth); | 705 | if (unlikely(ppos > depth)) { |
| 706 | put_bh(bh); | ||
| 707 | EXT4_ERROR_INODE(inode, | ||
| 708 | "ppos %d > depth %d", ppos, depth); | ||
| 709 | goto err; | ||
| 710 | } | ||
| 707 | path[ppos].p_bh = bh; | 711 | path[ppos].p_bh = bh; |
| 708 | path[ppos].p_hdr = eh; | 712 | path[ppos].p_hdr = eh; |
| 709 | i--; | 713 | i--; |
| @@ -749,7 +753,12 @@ int ext4_ext_insert_index(handle_t *handle, struct inode *inode, | |||
| 749 | if (err) | 753 | if (err) |
| 750 | return err; | 754 | return err; |
| 751 | 755 | ||
| 752 | BUG_ON(logical == le32_to_cpu(curp->p_idx->ei_block)); | 756 | if (unlikely(logical == le32_to_cpu(curp->p_idx->ei_block))) { |
| 757 | EXT4_ERROR_INODE(inode, | ||
| 758 | "logical %d == ei_block %d!", | ||
| 759 | logical, le32_to_cpu(curp->p_idx->ei_block)); | ||
| 760 | return -EIO; | ||
| 761 | } | ||
| 753 | len = EXT_MAX_INDEX(curp->p_hdr) - curp->p_idx; | 762 | len = EXT_MAX_INDEX(curp->p_hdr) - curp->p_idx; |
| 754 | if (logical > le32_to_cpu(curp->p_idx->ei_block)) { | 763 | if (logical > le32_to_cpu(curp->p_idx->ei_block)) { |
| 755 | /* insert after */ | 764 | /* insert after */ |
| @@ -779,9 +788,17 @@ int ext4_ext_insert_index(handle_t *handle, struct inode *inode, | |||
| 779 | ext4_idx_store_pblock(ix, ptr); | 788 | ext4_idx_store_pblock(ix, ptr); |
| 780 | le16_add_cpu(&curp->p_hdr->eh_entries, 1); | 789 | le16_add_cpu(&curp->p_hdr->eh_entries, 1); |
| 781 | 790 | ||
| 782 | BUG_ON(le16_to_cpu(curp->p_hdr->eh_entries) | 791 | if (unlikely(le16_to_cpu(curp->p_hdr->eh_entries) |
| 783 | > le16_to_cpu(curp->p_hdr->eh_max)); | 792 | > le16_to_cpu(curp->p_hdr->eh_max))) { |
| 784 | BUG_ON(ix > EXT_LAST_INDEX(curp->p_hdr)); | 793 | EXT4_ERROR_INODE(inode, |
| 794 | "logical %d == ei_block %d!", | ||
| 795 | logical, le32_to_cpu(curp->p_idx->ei_block)); | ||
| 796 | return -EIO; | ||
| 797 | } | ||
| 798 | if (unlikely(ix > EXT_LAST_INDEX(curp->p_hdr))) { | ||
| 799 | EXT4_ERROR_INODE(inode, "ix > EXT_LAST_INDEX!"); | ||
| 800 | return -EIO; | ||
| 801 | } | ||
| 785 | 802 | ||
| 786 | err = ext4_ext_dirty(handle, inode, curp); | 803 | err = ext4_ext_dirty(handle, inode, curp); |
| 787 | ext4_std_error(inode->i_sb, err); | 804 | ext4_std_error(inode->i_sb, err); |
| @@ -819,7 +836,10 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
| 819 | 836 | ||
| 820 | /* if current leaf will be split, then we should use | 837 | /* if current leaf will be split, then we should use |
| 821 | * border from split point */ | 838 | * border from split point */ |
| 822 | BUG_ON(path[depth].p_ext > EXT_MAX_EXTENT(path[depth].p_hdr)); | 839 | if (unlikely(path[depth].p_ext > EXT_MAX_EXTENT(path[depth].p_hdr))) { |
| 840 | EXT4_ERROR_INODE(inode, "p_ext > EXT_MAX_EXTENT!"); | ||
| 841 | return -EIO; | ||
| 842 | } | ||
| 823 | if (path[depth].p_ext != EXT_MAX_EXTENT(path[depth].p_hdr)) { | 843 | if (path[depth].p_ext != EXT_MAX_EXTENT(path[depth].p_hdr)) { |
| 824 | border = path[depth].p_ext[1].ee_block; | 844 | border = path[depth].p_ext[1].ee_block; |
| 825 | ext_debug("leaf will be split." | 845 | ext_debug("leaf will be split." |
| @@ -860,7 +880,11 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
| 860 | 880 | ||
| 861 | /* initialize new leaf */ | 881 | /* initialize new leaf */ |
| 862 | newblock = ablocks[--a]; | 882 | newblock = ablocks[--a]; |
| 863 | BUG_ON(newblock == 0); | 883 | if (unlikely(newblock == 0)) { |
| 884 | EXT4_ERROR_INODE(inode, "newblock == 0!"); | ||
| 885 | err = -EIO; | ||
| 886 | goto cleanup; | ||
| 887 | } | ||
| 864 | bh = sb_getblk(inode->i_sb, newblock); | 888 | bh = sb_getblk(inode->i_sb, newblock); |
| 865 | if (!bh) { | 889 | if (!bh) { |
| 866 | err = -EIO; | 890 | err = -EIO; |
| @@ -880,7 +904,14 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
| 880 | ex = EXT_FIRST_EXTENT(neh); | 904 | ex = EXT_FIRST_EXTENT(neh); |
| 881 | 905 | ||
| 882 | /* move remainder of path[depth] to the new leaf */ | 906 | /* move remainder of path[depth] to the new leaf */ |
| 883 | BUG_ON(path[depth].p_hdr->eh_entries != path[depth].p_hdr->eh_max); | 907 | if (unlikely(path[depth].p_hdr->eh_entries != |
| 908 | path[depth].p_hdr->eh_max)) { | ||
| 909 | EXT4_ERROR_INODE(inode, "eh_entries %d != eh_max %d!", | ||
| 910 | path[depth].p_hdr->eh_entries, | ||
| 911 | path[depth].p_hdr->eh_max); | ||
| 912 | err = -EIO; | ||
| 913 | goto cleanup; | ||
| 914 | } | ||
| 884 | /* start copy from next extent */ | 915 | /* start copy from next extent */ |
| 885 | /* TODO: we could do it by single memmove */ | 916 | /* TODO: we could do it by single memmove */ |
| 886 | m = 0; | 917 | m = 0; |
| @@ -927,7 +958,11 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
| 927 | 958 | ||
| 928 | /* create intermediate indexes */ | 959 | /* create intermediate indexes */ |
| 929 | k = depth - at - 1; | 960 | k = depth - at - 1; |
| 930 | BUG_ON(k < 0); | 961 | if (unlikely(k < 0)) { |
| 962 | EXT4_ERROR_INODE(inode, "k %d < 0!", k); | ||
| 963 | err = -EIO; | ||
| 964 | goto cleanup; | ||
| 965 | } | ||
| 931 | if (k) | 966 | if (k) |
| 932 | ext_debug("create %d intermediate indices\n", k); | 967 | ext_debug("create %d intermediate indices\n", k); |
| 933 | /* insert new index into current index block */ | 968 | /* insert new index into current index block */ |
| @@ -964,8 +999,14 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
| 964 | 999 | ||
| 965 | ext_debug("cur 0x%p, last 0x%p\n", path[i].p_idx, | 1000 | ext_debug("cur 0x%p, last 0x%p\n", path[i].p_idx, |
| 966 | EXT_MAX_INDEX(path[i].p_hdr)); | 1001 | EXT_MAX_INDEX(path[i].p_hdr)); |
| 967 | BUG_ON(EXT_MAX_INDEX(path[i].p_hdr) != | 1002 | if (unlikely(EXT_MAX_INDEX(path[i].p_hdr) != |
| 968 | EXT_LAST_INDEX(path[i].p_hdr)); | 1003 | EXT_LAST_INDEX(path[i].p_hdr))) { |
| 1004 | EXT4_ERROR_INODE(inode, | ||
| 1005 | "EXT_MAX_INDEX != EXT_LAST_INDEX ee_block %d!", | ||
| 1006 | le32_to_cpu(path[i].p_ext->ee_block)); | ||
| 1007 | err = -EIO; | ||
| 1008 | goto cleanup; | ||
| 1009 | } | ||
| 969 | while (path[i].p_idx <= EXT_MAX_INDEX(path[i].p_hdr)) { | 1010 | while (path[i].p_idx <= EXT_MAX_INDEX(path[i].p_hdr)) { |
| 970 | ext_debug("%d: move %d:%llu in new index %llu\n", i, | 1011 | ext_debug("%d: move %d:%llu in new index %llu\n", i, |
| 971 | le32_to_cpu(path[i].p_idx->ei_block), | 1012 | le32_to_cpu(path[i].p_idx->ei_block), |
| @@ -1203,7 +1244,10 @@ ext4_ext_search_left(struct inode *inode, struct ext4_ext_path *path, | |||
| 1203 | struct ext4_extent *ex; | 1244 | struct ext4_extent *ex; |
| 1204 | int depth, ee_len; | 1245 | int depth, ee_len; |
| 1205 | 1246 | ||
| 1206 | BUG_ON(path == NULL); | 1247 | if (unlikely(path == NULL)) { |
| 1248 | EXT4_ERROR_INODE(inode, "path == NULL *logical %d!", *logical); | ||
| 1249 | return -EIO; | ||
| 1250 | } | ||
| 1207 | depth = path->p_depth; | 1251 | depth = path->p_depth; |
| 1208 | *phys = 0; | 1252 | *phys = 0; |
| 1209 | 1253 | ||
| @@ -1217,15 +1261,33 @@ ext4_ext_search_left(struct inode *inode, struct ext4_ext_path *path, | |||
| 1217 | ex = path[depth].p_ext; | 1261 | ex = path[depth].p_ext; |
| 1218 | ee_len = ext4_ext_get_actual_len(ex); | 1262 | ee_len = ext4_ext_get_actual_len(ex); |
| 1219 | if (*logical < le32_to_cpu(ex->ee_block)) { | 1263 | if (*logical < le32_to_cpu(ex->ee_block)) { |
| 1220 | BUG_ON(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex); | 1264 | if (unlikely(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex)) { |
| 1265 | EXT4_ERROR_INODE(inode, | ||
| 1266 | "EXT_FIRST_EXTENT != ex *logical %d ee_block %d!", | ||
| 1267 | *logical, le32_to_cpu(ex->ee_block)); | ||
| 1268 | return -EIO; | ||
| 1269 | } | ||
| 1221 | while (--depth >= 0) { | 1270 | while (--depth >= 0) { |
| 1222 | ix = path[depth].p_idx; | 1271 | ix = path[depth].p_idx; |
| 1223 | BUG_ON(ix != EXT_FIRST_INDEX(path[depth].p_hdr)); | 1272 | if (unlikely(ix != EXT_FIRST_INDEX(path[depth].p_hdr))) { |
| 1273 | EXT4_ERROR_INODE(inode, | ||
| 1274 | "ix (%d) != EXT_FIRST_INDEX (%d) (depth %d)!", | ||
| 1275 | ix != NULL ? ix->ei_block : 0, | ||
| 1276 | EXT_FIRST_INDEX(path[depth].p_hdr) != NULL ? | ||
| 1277 | EXT_FIRST_INDEX(path[depth].p_hdr)->ei_block : 0, | ||
| 1278 | depth); | ||
| 1279 | return -EIO; | ||
| 1280 | } | ||
| 1224 | } | 1281 | } |
| 1225 | return 0; | 1282 | return 0; |
| 1226 | } | 1283 | } |
| 1227 | 1284 | ||
| 1228 | BUG_ON(*logical < (le32_to_cpu(ex->ee_block) + ee_len)); | 1285 | if (unlikely(*logical < (le32_to_cpu(ex->ee_block) + ee_len))) { |
| 1286 | EXT4_ERROR_INODE(inode, | ||
| 1287 | "logical %d < ee_block %d + ee_len %d!", | ||
| 1288 | *logical, le32_to_cpu(ex->ee_block), ee_len); | ||
| 1289 | return -EIO; | ||
| 1290 | } | ||
| 1229 | 1291 | ||
| 1230 | *logical = le32_to_cpu(ex->ee_block) + ee_len - 1; | 1292 | *logical = le32_to_cpu(ex->ee_block) + ee_len - 1; |
| 1231 | *phys = ext_pblock(ex) + ee_len - 1; | 1293 | *phys = ext_pblock(ex) + ee_len - 1; |
| @@ -1251,7 +1313,10 @@ ext4_ext_search_right(struct inode *inode, struct ext4_ext_path *path, | |||
| 1251 | int depth; /* Note, NOT eh_depth; depth from top of tree */ | 1313 | int depth; /* Note, NOT eh_depth; depth from top of tree */ |
| 1252 | int ee_len; | 1314 | int ee_len; |
| 1253 | 1315 | ||
| 1254 | BUG_ON(path == NULL); | 1316 | if (unlikely(path == NULL)) { |
| 1317 | EXT4_ERROR_INODE(inode, "path == NULL *logical %d!", *logical); | ||
| 1318 | return -EIO; | ||
| 1319 | } | ||
| 1255 | depth = path->p_depth; | 1320 | depth = path->p_depth; |
| 1256 | *phys = 0; | 1321 | *phys = 0; |
| 1257 | 1322 | ||
| @@ -1265,17 +1330,32 @@ ext4_ext_search_right(struct inode *inode, struct ext4_ext_path *path, | |||
| 1265 | ex = path[depth].p_ext; | 1330 | ex = path[depth].p_ext; |
| 1266 | ee_len = ext4_ext_get_actual_len(ex); | 1331 | ee_len = ext4_ext_get_actual_len(ex); |
| 1267 | if (*logical < le32_to_cpu(ex->ee_block)) { | 1332 | if (*logical < le32_to_cpu(ex->ee_block)) { |
| 1268 | BUG_ON(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex); | 1333 | if (unlikely(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex)) { |
| 1334 | EXT4_ERROR_INODE(inode, | ||
| 1335 | "first_extent(path[%d].p_hdr) != ex", | ||
| 1336 | depth); | ||
| 1337 | return -EIO; | ||
| 1338 | } | ||
| 1269 | while (--depth >= 0) { | 1339 | while (--depth >= 0) { |
| 1270 | ix = path[depth].p_idx; | 1340 | ix = path[depth].p_idx; |
| 1271 | BUG_ON(ix != EXT_FIRST_INDEX(path[depth].p_hdr)); | 1341 | if (unlikely(ix != EXT_FIRST_INDEX(path[depth].p_hdr))) { |
| 1342 | EXT4_ERROR_INODE(inode, | ||
| 1343 | "ix != EXT_FIRST_INDEX *logical %d!", | ||
| 1344 | *logical); | ||
| 1345 | return -EIO; | ||
| 1346 | } | ||
| 1272 | } | 1347 | } |
| 1273 | *logical = le32_to_cpu(ex->ee_block); | 1348 | *logical = le32_to_cpu(ex->ee_block); |
| 1274 | *phys = ext_pblock(ex); | 1349 | *phys = ext_pblock(ex); |
| 1275 | return 0; | 1350 | return 0; |
| 1276 | } | 1351 | } |
| 1277 | 1352 | ||
| 1278 | BUG_ON(*logical < (le32_to_cpu(ex->ee_block) + ee_len)); | 1353 | if (unlikely(*logical < (le32_to_cpu(ex->ee_block) + ee_len))) { |
| 1354 | EXT4_ERROR_INODE(inode, | ||
| 1355 | "logical %d < ee_block %d + ee_len %d!", | ||
| 1356 | *logical, le32_to_cpu(ex->ee_block), ee_len); | ||
| 1357 | return -EIO; | ||
| 1358 | } | ||
| 1279 | 1359 | ||
| 1280 | if (ex != EXT_LAST_EXTENT(path[depth].p_hdr)) { | 1360 | if (ex != EXT_LAST_EXTENT(path[depth].p_hdr)) { |
| 1281 | /* next allocated block in this leaf */ | 1361 | /* next allocated block in this leaf */ |
| @@ -1414,8 +1494,12 @@ static int ext4_ext_correct_indexes(handle_t *handle, struct inode *inode, | |||
| 1414 | 1494 | ||
| 1415 | eh = path[depth].p_hdr; | 1495 | eh = path[depth].p_hdr; |
| 1416 | ex = path[depth].p_ext; | 1496 | ex = path[depth].p_ext; |
| 1417 | BUG_ON(ex == NULL); | 1497 | |
| 1418 | BUG_ON(eh == NULL); | 1498 | if (unlikely(ex == NULL || eh == NULL)) { |
| 1499 | EXT4_ERROR_INODE(inode, | ||
| 1500 | "ex %p == NULL or eh %p == NULL", ex, eh); | ||
| 1501 | return -EIO; | ||
| 1502 | } | ||
| 1419 | 1503 | ||
| 1420 | if (depth == 0) { | 1504 | if (depth == 0) { |
| 1421 | /* there is no tree at all */ | 1505 | /* there is no tree at all */ |
| @@ -1538,8 +1622,9 @@ int ext4_ext_try_to_merge(struct inode *inode, | |||
| 1538 | merge_done = 1; | 1622 | merge_done = 1; |
| 1539 | WARN_ON(eh->eh_entries == 0); | 1623 | WARN_ON(eh->eh_entries == 0); |
| 1540 | if (!eh->eh_entries) | 1624 | if (!eh->eh_entries) |
| 1541 | ext4_error(inode->i_sb, "ext4_ext_try_to_merge", | 1625 | ext4_error(inode->i_sb, |
| 1542 | "inode#%lu, eh->eh_entries = 0!", inode->i_ino); | 1626 | "inode#%lu, eh->eh_entries = 0!", |
| 1627 | inode->i_ino); | ||
| 1543 | } | 1628 | } |
| 1544 | 1629 | ||
| 1545 | return merge_done; | 1630 | return merge_done; |
| @@ -1612,13 +1697,19 @@ int ext4_ext_insert_extent(handle_t *handle, struct inode *inode, | |||
| 1612 | ext4_lblk_t next; | 1697 | ext4_lblk_t next; |
| 1613 | unsigned uninitialized = 0; | 1698 | unsigned uninitialized = 0; |
| 1614 | 1699 | ||
| 1615 | BUG_ON(ext4_ext_get_actual_len(newext) == 0); | 1700 | if (unlikely(ext4_ext_get_actual_len(newext) == 0)) { |
| 1701 | EXT4_ERROR_INODE(inode, "ext4_ext_get_actual_len(newext) == 0"); | ||
| 1702 | return -EIO; | ||
| 1703 | } | ||
| 1616 | depth = ext_depth(inode); | 1704 | depth = ext_depth(inode); |
| 1617 | ex = path[depth].p_ext; | 1705 | ex = path[depth].p_ext; |
| 1618 | BUG_ON(path[depth].p_hdr == NULL); | 1706 | if (unlikely(path[depth].p_hdr == NULL)) { |
| 1707 | EXT4_ERROR_INODE(inode, "path[%d].p_hdr == NULL", depth); | ||
| 1708 | return -EIO; | ||
| 1709 | } | ||
| 1619 | 1710 | ||
| 1620 | /* try to insert block into found extent and return */ | 1711 | /* try to insert block into found extent and return */ |
| 1621 | if (ex && (flag != EXT4_GET_BLOCKS_DIO_CREATE_EXT) | 1712 | if (ex && !(flag & EXT4_GET_BLOCKS_PRE_IO) |
| 1622 | && ext4_can_extents_be_merged(inode, ex, newext)) { | 1713 | && ext4_can_extents_be_merged(inode, ex, newext)) { |
| 1623 | ext_debug("append [%d]%d block to %d:[%d]%d (from %llu)\n", | 1714 | ext_debug("append [%d]%d block to %d:[%d]%d (from %llu)\n", |
| 1624 | ext4_ext_is_uninitialized(newext), | 1715 | ext4_ext_is_uninitialized(newext), |
| @@ -1739,7 +1830,7 @@ has_space: | |||
| 1739 | 1830 | ||
| 1740 | merge: | 1831 | merge: |
| 1741 | /* try to merge extents to the right */ | 1832 | /* try to merge extents to the right */ |
| 1742 | if (flag != EXT4_GET_BLOCKS_DIO_CREATE_EXT) | 1833 | if (!(flag & EXT4_GET_BLOCKS_PRE_IO)) |
| 1743 | ext4_ext_try_to_merge(inode, path, nearex); | 1834 | ext4_ext_try_to_merge(inode, path, nearex); |
| 1744 | 1835 | ||
| 1745 | /* try to merge extents to the left */ | 1836 | /* try to merge extents to the left */ |
| @@ -1787,7 +1878,11 @@ int ext4_ext_walk_space(struct inode *inode, ext4_lblk_t block, | |||
| 1787 | } | 1878 | } |
| 1788 | 1879 | ||
| 1789 | depth = ext_depth(inode); | 1880 | depth = ext_depth(inode); |
| 1790 | BUG_ON(path[depth].p_hdr == NULL); | 1881 | if (unlikely(path[depth].p_hdr == NULL)) { |
| 1882 | EXT4_ERROR_INODE(inode, "path[%d].p_hdr == NULL", depth); | ||
| 1883 | err = -EIO; | ||
| 1884 | break; | ||
| 1885 | } | ||
| 1791 | ex = path[depth].p_ext; | 1886 | ex = path[depth].p_ext; |
| 1792 | next = ext4_ext_next_allocated_block(path); | 1887 | next = ext4_ext_next_allocated_block(path); |
| 1793 | 1888 | ||
| @@ -1838,7 +1933,11 @@ int ext4_ext_walk_space(struct inode *inode, ext4_lblk_t block, | |||
| 1838 | cbex.ec_type = EXT4_EXT_CACHE_EXTENT; | 1933 | cbex.ec_type = EXT4_EXT_CACHE_EXTENT; |
| 1839 | } | 1934 | } |
| 1840 | 1935 | ||
| 1841 | BUG_ON(cbex.ec_len == 0); | 1936 | if (unlikely(cbex.ec_len == 0)) { |
| 1937 | EXT4_ERROR_INODE(inode, "cbex.ec_len == 0"); | ||
| 1938 | err = -EIO; | ||
| 1939 | break; | ||
| 1940 | } | ||
| 1842 | err = func(inode, path, &cbex, ex, cbdata); | 1941 | err = func(inode, path, &cbex, ex, cbdata); |
| 1843 | ext4_ext_drop_refs(path); | 1942 | ext4_ext_drop_refs(path); |
| 1844 | 1943 | ||
| @@ -1952,7 +2051,7 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block, | |||
| 1952 | 2051 | ||
| 1953 | BUG_ON(cex->ec_type != EXT4_EXT_CACHE_GAP && | 2052 | BUG_ON(cex->ec_type != EXT4_EXT_CACHE_GAP && |
| 1954 | cex->ec_type != EXT4_EXT_CACHE_EXTENT); | 2053 | cex->ec_type != EXT4_EXT_CACHE_EXTENT); |
| 1955 | if (block >= cex->ec_block && block < cex->ec_block + cex->ec_len) { | 2054 | if (in_range(block, cex->ec_block, cex->ec_len)) { |
| 1956 | ex->ee_block = cpu_to_le32(cex->ec_block); | 2055 | ex->ee_block = cpu_to_le32(cex->ec_block); |
| 1957 | ext4_ext_store_pblock(ex, cex->ec_start); | 2056 | ext4_ext_store_pblock(ex, cex->ec_start); |
| 1958 | ex->ee_len = cpu_to_le16(cex->ec_len); | 2057 | ex->ee_len = cpu_to_le16(cex->ec_len); |
| @@ -1981,7 +2080,10 @@ static int ext4_ext_rm_idx(handle_t *handle, struct inode *inode, | |||
| 1981 | /* free index block */ | 2080 | /* free index block */ |
| 1982 | path--; | 2081 | path--; |
| 1983 | leaf = idx_pblock(path->p_idx); | 2082 | leaf = idx_pblock(path->p_idx); |
| 1984 | BUG_ON(path->p_hdr->eh_entries == 0); | 2083 | if (unlikely(path->p_hdr->eh_entries == 0)) { |
| 2084 | EXT4_ERROR_INODE(inode, "path->p_hdr->eh_entries == 0"); | ||
| 2085 | return -EIO; | ||
| 2086 | } | ||
| 1985 | err = ext4_ext_get_access(handle, inode, path); | 2087 | err = ext4_ext_get_access(handle, inode, path); |
| 1986 | if (err) | 2088 | if (err) |
| 1987 | return err; | 2089 | return err; |
| @@ -2119,8 +2221,10 @@ ext4_ext_rm_leaf(handle_t *handle, struct inode *inode, | |||
| 2119 | if (!path[depth].p_hdr) | 2221 | if (!path[depth].p_hdr) |
| 2120 | path[depth].p_hdr = ext_block_hdr(path[depth].p_bh); | 2222 | path[depth].p_hdr = ext_block_hdr(path[depth].p_bh); |
| 2121 | eh = path[depth].p_hdr; | 2223 | eh = path[depth].p_hdr; |
| 2122 | BUG_ON(eh == NULL); | 2224 | if (unlikely(path[depth].p_hdr == NULL)) { |
| 2123 | 2225 | EXT4_ERROR_INODE(inode, "path[%d].p_hdr == NULL", depth); | |
| 2226 | return -EIO; | ||
| 2227 | } | ||
| 2124 | /* find where to start removing */ | 2228 | /* find where to start removing */ |
| 2125 | ex = EXT_LAST_EXTENT(eh); | 2229 | ex = EXT_LAST_EXTENT(eh); |
| 2126 | 2230 | ||
| @@ -2983,7 +3087,7 @@ fix_extent_len: | |||
| 2983 | ext4_ext_dirty(handle, inode, path + depth); | 3087 | ext4_ext_dirty(handle, inode, path + depth); |
| 2984 | return err; | 3088 | return err; |
| 2985 | } | 3089 | } |
| 2986 | static int ext4_convert_unwritten_extents_dio(handle_t *handle, | 3090 | static int ext4_convert_unwritten_extents_endio(handle_t *handle, |
| 2987 | struct inode *inode, | 3091 | struct inode *inode, |
| 2988 | struct ext4_ext_path *path) | 3092 | struct ext4_ext_path *path) |
| 2989 | { | 3093 | { |
| @@ -3063,8 +3167,8 @@ ext4_ext_handle_uninitialized_extents(handle_t *handle, struct inode *inode, | |||
| 3063 | flags, allocated); | 3167 | flags, allocated); |
| 3064 | ext4_ext_show_leaf(inode, path); | 3168 | ext4_ext_show_leaf(inode, path); |
| 3065 | 3169 | ||
| 3066 | /* DIO get_block() before submit the IO, split the extent */ | 3170 | /* get_block() before submit the IO, split the extent */ |
| 3067 | if (flags == EXT4_GET_BLOCKS_DIO_CREATE_EXT) { | 3171 | if ((flags & EXT4_GET_BLOCKS_PRE_IO)) { |
| 3068 | ret = ext4_split_unwritten_extents(handle, | 3172 | ret = ext4_split_unwritten_extents(handle, |
| 3069 | inode, path, iblock, | 3173 | inode, path, iblock, |
| 3070 | max_blocks, flags); | 3174 | max_blocks, flags); |
| @@ -3074,14 +3178,16 @@ ext4_ext_handle_uninitialized_extents(handle_t *handle, struct inode *inode, | |||
| 3074 | * completed | 3178 | * completed |
| 3075 | */ | 3179 | */ |
| 3076 | if (io) | 3180 | if (io) |
| 3077 | io->flag = DIO_AIO_UNWRITTEN; | 3181 | io->flag = EXT4_IO_UNWRITTEN; |
| 3078 | else | 3182 | else |
| 3079 | EXT4_I(inode)->i_state |= EXT4_STATE_DIO_UNWRITTEN; | 3183 | ext4_set_inode_state(inode, EXT4_STATE_DIO_UNWRITTEN); |
| 3184 | if (ext4_should_dioread_nolock(inode)) | ||
| 3185 | set_buffer_uninit(bh_result); | ||
| 3080 | goto out; | 3186 | goto out; |
| 3081 | } | 3187 | } |
| 3082 | /* async DIO end_io complete, convert the filled extent to written */ | 3188 | /* IO end_io complete, convert the filled extent to written */ |
| 3083 | if (flags == EXT4_GET_BLOCKS_DIO_CONVERT_EXT) { | 3189 | if ((flags & EXT4_GET_BLOCKS_CONVERT)) { |
| 3084 | ret = ext4_convert_unwritten_extents_dio(handle, inode, | 3190 | ret = ext4_convert_unwritten_extents_endio(handle, inode, |
| 3085 | path); | 3191 | path); |
| 3086 | if (ret >= 0) | 3192 | if (ret >= 0) |
| 3087 | ext4_update_inode_fsync_trans(handle, inode, 1); | 3193 | ext4_update_inode_fsync_trans(handle, inode, 1); |
| @@ -3185,7 +3291,7 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
| 3185 | { | 3291 | { |
| 3186 | struct ext4_ext_path *path = NULL; | 3292 | struct ext4_ext_path *path = NULL; |
| 3187 | struct ext4_extent_header *eh; | 3293 | struct ext4_extent_header *eh; |
| 3188 | struct ext4_extent newex, *ex; | 3294 | struct ext4_extent newex, *ex, *last_ex; |
| 3189 | ext4_fsblk_t newblock; | 3295 | ext4_fsblk_t newblock; |
| 3190 | int err = 0, depth, ret, cache_type; | 3296 | int err = 0, depth, ret, cache_type; |
| 3191 | unsigned int allocated = 0; | 3297 | unsigned int allocated = 0; |
| @@ -3237,10 +3343,10 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
| 3237 | * this situation is possible, though, _during_ tree modification; | 3343 | * this situation is possible, though, _during_ tree modification; |
| 3238 | * this is why assert can't be put in ext4_ext_find_extent() | 3344 | * this is why assert can't be put in ext4_ext_find_extent() |
| 3239 | */ | 3345 | */ |
| 3240 | if (path[depth].p_ext == NULL && depth != 0) { | 3346 | if (unlikely(path[depth].p_ext == NULL && depth != 0)) { |
| 3241 | ext4_error(inode->i_sb, __func__, "bad extent address " | 3347 | EXT4_ERROR_INODE(inode, "bad extent address " |
| 3242 | "inode: %lu, iblock: %d, depth: %d", | 3348 | "iblock: %d, depth: %d pblock %lld", |
| 3243 | inode->i_ino, iblock, depth); | 3349 | iblock, depth, path[depth].p_block); |
| 3244 | err = -EIO; | 3350 | err = -EIO; |
| 3245 | goto out2; | 3351 | goto out2; |
| 3246 | } | 3352 | } |
| @@ -3258,7 +3364,7 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
| 3258 | */ | 3364 | */ |
| 3259 | ee_len = ext4_ext_get_actual_len(ex); | 3365 | ee_len = ext4_ext_get_actual_len(ex); |
| 3260 | /* if found extent covers block, simply return it */ | 3366 | /* if found extent covers block, simply return it */ |
| 3261 | if (iblock >= ee_block && iblock < ee_block + ee_len) { | 3367 | if (in_range(iblock, ee_block, ee_len)) { |
| 3262 | newblock = iblock - ee_block + ee_start; | 3368 | newblock = iblock - ee_block + ee_start; |
| 3263 | /* number of remaining blocks in the extent */ | 3369 | /* number of remaining blocks in the extent */ |
| 3264 | allocated = ee_len - (iblock - ee_block); | 3370 | allocated = ee_len - (iblock - ee_block); |
| @@ -3350,21 +3456,35 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
| 3350 | if (flags & EXT4_GET_BLOCKS_UNINIT_EXT){ | 3456 | if (flags & EXT4_GET_BLOCKS_UNINIT_EXT){ |
| 3351 | ext4_ext_mark_uninitialized(&newex); | 3457 | ext4_ext_mark_uninitialized(&newex); |
| 3352 | /* | 3458 | /* |
| 3353 | * io_end structure was created for every async | 3459 | * io_end structure was created for every IO write to an |
| 3354 | * direct IO write to the middle of the file. | 3460 | * uninitialized extent. To avoid unecessary conversion, |
| 3355 | * To avoid unecessary convertion for every aio dio rewrite | 3461 | * here we flag the IO that really needs the conversion. |
| 3356 | * to the mid of file, here we flag the IO that is really | ||
| 3357 | * need the convertion. | ||
| 3358 | * For non asycn direct IO case, flag the inode state | 3462 | * For non asycn direct IO case, flag the inode state |
| 3359 | * that we need to perform convertion when IO is done. | 3463 | * that we need to perform convertion when IO is done. |
| 3360 | */ | 3464 | */ |
| 3361 | if (flags == EXT4_GET_BLOCKS_DIO_CREATE_EXT) { | 3465 | if ((flags & EXT4_GET_BLOCKS_PRE_IO)) { |
| 3362 | if (io) | 3466 | if (io) |
| 3363 | io->flag = DIO_AIO_UNWRITTEN; | 3467 | io->flag = EXT4_IO_UNWRITTEN; |
| 3364 | else | 3468 | else |
| 3365 | EXT4_I(inode)->i_state |= | 3469 | ext4_set_inode_state(inode, |
| 3366 | EXT4_STATE_DIO_UNWRITTEN;; | 3470 | EXT4_STATE_DIO_UNWRITTEN); |
| 3471 | } | ||
| 3472 | if (ext4_should_dioread_nolock(inode)) | ||
| 3473 | set_buffer_uninit(bh_result); | ||
| 3474 | } | ||
| 3475 | |||
| 3476 | if (unlikely(EXT4_I(inode)->i_flags & EXT4_EOFBLOCKS_FL)) { | ||
| 3477 | if (unlikely(!eh->eh_entries)) { | ||
| 3478 | EXT4_ERROR_INODE(inode, | ||
| 3479 | "eh->eh_entries == 0 ee_block %d", | ||
| 3480 | ex->ee_block); | ||
| 3481 | err = -EIO; | ||
| 3482 | goto out2; | ||
| 3367 | } | 3483 | } |
| 3484 | last_ex = EXT_LAST_EXTENT(eh); | ||
| 3485 | if (iblock + ar.len > le32_to_cpu(last_ex->ee_block) | ||
| 3486 | + ext4_ext_get_actual_len(last_ex)) | ||
| 3487 | EXT4_I(inode)->i_flags &= ~EXT4_EOFBLOCKS_FL; | ||
| 3368 | } | 3488 | } |
| 3369 | err = ext4_ext_insert_extent(handle, inode, path, &newex, flags); | 3489 | err = ext4_ext_insert_extent(handle, inode, path, &newex, flags); |
| 3370 | if (err) { | 3490 | if (err) { |
| @@ -3499,6 +3619,13 @@ static void ext4_falloc_update_inode(struct inode *inode, | |||
| 3499 | i_size_write(inode, new_size); | 3619 | i_size_write(inode, new_size); |
| 3500 | if (new_size > EXT4_I(inode)->i_disksize) | 3620 | if (new_size > EXT4_I(inode)->i_disksize) |
| 3501 | ext4_update_i_disksize(inode, new_size); | 3621 | ext4_update_i_disksize(inode, new_size); |
| 3622 | } else { | ||
| 3623 | /* | ||
| 3624 | * Mark that we allocate beyond EOF so the subsequent truncate | ||
| 3625 | * can proceed even if the new size is the same as i_size. | ||
| 3626 | */ | ||
| 3627 | if (new_size > i_size_read(inode)) | ||
| 3628 | EXT4_I(inode)->i_flags |= EXT4_EOFBLOCKS_FL; | ||
| 3502 | } | 3629 | } |
| 3503 | 3630 | ||
| 3504 | } | 3631 | } |
| @@ -3603,7 +3730,7 @@ retry: | |||
| 3603 | * Returns 0 on success. | 3730 | * Returns 0 on success. |
| 3604 | */ | 3731 | */ |
| 3605 | int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset, | 3732 | int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset, |
| 3606 | loff_t len) | 3733 | ssize_t len) |
| 3607 | { | 3734 | { |
| 3608 | handle_t *handle; | 3735 | handle_t *handle; |
| 3609 | ext4_lblk_t block; | 3736 | ext4_lblk_t block; |
| @@ -3635,7 +3762,7 @@ int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset, | |||
| 3635 | map_bh.b_state = 0; | 3762 | map_bh.b_state = 0; |
| 3636 | ret = ext4_get_blocks(handle, inode, block, | 3763 | ret = ext4_get_blocks(handle, inode, block, |
| 3637 | max_blocks, &map_bh, | 3764 | max_blocks, &map_bh, |
| 3638 | EXT4_GET_BLOCKS_DIO_CONVERT_EXT); | 3765 | EXT4_GET_BLOCKS_IO_CONVERT_EXT); |
| 3639 | if (ret <= 0) { | 3766 | if (ret <= 0) { |
| 3640 | WARN_ON(ret <= 0); | 3767 | WARN_ON(ret <= 0); |
| 3641 | printk(KERN_ERR "%s: ext4_ext_get_blocks " | 3768 | printk(KERN_ERR "%s: ext4_ext_get_blocks " |
| @@ -3739,7 +3866,7 @@ static int ext4_xattr_fiemap(struct inode *inode, | |||
| 3739 | int error = 0; | 3866 | int error = 0; |
| 3740 | 3867 | ||
| 3741 | /* in-inode? */ | 3868 | /* in-inode? */ |
| 3742 | if (EXT4_I(inode)->i_state & EXT4_STATE_XATTR) { | 3869 | if (ext4_test_inode_state(inode, EXT4_STATE_XATTR)) { |
| 3743 | struct ext4_iloc iloc; | 3870 | struct ext4_iloc iloc; |
| 3744 | int offset; /* offset of xattr in inode */ | 3871 | int offset; /* offset of xattr in inode */ |
| 3745 | 3872 | ||
| @@ -3767,7 +3894,6 @@ int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
| 3767 | __u64 start, __u64 len) | 3894 | __u64 start, __u64 len) |
| 3768 | { | 3895 | { |
| 3769 | ext4_lblk_t start_blk; | 3896 | ext4_lblk_t start_blk; |
| 3770 | ext4_lblk_t len_blks; | ||
| 3771 | int error = 0; | 3897 | int error = 0; |
| 3772 | 3898 | ||
| 3773 | /* fallback to generic here if not in extents fmt */ | 3899 | /* fallback to generic here if not in extents fmt */ |
| @@ -3781,8 +3907,14 @@ int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
| 3781 | if (fieinfo->fi_flags & FIEMAP_FLAG_XATTR) { | 3907 | if (fieinfo->fi_flags & FIEMAP_FLAG_XATTR) { |
| 3782 | error = ext4_xattr_fiemap(inode, fieinfo); | 3908 | error = ext4_xattr_fiemap(inode, fieinfo); |
| 3783 | } else { | 3909 | } else { |
| 3910 | ext4_lblk_t len_blks; | ||
| 3911 | __u64 last_blk; | ||
| 3912 | |||
| 3784 | start_blk = start >> inode->i_sb->s_blocksize_bits; | 3913 | start_blk = start >> inode->i_sb->s_blocksize_bits; |
| 3785 | len_blks = len >> inode->i_sb->s_blocksize_bits; | 3914 | last_blk = (start + len - 1) >> inode->i_sb->s_blocksize_bits; |
| 3915 | if (last_blk >= EXT_MAX_BLOCK) | ||
| 3916 | last_blk = EXT_MAX_BLOCK-1; | ||
| 3917 | len_blks = ((ext4_lblk_t) last_blk) - start_blk + 1; | ||
| 3786 | 3918 | ||
| 3787 | /* | 3919 | /* |
| 3788 | * Walk the extent tree gathering extent information. | 3920 | * Walk the extent tree gathering extent information. |
