diff options
Diffstat (limited to 'fs/ext4/extents.c')
-rw-r--r-- | fs/ext4/extents.c | 260 |
1 files changed, 196 insertions, 64 deletions
diff --git a/fs/ext4/extents.c b/fs/ext4/extents.c index 765a4826b118..94c8ee81f5e1 100644 --- a/fs/ext4/extents.c +++ b/fs/ext4/extents.c | |||
@@ -195,8 +195,7 @@ static ext4_fsblk_t ext4_ext_find_goal(struct inode *inode, | |||
195 | if (S_ISREG(inode->i_mode)) | 195 | if (S_ISREG(inode->i_mode)) |
196 | block_group++; | 196 | block_group++; |
197 | } | 197 | } |
198 | bg_start = (block_group * EXT4_BLOCKS_PER_GROUP(inode->i_sb)) + | 198 | bg_start = ext4_group_first_block_no(inode->i_sb, block_group); |
199 | le32_to_cpu(EXT4_SB(inode->i_sb)->s_es->s_first_data_block); | ||
200 | last_block = ext4_blocks_count(EXT4_SB(inode->i_sb)->s_es) - 1; | 199 | last_block = ext4_blocks_count(EXT4_SB(inode->i_sb)->s_es) - 1; |
201 | 200 | ||
202 | /* | 201 | /* |
@@ -440,7 +439,7 @@ static int __ext4_ext_check(const char *function, struct inode *inode, | |||
440 | return 0; | 439 | return 0; |
441 | 440 | ||
442 | corrupted: | 441 | corrupted: |
443 | ext4_error(inode->i_sb, function, | 442 | __ext4_error(inode->i_sb, function, |
444 | "bad header/extent in inode #%lu: %s - magic %x, " | 443 | "bad header/extent in inode #%lu: %s - magic %x, " |
445 | "entries %u, max %u(%u), depth %u(%u)", | 444 | "entries %u, max %u(%u), depth %u(%u)", |
446 | inode->i_ino, error_msg, le16_to_cpu(eh->eh_magic), | 445 | inode->i_ino, error_msg, le16_to_cpu(eh->eh_magic), |
@@ -703,7 +702,12 @@ ext4_ext_find_extent(struct inode *inode, ext4_lblk_t block, | |||
703 | } | 702 | } |
704 | eh = ext_block_hdr(bh); | 703 | eh = ext_block_hdr(bh); |
705 | ppos++; | 704 | ppos++; |
706 | BUG_ON(ppos > depth); | 705 | if (unlikely(ppos > depth)) { |
706 | put_bh(bh); | ||
707 | EXT4_ERROR_INODE(inode, | ||
708 | "ppos %d > depth %d", ppos, depth); | ||
709 | goto err; | ||
710 | } | ||
707 | path[ppos].p_bh = bh; | 711 | path[ppos].p_bh = bh; |
708 | path[ppos].p_hdr = eh; | 712 | path[ppos].p_hdr = eh; |
709 | i--; | 713 | i--; |
@@ -749,7 +753,12 @@ int ext4_ext_insert_index(handle_t *handle, struct inode *inode, | |||
749 | if (err) | 753 | if (err) |
750 | return err; | 754 | return err; |
751 | 755 | ||
752 | BUG_ON(logical == le32_to_cpu(curp->p_idx->ei_block)); | 756 | if (unlikely(logical == le32_to_cpu(curp->p_idx->ei_block))) { |
757 | EXT4_ERROR_INODE(inode, | ||
758 | "logical %d == ei_block %d!", | ||
759 | logical, le32_to_cpu(curp->p_idx->ei_block)); | ||
760 | return -EIO; | ||
761 | } | ||
753 | len = EXT_MAX_INDEX(curp->p_hdr) - curp->p_idx; | 762 | len = EXT_MAX_INDEX(curp->p_hdr) - curp->p_idx; |
754 | if (logical > le32_to_cpu(curp->p_idx->ei_block)) { | 763 | if (logical > le32_to_cpu(curp->p_idx->ei_block)) { |
755 | /* insert after */ | 764 | /* insert after */ |
@@ -779,9 +788,17 @@ int ext4_ext_insert_index(handle_t *handle, struct inode *inode, | |||
779 | ext4_idx_store_pblock(ix, ptr); | 788 | ext4_idx_store_pblock(ix, ptr); |
780 | le16_add_cpu(&curp->p_hdr->eh_entries, 1); | 789 | le16_add_cpu(&curp->p_hdr->eh_entries, 1); |
781 | 790 | ||
782 | BUG_ON(le16_to_cpu(curp->p_hdr->eh_entries) | 791 | if (unlikely(le16_to_cpu(curp->p_hdr->eh_entries) |
783 | > le16_to_cpu(curp->p_hdr->eh_max)); | 792 | > le16_to_cpu(curp->p_hdr->eh_max))) { |
784 | BUG_ON(ix > EXT_LAST_INDEX(curp->p_hdr)); | 793 | EXT4_ERROR_INODE(inode, |
794 | "logical %d == ei_block %d!", | ||
795 | logical, le32_to_cpu(curp->p_idx->ei_block)); | ||
796 | return -EIO; | ||
797 | } | ||
798 | if (unlikely(ix > EXT_LAST_INDEX(curp->p_hdr))) { | ||
799 | EXT4_ERROR_INODE(inode, "ix > EXT_LAST_INDEX!"); | ||
800 | return -EIO; | ||
801 | } | ||
785 | 802 | ||
786 | err = ext4_ext_dirty(handle, inode, curp); | 803 | err = ext4_ext_dirty(handle, inode, curp); |
787 | ext4_std_error(inode->i_sb, err); | 804 | ext4_std_error(inode->i_sb, err); |
@@ -819,7 +836,10 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
819 | 836 | ||
820 | /* if current leaf will be split, then we should use | 837 | /* if current leaf will be split, then we should use |
821 | * border from split point */ | 838 | * border from split point */ |
822 | BUG_ON(path[depth].p_ext > EXT_MAX_EXTENT(path[depth].p_hdr)); | 839 | if (unlikely(path[depth].p_ext > EXT_MAX_EXTENT(path[depth].p_hdr))) { |
840 | EXT4_ERROR_INODE(inode, "p_ext > EXT_MAX_EXTENT!"); | ||
841 | return -EIO; | ||
842 | } | ||
823 | if (path[depth].p_ext != EXT_MAX_EXTENT(path[depth].p_hdr)) { | 843 | if (path[depth].p_ext != EXT_MAX_EXTENT(path[depth].p_hdr)) { |
824 | border = path[depth].p_ext[1].ee_block; | 844 | border = path[depth].p_ext[1].ee_block; |
825 | ext_debug("leaf will be split." | 845 | ext_debug("leaf will be split." |
@@ -860,7 +880,11 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
860 | 880 | ||
861 | /* initialize new leaf */ | 881 | /* initialize new leaf */ |
862 | newblock = ablocks[--a]; | 882 | newblock = ablocks[--a]; |
863 | BUG_ON(newblock == 0); | 883 | if (unlikely(newblock == 0)) { |
884 | EXT4_ERROR_INODE(inode, "newblock == 0!"); | ||
885 | err = -EIO; | ||
886 | goto cleanup; | ||
887 | } | ||
864 | bh = sb_getblk(inode->i_sb, newblock); | 888 | bh = sb_getblk(inode->i_sb, newblock); |
865 | if (!bh) { | 889 | if (!bh) { |
866 | err = -EIO; | 890 | err = -EIO; |
@@ -880,7 +904,14 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
880 | ex = EXT_FIRST_EXTENT(neh); | 904 | ex = EXT_FIRST_EXTENT(neh); |
881 | 905 | ||
882 | /* move remainder of path[depth] to the new leaf */ | 906 | /* move remainder of path[depth] to the new leaf */ |
883 | BUG_ON(path[depth].p_hdr->eh_entries != path[depth].p_hdr->eh_max); | 907 | if (unlikely(path[depth].p_hdr->eh_entries != |
908 | path[depth].p_hdr->eh_max)) { | ||
909 | EXT4_ERROR_INODE(inode, "eh_entries %d != eh_max %d!", | ||
910 | path[depth].p_hdr->eh_entries, | ||
911 | path[depth].p_hdr->eh_max); | ||
912 | err = -EIO; | ||
913 | goto cleanup; | ||
914 | } | ||
884 | /* start copy from next extent */ | 915 | /* start copy from next extent */ |
885 | /* TODO: we could do it by single memmove */ | 916 | /* TODO: we could do it by single memmove */ |
886 | m = 0; | 917 | m = 0; |
@@ -927,7 +958,11 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
927 | 958 | ||
928 | /* create intermediate indexes */ | 959 | /* create intermediate indexes */ |
929 | k = depth - at - 1; | 960 | k = depth - at - 1; |
930 | BUG_ON(k < 0); | 961 | if (unlikely(k < 0)) { |
962 | EXT4_ERROR_INODE(inode, "k %d < 0!", k); | ||
963 | err = -EIO; | ||
964 | goto cleanup; | ||
965 | } | ||
931 | if (k) | 966 | if (k) |
932 | ext_debug("create %d intermediate indices\n", k); | 967 | ext_debug("create %d intermediate indices\n", k); |
933 | /* insert new index into current index block */ | 968 | /* insert new index into current index block */ |
@@ -964,8 +999,14 @@ static int ext4_ext_split(handle_t *handle, struct inode *inode, | |||
964 | 999 | ||
965 | ext_debug("cur 0x%p, last 0x%p\n", path[i].p_idx, | 1000 | ext_debug("cur 0x%p, last 0x%p\n", path[i].p_idx, |
966 | EXT_MAX_INDEX(path[i].p_hdr)); | 1001 | EXT_MAX_INDEX(path[i].p_hdr)); |
967 | BUG_ON(EXT_MAX_INDEX(path[i].p_hdr) != | 1002 | if (unlikely(EXT_MAX_INDEX(path[i].p_hdr) != |
968 | EXT_LAST_INDEX(path[i].p_hdr)); | 1003 | EXT_LAST_INDEX(path[i].p_hdr))) { |
1004 | EXT4_ERROR_INODE(inode, | ||
1005 | "EXT_MAX_INDEX != EXT_LAST_INDEX ee_block %d!", | ||
1006 | le32_to_cpu(path[i].p_ext->ee_block)); | ||
1007 | err = -EIO; | ||
1008 | goto cleanup; | ||
1009 | } | ||
969 | while (path[i].p_idx <= EXT_MAX_INDEX(path[i].p_hdr)) { | 1010 | while (path[i].p_idx <= EXT_MAX_INDEX(path[i].p_hdr)) { |
970 | ext_debug("%d: move %d:%llu in new index %llu\n", i, | 1011 | ext_debug("%d: move %d:%llu in new index %llu\n", i, |
971 | le32_to_cpu(path[i].p_idx->ei_block), | 1012 | le32_to_cpu(path[i].p_idx->ei_block), |
@@ -1203,7 +1244,10 @@ ext4_ext_search_left(struct inode *inode, struct ext4_ext_path *path, | |||
1203 | struct ext4_extent *ex; | 1244 | struct ext4_extent *ex; |
1204 | int depth, ee_len; | 1245 | int depth, ee_len; |
1205 | 1246 | ||
1206 | BUG_ON(path == NULL); | 1247 | if (unlikely(path == NULL)) { |
1248 | EXT4_ERROR_INODE(inode, "path == NULL *logical %d!", *logical); | ||
1249 | return -EIO; | ||
1250 | } | ||
1207 | depth = path->p_depth; | 1251 | depth = path->p_depth; |
1208 | *phys = 0; | 1252 | *phys = 0; |
1209 | 1253 | ||
@@ -1217,15 +1261,33 @@ ext4_ext_search_left(struct inode *inode, struct ext4_ext_path *path, | |||
1217 | ex = path[depth].p_ext; | 1261 | ex = path[depth].p_ext; |
1218 | ee_len = ext4_ext_get_actual_len(ex); | 1262 | ee_len = ext4_ext_get_actual_len(ex); |
1219 | if (*logical < le32_to_cpu(ex->ee_block)) { | 1263 | if (*logical < le32_to_cpu(ex->ee_block)) { |
1220 | BUG_ON(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex); | 1264 | if (unlikely(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex)) { |
1265 | EXT4_ERROR_INODE(inode, | ||
1266 | "EXT_FIRST_EXTENT != ex *logical %d ee_block %d!", | ||
1267 | *logical, le32_to_cpu(ex->ee_block)); | ||
1268 | return -EIO; | ||
1269 | } | ||
1221 | while (--depth >= 0) { | 1270 | while (--depth >= 0) { |
1222 | ix = path[depth].p_idx; | 1271 | ix = path[depth].p_idx; |
1223 | BUG_ON(ix != EXT_FIRST_INDEX(path[depth].p_hdr)); | 1272 | if (unlikely(ix != EXT_FIRST_INDEX(path[depth].p_hdr))) { |
1273 | EXT4_ERROR_INODE(inode, | ||
1274 | "ix (%d) != EXT_FIRST_INDEX (%d) (depth %d)!", | ||
1275 | ix != NULL ? ix->ei_block : 0, | ||
1276 | EXT_FIRST_INDEX(path[depth].p_hdr) != NULL ? | ||
1277 | EXT_FIRST_INDEX(path[depth].p_hdr)->ei_block : 0, | ||
1278 | depth); | ||
1279 | return -EIO; | ||
1280 | } | ||
1224 | } | 1281 | } |
1225 | return 0; | 1282 | return 0; |
1226 | } | 1283 | } |
1227 | 1284 | ||
1228 | BUG_ON(*logical < (le32_to_cpu(ex->ee_block) + ee_len)); | 1285 | if (unlikely(*logical < (le32_to_cpu(ex->ee_block) + ee_len))) { |
1286 | EXT4_ERROR_INODE(inode, | ||
1287 | "logical %d < ee_block %d + ee_len %d!", | ||
1288 | *logical, le32_to_cpu(ex->ee_block), ee_len); | ||
1289 | return -EIO; | ||
1290 | } | ||
1229 | 1291 | ||
1230 | *logical = le32_to_cpu(ex->ee_block) + ee_len - 1; | 1292 | *logical = le32_to_cpu(ex->ee_block) + ee_len - 1; |
1231 | *phys = ext_pblock(ex) + ee_len - 1; | 1293 | *phys = ext_pblock(ex) + ee_len - 1; |
@@ -1251,7 +1313,10 @@ ext4_ext_search_right(struct inode *inode, struct ext4_ext_path *path, | |||
1251 | int depth; /* Note, NOT eh_depth; depth from top of tree */ | 1313 | int depth; /* Note, NOT eh_depth; depth from top of tree */ |
1252 | int ee_len; | 1314 | int ee_len; |
1253 | 1315 | ||
1254 | BUG_ON(path == NULL); | 1316 | if (unlikely(path == NULL)) { |
1317 | EXT4_ERROR_INODE(inode, "path == NULL *logical %d!", *logical); | ||
1318 | return -EIO; | ||
1319 | } | ||
1255 | depth = path->p_depth; | 1320 | depth = path->p_depth; |
1256 | *phys = 0; | 1321 | *phys = 0; |
1257 | 1322 | ||
@@ -1265,17 +1330,32 @@ ext4_ext_search_right(struct inode *inode, struct ext4_ext_path *path, | |||
1265 | ex = path[depth].p_ext; | 1330 | ex = path[depth].p_ext; |
1266 | ee_len = ext4_ext_get_actual_len(ex); | 1331 | ee_len = ext4_ext_get_actual_len(ex); |
1267 | if (*logical < le32_to_cpu(ex->ee_block)) { | 1332 | if (*logical < le32_to_cpu(ex->ee_block)) { |
1268 | BUG_ON(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex); | 1333 | if (unlikely(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex)) { |
1334 | EXT4_ERROR_INODE(inode, | ||
1335 | "first_extent(path[%d].p_hdr) != ex", | ||
1336 | depth); | ||
1337 | return -EIO; | ||
1338 | } | ||
1269 | while (--depth >= 0) { | 1339 | while (--depth >= 0) { |
1270 | ix = path[depth].p_idx; | 1340 | ix = path[depth].p_idx; |
1271 | BUG_ON(ix != EXT_FIRST_INDEX(path[depth].p_hdr)); | 1341 | if (unlikely(ix != EXT_FIRST_INDEX(path[depth].p_hdr))) { |
1342 | EXT4_ERROR_INODE(inode, | ||
1343 | "ix != EXT_FIRST_INDEX *logical %d!", | ||
1344 | *logical); | ||
1345 | return -EIO; | ||
1346 | } | ||
1272 | } | 1347 | } |
1273 | *logical = le32_to_cpu(ex->ee_block); | 1348 | *logical = le32_to_cpu(ex->ee_block); |
1274 | *phys = ext_pblock(ex); | 1349 | *phys = ext_pblock(ex); |
1275 | return 0; | 1350 | return 0; |
1276 | } | 1351 | } |
1277 | 1352 | ||
1278 | BUG_ON(*logical < (le32_to_cpu(ex->ee_block) + ee_len)); | 1353 | if (unlikely(*logical < (le32_to_cpu(ex->ee_block) + ee_len))) { |
1354 | EXT4_ERROR_INODE(inode, | ||
1355 | "logical %d < ee_block %d + ee_len %d!", | ||
1356 | *logical, le32_to_cpu(ex->ee_block), ee_len); | ||
1357 | return -EIO; | ||
1358 | } | ||
1279 | 1359 | ||
1280 | if (ex != EXT_LAST_EXTENT(path[depth].p_hdr)) { | 1360 | if (ex != EXT_LAST_EXTENT(path[depth].p_hdr)) { |
1281 | /* next allocated block in this leaf */ | 1361 | /* next allocated block in this leaf */ |
@@ -1414,8 +1494,12 @@ static int ext4_ext_correct_indexes(handle_t *handle, struct inode *inode, | |||
1414 | 1494 | ||
1415 | eh = path[depth].p_hdr; | 1495 | eh = path[depth].p_hdr; |
1416 | ex = path[depth].p_ext; | 1496 | ex = path[depth].p_ext; |
1417 | BUG_ON(ex == NULL); | 1497 | |
1418 | BUG_ON(eh == NULL); | 1498 | if (unlikely(ex == NULL || eh == NULL)) { |
1499 | EXT4_ERROR_INODE(inode, | ||
1500 | "ex %p == NULL or eh %p == NULL", ex, eh); | ||
1501 | return -EIO; | ||
1502 | } | ||
1419 | 1503 | ||
1420 | if (depth == 0) { | 1504 | if (depth == 0) { |
1421 | /* there is no tree at all */ | 1505 | /* there is no tree at all */ |
@@ -1538,8 +1622,9 @@ int ext4_ext_try_to_merge(struct inode *inode, | |||
1538 | merge_done = 1; | 1622 | merge_done = 1; |
1539 | WARN_ON(eh->eh_entries == 0); | 1623 | WARN_ON(eh->eh_entries == 0); |
1540 | if (!eh->eh_entries) | 1624 | if (!eh->eh_entries) |
1541 | ext4_error(inode->i_sb, "ext4_ext_try_to_merge", | 1625 | ext4_error(inode->i_sb, |
1542 | "inode#%lu, eh->eh_entries = 0!", inode->i_ino); | 1626 | "inode#%lu, eh->eh_entries = 0!", |
1627 | inode->i_ino); | ||
1543 | } | 1628 | } |
1544 | 1629 | ||
1545 | return merge_done; | 1630 | return merge_done; |
@@ -1612,13 +1697,19 @@ int ext4_ext_insert_extent(handle_t *handle, struct inode *inode, | |||
1612 | ext4_lblk_t next; | 1697 | ext4_lblk_t next; |
1613 | unsigned uninitialized = 0; | 1698 | unsigned uninitialized = 0; |
1614 | 1699 | ||
1615 | BUG_ON(ext4_ext_get_actual_len(newext) == 0); | 1700 | if (unlikely(ext4_ext_get_actual_len(newext) == 0)) { |
1701 | EXT4_ERROR_INODE(inode, "ext4_ext_get_actual_len(newext) == 0"); | ||
1702 | return -EIO; | ||
1703 | } | ||
1616 | depth = ext_depth(inode); | 1704 | depth = ext_depth(inode); |
1617 | ex = path[depth].p_ext; | 1705 | ex = path[depth].p_ext; |
1618 | BUG_ON(path[depth].p_hdr == NULL); | 1706 | if (unlikely(path[depth].p_hdr == NULL)) { |
1707 | EXT4_ERROR_INODE(inode, "path[%d].p_hdr == NULL", depth); | ||
1708 | return -EIO; | ||
1709 | } | ||
1619 | 1710 | ||
1620 | /* try to insert block into found extent and return */ | 1711 | /* try to insert block into found extent and return */ |
1621 | if (ex && (flag != EXT4_GET_BLOCKS_DIO_CREATE_EXT) | 1712 | if (ex && !(flag & EXT4_GET_BLOCKS_PRE_IO) |
1622 | && ext4_can_extents_be_merged(inode, ex, newext)) { | 1713 | && ext4_can_extents_be_merged(inode, ex, newext)) { |
1623 | ext_debug("append [%d]%d block to %d:[%d]%d (from %llu)\n", | 1714 | ext_debug("append [%d]%d block to %d:[%d]%d (from %llu)\n", |
1624 | ext4_ext_is_uninitialized(newext), | 1715 | ext4_ext_is_uninitialized(newext), |
@@ -1739,7 +1830,7 @@ has_space: | |||
1739 | 1830 | ||
1740 | merge: | 1831 | merge: |
1741 | /* try to merge extents to the right */ | 1832 | /* try to merge extents to the right */ |
1742 | if (flag != EXT4_GET_BLOCKS_DIO_CREATE_EXT) | 1833 | if (!(flag & EXT4_GET_BLOCKS_PRE_IO)) |
1743 | ext4_ext_try_to_merge(inode, path, nearex); | 1834 | ext4_ext_try_to_merge(inode, path, nearex); |
1744 | 1835 | ||
1745 | /* try to merge extents to the left */ | 1836 | /* try to merge extents to the left */ |
@@ -1787,7 +1878,11 @@ int ext4_ext_walk_space(struct inode *inode, ext4_lblk_t block, | |||
1787 | } | 1878 | } |
1788 | 1879 | ||
1789 | depth = ext_depth(inode); | 1880 | depth = ext_depth(inode); |
1790 | BUG_ON(path[depth].p_hdr == NULL); | 1881 | if (unlikely(path[depth].p_hdr == NULL)) { |
1882 | EXT4_ERROR_INODE(inode, "path[%d].p_hdr == NULL", depth); | ||
1883 | err = -EIO; | ||
1884 | break; | ||
1885 | } | ||
1791 | ex = path[depth].p_ext; | 1886 | ex = path[depth].p_ext; |
1792 | next = ext4_ext_next_allocated_block(path); | 1887 | next = ext4_ext_next_allocated_block(path); |
1793 | 1888 | ||
@@ -1838,7 +1933,11 @@ int ext4_ext_walk_space(struct inode *inode, ext4_lblk_t block, | |||
1838 | cbex.ec_type = EXT4_EXT_CACHE_EXTENT; | 1933 | cbex.ec_type = EXT4_EXT_CACHE_EXTENT; |
1839 | } | 1934 | } |
1840 | 1935 | ||
1841 | BUG_ON(cbex.ec_len == 0); | 1936 | if (unlikely(cbex.ec_len == 0)) { |
1937 | EXT4_ERROR_INODE(inode, "cbex.ec_len == 0"); | ||
1938 | err = -EIO; | ||
1939 | break; | ||
1940 | } | ||
1842 | err = func(inode, path, &cbex, ex, cbdata); | 1941 | err = func(inode, path, &cbex, ex, cbdata); |
1843 | ext4_ext_drop_refs(path); | 1942 | ext4_ext_drop_refs(path); |
1844 | 1943 | ||
@@ -1952,7 +2051,7 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block, | |||
1952 | 2051 | ||
1953 | BUG_ON(cex->ec_type != EXT4_EXT_CACHE_GAP && | 2052 | BUG_ON(cex->ec_type != EXT4_EXT_CACHE_GAP && |
1954 | cex->ec_type != EXT4_EXT_CACHE_EXTENT); | 2053 | cex->ec_type != EXT4_EXT_CACHE_EXTENT); |
1955 | if (block >= cex->ec_block && block < cex->ec_block + cex->ec_len) { | 2054 | if (in_range(block, cex->ec_block, cex->ec_len)) { |
1956 | ex->ee_block = cpu_to_le32(cex->ec_block); | 2055 | ex->ee_block = cpu_to_le32(cex->ec_block); |
1957 | ext4_ext_store_pblock(ex, cex->ec_start); | 2056 | ext4_ext_store_pblock(ex, cex->ec_start); |
1958 | ex->ee_len = cpu_to_le16(cex->ec_len); | 2057 | ex->ee_len = cpu_to_le16(cex->ec_len); |
@@ -1981,7 +2080,10 @@ static int ext4_ext_rm_idx(handle_t *handle, struct inode *inode, | |||
1981 | /* free index block */ | 2080 | /* free index block */ |
1982 | path--; | 2081 | path--; |
1983 | leaf = idx_pblock(path->p_idx); | 2082 | leaf = idx_pblock(path->p_idx); |
1984 | BUG_ON(path->p_hdr->eh_entries == 0); | 2083 | if (unlikely(path->p_hdr->eh_entries == 0)) { |
2084 | EXT4_ERROR_INODE(inode, "path->p_hdr->eh_entries == 0"); | ||
2085 | return -EIO; | ||
2086 | } | ||
1985 | err = ext4_ext_get_access(handle, inode, path); | 2087 | err = ext4_ext_get_access(handle, inode, path); |
1986 | if (err) | 2088 | if (err) |
1987 | return err; | 2089 | return err; |
@@ -2119,8 +2221,10 @@ ext4_ext_rm_leaf(handle_t *handle, struct inode *inode, | |||
2119 | if (!path[depth].p_hdr) | 2221 | if (!path[depth].p_hdr) |
2120 | path[depth].p_hdr = ext_block_hdr(path[depth].p_bh); | 2222 | path[depth].p_hdr = ext_block_hdr(path[depth].p_bh); |
2121 | eh = path[depth].p_hdr; | 2223 | eh = path[depth].p_hdr; |
2122 | BUG_ON(eh == NULL); | 2224 | if (unlikely(path[depth].p_hdr == NULL)) { |
2123 | 2225 | EXT4_ERROR_INODE(inode, "path[%d].p_hdr == NULL", depth); | |
2226 | return -EIO; | ||
2227 | } | ||
2124 | /* find where to start removing */ | 2228 | /* find where to start removing */ |
2125 | ex = EXT_LAST_EXTENT(eh); | 2229 | ex = EXT_LAST_EXTENT(eh); |
2126 | 2230 | ||
@@ -2983,7 +3087,7 @@ fix_extent_len: | |||
2983 | ext4_ext_dirty(handle, inode, path + depth); | 3087 | ext4_ext_dirty(handle, inode, path + depth); |
2984 | return err; | 3088 | return err; |
2985 | } | 3089 | } |
2986 | static int ext4_convert_unwritten_extents_dio(handle_t *handle, | 3090 | static int ext4_convert_unwritten_extents_endio(handle_t *handle, |
2987 | struct inode *inode, | 3091 | struct inode *inode, |
2988 | struct ext4_ext_path *path) | 3092 | struct ext4_ext_path *path) |
2989 | { | 3093 | { |
@@ -3063,8 +3167,8 @@ ext4_ext_handle_uninitialized_extents(handle_t *handle, struct inode *inode, | |||
3063 | flags, allocated); | 3167 | flags, allocated); |
3064 | ext4_ext_show_leaf(inode, path); | 3168 | ext4_ext_show_leaf(inode, path); |
3065 | 3169 | ||
3066 | /* DIO get_block() before submit the IO, split the extent */ | 3170 | /* get_block() before submit the IO, split the extent */ |
3067 | if (flags == EXT4_GET_BLOCKS_DIO_CREATE_EXT) { | 3171 | if ((flags & EXT4_GET_BLOCKS_PRE_IO)) { |
3068 | ret = ext4_split_unwritten_extents(handle, | 3172 | ret = ext4_split_unwritten_extents(handle, |
3069 | inode, path, iblock, | 3173 | inode, path, iblock, |
3070 | max_blocks, flags); | 3174 | max_blocks, flags); |
@@ -3074,14 +3178,16 @@ ext4_ext_handle_uninitialized_extents(handle_t *handle, struct inode *inode, | |||
3074 | * completed | 3178 | * completed |
3075 | */ | 3179 | */ |
3076 | if (io) | 3180 | if (io) |
3077 | io->flag = DIO_AIO_UNWRITTEN; | 3181 | io->flag = EXT4_IO_UNWRITTEN; |
3078 | else | 3182 | else |
3079 | EXT4_I(inode)->i_state |= EXT4_STATE_DIO_UNWRITTEN; | 3183 | ext4_set_inode_state(inode, EXT4_STATE_DIO_UNWRITTEN); |
3184 | if (ext4_should_dioread_nolock(inode)) | ||
3185 | set_buffer_uninit(bh_result); | ||
3080 | goto out; | 3186 | goto out; |
3081 | } | 3187 | } |
3082 | /* async DIO end_io complete, convert the filled extent to written */ | 3188 | /* IO end_io complete, convert the filled extent to written */ |
3083 | if (flags == EXT4_GET_BLOCKS_DIO_CONVERT_EXT) { | 3189 | if ((flags & EXT4_GET_BLOCKS_CONVERT)) { |
3084 | ret = ext4_convert_unwritten_extents_dio(handle, inode, | 3190 | ret = ext4_convert_unwritten_extents_endio(handle, inode, |
3085 | path); | 3191 | path); |
3086 | if (ret >= 0) | 3192 | if (ret >= 0) |
3087 | ext4_update_inode_fsync_trans(handle, inode, 1); | 3193 | ext4_update_inode_fsync_trans(handle, inode, 1); |
@@ -3185,7 +3291,7 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
3185 | { | 3291 | { |
3186 | struct ext4_ext_path *path = NULL; | 3292 | struct ext4_ext_path *path = NULL; |
3187 | struct ext4_extent_header *eh; | 3293 | struct ext4_extent_header *eh; |
3188 | struct ext4_extent newex, *ex; | 3294 | struct ext4_extent newex, *ex, *last_ex; |
3189 | ext4_fsblk_t newblock; | 3295 | ext4_fsblk_t newblock; |
3190 | int err = 0, depth, ret, cache_type; | 3296 | int err = 0, depth, ret, cache_type; |
3191 | unsigned int allocated = 0; | 3297 | unsigned int allocated = 0; |
@@ -3237,10 +3343,10 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
3237 | * this situation is possible, though, _during_ tree modification; | 3343 | * this situation is possible, though, _during_ tree modification; |
3238 | * this is why assert can't be put in ext4_ext_find_extent() | 3344 | * this is why assert can't be put in ext4_ext_find_extent() |
3239 | */ | 3345 | */ |
3240 | if (path[depth].p_ext == NULL && depth != 0) { | 3346 | if (unlikely(path[depth].p_ext == NULL && depth != 0)) { |
3241 | ext4_error(inode->i_sb, __func__, "bad extent address " | 3347 | EXT4_ERROR_INODE(inode, "bad extent address " |
3242 | "inode: %lu, iblock: %d, depth: %d", | 3348 | "iblock: %d, depth: %d pblock %lld", |
3243 | inode->i_ino, iblock, depth); | 3349 | iblock, depth, path[depth].p_block); |
3244 | err = -EIO; | 3350 | err = -EIO; |
3245 | goto out2; | 3351 | goto out2; |
3246 | } | 3352 | } |
@@ -3258,7 +3364,7 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
3258 | */ | 3364 | */ |
3259 | ee_len = ext4_ext_get_actual_len(ex); | 3365 | ee_len = ext4_ext_get_actual_len(ex); |
3260 | /* if found extent covers block, simply return it */ | 3366 | /* if found extent covers block, simply return it */ |
3261 | if (iblock >= ee_block && iblock < ee_block + ee_len) { | 3367 | if (in_range(iblock, ee_block, ee_len)) { |
3262 | newblock = iblock - ee_block + ee_start; | 3368 | newblock = iblock - ee_block + ee_start; |
3263 | /* number of remaining blocks in the extent */ | 3369 | /* number of remaining blocks in the extent */ |
3264 | allocated = ee_len - (iblock - ee_block); | 3370 | allocated = ee_len - (iblock - ee_block); |
@@ -3350,21 +3456,35 @@ int ext4_ext_get_blocks(handle_t *handle, struct inode *inode, | |||
3350 | if (flags & EXT4_GET_BLOCKS_UNINIT_EXT){ | 3456 | if (flags & EXT4_GET_BLOCKS_UNINIT_EXT){ |
3351 | ext4_ext_mark_uninitialized(&newex); | 3457 | ext4_ext_mark_uninitialized(&newex); |
3352 | /* | 3458 | /* |
3353 | * io_end structure was created for every async | 3459 | * io_end structure was created for every IO write to an |
3354 | * direct IO write to the middle of the file. | 3460 | * uninitialized extent. To avoid unecessary conversion, |
3355 | * To avoid unecessary convertion for every aio dio rewrite | 3461 | * here we flag the IO that really needs the conversion. |
3356 | * to the mid of file, here we flag the IO that is really | ||
3357 | * need the convertion. | ||
3358 | * For non asycn direct IO case, flag the inode state | 3462 | * For non asycn direct IO case, flag the inode state |
3359 | * that we need to perform convertion when IO is done. | 3463 | * that we need to perform convertion when IO is done. |
3360 | */ | 3464 | */ |
3361 | if (flags == EXT4_GET_BLOCKS_DIO_CREATE_EXT) { | 3465 | if ((flags & EXT4_GET_BLOCKS_PRE_IO)) { |
3362 | if (io) | 3466 | if (io) |
3363 | io->flag = DIO_AIO_UNWRITTEN; | 3467 | io->flag = EXT4_IO_UNWRITTEN; |
3364 | else | 3468 | else |
3365 | EXT4_I(inode)->i_state |= | 3469 | ext4_set_inode_state(inode, |
3366 | EXT4_STATE_DIO_UNWRITTEN;; | 3470 | EXT4_STATE_DIO_UNWRITTEN); |
3471 | } | ||
3472 | if (ext4_should_dioread_nolock(inode)) | ||
3473 | set_buffer_uninit(bh_result); | ||
3474 | } | ||
3475 | |||
3476 | if (unlikely(EXT4_I(inode)->i_flags & EXT4_EOFBLOCKS_FL)) { | ||
3477 | if (unlikely(!eh->eh_entries)) { | ||
3478 | EXT4_ERROR_INODE(inode, | ||
3479 | "eh->eh_entries == 0 ee_block %d", | ||
3480 | ex->ee_block); | ||
3481 | err = -EIO; | ||
3482 | goto out2; | ||
3367 | } | 3483 | } |
3484 | last_ex = EXT_LAST_EXTENT(eh); | ||
3485 | if (iblock + ar.len > le32_to_cpu(last_ex->ee_block) | ||
3486 | + ext4_ext_get_actual_len(last_ex)) | ||
3487 | EXT4_I(inode)->i_flags &= ~EXT4_EOFBLOCKS_FL; | ||
3368 | } | 3488 | } |
3369 | err = ext4_ext_insert_extent(handle, inode, path, &newex, flags); | 3489 | err = ext4_ext_insert_extent(handle, inode, path, &newex, flags); |
3370 | if (err) { | 3490 | if (err) { |
@@ -3499,6 +3619,13 @@ static void ext4_falloc_update_inode(struct inode *inode, | |||
3499 | i_size_write(inode, new_size); | 3619 | i_size_write(inode, new_size); |
3500 | if (new_size > EXT4_I(inode)->i_disksize) | 3620 | if (new_size > EXT4_I(inode)->i_disksize) |
3501 | ext4_update_i_disksize(inode, new_size); | 3621 | ext4_update_i_disksize(inode, new_size); |
3622 | } else { | ||
3623 | /* | ||
3624 | * Mark that we allocate beyond EOF so the subsequent truncate | ||
3625 | * can proceed even if the new size is the same as i_size. | ||
3626 | */ | ||
3627 | if (new_size > i_size_read(inode)) | ||
3628 | EXT4_I(inode)->i_flags |= EXT4_EOFBLOCKS_FL; | ||
3502 | } | 3629 | } |
3503 | 3630 | ||
3504 | } | 3631 | } |
@@ -3603,7 +3730,7 @@ retry: | |||
3603 | * Returns 0 on success. | 3730 | * Returns 0 on success. |
3604 | */ | 3731 | */ |
3605 | int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset, | 3732 | int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset, |
3606 | loff_t len) | 3733 | ssize_t len) |
3607 | { | 3734 | { |
3608 | handle_t *handle; | 3735 | handle_t *handle; |
3609 | ext4_lblk_t block; | 3736 | ext4_lblk_t block; |
@@ -3635,7 +3762,7 @@ int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset, | |||
3635 | map_bh.b_state = 0; | 3762 | map_bh.b_state = 0; |
3636 | ret = ext4_get_blocks(handle, inode, block, | 3763 | ret = ext4_get_blocks(handle, inode, block, |
3637 | max_blocks, &map_bh, | 3764 | max_blocks, &map_bh, |
3638 | EXT4_GET_BLOCKS_DIO_CONVERT_EXT); | 3765 | EXT4_GET_BLOCKS_IO_CONVERT_EXT); |
3639 | if (ret <= 0) { | 3766 | if (ret <= 0) { |
3640 | WARN_ON(ret <= 0); | 3767 | WARN_ON(ret <= 0); |
3641 | printk(KERN_ERR "%s: ext4_ext_get_blocks " | 3768 | printk(KERN_ERR "%s: ext4_ext_get_blocks " |
@@ -3739,7 +3866,7 @@ static int ext4_xattr_fiemap(struct inode *inode, | |||
3739 | int error = 0; | 3866 | int error = 0; |
3740 | 3867 | ||
3741 | /* in-inode? */ | 3868 | /* in-inode? */ |
3742 | if (EXT4_I(inode)->i_state & EXT4_STATE_XATTR) { | 3869 | if (ext4_test_inode_state(inode, EXT4_STATE_XATTR)) { |
3743 | struct ext4_iloc iloc; | 3870 | struct ext4_iloc iloc; |
3744 | int offset; /* offset of xattr in inode */ | 3871 | int offset; /* offset of xattr in inode */ |
3745 | 3872 | ||
@@ -3767,7 +3894,6 @@ int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
3767 | __u64 start, __u64 len) | 3894 | __u64 start, __u64 len) |
3768 | { | 3895 | { |
3769 | ext4_lblk_t start_blk; | 3896 | ext4_lblk_t start_blk; |
3770 | ext4_lblk_t len_blks; | ||
3771 | int error = 0; | 3897 | int error = 0; |
3772 | 3898 | ||
3773 | /* fallback to generic here if not in extents fmt */ | 3899 | /* fallback to generic here if not in extents fmt */ |
@@ -3781,8 +3907,14 @@ int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
3781 | if (fieinfo->fi_flags & FIEMAP_FLAG_XATTR) { | 3907 | if (fieinfo->fi_flags & FIEMAP_FLAG_XATTR) { |
3782 | error = ext4_xattr_fiemap(inode, fieinfo); | 3908 | error = ext4_xattr_fiemap(inode, fieinfo); |
3783 | } else { | 3909 | } else { |
3910 | ext4_lblk_t len_blks; | ||
3911 | __u64 last_blk; | ||
3912 | |||
3784 | start_blk = start >> inode->i_sb->s_blocksize_bits; | 3913 | start_blk = start >> inode->i_sb->s_blocksize_bits; |
3785 | len_blks = len >> inode->i_sb->s_blocksize_bits; | 3914 | last_blk = (start + len - 1) >> inode->i_sb->s_blocksize_bits; |
3915 | if (last_blk >= EXT_MAX_BLOCK) | ||
3916 | last_blk = EXT_MAX_BLOCK-1; | ||
3917 | len_blks = ((ext4_lblk_t) last_blk) - start_blk + 1; | ||
3786 | 3918 | ||
3787 | /* | 3919 | /* |
3788 | * Walk the extent tree gathering extent information. | 3920 | * Walk the extent tree gathering extent information. |