diff options
author | Chris Mason <chris.mason@oracle.com> | 2009-09-02 15:04:12 -0400 |
---|---|---|
committer | Chris Mason <chris.mason@oracle.com> | 2009-09-11 13:31:06 -0400 |
commit | 2c64c53d8d30d43d0670482503a3914dfd3d6d46 (patch) | |
tree | cfe7bfa8f5dc4078f8c995f305229622e25572b1 /fs/btrfs | |
parent | 1edbb734b4e010974c41d2859d22a43d04f5f1cf (diff) |
Btrfs: cache values for locking extents
Many of the btrfs extent state tree users follow the same pattern.
They lock an extent range in the tree, do some operation and then
unlock.
This translates to at least 2 rbtree searches, and maybe more if they
are doing operations on the extent state tree. A locked extent
in the tree isn't going to be merged or changed, and so we can
safely return the extent state structure as a cached handle.
This changes set_extent_bit to give back a cached handle, and also
changes both set_extent_bit and clear_extent_bit to use the cached
handle if it is available.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
Diffstat (limited to 'fs/btrfs')
-rw-r--r-- | fs/btrfs/extent_io.c | 125 | ||||
-rw-r--r-- | fs/btrfs/extent_io.h | 5 | ||||
-rw-r--r-- | fs/btrfs/inode.c | 6 |
3 files changed, 100 insertions, 36 deletions
diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c index 7c70613eb72c..c7a5e860fe21 100644 --- a/fs/btrfs/extent_io.c +++ b/fs/btrfs/extent_io.c | |||
@@ -471,10 +471,14 @@ static int clear_state_bit(struct extent_io_tree *tree, | |||
471 | * bits were already set, or zero if none of the bits were already set. | 471 | * bits were already set, or zero if none of the bits were already set. |
472 | */ | 472 | */ |
473 | int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, | 473 | int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, |
474 | int bits, int wake, int delete, gfp_t mask) | 474 | int bits, int wake, int delete, |
475 | struct extent_state **cached_state, | ||
476 | gfp_t mask) | ||
475 | { | 477 | { |
476 | struct extent_state *state; | 478 | struct extent_state *state; |
479 | struct extent_state *cached; | ||
477 | struct extent_state *prealloc = NULL; | 480 | struct extent_state *prealloc = NULL; |
481 | struct rb_node *next_node; | ||
478 | struct rb_node *node; | 482 | struct rb_node *node; |
479 | u64 last_end; | 483 | u64 last_end; |
480 | int err; | 484 | int err; |
@@ -488,6 +492,17 @@ again: | |||
488 | } | 492 | } |
489 | 493 | ||
490 | spin_lock(&tree->lock); | 494 | spin_lock(&tree->lock); |
495 | if (cached_state) { | ||
496 | cached = *cached_state; | ||
497 | *cached_state = NULL; | ||
498 | if (cached->tree && cached->start == start) { | ||
499 | atomic_dec(&cached->refs); | ||
500 | state = cached; | ||
501 | last_end = state->end; | ||
502 | goto found; | ||
503 | } | ||
504 | free_extent_state(cached); | ||
505 | } | ||
491 | /* | 506 | /* |
492 | * this search will find the extents that end after | 507 | * this search will find the extents that end after |
493 | * our range starts | 508 | * our range starts |
@@ -496,6 +511,7 @@ again: | |||
496 | if (!node) | 511 | if (!node) |
497 | goto out; | 512 | goto out; |
498 | state = rb_entry(node, struct extent_state, rb_node); | 513 | state = rb_entry(node, struct extent_state, rb_node); |
514 | hit_next: | ||
499 | if (state->start > end) | 515 | if (state->start > end) |
500 | goto out; | 516 | goto out; |
501 | WARN_ON(state->end < start); | 517 | WARN_ON(state->end < start); |
@@ -555,11 +571,21 @@ again: | |||
555 | prealloc = NULL; | 571 | prealloc = NULL; |
556 | goto out; | 572 | goto out; |
557 | } | 573 | } |
558 | 574 | found: | |
575 | if (state->end < end && prealloc && !need_resched()) | ||
576 | next_node = rb_next(&state->rb_node); | ||
577 | else | ||
578 | next_node = NULL; | ||
559 | set |= clear_state_bit(tree, state, bits, wake, delete); | 579 | set |= clear_state_bit(tree, state, bits, wake, delete); |
560 | if (last_end == (u64)-1) | 580 | if (last_end == (u64)-1) |
561 | goto out; | 581 | goto out; |
562 | start = last_end + 1; | 582 | start = last_end + 1; |
583 | if (start <= end && next_node) { | ||
584 | state = rb_entry(next_node, struct extent_state, | ||
585 | rb_node); | ||
586 | if (state->start == start) | ||
587 | goto hit_next; | ||
588 | } | ||
563 | goto search_again; | 589 | goto search_again; |
564 | 590 | ||
565 | out: | 591 | out: |
@@ -653,6 +679,17 @@ static void set_state_bits(struct extent_io_tree *tree, | |||
653 | state->state |= bits; | 679 | state->state |= bits; |
654 | } | 680 | } |
655 | 681 | ||
682 | static void cache_state(struct extent_state *state, | ||
683 | struct extent_state **cached_ptr) | ||
684 | { | ||
685 | if (cached_ptr && !(*cached_ptr)) { | ||
686 | if (state->state & (EXTENT_IOBITS | EXTENT_BOUNDARY)) { | ||
687 | *cached_ptr = state; | ||
688 | atomic_inc(&state->refs); | ||
689 | } | ||
690 | } | ||
691 | } | ||
692 | |||
656 | /* | 693 | /* |
657 | * set some bits on a range in the tree. This may require allocations or | 694 | * set some bits on a range in the tree. This may require allocations or |
658 | * sleeping, so the gfp mask is used to indicate what is allowed. | 695 | * sleeping, so the gfp mask is used to indicate what is allowed. |
@@ -666,6 +703,7 @@ static void set_state_bits(struct extent_io_tree *tree, | |||
666 | 703 | ||
667 | static int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, | 704 | static int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, |
668 | int bits, int exclusive_bits, u64 *failed_start, | 705 | int bits, int exclusive_bits, u64 *failed_start, |
706 | struct extent_state **cached_state, | ||
669 | gfp_t mask) | 707 | gfp_t mask) |
670 | { | 708 | { |
671 | struct extent_state *state; | 709 | struct extent_state *state; |
@@ -712,6 +750,7 @@ hit_next: | |||
712 | goto out; | 750 | goto out; |
713 | } | 751 | } |
714 | set_state_bits(tree, state, bits); | 752 | set_state_bits(tree, state, bits); |
753 | cache_state(state, cached_state); | ||
715 | merge_state(tree, state); | 754 | merge_state(tree, state); |
716 | if (last_end == (u64)-1) | 755 | if (last_end == (u64)-1) |
717 | goto out; | 756 | goto out; |
@@ -758,6 +797,7 @@ hit_next: | |||
758 | goto out; | 797 | goto out; |
759 | if (state->end <= end) { | 798 | if (state->end <= end) { |
760 | set_state_bits(tree, state, bits); | 799 | set_state_bits(tree, state, bits); |
800 | cache_state(state, cached_state); | ||
761 | merge_state(tree, state); | 801 | merge_state(tree, state); |
762 | if (last_end == (u64)-1) | 802 | if (last_end == (u64)-1) |
763 | goto out; | 803 | goto out; |
@@ -782,6 +822,7 @@ hit_next: | |||
782 | this_end = last_start - 1; | 822 | this_end = last_start - 1; |
783 | err = insert_state(tree, prealloc, start, this_end, | 823 | err = insert_state(tree, prealloc, start, this_end, |
784 | bits); | 824 | bits); |
825 | cache_state(prealloc, cached_state); | ||
785 | prealloc = NULL; | 826 | prealloc = NULL; |
786 | BUG_ON(err == -EEXIST); | 827 | BUG_ON(err == -EEXIST); |
787 | if (err) | 828 | if (err) |
@@ -805,6 +846,7 @@ hit_next: | |||
805 | BUG_ON(err == -EEXIST); | 846 | BUG_ON(err == -EEXIST); |
806 | 847 | ||
807 | set_state_bits(tree, prealloc, bits); | 848 | set_state_bits(tree, prealloc, bits); |
849 | cache_state(prealloc, cached_state); | ||
808 | merge_state(tree, prealloc); | 850 | merge_state(tree, prealloc); |
809 | prealloc = NULL; | 851 | prealloc = NULL; |
810 | goto out; | 852 | goto out; |
@@ -833,26 +875,27 @@ int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, | |||
833 | gfp_t mask) | 875 | gfp_t mask) |
834 | { | 876 | { |
835 | return set_extent_bit(tree, start, end, EXTENT_DIRTY, 0, NULL, | 877 | return set_extent_bit(tree, start, end, EXTENT_DIRTY, 0, NULL, |
836 | mask); | 878 | NULL, mask); |
837 | } | 879 | } |
838 | 880 | ||
839 | int set_extent_ordered(struct extent_io_tree *tree, u64 start, u64 end, | 881 | int set_extent_ordered(struct extent_io_tree *tree, u64 start, u64 end, |
840 | gfp_t mask) | 882 | gfp_t mask) |
841 | { | 883 | { |
842 | return set_extent_bit(tree, start, end, EXTENT_ORDERED, 0, NULL, mask); | 884 | return set_extent_bit(tree, start, end, EXTENT_ORDERED, 0, NULL, NULL, |
885 | mask); | ||
843 | } | 886 | } |
844 | 887 | ||
845 | int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, | 888 | int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
846 | int bits, gfp_t mask) | 889 | int bits, gfp_t mask) |
847 | { | 890 | { |
848 | return set_extent_bit(tree, start, end, bits, 0, NULL, | 891 | return set_extent_bit(tree, start, end, bits, 0, NULL, |
849 | mask); | 892 | NULL, mask); |
850 | } | 893 | } |
851 | 894 | ||
852 | int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, | 895 | int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
853 | int bits, gfp_t mask) | 896 | int bits, gfp_t mask) |
854 | { | 897 | { |
855 | return clear_extent_bit(tree, start, end, bits, 0, 0, mask); | 898 | return clear_extent_bit(tree, start, end, bits, 0, 0, NULL, mask); |
856 | } | 899 | } |
857 | 900 | ||
858 | int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, | 901 | int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, |
@@ -860,46 +903,50 @@ int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, | |||
860 | { | 903 | { |
861 | return set_extent_bit(tree, start, end, | 904 | return set_extent_bit(tree, start, end, |
862 | EXTENT_DELALLOC | EXTENT_DIRTY | EXTENT_UPTODATE, | 905 | EXTENT_DELALLOC | EXTENT_DIRTY | EXTENT_UPTODATE, |
863 | 0, NULL, mask); | 906 | 0, NULL, NULL, mask); |
864 | } | 907 | } |
865 | 908 | ||
866 | int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, | 909 | int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, |
867 | gfp_t mask) | 910 | gfp_t mask) |
868 | { | 911 | { |
869 | return clear_extent_bit(tree, start, end, | 912 | return clear_extent_bit(tree, start, end, |
870 | EXTENT_DIRTY | EXTENT_DELALLOC, 0, 0, mask); | 913 | EXTENT_DIRTY | EXTENT_DELALLOC, 0, 0, |
914 | NULL, mask); | ||
871 | } | 915 | } |
872 | 916 | ||
873 | int clear_extent_ordered(struct extent_io_tree *tree, u64 start, u64 end, | 917 | int clear_extent_ordered(struct extent_io_tree *tree, u64 start, u64 end, |
874 | gfp_t mask) | 918 | gfp_t mask) |
875 | { | 919 | { |
876 | return clear_extent_bit(tree, start, end, EXTENT_ORDERED, 1, 0, mask); | 920 | return clear_extent_bit(tree, start, end, EXTENT_ORDERED, 1, 0, |
921 | NULL, mask); | ||
877 | } | 922 | } |
878 | 923 | ||
879 | int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, | 924 | int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, |
880 | gfp_t mask) | 925 | gfp_t mask) |
881 | { | 926 | { |
882 | return set_extent_bit(tree, start, end, EXTENT_NEW, 0, NULL, | 927 | return set_extent_bit(tree, start, end, EXTENT_NEW, 0, NULL, |
883 | mask); | 928 | NULL, mask); |
884 | } | 929 | } |
885 | 930 | ||
886 | static int clear_extent_new(struct extent_io_tree *tree, u64 start, u64 end, | 931 | static int clear_extent_new(struct extent_io_tree *tree, u64 start, u64 end, |
887 | gfp_t mask) | 932 | gfp_t mask) |
888 | { | 933 | { |
889 | return clear_extent_bit(tree, start, end, EXTENT_NEW, 0, 0, mask); | 934 | return clear_extent_bit(tree, start, end, EXTENT_NEW, 0, 0, |
935 | NULL, mask); | ||
890 | } | 936 | } |
891 | 937 | ||
892 | int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, | 938 | int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, |
893 | gfp_t mask) | 939 | gfp_t mask) |
894 | { | 940 | { |
895 | return set_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, NULL, | 941 | return set_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, NULL, |
896 | mask); | 942 | NULL, mask); |
897 | } | 943 | } |
898 | 944 | ||
899 | static int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, | 945 | static int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, |
900 | u64 end, gfp_t mask) | 946 | u64 end, gfp_t mask) |
901 | { | 947 | { |
902 | return clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0, mask); | 948 | return clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0, |
949 | NULL, mask); | ||
903 | } | 950 | } |
904 | 951 | ||
905 | int wait_on_extent_writeback(struct extent_io_tree *tree, u64 start, u64 end) | 952 | int wait_on_extent_writeback(struct extent_io_tree *tree, u64 start, u64 end) |
@@ -912,13 +959,14 @@ int wait_on_extent_writeback(struct extent_io_tree *tree, u64 start, u64 end) | |||
912 | * us if waiting is desired. | 959 | * us if waiting is desired. |
913 | */ | 960 | */ |
914 | int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, | 961 | int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
915 | int bits, gfp_t mask) | 962 | int bits, struct extent_state **cached_state, gfp_t mask) |
916 | { | 963 | { |
917 | int err; | 964 | int err; |
918 | u64 failed_start; | 965 | u64 failed_start; |
919 | while (1) { | 966 | while (1) { |
920 | err = set_extent_bit(tree, start, end, EXTENT_LOCKED | bits, | 967 | err = set_extent_bit(tree, start, end, EXTENT_LOCKED | bits, |
921 | EXTENT_LOCKED, &failed_start, mask); | 968 | EXTENT_LOCKED, &failed_start, |
969 | cached_state, mask); | ||
922 | if (err == -EEXIST && (mask & __GFP_WAIT)) { | 970 | if (err == -EEXIST && (mask & __GFP_WAIT)) { |
923 | wait_extent_bit(tree, failed_start, end, EXTENT_LOCKED); | 971 | wait_extent_bit(tree, failed_start, end, EXTENT_LOCKED); |
924 | start = failed_start; | 972 | start = failed_start; |
@@ -932,7 +980,7 @@ int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, | |||
932 | 980 | ||
933 | int lock_extent(struct extent_io_tree *tree, u64 start, u64 end, gfp_t mask) | 981 | int lock_extent(struct extent_io_tree *tree, u64 start, u64 end, gfp_t mask) |
934 | { | 982 | { |
935 | return lock_extent_bits(tree, start, end, 0, mask); | 983 | return lock_extent_bits(tree, start, end, 0, NULL, mask); |
936 | } | 984 | } |
937 | 985 | ||
938 | int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end, | 986 | int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end, |
@@ -941,21 +989,29 @@ int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end, | |||
941 | int err; | 989 | int err; |
942 | u64 failed_start; | 990 | u64 failed_start; |
943 | 991 | ||
944 | err = set_extent_bit(tree, start, end, EXTENT_LOCKED, 1, | 992 | err = set_extent_bit(tree, start, end, EXTENT_LOCKED, EXTENT_LOCKED, |
945 | &failed_start, mask); | 993 | &failed_start, NULL, mask); |
946 | if (err == -EEXIST) { | 994 | if (err == -EEXIST) { |
947 | if (failed_start > start) | 995 | if (failed_start > start) |
948 | clear_extent_bit(tree, start, failed_start - 1, | 996 | clear_extent_bit(tree, start, failed_start - 1, |
949 | EXTENT_LOCKED, 1, 0, mask); | 997 | EXTENT_LOCKED, 1, 0, NULL, mask); |
950 | return 0; | 998 | return 0; |
951 | } | 999 | } |
952 | return 1; | 1000 | return 1; |
953 | } | 1001 | } |
954 | 1002 | ||
1003 | int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end, | ||
1004 | struct extent_state **cached, gfp_t mask) | ||
1005 | { | ||
1006 | return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached, | ||
1007 | mask); | ||
1008 | } | ||
1009 | |||
955 | int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end, | 1010 | int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end, |
956 | gfp_t mask) | 1011 | gfp_t mask) |
957 | { | 1012 | { |
958 | return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, mask); | 1013 | return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, NULL, |
1014 | mask); | ||
959 | } | 1015 | } |
960 | 1016 | ||
961 | /* | 1017 | /* |
@@ -1323,7 +1379,7 @@ int extent_clear_unlock_delalloc(struct inode *inode, | |||
1323 | if (clear_delalloc) | 1379 | if (clear_delalloc) |
1324 | clear_bits |= EXTENT_DELALLOC; | 1380 | clear_bits |= EXTENT_DELALLOC; |
1325 | 1381 | ||
1326 | clear_extent_bit(tree, start, end, clear_bits, 1, 0, GFP_NOFS); | 1382 | clear_extent_bit(tree, start, end, clear_bits, 1, 0, NULL, GFP_NOFS); |
1327 | if (!(unlock_pages || clear_dirty || set_writeback || end_writeback)) | 1383 | if (!(unlock_pages || clear_dirty || set_writeback || end_writeback)) |
1328 | return 0; | 1384 | return 0; |
1329 | 1385 | ||
@@ -2071,6 +2127,7 @@ static int __extent_writepage(struct page *page, struct writeback_control *wbc, | |||
2071 | u64 iosize; | 2127 | u64 iosize; |
2072 | u64 unlock_start; | 2128 | u64 unlock_start; |
2073 | sector_t sector; | 2129 | sector_t sector; |
2130 | struct extent_state *cached_state = NULL; | ||
2074 | struct extent_map *em; | 2131 | struct extent_map *em; |
2075 | struct block_device *bdev; | 2132 | struct block_device *bdev; |
2076 | int ret; | 2133 | int ret; |
@@ -2162,7 +2219,7 @@ static int __extent_writepage(struct page *page, struct writeback_control *wbc, | |||
2162 | goto done_unlocked; | 2219 | goto done_unlocked; |
2163 | } | 2220 | } |
2164 | } | 2221 | } |
2165 | lock_extent(tree, start, page_end, GFP_NOFS); | 2222 | lock_extent_bits(tree, start, page_end, 0, &cached_state, GFP_NOFS); |
2166 | 2223 | ||
2167 | unlock_start = start; | 2224 | unlock_start = start; |
2168 | 2225 | ||
@@ -2170,7 +2227,8 @@ static int __extent_writepage(struct page *page, struct writeback_control *wbc, | |||
2170 | ret = tree->ops->writepage_start_hook(page, start, | 2227 | ret = tree->ops->writepage_start_hook(page, start, |
2171 | page_end); | 2228 | page_end); |
2172 | if (ret == -EAGAIN) { | 2229 | if (ret == -EAGAIN) { |
2173 | unlock_extent(tree, start, page_end, GFP_NOFS); | 2230 | unlock_extent_cached(tree, start, page_end, |
2231 | &cached_state, GFP_NOFS); | ||
2174 | redirty_page_for_writepage(wbc, page); | 2232 | redirty_page_for_writepage(wbc, page); |
2175 | update_nr_written(page, wbc, nr_written); | 2233 | update_nr_written(page, wbc, nr_written); |
2176 | unlock_page(page); | 2234 | unlock_page(page); |
@@ -2192,7 +2250,7 @@ static int __extent_writepage(struct page *page, struct writeback_control *wbc, | |||
2192 | if (last_byte <= start) { | 2250 | if (last_byte <= start) { |
2193 | clear_extent_bit(tree, start, page_end, | 2251 | clear_extent_bit(tree, start, page_end, |
2194 | EXTENT_LOCKED | EXTENT_DIRTY, | 2252 | EXTENT_LOCKED | EXTENT_DIRTY, |
2195 | 1, 0, GFP_NOFS); | 2253 | 1, 0, NULL, GFP_NOFS); |
2196 | if (tree->ops && tree->ops->writepage_end_io_hook) | 2254 | if (tree->ops && tree->ops->writepage_end_io_hook) |
2197 | tree->ops->writepage_end_io_hook(page, start, | 2255 | tree->ops->writepage_end_io_hook(page, start, |
2198 | page_end, NULL, 1); | 2256 | page_end, NULL, 1); |
@@ -2204,7 +2262,8 @@ static int __extent_writepage(struct page *page, struct writeback_control *wbc, | |||
2204 | 2262 | ||
2205 | while (cur <= end) { | 2263 | while (cur <= end) { |
2206 | if (cur >= last_byte) { | 2264 | if (cur >= last_byte) { |
2207 | unlock_extent(tree, unlock_start, page_end, GFP_NOFS); | 2265 | unlock_extent_cached(tree, unlock_start, page_end, |
2266 | &cached_state, GFP_NOFS); | ||
2208 | if (tree->ops && tree->ops->writepage_end_io_hook) | 2267 | if (tree->ops && tree->ops->writepage_end_io_hook) |
2209 | tree->ops->writepage_end_io_hook(page, cur, | 2268 | tree->ops->writepage_end_io_hook(page, cur, |
2210 | page_end, NULL, 1); | 2269 | page_end, NULL, 1); |
@@ -2236,8 +2295,9 @@ static int __extent_writepage(struct page *page, struct writeback_control *wbc, | |||
2236 | */ | 2295 | */ |
2237 | if (compressed || block_start == EXTENT_MAP_HOLE || | 2296 | if (compressed || block_start == EXTENT_MAP_HOLE || |
2238 | block_start == EXTENT_MAP_INLINE) { | 2297 | block_start == EXTENT_MAP_INLINE) { |
2239 | unlock_extent(tree, unlock_start, cur + iosize - 1, | 2298 | unlock_extent_cached(tree, unlock_start, |
2240 | GFP_NOFS); | 2299 | cur + iosize - 1, &cached_state, |
2300 | GFP_NOFS); | ||
2241 | 2301 | ||
2242 | /* | 2302 | /* |
2243 | * end_io notification does not happen here for | 2303 | * end_io notification does not happen here for |
@@ -2307,11 +2367,14 @@ done: | |||
2307 | end_page_writeback(page); | 2367 | end_page_writeback(page); |
2308 | } | 2368 | } |
2309 | if (unlock_start <= page_end) | 2369 | if (unlock_start <= page_end) |
2310 | unlock_extent(tree, unlock_start, page_end, GFP_NOFS); | 2370 | unlock_extent_cached(tree, unlock_start, page_end, |
2371 | &cached_state, GFP_NOFS); | ||
2311 | unlock_page(page); | 2372 | unlock_page(page); |
2312 | 2373 | ||
2313 | done_unlocked: | 2374 | done_unlocked: |
2314 | 2375 | ||
2376 | /* drop our reference on any cached states */ | ||
2377 | free_extent_state(cached_state); | ||
2315 | return 0; | 2378 | return 0; |
2316 | } | 2379 | } |
2317 | 2380 | ||
@@ -2599,7 +2662,7 @@ int extent_invalidatepage(struct extent_io_tree *tree, | |||
2599 | wait_on_page_writeback(page); | 2662 | wait_on_page_writeback(page); |
2600 | clear_extent_bit(tree, start, end, | 2663 | clear_extent_bit(tree, start, end, |
2601 | EXTENT_LOCKED | EXTENT_DIRTY | EXTENT_DELALLOC, | 2664 | EXTENT_LOCKED | EXTENT_DIRTY | EXTENT_DELALLOC, |
2602 | 1, 1, GFP_NOFS); | 2665 | 1, 1, NULL, GFP_NOFS); |
2603 | return 0; | 2666 | return 0; |
2604 | } | 2667 | } |
2605 | 2668 | ||
@@ -2693,7 +2756,7 @@ int extent_prepare_write(struct extent_io_tree *tree, | |||
2693 | */ | 2756 | */ |
2694 | set_extent_bit(tree, block_start, | 2757 | set_extent_bit(tree, block_start, |
2695 | block_start + iosize - 1, | 2758 | block_start + iosize - 1, |
2696 | EXTENT_LOCKED, 0, NULL, GFP_NOFS); | 2759 | EXTENT_LOCKED, 0, NULL, NULL, GFP_NOFS); |
2697 | ret = submit_extent_page(READ, tree, page, | 2760 | ret = submit_extent_page(READ, tree, page, |
2698 | sector, iosize, page_offset, em->bdev, | 2761 | sector, iosize, page_offset, em->bdev, |
2699 | NULL, 1, | 2762 | NULL, 1, |
@@ -2740,7 +2803,7 @@ int try_release_extent_state(struct extent_map_tree *map, | |||
2740 | if ((mask & GFP_NOFS) == GFP_NOFS) | 2803 | if ((mask & GFP_NOFS) == GFP_NOFS) |
2741 | mask = GFP_NOFS; | 2804 | mask = GFP_NOFS; |
2742 | clear_extent_bit(tree, start, end, EXTENT_UPTODATE, | 2805 | clear_extent_bit(tree, start, end, EXTENT_UPTODATE, |
2743 | 1, 1, mask); | 2806 | 1, 1, NULL, mask); |
2744 | } | 2807 | } |
2745 | return ret; | 2808 | return ret; |
2746 | } | 2809 | } |
diff --git a/fs/btrfs/extent_io.h b/fs/btrfs/extent_io.h index 88d134d01fbc..c8ead2b8c4c9 100644 --- a/fs/btrfs/extent_io.h +++ b/fs/btrfs/extent_io.h | |||
@@ -143,7 +143,7 @@ int try_release_extent_state(struct extent_map_tree *map, | |||
143 | gfp_t mask); | 143 | gfp_t mask); |
144 | int lock_extent(struct extent_io_tree *tree, u64 start, u64 end, gfp_t mask); | 144 | int lock_extent(struct extent_io_tree *tree, u64 start, u64 end, gfp_t mask); |
145 | int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, | 145 | int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
146 | int bits, gfp_t mask); | 146 | int bits, struct extent_state **cached, gfp_t mask); |
147 | int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end, gfp_t mask); | 147 | int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end, gfp_t mask); |
148 | int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end, | 148 | int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end, |
149 | gfp_t mask); | 149 | gfp_t mask); |
@@ -161,7 +161,8 @@ int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, | |||
161 | int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, | 161 | int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
162 | int bits, gfp_t mask); | 162 | int bits, gfp_t mask); |
163 | int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, | 163 | int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, |
164 | int bits, int wake, int delete, gfp_t mask); | 164 | int bits, int wake, int delete, struct extent_state **cached, |
165 | gfp_t mask); | ||
165 | int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, | 166 | int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
166 | int bits, gfp_t mask); | 167 | int bits, gfp_t mask); |
167 | int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, | 168 | int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, |
diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c index f1df11718618..e494545c4202 100644 --- a/fs/btrfs/inode.c +++ b/fs/btrfs/inode.c | |||
@@ -854,7 +854,7 @@ static int cow_file_range_async(struct inode *inode, struct page *locked_page, | |||
854 | int limit = 10 * 1024 * 1042; | 854 | int limit = 10 * 1024 * 1042; |
855 | 855 | ||
856 | clear_extent_bit(&BTRFS_I(inode)->io_tree, start, end, EXTENT_LOCKED | | 856 | clear_extent_bit(&BTRFS_I(inode)->io_tree, start, end, EXTENT_LOCKED | |
857 | EXTENT_DELALLOC, 1, 0, GFP_NOFS); | 857 | EXTENT_DELALLOC, 1, 0, NULL, GFP_NOFS); |
858 | while (start < end) { | 858 | while (start < end) { |
859 | async_cow = kmalloc(sizeof(*async_cow), GFP_NOFS); | 859 | async_cow = kmalloc(sizeof(*async_cow), GFP_NOFS); |
860 | async_cow->inode = inode; | 860 | async_cow->inode = inode; |
@@ -4420,7 +4420,7 @@ static void btrfs_invalidatepage(struct page *page, unsigned long offset) | |||
4420 | */ | 4420 | */ |
4421 | clear_extent_bit(tree, page_start, page_end, | 4421 | clear_extent_bit(tree, page_start, page_end, |
4422 | EXTENT_DIRTY | EXTENT_DELALLOC | | 4422 | EXTENT_DIRTY | EXTENT_DELALLOC | |
4423 | EXTENT_LOCKED, 1, 0, GFP_NOFS); | 4423 | EXTENT_LOCKED, 1, 0, NULL, GFP_NOFS); |
4424 | btrfs_finish_ordered_io(page->mapping->host, | 4424 | btrfs_finish_ordered_io(page->mapping->host, |
4425 | page_start, page_end); | 4425 | page_start, page_end); |
4426 | btrfs_put_ordered_extent(ordered); | 4426 | btrfs_put_ordered_extent(ordered); |
@@ -4429,7 +4429,7 @@ static void btrfs_invalidatepage(struct page *page, unsigned long offset) | |||
4429 | clear_extent_bit(tree, page_start, page_end, | 4429 | clear_extent_bit(tree, page_start, page_end, |
4430 | EXTENT_LOCKED | EXTENT_DIRTY | EXTENT_DELALLOC | | 4430 | EXTENT_LOCKED | EXTENT_DIRTY | EXTENT_DELALLOC | |
4431 | EXTENT_ORDERED, | 4431 | EXTENT_ORDERED, |
4432 | 1, 1, GFP_NOFS); | 4432 | 1, 1, NULL, GFP_NOFS); |
4433 | __btrfs_releasepage(page, GFP_NOFS); | 4433 | __btrfs_releasepage(page, GFP_NOFS); |
4434 | 4434 | ||
4435 | ClearPageChecked(page); | 4435 | ClearPageChecked(page); |