diff options
Diffstat (limited to 'fs/btrfs')
-rw-r--r-- | fs/btrfs/Makefile | 2 | ||||
-rw-r--r-- | fs/btrfs/acl.c | 5 | ||||
-rw-r--r-- | fs/btrfs/btrfs_inode.h | 14 | ||||
-rw-r--r-- | fs/btrfs/compression.c | 5 | ||||
-rw-r--r-- | fs/btrfs/ctree.c | 18 | ||||
-rw-r--r-- | fs/btrfs/ctree.h | 58 | ||||
-rw-r--r-- | fs/btrfs/delayed-inode.c | 1695 | ||||
-rw-r--r-- | fs/btrfs/delayed-inode.h | 141 | ||||
-rw-r--r-- | fs/btrfs/dir-item.c | 34 | ||||
-rw-r--r-- | fs/btrfs/disk-io.c | 69 | ||||
-rw-r--r-- | fs/btrfs/disk-io.h | 1 | ||||
-rw-r--r-- | fs/btrfs/export.c | 25 | ||||
-rw-r--r-- | fs/btrfs/extent-tree.c | 95 | ||||
-rw-r--r-- | fs/btrfs/extent_io.c | 4 | ||||
-rw-r--r-- | fs/btrfs/file-item.c | 5 | ||||
-rw-r--r-- | fs/btrfs/file.c | 27 | ||||
-rw-r--r-- | fs/btrfs/free-space-cache.c | 973 | ||||
-rw-r--r-- | fs/btrfs/free-space-cache.h | 47 | ||||
-rw-r--r-- | fs/btrfs/inode-map.c | 428 | ||||
-rw-r--r-- | fs/btrfs/inode-map.h | 13 | ||||
-rw-r--r-- | fs/btrfs/inode.c | 405 | ||||
-rw-r--r-- | fs/btrfs/ioctl.c | 46 | ||||
-rw-r--r-- | fs/btrfs/relocation.c | 27 | ||||
-rw-r--r-- | fs/btrfs/super.c | 10 | ||||
-rw-r--r-- | fs/btrfs/sysfs.c | 12 | ||||
-rw-r--r-- | fs/btrfs/transaction.c | 56 | ||||
-rw-r--r-- | fs/btrfs/transaction.h | 2 | ||||
-rw-r--r-- | fs/btrfs/tree-log.c | 61 | ||||
-rw-r--r-- | fs/btrfs/xattr.c | 8 |
29 files changed, 3533 insertions, 753 deletions
diff --git a/fs/btrfs/Makefile b/fs/btrfs/Makefile index 31610ea73ae..a8411c22313 100644 --- a/fs/btrfs/Makefile +++ b/fs/btrfs/Makefile | |||
@@ -7,4 +7,4 @@ btrfs-y += super.o ctree.o extent-tree.o print-tree.o root-tree.o dir-item.o \ | |||
7 | extent_map.o sysfs.o struct-funcs.o xattr.o ordered-data.o \ | 7 | extent_map.o sysfs.o struct-funcs.o xattr.o ordered-data.o \ |
8 | extent_io.o volumes.o async-thread.o ioctl.o locking.o orphan.o \ | 8 | extent_io.o volumes.o async-thread.o ioctl.o locking.o orphan.o \ |
9 | export.o tree-log.o acl.o free-space-cache.o zlib.o lzo.o \ | 9 | export.o tree-log.o acl.o free-space-cache.o zlib.o lzo.o \ |
10 | compression.o delayed-ref.o relocation.o | 10 | compression.o delayed-ref.o relocation.o delayed-inode.o |
diff --git a/fs/btrfs/acl.c b/fs/btrfs/acl.c index 1a21c99a91b..f66fc995973 100644 --- a/fs/btrfs/acl.c +++ b/fs/btrfs/acl.c | |||
@@ -178,12 +178,13 @@ static int btrfs_xattr_acl_set(struct dentry *dentry, const char *name, | |||
178 | 178 | ||
179 | if (value) { | 179 | if (value) { |
180 | acl = posix_acl_from_xattr(value, size); | 180 | acl = posix_acl_from_xattr(value, size); |
181 | if (IS_ERR(acl)) | ||
182 | return PTR_ERR(acl); | ||
183 | |||
181 | if (acl) { | 184 | if (acl) { |
182 | ret = posix_acl_valid(acl); | 185 | ret = posix_acl_valid(acl); |
183 | if (ret) | 186 | if (ret) |
184 | goto out; | 187 | goto out; |
185 | } else if (IS_ERR(acl)) { | ||
186 | return PTR_ERR(acl); | ||
187 | } | 188 | } |
188 | } | 189 | } |
189 | 190 | ||
diff --git a/fs/btrfs/btrfs_inode.h b/fs/btrfs/btrfs_inode.h index 57c3bb2884c..d0b0e43a6a8 100644 --- a/fs/btrfs/btrfs_inode.h +++ b/fs/btrfs/btrfs_inode.h | |||
@@ -22,6 +22,7 @@ | |||
22 | #include "extent_map.h" | 22 | #include "extent_map.h" |
23 | #include "extent_io.h" | 23 | #include "extent_io.h" |
24 | #include "ordered-data.h" | 24 | #include "ordered-data.h" |
25 | #include "delayed-inode.h" | ||
25 | 26 | ||
26 | /* in memory btrfs inode */ | 27 | /* in memory btrfs inode */ |
27 | struct btrfs_inode { | 28 | struct btrfs_inode { |
@@ -158,14 +159,27 @@ struct btrfs_inode { | |||
158 | */ | 159 | */ |
159 | unsigned force_compress:4; | 160 | unsigned force_compress:4; |
160 | 161 | ||
162 | struct btrfs_delayed_node *delayed_node; | ||
163 | |||
161 | struct inode vfs_inode; | 164 | struct inode vfs_inode; |
162 | }; | 165 | }; |
163 | 166 | ||
167 | extern unsigned char btrfs_filetype_table[]; | ||
168 | |||
164 | static inline struct btrfs_inode *BTRFS_I(struct inode *inode) | 169 | static inline struct btrfs_inode *BTRFS_I(struct inode *inode) |
165 | { | 170 | { |
166 | return container_of(inode, struct btrfs_inode, vfs_inode); | 171 | return container_of(inode, struct btrfs_inode, vfs_inode); |
167 | } | 172 | } |
168 | 173 | ||
174 | static inline u64 btrfs_ino(struct inode *inode) | ||
175 | { | ||
176 | u64 ino = BTRFS_I(inode)->location.objectid; | ||
177 | |||
178 | if (ino <= BTRFS_FIRST_FREE_OBJECTID) | ||
179 | ino = inode->i_ino; | ||
180 | return ino; | ||
181 | } | ||
182 | |||
169 | static inline void btrfs_i_size_write(struct inode *inode, u64 size) | 183 | static inline void btrfs_i_size_write(struct inode *inode, u64 size) |
170 | { | 184 | { |
171 | i_size_write(inode, size); | 185 | i_size_write(inode, size); |
diff --git a/fs/btrfs/compression.c b/fs/btrfs/compression.c index d4cd0f0cd69..bfe42b03eaf 100644 --- a/fs/btrfs/compression.c +++ b/fs/btrfs/compression.c | |||
@@ -125,9 +125,10 @@ static int check_compressed_csum(struct inode *inode, | |||
125 | kunmap_atomic(kaddr, KM_USER0); | 125 | kunmap_atomic(kaddr, KM_USER0); |
126 | 126 | ||
127 | if (csum != *cb_sum) { | 127 | if (csum != *cb_sum) { |
128 | printk(KERN_INFO "btrfs csum failed ino %lu " | 128 | printk(KERN_INFO "btrfs csum failed ino %llu " |
129 | "extent %llu csum %u " | 129 | "extent %llu csum %u " |
130 | "wanted %u mirror %d\n", inode->i_ino, | 130 | "wanted %u mirror %d\n", |
131 | (unsigned long long)btrfs_ino(inode), | ||
131 | (unsigned long long)disk_start, | 132 | (unsigned long long)disk_start, |
132 | csum, *cb_sum, cb->mirror_num); | 133 | csum, *cb_sum, cb->mirror_num); |
133 | ret = -EIO; | 134 | ret = -EIO; |
diff --git a/fs/btrfs/ctree.c b/fs/btrfs/ctree.c index fad8f23d70f..b6cbeed226b 100644 --- a/fs/btrfs/ctree.c +++ b/fs/btrfs/ctree.c | |||
@@ -38,11 +38,6 @@ static int balance_node_right(struct btrfs_trans_handle *trans, | |||
38 | struct extent_buffer *src_buf); | 38 | struct extent_buffer *src_buf); |
39 | static int del_ptr(struct btrfs_trans_handle *trans, struct btrfs_root *root, | 39 | static int del_ptr(struct btrfs_trans_handle *trans, struct btrfs_root *root, |
40 | struct btrfs_path *path, int level, int slot); | 40 | struct btrfs_path *path, int level, int slot); |
41 | static int setup_items_for_insert(struct btrfs_trans_handle *trans, | ||
42 | struct btrfs_root *root, struct btrfs_path *path, | ||
43 | struct btrfs_key *cpu_key, u32 *data_size, | ||
44 | u32 total_data, u32 total_size, int nr); | ||
45 | |||
46 | 41 | ||
47 | struct btrfs_path *btrfs_alloc_path(void) | 42 | struct btrfs_path *btrfs_alloc_path(void) |
48 | { | 43 | { |
@@ -74,8 +69,8 @@ noinline void btrfs_set_path_blocking(struct btrfs_path *p) | |||
74 | * retake all the spinlocks in the path. You can safely use NULL | 69 | * retake all the spinlocks in the path. You can safely use NULL |
75 | * for held | 70 | * for held |
76 | */ | 71 | */ |
77 | static noinline void btrfs_clear_path_blocking(struct btrfs_path *p, | 72 | noinline void btrfs_clear_path_blocking(struct btrfs_path *p, |
78 | struct extent_buffer *held) | 73 | struct extent_buffer *held) |
79 | { | 74 | { |
80 | int i; | 75 | int i; |
81 | 76 | ||
@@ -3559,11 +3554,10 @@ out: | |||
3559 | * to save stack depth by doing the bulk of the work in a function | 3554 | * to save stack depth by doing the bulk of the work in a function |
3560 | * that doesn't call btrfs_search_slot | 3555 | * that doesn't call btrfs_search_slot |
3561 | */ | 3556 | */ |
3562 | static noinline_for_stack int | 3557 | int setup_items_for_insert(struct btrfs_trans_handle *trans, |
3563 | setup_items_for_insert(struct btrfs_trans_handle *trans, | 3558 | struct btrfs_root *root, struct btrfs_path *path, |
3564 | struct btrfs_root *root, struct btrfs_path *path, | 3559 | struct btrfs_key *cpu_key, u32 *data_size, |
3565 | struct btrfs_key *cpu_key, u32 *data_size, | 3560 | u32 total_data, u32 total_size, int nr) |
3566 | u32 total_data, u32 total_size, int nr) | ||
3567 | { | 3561 | { |
3568 | struct btrfs_item *item; | 3562 | struct btrfs_item *item; |
3569 | int i; | 3563 | int i; |
diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h index 343304dec6d..e7d40791ec9 100644 --- a/fs/btrfs/ctree.h +++ b/fs/btrfs/ctree.h | |||
@@ -105,6 +105,12 @@ struct btrfs_ordered_sum; | |||
105 | /* For storing free space cache */ | 105 | /* For storing free space cache */ |
106 | #define BTRFS_FREE_SPACE_OBJECTID -11ULL | 106 | #define BTRFS_FREE_SPACE_OBJECTID -11ULL |
107 | 107 | ||
108 | /* | ||
109 | * The inode number assigned to the special inode for sotring | ||
110 | * free ino cache | ||
111 | */ | ||
112 | #define BTRFS_FREE_INO_OBJECTID -12ULL | ||
113 | |||
108 | /* dummy objectid represents multiple objectids */ | 114 | /* dummy objectid represents multiple objectids */ |
109 | #define BTRFS_MULTIPLE_OBJECTIDS -255ULL | 115 | #define BTRFS_MULTIPLE_OBJECTIDS -255ULL |
110 | 116 | ||
@@ -830,9 +836,6 @@ struct btrfs_block_group_cache { | |||
830 | u64 bytes_super; | 836 | u64 bytes_super; |
831 | u64 flags; | 837 | u64 flags; |
832 | u64 sectorsize; | 838 | u64 sectorsize; |
833 | int extents_thresh; | ||
834 | int free_extents; | ||
835 | int total_bitmaps; | ||
836 | unsigned int ro:1; | 839 | unsigned int ro:1; |
837 | unsigned int dirty:1; | 840 | unsigned int dirty:1; |
838 | unsigned int iref:1; | 841 | unsigned int iref:1; |
@@ -847,9 +850,7 @@ struct btrfs_block_group_cache { | |||
847 | struct btrfs_space_info *space_info; | 850 | struct btrfs_space_info *space_info; |
848 | 851 | ||
849 | /* free space cache stuff */ | 852 | /* free space cache stuff */ |
850 | spinlock_t tree_lock; | 853 | struct btrfs_free_space_ctl *free_space_ctl; |
851 | struct rb_root free_space_offset; | ||
852 | u64 free_space; | ||
853 | 854 | ||
854 | /* block group cache stuff */ | 855 | /* block group cache stuff */ |
855 | struct rb_node cache_node; | 856 | struct rb_node cache_node; |
@@ -869,6 +870,7 @@ struct btrfs_block_group_cache { | |||
869 | struct reloc_control; | 870 | struct reloc_control; |
870 | struct btrfs_device; | 871 | struct btrfs_device; |
871 | struct btrfs_fs_devices; | 872 | struct btrfs_fs_devices; |
873 | struct btrfs_delayed_root; | ||
872 | struct btrfs_fs_info { | 874 | struct btrfs_fs_info { |
873 | u8 fsid[BTRFS_FSID_SIZE]; | 875 | u8 fsid[BTRFS_FSID_SIZE]; |
874 | u8 chunk_tree_uuid[BTRFS_UUID_SIZE]; | 876 | u8 chunk_tree_uuid[BTRFS_UUID_SIZE]; |
@@ -895,7 +897,10 @@ struct btrfs_fs_info { | |||
895 | /* logical->physical extent mapping */ | 897 | /* logical->physical extent mapping */ |
896 | struct btrfs_mapping_tree mapping_tree; | 898 | struct btrfs_mapping_tree mapping_tree; |
897 | 899 | ||
898 | /* block reservation for extent, checksum and root tree */ | 900 | /* |
901 | * block reservation for extent, checksum, root tree and | ||
902 | * delayed dir index item | ||
903 | */ | ||
899 | struct btrfs_block_rsv global_block_rsv; | 904 | struct btrfs_block_rsv global_block_rsv; |
900 | /* block reservation for delay allocation */ | 905 | /* block reservation for delay allocation */ |
901 | struct btrfs_block_rsv delalloc_block_rsv; | 906 | struct btrfs_block_rsv delalloc_block_rsv; |
@@ -1022,6 +1027,7 @@ struct btrfs_fs_info { | |||
1022 | * for the sys_munmap function call path | 1027 | * for the sys_munmap function call path |
1023 | */ | 1028 | */ |
1024 | struct btrfs_workers fixup_workers; | 1029 | struct btrfs_workers fixup_workers; |
1030 | struct btrfs_workers delayed_workers; | ||
1025 | struct task_struct *transaction_kthread; | 1031 | struct task_struct *transaction_kthread; |
1026 | struct task_struct *cleaner_kthread; | 1032 | struct task_struct *cleaner_kthread; |
1027 | int thread_pool_size; | 1033 | int thread_pool_size; |
@@ -1079,6 +1085,8 @@ struct btrfs_fs_info { | |||
1079 | 1085 | ||
1080 | /* filesystem state */ | 1086 | /* filesystem state */ |
1081 | u64 fs_state; | 1087 | u64 fs_state; |
1088 | |||
1089 | struct btrfs_delayed_root *delayed_root; | ||
1082 | }; | 1090 | }; |
1083 | 1091 | ||
1084 | /* | 1092 | /* |
@@ -1107,6 +1115,16 @@ struct btrfs_root { | |||
1107 | spinlock_t accounting_lock; | 1115 | spinlock_t accounting_lock; |
1108 | struct btrfs_block_rsv *block_rsv; | 1116 | struct btrfs_block_rsv *block_rsv; |
1109 | 1117 | ||
1118 | /* free ino cache stuff */ | ||
1119 | struct mutex fs_commit_mutex; | ||
1120 | struct btrfs_free_space_ctl *free_ino_ctl; | ||
1121 | enum btrfs_caching_type cached; | ||
1122 | spinlock_t cache_lock; | ||
1123 | wait_queue_head_t cache_wait; | ||
1124 | struct btrfs_free_space_ctl *free_ino_pinned; | ||
1125 | u64 cache_progress; | ||
1126 | struct inode *cache_inode; | ||
1127 | |||
1110 | struct mutex log_mutex; | 1128 | struct mutex log_mutex; |
1111 | wait_queue_head_t log_writer_wait; | 1129 | wait_queue_head_t log_writer_wait; |
1112 | wait_queue_head_t log_commit_wait[2]; | 1130 | wait_queue_head_t log_commit_wait[2]; |
@@ -1162,6 +1180,11 @@ struct btrfs_root { | |||
1162 | struct rb_root inode_tree; | 1180 | struct rb_root inode_tree; |
1163 | 1181 | ||
1164 | /* | 1182 | /* |
1183 | * radix tree that keeps track of delayed nodes of every inode, | ||
1184 | * protected by inode_lock | ||
1185 | */ | ||
1186 | struct radix_tree_root delayed_nodes_tree; | ||
1187 | /* | ||
1165 | * right now this just gets used so that a root has its own devid | 1188 | * right now this just gets used so that a root has its own devid |
1166 | * for stat. It may be used for more later | 1189 | * for stat. It may be used for more later |
1167 | */ | 1190 | */ |
@@ -2034,6 +2057,13 @@ static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info) | |||
2034 | } | 2057 | } |
2035 | 2058 | ||
2036 | /* extent-tree.c */ | 2059 | /* extent-tree.c */ |
2060 | static inline u64 btrfs_calc_trans_metadata_size(struct btrfs_root *root, | ||
2061 | int num_items) | ||
2062 | { | ||
2063 | return (root->leafsize + root->nodesize * (BTRFS_MAX_LEVEL - 1)) * | ||
2064 | 3 * num_items; | ||
2065 | } | ||
2066 | |||
2037 | void btrfs_put_block_group(struct btrfs_block_group_cache *cache); | 2067 | void btrfs_put_block_group(struct btrfs_block_group_cache *cache); |
2038 | int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans, | 2068 | int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans, |
2039 | struct btrfs_root *root, unsigned long count); | 2069 | struct btrfs_root *root, unsigned long count); |
@@ -2226,6 +2256,8 @@ void btrfs_release_path(struct btrfs_path *p); | |||
2226 | struct btrfs_path *btrfs_alloc_path(void); | 2256 | struct btrfs_path *btrfs_alloc_path(void); |
2227 | void btrfs_free_path(struct btrfs_path *p); | 2257 | void btrfs_free_path(struct btrfs_path *p); |
2228 | void btrfs_set_path_blocking(struct btrfs_path *p); | 2258 | void btrfs_set_path_blocking(struct btrfs_path *p); |
2259 | void btrfs_clear_path_blocking(struct btrfs_path *p, | ||
2260 | struct extent_buffer *held); | ||
2229 | void btrfs_unlock_up_safe(struct btrfs_path *p, int level); | 2261 | void btrfs_unlock_up_safe(struct btrfs_path *p, int level); |
2230 | 2262 | ||
2231 | int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root, | 2263 | int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root, |
@@ -2237,6 +2269,10 @@ static inline int btrfs_del_item(struct btrfs_trans_handle *trans, | |||
2237 | return btrfs_del_items(trans, root, path, path->slots[0], 1); | 2269 | return btrfs_del_items(trans, root, path, path->slots[0], 1); |
2238 | } | 2270 | } |
2239 | 2271 | ||
2272 | int setup_items_for_insert(struct btrfs_trans_handle *trans, | ||
2273 | struct btrfs_root *root, struct btrfs_path *path, | ||
2274 | struct btrfs_key *cpu_key, u32 *data_size, | ||
2275 | u32 total_data, u32 total_size, int nr); | ||
2240 | int btrfs_insert_item(struct btrfs_trans_handle *trans, struct btrfs_root | 2276 | int btrfs_insert_item(struct btrfs_trans_handle *trans, struct btrfs_root |
2241 | *root, struct btrfs_key *key, void *data, u32 data_size); | 2277 | *root, struct btrfs_key *key, void *data, u32 data_size); |
2242 | int btrfs_insert_empty_items(struct btrfs_trans_handle *trans, | 2278 | int btrfs_insert_empty_items(struct btrfs_trans_handle *trans, |
@@ -2293,7 +2329,7 @@ void btrfs_check_and_init_root_item(struct btrfs_root_item *item); | |||
2293 | /* dir-item.c */ | 2329 | /* dir-item.c */ |
2294 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, | 2330 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, |
2295 | struct btrfs_root *root, const char *name, | 2331 | struct btrfs_root *root, const char *name, |
2296 | int name_len, u64 dir, | 2332 | int name_len, struct inode *dir, |
2297 | struct btrfs_key *location, u8 type, u64 index); | 2333 | struct btrfs_key *location, u8 type, u64 index); |
2298 | struct btrfs_dir_item *btrfs_lookup_dir_item(struct btrfs_trans_handle *trans, | 2334 | struct btrfs_dir_item *btrfs_lookup_dir_item(struct btrfs_trans_handle *trans, |
2299 | struct btrfs_root *root, | 2335 | struct btrfs_root *root, |
@@ -2338,12 +2374,6 @@ int btrfs_del_orphan_item(struct btrfs_trans_handle *trans, | |||
2338 | struct btrfs_root *root, u64 offset); | 2374 | struct btrfs_root *root, u64 offset); |
2339 | int btrfs_find_orphan_item(struct btrfs_root *root, u64 offset); | 2375 | int btrfs_find_orphan_item(struct btrfs_root *root, u64 offset); |
2340 | 2376 | ||
2341 | /* inode-map.c */ | ||
2342 | int btrfs_find_free_objectid(struct btrfs_trans_handle *trans, | ||
2343 | struct btrfs_root *fs_root, | ||
2344 | u64 dirid, u64 *objectid); | ||
2345 | int btrfs_find_highest_inode(struct btrfs_root *fs_root, u64 *objectid); | ||
2346 | |||
2347 | /* inode-item.c */ | 2377 | /* inode-item.c */ |
2348 | int btrfs_insert_inode_ref(struct btrfs_trans_handle *trans, | 2378 | int btrfs_insert_inode_ref(struct btrfs_trans_handle *trans, |
2349 | struct btrfs_root *root, | 2379 | struct btrfs_root *root, |
diff --git a/fs/btrfs/delayed-inode.c b/fs/btrfs/delayed-inode.c new file mode 100644 index 00000000000..01e29503a54 --- /dev/null +++ b/fs/btrfs/delayed-inode.c | |||
@@ -0,0 +1,1695 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2011 Fujitsu. All rights reserved. | ||
3 | * Written by Miao Xie <miaox@cn.fujitsu.com> | ||
4 | * | ||
5 | * This program is free software; you can redistribute it and/or | ||
6 | * modify it under the terms of the GNU General Public | ||
7 | * License v2 as published by the Free Software Foundation. | ||
8 | * | ||
9 | * This program is distributed in the hope that it will be useful, | ||
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | ||
12 | * General Public License for more details. | ||
13 | * | ||
14 | * You should have received a copy of the GNU General Public | ||
15 | * License along with this program; if not, write to the | ||
16 | * Free Software Foundation, Inc., 59 Temple Place - Suite 330, | ||
17 | * Boston, MA 021110-1307, USA. | ||
18 | */ | ||
19 | |||
20 | #include <linux/slab.h> | ||
21 | #include "delayed-inode.h" | ||
22 | #include "disk-io.h" | ||
23 | #include "transaction.h" | ||
24 | |||
25 | #define BTRFS_DELAYED_WRITEBACK 400 | ||
26 | #define BTRFS_DELAYED_BACKGROUND 100 | ||
27 | |||
28 | static struct kmem_cache *delayed_node_cache; | ||
29 | |||
30 | int __init btrfs_delayed_inode_init(void) | ||
31 | { | ||
32 | delayed_node_cache = kmem_cache_create("delayed_node", | ||
33 | sizeof(struct btrfs_delayed_node), | ||
34 | 0, | ||
35 | SLAB_RECLAIM_ACCOUNT | SLAB_MEM_SPREAD, | ||
36 | NULL); | ||
37 | if (!delayed_node_cache) | ||
38 | return -ENOMEM; | ||
39 | return 0; | ||
40 | } | ||
41 | |||
42 | void btrfs_delayed_inode_exit(void) | ||
43 | { | ||
44 | if (delayed_node_cache) | ||
45 | kmem_cache_destroy(delayed_node_cache); | ||
46 | } | ||
47 | |||
48 | static inline void btrfs_init_delayed_node( | ||
49 | struct btrfs_delayed_node *delayed_node, | ||
50 | struct btrfs_root *root, u64 inode_id) | ||
51 | { | ||
52 | delayed_node->root = root; | ||
53 | delayed_node->inode_id = inode_id; | ||
54 | atomic_set(&delayed_node->refs, 0); | ||
55 | delayed_node->count = 0; | ||
56 | delayed_node->in_list = 0; | ||
57 | delayed_node->inode_dirty = 0; | ||
58 | delayed_node->ins_root = RB_ROOT; | ||
59 | delayed_node->del_root = RB_ROOT; | ||
60 | mutex_init(&delayed_node->mutex); | ||
61 | delayed_node->index_cnt = 0; | ||
62 | INIT_LIST_HEAD(&delayed_node->n_list); | ||
63 | INIT_LIST_HEAD(&delayed_node->p_list); | ||
64 | delayed_node->bytes_reserved = 0; | ||
65 | } | ||
66 | |||
67 | static inline int btrfs_is_continuous_delayed_item( | ||
68 | struct btrfs_delayed_item *item1, | ||
69 | struct btrfs_delayed_item *item2) | ||
70 | { | ||
71 | if (item1->key.type == BTRFS_DIR_INDEX_KEY && | ||
72 | item1->key.objectid == item2->key.objectid && | ||
73 | item1->key.type == item2->key.type && | ||
74 | item1->key.offset + 1 == item2->key.offset) | ||
75 | return 1; | ||
76 | return 0; | ||
77 | } | ||
78 | |||
79 | static inline struct btrfs_delayed_root *btrfs_get_delayed_root( | ||
80 | struct btrfs_root *root) | ||
81 | { | ||
82 | return root->fs_info->delayed_root; | ||
83 | } | ||
84 | |||
85 | static struct btrfs_delayed_node *btrfs_get_or_create_delayed_node( | ||
86 | struct inode *inode) | ||
87 | { | ||
88 | struct btrfs_delayed_node *node; | ||
89 | struct btrfs_inode *btrfs_inode = BTRFS_I(inode); | ||
90 | struct btrfs_root *root = btrfs_inode->root; | ||
91 | u64 ino = btrfs_ino(inode); | ||
92 | int ret; | ||
93 | |||
94 | again: | ||
95 | node = ACCESS_ONCE(btrfs_inode->delayed_node); | ||
96 | if (node) { | ||
97 | atomic_inc(&node->refs); /* can be accessed */ | ||
98 | return node; | ||
99 | } | ||
100 | |||
101 | spin_lock(&root->inode_lock); | ||
102 | node = radix_tree_lookup(&root->delayed_nodes_tree, ino); | ||
103 | if (node) { | ||
104 | if (btrfs_inode->delayed_node) { | ||
105 | spin_unlock(&root->inode_lock); | ||
106 | goto again; | ||
107 | } | ||
108 | btrfs_inode->delayed_node = node; | ||
109 | atomic_inc(&node->refs); /* can be accessed */ | ||
110 | atomic_inc(&node->refs); /* cached in the inode */ | ||
111 | spin_unlock(&root->inode_lock); | ||
112 | return node; | ||
113 | } | ||
114 | spin_unlock(&root->inode_lock); | ||
115 | |||
116 | node = kmem_cache_alloc(delayed_node_cache, GFP_NOFS); | ||
117 | if (!node) | ||
118 | return ERR_PTR(-ENOMEM); | ||
119 | btrfs_init_delayed_node(node, root, ino); | ||
120 | |||
121 | atomic_inc(&node->refs); /* cached in the btrfs inode */ | ||
122 | atomic_inc(&node->refs); /* can be accessed */ | ||
123 | |||
124 | ret = radix_tree_preload(GFP_NOFS & ~__GFP_HIGHMEM); | ||
125 | if (ret) { | ||
126 | kmem_cache_free(delayed_node_cache, node); | ||
127 | return ERR_PTR(ret); | ||
128 | } | ||
129 | |||
130 | spin_lock(&root->inode_lock); | ||
131 | ret = radix_tree_insert(&root->delayed_nodes_tree, ino, node); | ||
132 | if (ret == -EEXIST) { | ||
133 | kmem_cache_free(delayed_node_cache, node); | ||
134 | spin_unlock(&root->inode_lock); | ||
135 | radix_tree_preload_end(); | ||
136 | goto again; | ||
137 | } | ||
138 | btrfs_inode->delayed_node = node; | ||
139 | spin_unlock(&root->inode_lock); | ||
140 | radix_tree_preload_end(); | ||
141 | |||
142 | return node; | ||
143 | } | ||
144 | |||
145 | /* | ||
146 | * Call it when holding delayed_node->mutex | ||
147 | * | ||
148 | * If mod = 1, add this node into the prepared list. | ||
149 | */ | ||
150 | static void btrfs_queue_delayed_node(struct btrfs_delayed_root *root, | ||
151 | struct btrfs_delayed_node *node, | ||
152 | int mod) | ||
153 | { | ||
154 | spin_lock(&root->lock); | ||
155 | if (node->in_list) { | ||
156 | if (!list_empty(&node->p_list)) | ||
157 | list_move_tail(&node->p_list, &root->prepare_list); | ||
158 | else if (mod) | ||
159 | list_add_tail(&node->p_list, &root->prepare_list); | ||
160 | } else { | ||
161 | list_add_tail(&node->n_list, &root->node_list); | ||
162 | list_add_tail(&node->p_list, &root->prepare_list); | ||
163 | atomic_inc(&node->refs); /* inserted into list */ | ||
164 | root->nodes++; | ||
165 | node->in_list = 1; | ||
166 | } | ||
167 | spin_unlock(&root->lock); | ||
168 | } | ||
169 | |||
170 | /* Call it when holding delayed_node->mutex */ | ||
171 | static void btrfs_dequeue_delayed_node(struct btrfs_delayed_root *root, | ||
172 | struct btrfs_delayed_node *node) | ||
173 | { | ||
174 | spin_lock(&root->lock); | ||
175 | if (node->in_list) { | ||
176 | root->nodes--; | ||
177 | atomic_dec(&node->refs); /* not in the list */ | ||
178 | list_del_init(&node->n_list); | ||
179 | if (!list_empty(&node->p_list)) | ||
180 | list_del_init(&node->p_list); | ||
181 | node->in_list = 0; | ||
182 | } | ||
183 | spin_unlock(&root->lock); | ||
184 | } | ||
185 | |||
186 | struct btrfs_delayed_node *btrfs_first_delayed_node( | ||
187 | struct btrfs_delayed_root *delayed_root) | ||
188 | { | ||
189 | struct list_head *p; | ||
190 | struct btrfs_delayed_node *node = NULL; | ||
191 | |||
192 | spin_lock(&delayed_root->lock); | ||
193 | if (list_empty(&delayed_root->node_list)) | ||
194 | goto out; | ||
195 | |||
196 | p = delayed_root->node_list.next; | ||
197 | node = list_entry(p, struct btrfs_delayed_node, n_list); | ||
198 | atomic_inc(&node->refs); | ||
199 | out: | ||
200 | spin_unlock(&delayed_root->lock); | ||
201 | |||
202 | return node; | ||
203 | } | ||
204 | |||
205 | struct btrfs_delayed_node *btrfs_next_delayed_node( | ||
206 | struct btrfs_delayed_node *node) | ||
207 | { | ||
208 | struct btrfs_delayed_root *delayed_root; | ||
209 | struct list_head *p; | ||
210 | struct btrfs_delayed_node *next = NULL; | ||
211 | |||
212 | delayed_root = node->root->fs_info->delayed_root; | ||
213 | spin_lock(&delayed_root->lock); | ||
214 | if (!node->in_list) { /* not in the list */ | ||
215 | if (list_empty(&delayed_root->node_list)) | ||
216 | goto out; | ||
217 | p = delayed_root->node_list.next; | ||
218 | } else if (list_is_last(&node->n_list, &delayed_root->node_list)) | ||
219 | goto out; | ||
220 | else | ||
221 | p = node->n_list.next; | ||
222 | |||
223 | next = list_entry(p, struct btrfs_delayed_node, n_list); | ||
224 | atomic_inc(&next->refs); | ||
225 | out: | ||
226 | spin_unlock(&delayed_root->lock); | ||
227 | |||
228 | return next; | ||
229 | } | ||
230 | |||
231 | static void __btrfs_release_delayed_node( | ||
232 | struct btrfs_delayed_node *delayed_node, | ||
233 | int mod) | ||
234 | { | ||
235 | struct btrfs_delayed_root *delayed_root; | ||
236 | |||
237 | if (!delayed_node) | ||
238 | return; | ||
239 | |||
240 | delayed_root = delayed_node->root->fs_info->delayed_root; | ||
241 | |||
242 | mutex_lock(&delayed_node->mutex); | ||
243 | if (delayed_node->count) | ||
244 | btrfs_queue_delayed_node(delayed_root, delayed_node, mod); | ||
245 | else | ||
246 | btrfs_dequeue_delayed_node(delayed_root, delayed_node); | ||
247 | mutex_unlock(&delayed_node->mutex); | ||
248 | |||
249 | if (atomic_dec_and_test(&delayed_node->refs)) { | ||
250 | struct btrfs_root *root = delayed_node->root; | ||
251 | spin_lock(&root->inode_lock); | ||
252 | if (atomic_read(&delayed_node->refs) == 0) { | ||
253 | radix_tree_delete(&root->delayed_nodes_tree, | ||
254 | delayed_node->inode_id); | ||
255 | kmem_cache_free(delayed_node_cache, delayed_node); | ||
256 | } | ||
257 | spin_unlock(&root->inode_lock); | ||
258 | } | ||
259 | } | ||
260 | |||
261 | static inline void btrfs_release_delayed_node(struct btrfs_delayed_node *node) | ||
262 | { | ||
263 | __btrfs_release_delayed_node(node, 0); | ||
264 | } | ||
265 | |||
266 | struct btrfs_delayed_node *btrfs_first_prepared_delayed_node( | ||
267 | struct btrfs_delayed_root *delayed_root) | ||
268 | { | ||
269 | struct list_head *p; | ||
270 | struct btrfs_delayed_node *node = NULL; | ||
271 | |||
272 | spin_lock(&delayed_root->lock); | ||
273 | if (list_empty(&delayed_root->prepare_list)) | ||
274 | goto out; | ||
275 | |||
276 | p = delayed_root->prepare_list.next; | ||
277 | list_del_init(p); | ||
278 | node = list_entry(p, struct btrfs_delayed_node, p_list); | ||
279 | atomic_inc(&node->refs); | ||
280 | out: | ||
281 | spin_unlock(&delayed_root->lock); | ||
282 | |||
283 | return node; | ||
284 | } | ||
285 | |||
286 | static inline void btrfs_release_prepared_delayed_node( | ||
287 | struct btrfs_delayed_node *node) | ||
288 | { | ||
289 | __btrfs_release_delayed_node(node, 1); | ||
290 | } | ||
291 | |||
292 | struct btrfs_delayed_item *btrfs_alloc_delayed_item(u32 data_len) | ||
293 | { | ||
294 | struct btrfs_delayed_item *item; | ||
295 | item = kmalloc(sizeof(*item) + data_len, GFP_NOFS); | ||
296 | if (item) { | ||
297 | item->data_len = data_len; | ||
298 | item->ins_or_del = 0; | ||
299 | item->bytes_reserved = 0; | ||
300 | item->block_rsv = NULL; | ||
301 | item->delayed_node = NULL; | ||
302 | atomic_set(&item->refs, 1); | ||
303 | } | ||
304 | return item; | ||
305 | } | ||
306 | |||
307 | /* | ||
308 | * __btrfs_lookup_delayed_item - look up the delayed item by key | ||
309 | * @delayed_node: pointer to the delayed node | ||
310 | * @key: the key to look up | ||
311 | * @prev: used to store the prev item if the right item isn't found | ||
312 | * @next: used to store the next item if the right item isn't found | ||
313 | * | ||
314 | * Note: if we don't find the right item, we will return the prev item and | ||
315 | * the next item. | ||
316 | */ | ||
317 | static struct btrfs_delayed_item *__btrfs_lookup_delayed_item( | ||
318 | struct rb_root *root, | ||
319 | struct btrfs_key *key, | ||
320 | struct btrfs_delayed_item **prev, | ||
321 | struct btrfs_delayed_item **next) | ||
322 | { | ||
323 | struct rb_node *node, *prev_node = NULL; | ||
324 | struct btrfs_delayed_item *delayed_item = NULL; | ||
325 | int ret = 0; | ||
326 | |||
327 | node = root->rb_node; | ||
328 | |||
329 | while (node) { | ||
330 | delayed_item = rb_entry(node, struct btrfs_delayed_item, | ||
331 | rb_node); | ||
332 | prev_node = node; | ||
333 | ret = btrfs_comp_cpu_keys(&delayed_item->key, key); | ||
334 | if (ret < 0) | ||
335 | node = node->rb_right; | ||
336 | else if (ret > 0) | ||
337 | node = node->rb_left; | ||
338 | else | ||
339 | return delayed_item; | ||
340 | } | ||
341 | |||
342 | if (prev) { | ||
343 | if (!prev_node) | ||
344 | *prev = NULL; | ||
345 | else if (ret < 0) | ||
346 | *prev = delayed_item; | ||
347 | else if ((node = rb_prev(prev_node)) != NULL) { | ||
348 | *prev = rb_entry(node, struct btrfs_delayed_item, | ||
349 | rb_node); | ||
350 | } else | ||
351 | *prev = NULL; | ||
352 | } | ||
353 | |||
354 | if (next) { | ||
355 | if (!prev_node) | ||
356 | *next = NULL; | ||
357 | else if (ret > 0) | ||
358 | *next = delayed_item; | ||
359 | else if ((node = rb_next(prev_node)) != NULL) { | ||
360 | *next = rb_entry(node, struct btrfs_delayed_item, | ||
361 | rb_node); | ||
362 | } else | ||
363 | *next = NULL; | ||
364 | } | ||
365 | return NULL; | ||
366 | } | ||
367 | |||
368 | struct btrfs_delayed_item *__btrfs_lookup_delayed_insertion_item( | ||
369 | struct btrfs_delayed_node *delayed_node, | ||
370 | struct btrfs_key *key) | ||
371 | { | ||
372 | struct btrfs_delayed_item *item; | ||
373 | |||
374 | item = __btrfs_lookup_delayed_item(&delayed_node->ins_root, key, | ||
375 | NULL, NULL); | ||
376 | return item; | ||
377 | } | ||
378 | |||
379 | struct btrfs_delayed_item *__btrfs_lookup_delayed_deletion_item( | ||
380 | struct btrfs_delayed_node *delayed_node, | ||
381 | struct btrfs_key *key) | ||
382 | { | ||
383 | struct btrfs_delayed_item *item; | ||
384 | |||
385 | item = __btrfs_lookup_delayed_item(&delayed_node->del_root, key, | ||
386 | NULL, NULL); | ||
387 | return item; | ||
388 | } | ||
389 | |||
390 | struct btrfs_delayed_item *__btrfs_search_delayed_insertion_item( | ||
391 | struct btrfs_delayed_node *delayed_node, | ||
392 | struct btrfs_key *key) | ||
393 | { | ||
394 | struct btrfs_delayed_item *item, *next; | ||
395 | |||
396 | item = __btrfs_lookup_delayed_item(&delayed_node->ins_root, key, | ||
397 | NULL, &next); | ||
398 | if (!item) | ||
399 | item = next; | ||
400 | |||
401 | return item; | ||
402 | } | ||
403 | |||
404 | struct btrfs_delayed_item *__btrfs_search_delayed_deletion_item( | ||
405 | struct btrfs_delayed_node *delayed_node, | ||
406 | struct btrfs_key *key) | ||
407 | { | ||
408 | struct btrfs_delayed_item *item, *next; | ||
409 | |||
410 | item = __btrfs_lookup_delayed_item(&delayed_node->del_root, key, | ||
411 | NULL, &next); | ||
412 | if (!item) | ||
413 | item = next; | ||
414 | |||
415 | return item; | ||
416 | } | ||
417 | |||
418 | static int __btrfs_add_delayed_item(struct btrfs_delayed_node *delayed_node, | ||
419 | struct btrfs_delayed_item *ins, | ||
420 | int action) | ||
421 | { | ||
422 | struct rb_node **p, *node; | ||
423 | struct rb_node *parent_node = NULL; | ||
424 | struct rb_root *root; | ||
425 | struct btrfs_delayed_item *item; | ||
426 | int cmp; | ||
427 | |||
428 | if (action == BTRFS_DELAYED_INSERTION_ITEM) | ||
429 | root = &delayed_node->ins_root; | ||
430 | else if (action == BTRFS_DELAYED_DELETION_ITEM) | ||
431 | root = &delayed_node->del_root; | ||
432 | else | ||
433 | BUG(); | ||
434 | p = &root->rb_node; | ||
435 | node = &ins->rb_node; | ||
436 | |||
437 | while (*p) { | ||
438 | parent_node = *p; | ||
439 | item = rb_entry(parent_node, struct btrfs_delayed_item, | ||
440 | rb_node); | ||
441 | |||
442 | cmp = btrfs_comp_cpu_keys(&item->key, &ins->key); | ||
443 | if (cmp < 0) | ||
444 | p = &(*p)->rb_right; | ||
445 | else if (cmp > 0) | ||
446 | p = &(*p)->rb_left; | ||
447 | else | ||
448 | return -EEXIST; | ||
449 | } | ||
450 | |||
451 | rb_link_node(node, parent_node, p); | ||
452 | rb_insert_color(node, root); | ||
453 | ins->delayed_node = delayed_node; | ||
454 | ins->ins_or_del = action; | ||
455 | |||
456 | if (ins->key.type == BTRFS_DIR_INDEX_KEY && | ||
457 | action == BTRFS_DELAYED_INSERTION_ITEM && | ||
458 | ins->key.offset >= delayed_node->index_cnt) | ||
459 | delayed_node->index_cnt = ins->key.offset + 1; | ||
460 | |||
461 | delayed_node->count++; | ||
462 | atomic_inc(&delayed_node->root->fs_info->delayed_root->items); | ||
463 | return 0; | ||
464 | } | ||
465 | |||
466 | static int __btrfs_add_delayed_insertion_item(struct btrfs_delayed_node *node, | ||
467 | struct btrfs_delayed_item *item) | ||
468 | { | ||
469 | return __btrfs_add_delayed_item(node, item, | ||
470 | BTRFS_DELAYED_INSERTION_ITEM); | ||
471 | } | ||
472 | |||
473 | static int __btrfs_add_delayed_deletion_item(struct btrfs_delayed_node *node, | ||
474 | struct btrfs_delayed_item *item) | ||
475 | { | ||
476 | return __btrfs_add_delayed_item(node, item, | ||
477 | BTRFS_DELAYED_DELETION_ITEM); | ||
478 | } | ||
479 | |||
480 | static void __btrfs_remove_delayed_item(struct btrfs_delayed_item *delayed_item) | ||
481 | { | ||
482 | struct rb_root *root; | ||
483 | struct btrfs_delayed_root *delayed_root; | ||
484 | |||
485 | delayed_root = delayed_item->delayed_node->root->fs_info->delayed_root; | ||
486 | |||
487 | BUG_ON(!delayed_root); | ||
488 | BUG_ON(delayed_item->ins_or_del != BTRFS_DELAYED_DELETION_ITEM && | ||
489 | delayed_item->ins_or_del != BTRFS_DELAYED_INSERTION_ITEM); | ||
490 | |||
491 | if (delayed_item->ins_or_del == BTRFS_DELAYED_INSERTION_ITEM) | ||
492 | root = &delayed_item->delayed_node->ins_root; | ||
493 | else | ||
494 | root = &delayed_item->delayed_node->del_root; | ||
495 | |||
496 | rb_erase(&delayed_item->rb_node, root); | ||
497 | delayed_item->delayed_node->count--; | ||
498 | atomic_dec(&delayed_root->items); | ||
499 | if (atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND && | ||
500 | waitqueue_active(&delayed_root->wait)) | ||
501 | wake_up(&delayed_root->wait); | ||
502 | } | ||
503 | |||
504 | static void btrfs_release_delayed_item(struct btrfs_delayed_item *item) | ||
505 | { | ||
506 | if (item) { | ||
507 | __btrfs_remove_delayed_item(item); | ||
508 | if (atomic_dec_and_test(&item->refs)) | ||
509 | kfree(item); | ||
510 | } | ||
511 | } | ||
512 | |||
513 | struct btrfs_delayed_item *__btrfs_first_delayed_insertion_item( | ||
514 | struct btrfs_delayed_node *delayed_node) | ||
515 | { | ||
516 | struct rb_node *p; | ||
517 | struct btrfs_delayed_item *item = NULL; | ||
518 | |||
519 | p = rb_first(&delayed_node->ins_root); | ||
520 | if (p) | ||
521 | item = rb_entry(p, struct btrfs_delayed_item, rb_node); | ||
522 | |||
523 | return item; | ||
524 | } | ||
525 | |||
526 | struct btrfs_delayed_item *__btrfs_first_delayed_deletion_item( | ||
527 | struct btrfs_delayed_node *delayed_node) | ||
528 | { | ||
529 | struct rb_node *p; | ||
530 | struct btrfs_delayed_item *item = NULL; | ||
531 | |||
532 | p = rb_first(&delayed_node->del_root); | ||
533 | if (p) | ||
534 | item = rb_entry(p, struct btrfs_delayed_item, rb_node); | ||
535 | |||
536 | return item; | ||
537 | } | ||
538 | |||
539 | struct btrfs_delayed_item *__btrfs_next_delayed_item( | ||
540 | struct btrfs_delayed_item *item) | ||
541 | { | ||
542 | struct rb_node *p; | ||
543 | struct btrfs_delayed_item *next = NULL; | ||
544 | |||
545 | p = rb_next(&item->rb_node); | ||
546 | if (p) | ||
547 | next = rb_entry(p, struct btrfs_delayed_item, rb_node); | ||
548 | |||
549 | return next; | ||
550 | } | ||
551 | |||
552 | static inline struct btrfs_delayed_node *btrfs_get_delayed_node( | ||
553 | struct inode *inode) | ||
554 | { | ||
555 | struct btrfs_inode *btrfs_inode = BTRFS_I(inode); | ||
556 | struct btrfs_delayed_node *delayed_node; | ||
557 | |||
558 | delayed_node = btrfs_inode->delayed_node; | ||
559 | if (delayed_node) | ||
560 | atomic_inc(&delayed_node->refs); | ||
561 | |||
562 | return delayed_node; | ||
563 | } | ||
564 | |||
565 | static inline struct btrfs_root *btrfs_get_fs_root(struct btrfs_root *root, | ||
566 | u64 root_id) | ||
567 | { | ||
568 | struct btrfs_key root_key; | ||
569 | |||
570 | if (root->objectid == root_id) | ||
571 | return root; | ||
572 | |||
573 | root_key.objectid = root_id; | ||
574 | root_key.type = BTRFS_ROOT_ITEM_KEY; | ||
575 | root_key.offset = (u64)-1; | ||
576 | return btrfs_read_fs_root_no_name(root->fs_info, &root_key); | ||
577 | } | ||
578 | |||
579 | static int btrfs_delayed_item_reserve_metadata(struct btrfs_trans_handle *trans, | ||
580 | struct btrfs_root *root, | ||
581 | struct btrfs_delayed_item *item) | ||
582 | { | ||
583 | struct btrfs_block_rsv *src_rsv; | ||
584 | struct btrfs_block_rsv *dst_rsv; | ||
585 | u64 num_bytes; | ||
586 | int ret; | ||
587 | |||
588 | if (!trans->bytes_reserved) | ||
589 | return 0; | ||
590 | |||
591 | src_rsv = trans->block_rsv; | ||
592 | dst_rsv = &root->fs_info->global_block_rsv; | ||
593 | |||
594 | num_bytes = btrfs_calc_trans_metadata_size(root, 1); | ||
595 | ret = btrfs_block_rsv_migrate(src_rsv, dst_rsv, num_bytes); | ||
596 | if (!ret) { | ||
597 | item->bytes_reserved = num_bytes; | ||
598 | item->block_rsv = dst_rsv; | ||
599 | } | ||
600 | |||
601 | return ret; | ||
602 | } | ||
603 | |||
604 | static void btrfs_delayed_item_release_metadata(struct btrfs_root *root, | ||
605 | struct btrfs_delayed_item *item) | ||
606 | { | ||
607 | if (!item->bytes_reserved) | ||
608 | return; | ||
609 | |||
610 | btrfs_block_rsv_release(root, item->block_rsv, | ||
611 | item->bytes_reserved); | ||
612 | } | ||
613 | |||
614 | static int btrfs_delayed_inode_reserve_metadata( | ||
615 | struct btrfs_trans_handle *trans, | ||
616 | struct btrfs_root *root, | ||
617 | struct btrfs_delayed_node *node) | ||
618 | { | ||
619 | struct btrfs_block_rsv *src_rsv; | ||
620 | struct btrfs_block_rsv *dst_rsv; | ||
621 | u64 num_bytes; | ||
622 | int ret; | ||
623 | |||
624 | if (!trans->bytes_reserved) | ||
625 | return 0; | ||
626 | |||
627 | src_rsv = trans->block_rsv; | ||
628 | dst_rsv = &root->fs_info->global_block_rsv; | ||
629 | |||
630 | num_bytes = btrfs_calc_trans_metadata_size(root, 1); | ||
631 | ret = btrfs_block_rsv_migrate(src_rsv, dst_rsv, num_bytes); | ||
632 | if (!ret) | ||
633 | node->bytes_reserved = num_bytes; | ||
634 | |||
635 | return ret; | ||
636 | } | ||
637 | |||
638 | static void btrfs_delayed_inode_release_metadata(struct btrfs_root *root, | ||
639 | struct btrfs_delayed_node *node) | ||
640 | { | ||
641 | struct btrfs_block_rsv *rsv; | ||
642 | |||
643 | if (!node->bytes_reserved) | ||
644 | return; | ||
645 | |||
646 | rsv = &root->fs_info->global_block_rsv; | ||
647 | btrfs_block_rsv_release(root, rsv, | ||
648 | node->bytes_reserved); | ||
649 | node->bytes_reserved = 0; | ||
650 | } | ||
651 | |||
652 | /* | ||
653 | * This helper will insert some continuous items into the same leaf according | ||
654 | * to the free space of the leaf. | ||
655 | */ | ||
656 | static int btrfs_batch_insert_items(struct btrfs_trans_handle *trans, | ||
657 | struct btrfs_root *root, | ||
658 | struct btrfs_path *path, | ||
659 | struct btrfs_delayed_item *item) | ||
660 | { | ||
661 | struct btrfs_delayed_item *curr, *next; | ||
662 | int free_space; | ||
663 | int total_data_size = 0, total_size = 0; | ||
664 | struct extent_buffer *leaf; | ||
665 | char *data_ptr; | ||
666 | struct btrfs_key *keys; | ||
667 | u32 *data_size; | ||
668 | struct list_head head; | ||
669 | int slot; | ||
670 | int nitems; | ||
671 | int i; | ||
672 | int ret = 0; | ||
673 | |||
674 | BUG_ON(!path->nodes[0]); | ||
675 | |||
676 | leaf = path->nodes[0]; | ||
677 | free_space = btrfs_leaf_free_space(root, leaf); | ||
678 | INIT_LIST_HEAD(&head); | ||
679 | |||
680 | next = item; | ||
681 | |||
682 | /* | ||
683 | * count the number of the continuous items that we can insert in batch | ||
684 | */ | ||
685 | while (total_size + next->data_len + sizeof(struct btrfs_item) <= | ||
686 | free_space) { | ||
687 | total_data_size += next->data_len; | ||
688 | total_size += next->data_len + sizeof(struct btrfs_item); | ||
689 | list_add_tail(&next->tree_list, &head); | ||
690 | nitems++; | ||
691 | |||
692 | curr = next; | ||
693 | next = __btrfs_next_delayed_item(curr); | ||
694 | if (!next) | ||
695 | break; | ||
696 | |||
697 | if (!btrfs_is_continuous_delayed_item(curr, next)) | ||
698 | break; | ||
699 | } | ||
700 | |||
701 | if (!nitems) { | ||
702 | ret = 0; | ||
703 | goto out; | ||
704 | } | ||
705 | |||
706 | /* | ||
707 | * we need allocate some memory space, but it might cause the task | ||
708 | * to sleep, so we set all locked nodes in the path to blocking locks | ||
709 | * first. | ||
710 | */ | ||
711 | btrfs_set_path_blocking(path); | ||
712 | |||
713 | keys = kmalloc(sizeof(struct btrfs_key) * nitems, GFP_NOFS); | ||
714 | if (!keys) { | ||
715 | ret = -ENOMEM; | ||
716 | goto out; | ||
717 | } | ||
718 | |||
719 | data_size = kmalloc(sizeof(u32) * nitems, GFP_NOFS); | ||
720 | if (!data_size) { | ||
721 | ret = -ENOMEM; | ||
722 | goto error; | ||
723 | } | ||
724 | |||
725 | /* get keys of all the delayed items */ | ||
726 | i = 0; | ||
727 | list_for_each_entry(next, &head, tree_list) { | ||
728 | keys[i] = next->key; | ||
729 | data_size[i] = next->data_len; | ||
730 | i++; | ||
731 | } | ||
732 | |||
733 | /* reset all the locked nodes in the patch to spinning locks. */ | ||
734 | btrfs_clear_path_blocking(path, NULL); | ||
735 | |||
736 | /* insert the keys of the items */ | ||
737 | ret = setup_items_for_insert(trans, root, path, keys, data_size, | ||
738 | total_data_size, total_size, nitems); | ||
739 | if (ret) | ||
740 | goto error; | ||
741 | |||
742 | /* insert the dir index items */ | ||
743 | slot = path->slots[0]; | ||
744 | list_for_each_entry_safe(curr, next, &head, tree_list) { | ||
745 | data_ptr = btrfs_item_ptr(leaf, slot, char); | ||
746 | write_extent_buffer(leaf, &curr->data, | ||
747 | (unsigned long)data_ptr, | ||
748 | curr->data_len); | ||
749 | slot++; | ||
750 | |||
751 | btrfs_delayed_item_release_metadata(root, curr); | ||
752 | |||
753 | list_del(&curr->tree_list); | ||
754 | btrfs_release_delayed_item(curr); | ||
755 | } | ||
756 | |||
757 | error: | ||
758 | kfree(data_size); | ||
759 | kfree(keys); | ||
760 | out: | ||
761 | return ret; | ||
762 | } | ||
763 | |||
764 | /* | ||
765 | * This helper can just do simple insertion that needn't extend item for new | ||
766 | * data, such as directory name index insertion, inode insertion. | ||
767 | */ | ||
768 | static int btrfs_insert_delayed_item(struct btrfs_trans_handle *trans, | ||
769 | struct btrfs_root *root, | ||
770 | struct btrfs_path *path, | ||
771 | struct btrfs_delayed_item *delayed_item) | ||
772 | { | ||
773 | struct extent_buffer *leaf; | ||
774 | struct btrfs_item *item; | ||
775 | char *ptr; | ||
776 | int ret; | ||
777 | |||
778 | ret = btrfs_insert_empty_item(trans, root, path, &delayed_item->key, | ||
779 | delayed_item->data_len); | ||
780 | if (ret < 0 && ret != -EEXIST) | ||
781 | return ret; | ||
782 | |||
783 | leaf = path->nodes[0]; | ||
784 | |||
785 | item = btrfs_item_nr(leaf, path->slots[0]); | ||
786 | ptr = btrfs_item_ptr(leaf, path->slots[0], char); | ||
787 | |||
788 | write_extent_buffer(leaf, delayed_item->data, (unsigned long)ptr, | ||
789 | delayed_item->data_len); | ||
790 | btrfs_mark_buffer_dirty(leaf); | ||
791 | |||
792 | btrfs_delayed_item_release_metadata(root, delayed_item); | ||
793 | return 0; | ||
794 | } | ||
795 | |||
796 | /* | ||
797 | * we insert an item first, then if there are some continuous items, we try | ||
798 | * to insert those items into the same leaf. | ||
799 | */ | ||
800 | static int btrfs_insert_delayed_items(struct btrfs_trans_handle *trans, | ||
801 | struct btrfs_path *path, | ||
802 | struct btrfs_root *root, | ||
803 | struct btrfs_delayed_node *node) | ||
804 | { | ||
805 | struct btrfs_delayed_item *curr, *prev; | ||
806 | int ret = 0; | ||
807 | |||
808 | do_again: | ||
809 | mutex_lock(&node->mutex); | ||
810 | curr = __btrfs_first_delayed_insertion_item(node); | ||
811 | if (!curr) | ||
812 | goto insert_end; | ||
813 | |||
814 | ret = btrfs_insert_delayed_item(trans, root, path, curr); | ||
815 | if (ret < 0) { | ||
816 | btrfs_release_path(path); | ||
817 | goto insert_end; | ||
818 | } | ||
819 | |||
820 | prev = curr; | ||
821 | curr = __btrfs_next_delayed_item(prev); | ||
822 | if (curr && btrfs_is_continuous_delayed_item(prev, curr)) { | ||
823 | /* insert the continuous items into the same leaf */ | ||
824 | path->slots[0]++; | ||
825 | btrfs_batch_insert_items(trans, root, path, curr); | ||
826 | } | ||
827 | btrfs_release_delayed_item(prev); | ||
828 | btrfs_mark_buffer_dirty(path->nodes[0]); | ||
829 | |||
830 | btrfs_release_path(path); | ||
831 | mutex_unlock(&node->mutex); | ||
832 | goto do_again; | ||
833 | |||
834 | insert_end: | ||
835 | mutex_unlock(&node->mutex); | ||
836 | return ret; | ||
837 | } | ||
838 | |||
839 | static int btrfs_batch_delete_items(struct btrfs_trans_handle *trans, | ||
840 | struct btrfs_root *root, | ||
841 | struct btrfs_path *path, | ||
842 | struct btrfs_delayed_item *item) | ||
843 | { | ||
844 | struct btrfs_delayed_item *curr, *next; | ||
845 | struct extent_buffer *leaf; | ||
846 | struct btrfs_key key; | ||
847 | struct list_head head; | ||
848 | int nitems, i, last_item; | ||
849 | int ret = 0; | ||
850 | |||
851 | BUG_ON(!path->nodes[0]); | ||
852 | |||
853 | leaf = path->nodes[0]; | ||
854 | |||
855 | i = path->slots[0]; | ||
856 | last_item = btrfs_header_nritems(leaf) - 1; | ||
857 | if (i > last_item) | ||
858 | return -ENOENT; /* FIXME: Is errno suitable? */ | ||
859 | |||
860 | next = item; | ||
861 | INIT_LIST_HEAD(&head); | ||
862 | btrfs_item_key_to_cpu(leaf, &key, i); | ||
863 | nitems = 0; | ||
864 | /* | ||
865 | * count the number of the dir index items that we can delete in batch | ||
866 | */ | ||
867 | while (btrfs_comp_cpu_keys(&next->key, &key) == 0) { | ||
868 | list_add_tail(&next->tree_list, &head); | ||
869 | nitems++; | ||
870 | |||
871 | curr = next; | ||
872 | next = __btrfs_next_delayed_item(curr); | ||
873 | if (!next) | ||
874 | break; | ||
875 | |||
876 | if (!btrfs_is_continuous_delayed_item(curr, next)) | ||
877 | break; | ||
878 | |||
879 | i++; | ||
880 | if (i > last_item) | ||
881 | break; | ||
882 | btrfs_item_key_to_cpu(leaf, &key, i); | ||
883 | } | ||
884 | |||
885 | if (!nitems) | ||
886 | return 0; | ||
887 | |||
888 | ret = btrfs_del_items(trans, root, path, path->slots[0], nitems); | ||
889 | if (ret) | ||
890 | goto out; | ||
891 | |||
892 | list_for_each_entry_safe(curr, next, &head, tree_list) { | ||
893 | btrfs_delayed_item_release_metadata(root, curr); | ||
894 | list_del(&curr->tree_list); | ||
895 | btrfs_release_delayed_item(curr); | ||
896 | } | ||
897 | |||
898 | out: | ||
899 | return ret; | ||
900 | } | ||
901 | |||
902 | static int btrfs_delete_delayed_items(struct btrfs_trans_handle *trans, | ||
903 | struct btrfs_path *path, | ||
904 | struct btrfs_root *root, | ||
905 | struct btrfs_delayed_node *node) | ||
906 | { | ||
907 | struct btrfs_delayed_item *curr, *prev; | ||
908 | int ret = 0; | ||
909 | |||
910 | do_again: | ||
911 | mutex_lock(&node->mutex); | ||
912 | curr = __btrfs_first_delayed_deletion_item(node); | ||
913 | if (!curr) | ||
914 | goto delete_fail; | ||
915 | |||
916 | ret = btrfs_search_slot(trans, root, &curr->key, path, -1, 1); | ||
917 | if (ret < 0) | ||
918 | goto delete_fail; | ||
919 | else if (ret > 0) { | ||
920 | /* | ||
921 | * can't find the item which the node points to, so this node | ||
922 | * is invalid, just drop it. | ||
923 | */ | ||
924 | prev = curr; | ||
925 | curr = __btrfs_next_delayed_item(prev); | ||
926 | btrfs_release_delayed_item(prev); | ||
927 | ret = 0; | ||
928 | btrfs_release_path(path); | ||
929 | if (curr) | ||
930 | goto do_again; | ||
931 | else | ||
932 | goto delete_fail; | ||
933 | } | ||
934 | |||
935 | btrfs_batch_delete_items(trans, root, path, curr); | ||
936 | btrfs_release_path(path); | ||
937 | mutex_unlock(&node->mutex); | ||
938 | goto do_again; | ||
939 | |||
940 | delete_fail: | ||
941 | btrfs_release_path(path); | ||
942 | mutex_unlock(&node->mutex); | ||
943 | return ret; | ||
944 | } | ||
945 | |||
946 | static void btrfs_release_delayed_inode(struct btrfs_delayed_node *delayed_node) | ||
947 | { | ||
948 | struct btrfs_delayed_root *delayed_root; | ||
949 | |||
950 | if (delayed_node && delayed_node->inode_dirty) { | ||
951 | BUG_ON(!delayed_node->root); | ||
952 | delayed_node->inode_dirty = 0; | ||
953 | delayed_node->count--; | ||
954 | |||
955 | delayed_root = delayed_node->root->fs_info->delayed_root; | ||
956 | atomic_dec(&delayed_root->items); | ||
957 | if (atomic_read(&delayed_root->items) < | ||
958 | BTRFS_DELAYED_BACKGROUND && | ||
959 | waitqueue_active(&delayed_root->wait)) | ||
960 | wake_up(&delayed_root->wait); | ||
961 | } | ||
962 | } | ||
963 | |||
964 | static int btrfs_update_delayed_inode(struct btrfs_trans_handle *trans, | ||
965 | struct btrfs_root *root, | ||
966 | struct btrfs_path *path, | ||
967 | struct btrfs_delayed_node *node) | ||
968 | { | ||
969 | struct btrfs_key key; | ||
970 | struct btrfs_inode_item *inode_item; | ||
971 | struct extent_buffer *leaf; | ||
972 | int ret; | ||
973 | |||
974 | mutex_lock(&node->mutex); | ||
975 | if (!node->inode_dirty) { | ||
976 | mutex_unlock(&node->mutex); | ||
977 | return 0; | ||
978 | } | ||
979 | |||
980 | key.objectid = node->inode_id; | ||
981 | btrfs_set_key_type(&key, BTRFS_INODE_ITEM_KEY); | ||
982 | key.offset = 0; | ||
983 | ret = btrfs_lookup_inode(trans, root, path, &key, 1); | ||
984 | if (ret > 0) { | ||
985 | btrfs_release_path(path); | ||
986 | mutex_unlock(&node->mutex); | ||
987 | return -ENOENT; | ||
988 | } else if (ret < 0) { | ||
989 | mutex_unlock(&node->mutex); | ||
990 | return ret; | ||
991 | } | ||
992 | |||
993 | btrfs_unlock_up_safe(path, 1); | ||
994 | leaf = path->nodes[0]; | ||
995 | inode_item = btrfs_item_ptr(leaf, path->slots[0], | ||
996 | struct btrfs_inode_item); | ||
997 | write_extent_buffer(leaf, &node->inode_item, (unsigned long)inode_item, | ||
998 | sizeof(struct btrfs_inode_item)); | ||
999 | btrfs_mark_buffer_dirty(leaf); | ||
1000 | btrfs_release_path(path); | ||
1001 | |||
1002 | btrfs_delayed_inode_release_metadata(root, node); | ||
1003 | btrfs_release_delayed_inode(node); | ||
1004 | mutex_unlock(&node->mutex); | ||
1005 | |||
1006 | return 0; | ||
1007 | } | ||
1008 | |||
1009 | /* Called when committing the transaction. */ | ||
1010 | int btrfs_run_delayed_items(struct btrfs_trans_handle *trans, | ||
1011 | struct btrfs_root *root) | ||
1012 | { | ||
1013 | struct btrfs_delayed_root *delayed_root; | ||
1014 | struct btrfs_delayed_node *curr_node, *prev_node; | ||
1015 | struct btrfs_path *path; | ||
1016 | int ret = 0; | ||
1017 | |||
1018 | path = btrfs_alloc_path(); | ||
1019 | if (!path) | ||
1020 | return -ENOMEM; | ||
1021 | path->leave_spinning = 1; | ||
1022 | |||
1023 | delayed_root = btrfs_get_delayed_root(root); | ||
1024 | |||
1025 | curr_node = btrfs_first_delayed_node(delayed_root); | ||
1026 | while (curr_node) { | ||
1027 | root = curr_node->root; | ||
1028 | ret = btrfs_insert_delayed_items(trans, path, root, | ||
1029 | curr_node); | ||
1030 | if (!ret) | ||
1031 | ret = btrfs_delete_delayed_items(trans, path, root, | ||
1032 | curr_node); | ||
1033 | if (!ret) | ||
1034 | ret = btrfs_update_delayed_inode(trans, root, path, | ||
1035 | curr_node); | ||
1036 | if (ret) { | ||
1037 | btrfs_release_delayed_node(curr_node); | ||
1038 | break; | ||
1039 | } | ||
1040 | |||
1041 | prev_node = curr_node; | ||
1042 | curr_node = btrfs_next_delayed_node(curr_node); | ||
1043 | btrfs_release_delayed_node(prev_node); | ||
1044 | } | ||
1045 | |||
1046 | btrfs_free_path(path); | ||
1047 | return ret; | ||
1048 | } | ||
1049 | |||
1050 | static int __btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans, | ||
1051 | struct btrfs_delayed_node *node) | ||
1052 | { | ||
1053 | struct btrfs_path *path; | ||
1054 | int ret; | ||
1055 | |||
1056 | path = btrfs_alloc_path(); | ||
1057 | if (!path) | ||
1058 | return -ENOMEM; | ||
1059 | path->leave_spinning = 1; | ||
1060 | |||
1061 | ret = btrfs_insert_delayed_items(trans, path, node->root, node); | ||
1062 | if (!ret) | ||
1063 | ret = btrfs_delete_delayed_items(trans, path, node->root, node); | ||
1064 | if (!ret) | ||
1065 | ret = btrfs_update_delayed_inode(trans, node->root, path, node); | ||
1066 | btrfs_free_path(path); | ||
1067 | |||
1068 | return ret; | ||
1069 | } | ||
1070 | |||
1071 | int btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans, | ||
1072 | struct inode *inode) | ||
1073 | { | ||
1074 | struct btrfs_delayed_node *delayed_node = btrfs_get_delayed_node(inode); | ||
1075 | int ret; | ||
1076 | |||
1077 | if (!delayed_node) | ||
1078 | return 0; | ||
1079 | |||
1080 | mutex_lock(&delayed_node->mutex); | ||
1081 | if (!delayed_node->count) { | ||
1082 | mutex_unlock(&delayed_node->mutex); | ||
1083 | btrfs_release_delayed_node(delayed_node); | ||
1084 | return 0; | ||
1085 | } | ||
1086 | mutex_unlock(&delayed_node->mutex); | ||
1087 | |||
1088 | ret = __btrfs_commit_inode_delayed_items(trans, delayed_node); | ||
1089 | btrfs_release_delayed_node(delayed_node); | ||
1090 | return ret; | ||
1091 | } | ||
1092 | |||
1093 | void btrfs_remove_delayed_node(struct inode *inode) | ||
1094 | { | ||
1095 | struct btrfs_delayed_node *delayed_node; | ||
1096 | |||
1097 | delayed_node = ACCESS_ONCE(BTRFS_I(inode)->delayed_node); | ||
1098 | if (!delayed_node) | ||
1099 | return; | ||
1100 | |||
1101 | BTRFS_I(inode)->delayed_node = NULL; | ||
1102 | btrfs_release_delayed_node(delayed_node); | ||
1103 | } | ||
1104 | |||
1105 | struct btrfs_async_delayed_node { | ||
1106 | struct btrfs_root *root; | ||
1107 | struct btrfs_delayed_node *delayed_node; | ||
1108 | struct btrfs_work work; | ||
1109 | }; | ||
1110 | |||
1111 | static void btrfs_async_run_delayed_node_done(struct btrfs_work *work) | ||
1112 | { | ||
1113 | struct btrfs_async_delayed_node *async_node; | ||
1114 | struct btrfs_trans_handle *trans; | ||
1115 | struct btrfs_path *path; | ||
1116 | struct btrfs_delayed_node *delayed_node = NULL; | ||
1117 | struct btrfs_root *root; | ||
1118 | unsigned long nr = 0; | ||
1119 | int need_requeue = 0; | ||
1120 | int ret; | ||
1121 | |||
1122 | async_node = container_of(work, struct btrfs_async_delayed_node, work); | ||
1123 | |||
1124 | path = btrfs_alloc_path(); | ||
1125 | if (!path) | ||
1126 | goto out; | ||
1127 | path->leave_spinning = 1; | ||
1128 | |||
1129 | delayed_node = async_node->delayed_node; | ||
1130 | root = delayed_node->root; | ||
1131 | |||
1132 | trans = btrfs_join_transaction(root, 0); | ||
1133 | if (IS_ERR(trans)) | ||
1134 | goto free_path; | ||
1135 | |||
1136 | ret = btrfs_insert_delayed_items(trans, path, root, delayed_node); | ||
1137 | if (!ret) | ||
1138 | ret = btrfs_delete_delayed_items(trans, path, root, | ||
1139 | delayed_node); | ||
1140 | |||
1141 | if (!ret) | ||
1142 | btrfs_update_delayed_inode(trans, root, path, delayed_node); | ||
1143 | |||
1144 | /* | ||
1145 | * Maybe new delayed items have been inserted, so we need requeue | ||
1146 | * the work. Besides that, we must dequeue the empty delayed nodes | ||
1147 | * to avoid the race between delayed items balance and the worker. | ||
1148 | * The race like this: | ||
1149 | * Task1 Worker thread | ||
1150 | * count == 0, needn't requeue | ||
1151 | * also needn't insert the | ||
1152 | * delayed node into prepare | ||
1153 | * list again. | ||
1154 | * add lots of delayed items | ||
1155 | * queue the delayed node | ||
1156 | * already in the list, | ||
1157 | * and not in the prepare | ||
1158 | * list, it means the delayed | ||
1159 | * node is being dealt with | ||
1160 | * by the worker. | ||
1161 | * do delayed items balance | ||
1162 | * the delayed node is being | ||
1163 | * dealt with by the worker | ||
1164 | * now, just wait. | ||
1165 | * the worker goto idle. | ||
1166 | * Task1 will sleep until the transaction is commited. | ||
1167 | */ | ||
1168 | mutex_lock(&delayed_node->mutex); | ||
1169 | if (delayed_node->count) | ||
1170 | need_requeue = 1; | ||
1171 | else | ||
1172 | btrfs_dequeue_delayed_node(root->fs_info->delayed_root, | ||
1173 | delayed_node); | ||
1174 | mutex_unlock(&delayed_node->mutex); | ||
1175 | |||
1176 | nr = trans->blocks_used; | ||
1177 | |||
1178 | btrfs_end_transaction_dmeta(trans, root); | ||
1179 | __btrfs_btree_balance_dirty(root, nr); | ||
1180 | free_path: | ||
1181 | btrfs_free_path(path); | ||
1182 | out: | ||
1183 | if (need_requeue) | ||
1184 | btrfs_requeue_work(&async_node->work); | ||
1185 | else { | ||
1186 | btrfs_release_prepared_delayed_node(delayed_node); | ||
1187 | kfree(async_node); | ||
1188 | } | ||
1189 | } | ||
1190 | |||
1191 | static int btrfs_wq_run_delayed_node(struct btrfs_delayed_root *delayed_root, | ||
1192 | struct btrfs_root *root, int all) | ||
1193 | { | ||
1194 | struct btrfs_async_delayed_node *async_node; | ||
1195 | struct btrfs_delayed_node *curr; | ||
1196 | int count = 0; | ||
1197 | |||
1198 | again: | ||
1199 | curr = btrfs_first_prepared_delayed_node(delayed_root); | ||
1200 | if (!curr) | ||
1201 | return 0; | ||
1202 | |||
1203 | async_node = kmalloc(sizeof(*async_node), GFP_NOFS); | ||
1204 | if (!async_node) { | ||
1205 | btrfs_release_prepared_delayed_node(curr); | ||
1206 | return -ENOMEM; | ||
1207 | } | ||
1208 | |||
1209 | async_node->root = root; | ||
1210 | async_node->delayed_node = curr; | ||
1211 | |||
1212 | async_node->work.func = btrfs_async_run_delayed_node_done; | ||
1213 | async_node->work.flags = 0; | ||
1214 | |||
1215 | btrfs_queue_worker(&root->fs_info->delayed_workers, &async_node->work); | ||
1216 | count++; | ||
1217 | |||
1218 | if (all || count < 4) | ||
1219 | goto again; | ||
1220 | |||
1221 | return 0; | ||
1222 | } | ||
1223 | |||
1224 | void btrfs_balance_delayed_items(struct btrfs_root *root) | ||
1225 | { | ||
1226 | struct btrfs_delayed_root *delayed_root; | ||
1227 | |||
1228 | delayed_root = btrfs_get_delayed_root(root); | ||
1229 | |||
1230 | if (atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND) | ||
1231 | return; | ||
1232 | |||
1233 | if (atomic_read(&delayed_root->items) >= BTRFS_DELAYED_WRITEBACK) { | ||
1234 | int ret; | ||
1235 | ret = btrfs_wq_run_delayed_node(delayed_root, root, 1); | ||
1236 | if (ret) | ||
1237 | return; | ||
1238 | |||
1239 | wait_event_interruptible_timeout( | ||
1240 | delayed_root->wait, | ||
1241 | (atomic_read(&delayed_root->items) < | ||
1242 | BTRFS_DELAYED_BACKGROUND), | ||
1243 | HZ); | ||
1244 | return; | ||
1245 | } | ||
1246 | |||
1247 | btrfs_wq_run_delayed_node(delayed_root, root, 0); | ||
1248 | } | ||
1249 | |||
1250 | int btrfs_insert_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
1251 | struct btrfs_root *root, const char *name, | ||
1252 | int name_len, struct inode *dir, | ||
1253 | struct btrfs_disk_key *disk_key, u8 type, | ||
1254 | u64 index) | ||
1255 | { | ||
1256 | struct btrfs_delayed_node *delayed_node; | ||
1257 | struct btrfs_delayed_item *delayed_item; | ||
1258 | struct btrfs_dir_item *dir_item; | ||
1259 | int ret; | ||
1260 | |||
1261 | delayed_node = btrfs_get_or_create_delayed_node(dir); | ||
1262 | if (IS_ERR(delayed_node)) | ||
1263 | return PTR_ERR(delayed_node); | ||
1264 | |||
1265 | delayed_item = btrfs_alloc_delayed_item(sizeof(*dir_item) + name_len); | ||
1266 | if (!delayed_item) { | ||
1267 | ret = -ENOMEM; | ||
1268 | goto release_node; | ||
1269 | } | ||
1270 | |||
1271 | ret = btrfs_delayed_item_reserve_metadata(trans, root, delayed_item); | ||
1272 | /* | ||
1273 | * we have reserved enough space when we start a new transaction, | ||
1274 | * so reserving metadata failure is impossible | ||
1275 | */ | ||
1276 | BUG_ON(ret); | ||
1277 | |||
1278 | delayed_item->key.objectid = btrfs_ino(dir); | ||
1279 | btrfs_set_key_type(&delayed_item->key, BTRFS_DIR_INDEX_KEY); | ||
1280 | delayed_item->key.offset = index; | ||
1281 | |||
1282 | dir_item = (struct btrfs_dir_item *)delayed_item->data; | ||
1283 | dir_item->location = *disk_key; | ||
1284 | dir_item->transid = cpu_to_le64(trans->transid); | ||
1285 | dir_item->data_len = 0; | ||
1286 | dir_item->name_len = cpu_to_le16(name_len); | ||
1287 | dir_item->type = type; | ||
1288 | memcpy((char *)(dir_item + 1), name, name_len); | ||
1289 | |||
1290 | mutex_lock(&delayed_node->mutex); | ||
1291 | ret = __btrfs_add_delayed_insertion_item(delayed_node, delayed_item); | ||
1292 | if (unlikely(ret)) { | ||
1293 | printk(KERN_ERR "err add delayed dir index item(name: %s) into " | ||
1294 | "the insertion tree of the delayed node" | ||
1295 | "(root id: %llu, inode id: %llu, errno: %d)\n", | ||
1296 | name, | ||
1297 | (unsigned long long)delayed_node->root->objectid, | ||
1298 | (unsigned long long)delayed_node->inode_id, | ||
1299 | ret); | ||
1300 | BUG(); | ||
1301 | } | ||
1302 | mutex_unlock(&delayed_node->mutex); | ||
1303 | |||
1304 | release_node: | ||
1305 | btrfs_release_delayed_node(delayed_node); | ||
1306 | return ret; | ||
1307 | } | ||
1308 | |||
1309 | static int btrfs_delete_delayed_insertion_item(struct btrfs_root *root, | ||
1310 | struct btrfs_delayed_node *node, | ||
1311 | struct btrfs_key *key) | ||
1312 | { | ||
1313 | struct btrfs_delayed_item *item; | ||
1314 | |||
1315 | mutex_lock(&node->mutex); | ||
1316 | item = __btrfs_lookup_delayed_insertion_item(node, key); | ||
1317 | if (!item) { | ||
1318 | mutex_unlock(&node->mutex); | ||
1319 | return 1; | ||
1320 | } | ||
1321 | |||
1322 | btrfs_delayed_item_release_metadata(root, item); | ||
1323 | btrfs_release_delayed_item(item); | ||
1324 | mutex_unlock(&node->mutex); | ||
1325 | return 0; | ||
1326 | } | ||
1327 | |||
1328 | int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
1329 | struct btrfs_root *root, struct inode *dir, | ||
1330 | u64 index) | ||
1331 | { | ||
1332 | struct btrfs_delayed_node *node; | ||
1333 | struct btrfs_delayed_item *item; | ||
1334 | struct btrfs_key item_key; | ||
1335 | int ret; | ||
1336 | |||
1337 | node = btrfs_get_or_create_delayed_node(dir); | ||
1338 | if (IS_ERR(node)) | ||
1339 | return PTR_ERR(node); | ||
1340 | |||
1341 | item_key.objectid = btrfs_ino(dir); | ||
1342 | btrfs_set_key_type(&item_key, BTRFS_DIR_INDEX_KEY); | ||
1343 | item_key.offset = index; | ||
1344 | |||
1345 | ret = btrfs_delete_delayed_insertion_item(root, node, &item_key); | ||
1346 | if (!ret) | ||
1347 | goto end; | ||
1348 | |||
1349 | item = btrfs_alloc_delayed_item(0); | ||
1350 | if (!item) { | ||
1351 | ret = -ENOMEM; | ||
1352 | goto end; | ||
1353 | } | ||
1354 | |||
1355 | item->key = item_key; | ||
1356 | |||
1357 | ret = btrfs_delayed_item_reserve_metadata(trans, root, item); | ||
1358 | /* | ||
1359 | * we have reserved enough space when we start a new transaction, | ||
1360 | * so reserving metadata failure is impossible. | ||
1361 | */ | ||
1362 | BUG_ON(ret); | ||
1363 | |||
1364 | mutex_lock(&node->mutex); | ||
1365 | ret = __btrfs_add_delayed_deletion_item(node, item); | ||
1366 | if (unlikely(ret)) { | ||
1367 | printk(KERN_ERR "err add delayed dir index item(index: %llu) " | ||
1368 | "into the deletion tree of the delayed node" | ||
1369 | "(root id: %llu, inode id: %llu, errno: %d)\n", | ||
1370 | (unsigned long long)index, | ||
1371 | (unsigned long long)node->root->objectid, | ||
1372 | (unsigned long long)node->inode_id, | ||
1373 | ret); | ||
1374 | BUG(); | ||
1375 | } | ||
1376 | mutex_unlock(&node->mutex); | ||
1377 | end: | ||
1378 | btrfs_release_delayed_node(node); | ||
1379 | return ret; | ||
1380 | } | ||
1381 | |||
1382 | int btrfs_inode_delayed_dir_index_count(struct inode *inode) | ||
1383 | { | ||
1384 | struct btrfs_delayed_node *delayed_node = BTRFS_I(inode)->delayed_node; | ||
1385 | int ret = 0; | ||
1386 | |||
1387 | if (!delayed_node) | ||
1388 | return -ENOENT; | ||
1389 | |||
1390 | /* | ||
1391 | * Since we have held i_mutex of this directory, it is impossible that | ||
1392 | * a new directory index is added into the delayed node and index_cnt | ||
1393 | * is updated now. So we needn't lock the delayed node. | ||
1394 | */ | ||
1395 | if (!delayed_node->index_cnt) | ||
1396 | return -EINVAL; | ||
1397 | |||
1398 | BTRFS_I(inode)->index_cnt = delayed_node->index_cnt; | ||
1399 | return ret; | ||
1400 | } | ||
1401 | |||
1402 | void btrfs_get_delayed_items(struct inode *inode, struct list_head *ins_list, | ||
1403 | struct list_head *del_list) | ||
1404 | { | ||
1405 | struct btrfs_delayed_node *delayed_node; | ||
1406 | struct btrfs_delayed_item *item; | ||
1407 | |||
1408 | delayed_node = btrfs_get_delayed_node(inode); | ||
1409 | if (!delayed_node) | ||
1410 | return; | ||
1411 | |||
1412 | mutex_lock(&delayed_node->mutex); | ||
1413 | item = __btrfs_first_delayed_insertion_item(delayed_node); | ||
1414 | while (item) { | ||
1415 | atomic_inc(&item->refs); | ||
1416 | list_add_tail(&item->readdir_list, ins_list); | ||
1417 | item = __btrfs_next_delayed_item(item); | ||
1418 | } | ||
1419 | |||
1420 | item = __btrfs_first_delayed_deletion_item(delayed_node); | ||
1421 | while (item) { | ||
1422 | atomic_inc(&item->refs); | ||
1423 | list_add_tail(&item->readdir_list, del_list); | ||
1424 | item = __btrfs_next_delayed_item(item); | ||
1425 | } | ||
1426 | mutex_unlock(&delayed_node->mutex); | ||
1427 | /* | ||
1428 | * This delayed node is still cached in the btrfs inode, so refs | ||
1429 | * must be > 1 now, and we needn't check it is going to be freed | ||
1430 | * or not. | ||
1431 | * | ||
1432 | * Besides that, this function is used to read dir, we do not | ||
1433 | * insert/delete delayed items in this period. So we also needn't | ||
1434 | * requeue or dequeue this delayed node. | ||
1435 | */ | ||
1436 | atomic_dec(&delayed_node->refs); | ||
1437 | } | ||
1438 | |||
1439 | void btrfs_put_delayed_items(struct list_head *ins_list, | ||
1440 | struct list_head *del_list) | ||
1441 | { | ||
1442 | struct btrfs_delayed_item *curr, *next; | ||
1443 | |||
1444 | list_for_each_entry_safe(curr, next, ins_list, readdir_list) { | ||
1445 | list_del(&curr->readdir_list); | ||
1446 | if (atomic_dec_and_test(&curr->refs)) | ||
1447 | kfree(curr); | ||
1448 | } | ||
1449 | |||
1450 | list_for_each_entry_safe(curr, next, del_list, readdir_list) { | ||
1451 | list_del(&curr->readdir_list); | ||
1452 | if (atomic_dec_and_test(&curr->refs)) | ||
1453 | kfree(curr); | ||
1454 | } | ||
1455 | } | ||
1456 | |||
1457 | int btrfs_should_delete_dir_index(struct list_head *del_list, | ||
1458 | u64 index) | ||
1459 | { | ||
1460 | struct btrfs_delayed_item *curr, *next; | ||
1461 | int ret; | ||
1462 | |||
1463 | if (list_empty(del_list)) | ||
1464 | return 0; | ||
1465 | |||
1466 | list_for_each_entry_safe(curr, next, del_list, readdir_list) { | ||
1467 | if (curr->key.offset > index) | ||
1468 | break; | ||
1469 | |||
1470 | list_del(&curr->readdir_list); | ||
1471 | ret = (curr->key.offset == index); | ||
1472 | |||
1473 | if (atomic_dec_and_test(&curr->refs)) | ||
1474 | kfree(curr); | ||
1475 | |||
1476 | if (ret) | ||
1477 | return 1; | ||
1478 | else | ||
1479 | continue; | ||
1480 | } | ||
1481 | return 0; | ||
1482 | } | ||
1483 | |||
1484 | /* | ||
1485 | * btrfs_readdir_delayed_dir_index - read dir info stored in the delayed tree | ||
1486 | * | ||
1487 | */ | ||
1488 | int btrfs_readdir_delayed_dir_index(struct file *filp, void *dirent, | ||
1489 | filldir_t filldir, | ||
1490 | struct list_head *ins_list) | ||
1491 | { | ||
1492 | struct btrfs_dir_item *di; | ||
1493 | struct btrfs_delayed_item *curr, *next; | ||
1494 | struct btrfs_key location; | ||
1495 | char *name; | ||
1496 | int name_len; | ||
1497 | int over = 0; | ||
1498 | unsigned char d_type; | ||
1499 | |||
1500 | if (list_empty(ins_list)) | ||
1501 | return 0; | ||
1502 | |||
1503 | /* | ||
1504 | * Changing the data of the delayed item is impossible. So | ||
1505 | * we needn't lock them. And we have held i_mutex of the | ||
1506 | * directory, nobody can delete any directory indexes now. | ||
1507 | */ | ||
1508 | list_for_each_entry_safe(curr, next, ins_list, readdir_list) { | ||
1509 | list_del(&curr->readdir_list); | ||
1510 | |||
1511 | if (curr->key.offset < filp->f_pos) { | ||
1512 | if (atomic_dec_and_test(&curr->refs)) | ||
1513 | kfree(curr); | ||
1514 | continue; | ||
1515 | } | ||
1516 | |||
1517 | filp->f_pos = curr->key.offset; | ||
1518 | |||
1519 | di = (struct btrfs_dir_item *)curr->data; | ||
1520 | name = (char *)(di + 1); | ||
1521 | name_len = le16_to_cpu(di->name_len); | ||
1522 | |||
1523 | d_type = btrfs_filetype_table[di->type]; | ||
1524 | btrfs_disk_key_to_cpu(&location, &di->location); | ||
1525 | |||
1526 | over = filldir(dirent, name, name_len, curr->key.offset, | ||
1527 | location.objectid, d_type); | ||
1528 | |||
1529 | if (atomic_dec_and_test(&curr->refs)) | ||
1530 | kfree(curr); | ||
1531 | |||
1532 | if (over) | ||
1533 | return 1; | ||
1534 | } | ||
1535 | return 0; | ||
1536 | } | ||
1537 | |||
1538 | BTRFS_SETGET_STACK_FUNCS(stack_inode_generation, struct btrfs_inode_item, | ||
1539 | generation, 64); | ||
1540 | BTRFS_SETGET_STACK_FUNCS(stack_inode_sequence, struct btrfs_inode_item, | ||
1541 | sequence, 64); | ||
1542 | BTRFS_SETGET_STACK_FUNCS(stack_inode_transid, struct btrfs_inode_item, | ||
1543 | transid, 64); | ||
1544 | BTRFS_SETGET_STACK_FUNCS(stack_inode_size, struct btrfs_inode_item, size, 64); | ||
1545 | BTRFS_SETGET_STACK_FUNCS(stack_inode_nbytes, struct btrfs_inode_item, | ||
1546 | nbytes, 64); | ||
1547 | BTRFS_SETGET_STACK_FUNCS(stack_inode_block_group, struct btrfs_inode_item, | ||
1548 | block_group, 64); | ||
1549 | BTRFS_SETGET_STACK_FUNCS(stack_inode_nlink, struct btrfs_inode_item, nlink, 32); | ||
1550 | BTRFS_SETGET_STACK_FUNCS(stack_inode_uid, struct btrfs_inode_item, uid, 32); | ||
1551 | BTRFS_SETGET_STACK_FUNCS(stack_inode_gid, struct btrfs_inode_item, gid, 32); | ||
1552 | BTRFS_SETGET_STACK_FUNCS(stack_inode_mode, struct btrfs_inode_item, mode, 32); | ||
1553 | BTRFS_SETGET_STACK_FUNCS(stack_inode_rdev, struct btrfs_inode_item, rdev, 64); | ||
1554 | BTRFS_SETGET_STACK_FUNCS(stack_inode_flags, struct btrfs_inode_item, flags, 64); | ||
1555 | |||
1556 | BTRFS_SETGET_STACK_FUNCS(stack_timespec_sec, struct btrfs_timespec, sec, 64); | ||
1557 | BTRFS_SETGET_STACK_FUNCS(stack_timespec_nsec, struct btrfs_timespec, nsec, 32); | ||
1558 | |||
1559 | static void fill_stack_inode_item(struct btrfs_trans_handle *trans, | ||
1560 | struct btrfs_inode_item *inode_item, | ||
1561 | struct inode *inode) | ||
1562 | { | ||
1563 | btrfs_set_stack_inode_uid(inode_item, inode->i_uid); | ||
1564 | btrfs_set_stack_inode_gid(inode_item, inode->i_gid); | ||
1565 | btrfs_set_stack_inode_size(inode_item, BTRFS_I(inode)->disk_i_size); | ||
1566 | btrfs_set_stack_inode_mode(inode_item, inode->i_mode); | ||
1567 | btrfs_set_stack_inode_nlink(inode_item, inode->i_nlink); | ||
1568 | btrfs_set_stack_inode_nbytes(inode_item, inode_get_bytes(inode)); | ||
1569 | btrfs_set_stack_inode_generation(inode_item, | ||
1570 | BTRFS_I(inode)->generation); | ||
1571 | btrfs_set_stack_inode_sequence(inode_item, BTRFS_I(inode)->sequence); | ||
1572 | btrfs_set_stack_inode_transid(inode_item, trans->transid); | ||
1573 | btrfs_set_stack_inode_rdev(inode_item, inode->i_rdev); | ||
1574 | btrfs_set_stack_inode_flags(inode_item, BTRFS_I(inode)->flags); | ||
1575 | btrfs_set_stack_inode_block_group(inode_item, | ||
1576 | BTRFS_I(inode)->block_group); | ||
1577 | |||
1578 | btrfs_set_stack_timespec_sec(btrfs_inode_atime(inode_item), | ||
1579 | inode->i_atime.tv_sec); | ||
1580 | btrfs_set_stack_timespec_nsec(btrfs_inode_atime(inode_item), | ||
1581 | inode->i_atime.tv_nsec); | ||
1582 | |||
1583 | btrfs_set_stack_timespec_sec(btrfs_inode_mtime(inode_item), | ||
1584 | inode->i_mtime.tv_sec); | ||
1585 | btrfs_set_stack_timespec_nsec(btrfs_inode_mtime(inode_item), | ||
1586 | inode->i_mtime.tv_nsec); | ||
1587 | |||
1588 | btrfs_set_stack_timespec_sec(btrfs_inode_ctime(inode_item), | ||
1589 | inode->i_ctime.tv_sec); | ||
1590 | btrfs_set_stack_timespec_nsec(btrfs_inode_ctime(inode_item), | ||
1591 | inode->i_ctime.tv_nsec); | ||
1592 | } | ||
1593 | |||
1594 | int btrfs_delayed_update_inode(struct btrfs_trans_handle *trans, | ||
1595 | struct btrfs_root *root, struct inode *inode) | ||
1596 | { | ||
1597 | struct btrfs_delayed_node *delayed_node; | ||
1598 | int ret; | ||
1599 | |||
1600 | delayed_node = btrfs_get_or_create_delayed_node(inode); | ||
1601 | if (IS_ERR(delayed_node)) | ||
1602 | return PTR_ERR(delayed_node); | ||
1603 | |||
1604 | mutex_lock(&delayed_node->mutex); | ||
1605 | if (delayed_node->inode_dirty) { | ||
1606 | fill_stack_inode_item(trans, &delayed_node->inode_item, inode); | ||
1607 | goto release_node; | ||
1608 | } | ||
1609 | |||
1610 | ret = btrfs_delayed_inode_reserve_metadata(trans, root, delayed_node); | ||
1611 | /* | ||
1612 | * we must reserve enough space when we start a new transaction, | ||
1613 | * so reserving metadata failure is impossible | ||
1614 | */ | ||
1615 | BUG_ON(ret); | ||
1616 | |||
1617 | fill_stack_inode_item(trans, &delayed_node->inode_item, inode); | ||
1618 | delayed_node->inode_dirty = 1; | ||
1619 | delayed_node->count++; | ||
1620 | atomic_inc(&root->fs_info->delayed_root->items); | ||
1621 | release_node: | ||
1622 | mutex_unlock(&delayed_node->mutex); | ||
1623 | btrfs_release_delayed_node(delayed_node); | ||
1624 | return ret; | ||
1625 | } | ||
1626 | |||
1627 | static void __btrfs_kill_delayed_node(struct btrfs_delayed_node *delayed_node) | ||
1628 | { | ||
1629 | struct btrfs_root *root = delayed_node->root; | ||
1630 | struct btrfs_delayed_item *curr_item, *prev_item; | ||
1631 | |||
1632 | mutex_lock(&delayed_node->mutex); | ||
1633 | curr_item = __btrfs_first_delayed_insertion_item(delayed_node); | ||
1634 | while (curr_item) { | ||
1635 | btrfs_delayed_item_release_metadata(root, curr_item); | ||
1636 | prev_item = curr_item; | ||
1637 | curr_item = __btrfs_next_delayed_item(prev_item); | ||
1638 | btrfs_release_delayed_item(prev_item); | ||
1639 | } | ||
1640 | |||
1641 | curr_item = __btrfs_first_delayed_deletion_item(delayed_node); | ||
1642 | while (curr_item) { | ||
1643 | btrfs_delayed_item_release_metadata(root, curr_item); | ||
1644 | prev_item = curr_item; | ||
1645 | curr_item = __btrfs_next_delayed_item(prev_item); | ||
1646 | btrfs_release_delayed_item(prev_item); | ||
1647 | } | ||
1648 | |||
1649 | if (delayed_node->inode_dirty) { | ||
1650 | btrfs_delayed_inode_release_metadata(root, delayed_node); | ||
1651 | btrfs_release_delayed_inode(delayed_node); | ||
1652 | } | ||
1653 | mutex_unlock(&delayed_node->mutex); | ||
1654 | } | ||
1655 | |||
1656 | void btrfs_kill_delayed_inode_items(struct inode *inode) | ||
1657 | { | ||
1658 | struct btrfs_delayed_node *delayed_node; | ||
1659 | |||
1660 | delayed_node = btrfs_get_delayed_node(inode); | ||
1661 | if (!delayed_node) | ||
1662 | return; | ||
1663 | |||
1664 | __btrfs_kill_delayed_node(delayed_node); | ||
1665 | btrfs_release_delayed_node(delayed_node); | ||
1666 | } | ||
1667 | |||
1668 | void btrfs_kill_all_delayed_nodes(struct btrfs_root *root) | ||
1669 | { | ||
1670 | u64 inode_id = 0; | ||
1671 | struct btrfs_delayed_node *delayed_nodes[8]; | ||
1672 | int i, n; | ||
1673 | |||
1674 | while (1) { | ||
1675 | spin_lock(&root->inode_lock); | ||
1676 | n = radix_tree_gang_lookup(&root->delayed_nodes_tree, | ||
1677 | (void **)delayed_nodes, inode_id, | ||
1678 | ARRAY_SIZE(delayed_nodes)); | ||
1679 | if (!n) { | ||
1680 | spin_unlock(&root->inode_lock); | ||
1681 | break; | ||
1682 | } | ||
1683 | |||
1684 | inode_id = delayed_nodes[n - 1]->inode_id + 1; | ||
1685 | |||
1686 | for (i = 0; i < n; i++) | ||
1687 | atomic_inc(&delayed_nodes[i]->refs); | ||
1688 | spin_unlock(&root->inode_lock); | ||
1689 | |||
1690 | for (i = 0; i < n; i++) { | ||
1691 | __btrfs_kill_delayed_node(delayed_nodes[i]); | ||
1692 | btrfs_release_delayed_node(delayed_nodes[i]); | ||
1693 | } | ||
1694 | } | ||
1695 | } | ||
diff --git a/fs/btrfs/delayed-inode.h b/fs/btrfs/delayed-inode.h new file mode 100644 index 00000000000..eb7d240aa64 --- /dev/null +++ b/fs/btrfs/delayed-inode.h | |||
@@ -0,0 +1,141 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2011 Fujitsu. All rights reserved. | ||
3 | * Written by Miao Xie <miaox@cn.fujitsu.com> | ||
4 | * | ||
5 | * This program is free software; you can redistribute it and/or | ||
6 | * modify it under the terms of the GNU General Public | ||
7 | * License v2 as published by the Free Software Foundation. | ||
8 | * | ||
9 | * This program is distributed in the hope that it will be useful, | ||
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | ||
12 | * General Public License for more details. | ||
13 | * | ||
14 | * You should have received a copy of the GNU General Public | ||
15 | * License along with this program; if not, write to the | ||
16 | * Free Software Foundation, Inc., 59 Temple Place - Suite 330, | ||
17 | * Boston, MA 021110-1307, USA. | ||
18 | */ | ||
19 | |||
20 | #ifndef __DELAYED_TREE_OPERATION_H | ||
21 | #define __DELAYED_TREE_OPERATION_H | ||
22 | |||
23 | #include <linux/rbtree.h> | ||
24 | #include <linux/spinlock.h> | ||
25 | #include <linux/mutex.h> | ||
26 | #include <linux/list.h> | ||
27 | #include <linux/wait.h> | ||
28 | #include <asm/atomic.h> | ||
29 | |||
30 | #include "ctree.h" | ||
31 | |||
32 | /* types of the delayed item */ | ||
33 | #define BTRFS_DELAYED_INSERTION_ITEM 1 | ||
34 | #define BTRFS_DELAYED_DELETION_ITEM 2 | ||
35 | |||
36 | struct btrfs_delayed_root { | ||
37 | spinlock_t lock; | ||
38 | struct list_head node_list; | ||
39 | /* | ||
40 | * Used for delayed nodes which is waiting to be dealt with by the | ||
41 | * worker. If the delayed node is inserted into the work queue, we | ||
42 | * drop it from this list. | ||
43 | */ | ||
44 | struct list_head prepare_list; | ||
45 | atomic_t items; /* for delayed items */ | ||
46 | int nodes; /* for delayed nodes */ | ||
47 | wait_queue_head_t wait; | ||
48 | }; | ||
49 | |||
50 | struct btrfs_delayed_node { | ||
51 | u64 inode_id; | ||
52 | u64 bytes_reserved; | ||
53 | struct btrfs_root *root; | ||
54 | /* Used to add the node into the delayed root's node list. */ | ||
55 | struct list_head n_list; | ||
56 | /* | ||
57 | * Used to add the node into the prepare list, the nodes in this list | ||
58 | * is waiting to be dealt with by the async worker. | ||
59 | */ | ||
60 | struct list_head p_list; | ||
61 | struct rb_root ins_root; | ||
62 | struct rb_root del_root; | ||
63 | struct mutex mutex; | ||
64 | struct btrfs_inode_item inode_item; | ||
65 | atomic_t refs; | ||
66 | u64 index_cnt; | ||
67 | bool in_list; | ||
68 | bool inode_dirty; | ||
69 | int count; | ||
70 | }; | ||
71 | |||
72 | struct btrfs_delayed_item { | ||
73 | struct rb_node rb_node; | ||
74 | struct btrfs_key key; | ||
75 | struct list_head tree_list; /* used for batch insert/delete items */ | ||
76 | struct list_head readdir_list; /* used for readdir items */ | ||
77 | u64 bytes_reserved; | ||
78 | struct btrfs_block_rsv *block_rsv; | ||
79 | struct btrfs_delayed_node *delayed_node; | ||
80 | atomic_t refs; | ||
81 | int ins_or_del; | ||
82 | u32 data_len; | ||
83 | char data[0]; | ||
84 | }; | ||
85 | |||
86 | static inline void btrfs_init_delayed_root( | ||
87 | struct btrfs_delayed_root *delayed_root) | ||
88 | { | ||
89 | atomic_set(&delayed_root->items, 0); | ||
90 | delayed_root->nodes = 0; | ||
91 | spin_lock_init(&delayed_root->lock); | ||
92 | init_waitqueue_head(&delayed_root->wait); | ||
93 | INIT_LIST_HEAD(&delayed_root->node_list); | ||
94 | INIT_LIST_HEAD(&delayed_root->prepare_list); | ||
95 | } | ||
96 | |||
97 | int btrfs_insert_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
98 | struct btrfs_root *root, const char *name, | ||
99 | int name_len, struct inode *dir, | ||
100 | struct btrfs_disk_key *disk_key, u8 type, | ||
101 | u64 index); | ||
102 | |||
103 | int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
104 | struct btrfs_root *root, struct inode *dir, | ||
105 | u64 index); | ||
106 | |||
107 | int btrfs_inode_delayed_dir_index_count(struct inode *inode); | ||
108 | |||
109 | int btrfs_run_delayed_items(struct btrfs_trans_handle *trans, | ||
110 | struct btrfs_root *root); | ||
111 | |||
112 | void btrfs_balance_delayed_items(struct btrfs_root *root); | ||
113 | |||
114 | int btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans, | ||
115 | struct inode *inode); | ||
116 | /* Used for evicting the inode. */ | ||
117 | void btrfs_remove_delayed_node(struct inode *inode); | ||
118 | void btrfs_kill_delayed_inode_items(struct inode *inode); | ||
119 | |||
120 | |||
121 | int btrfs_delayed_update_inode(struct btrfs_trans_handle *trans, | ||
122 | struct btrfs_root *root, struct inode *inode); | ||
123 | |||
124 | /* Used for drop dead root */ | ||
125 | void btrfs_kill_all_delayed_nodes(struct btrfs_root *root); | ||
126 | |||
127 | /* Used for readdir() */ | ||
128 | void btrfs_get_delayed_items(struct inode *inode, struct list_head *ins_list, | ||
129 | struct list_head *del_list); | ||
130 | void btrfs_put_delayed_items(struct list_head *ins_list, | ||
131 | struct list_head *del_list); | ||
132 | int btrfs_should_delete_dir_index(struct list_head *del_list, | ||
133 | u64 index); | ||
134 | int btrfs_readdir_delayed_dir_index(struct file *filp, void *dirent, | ||
135 | filldir_t filldir, | ||
136 | struct list_head *ins_list); | ||
137 | |||
138 | /* for init */ | ||
139 | int __init btrfs_delayed_inode_init(void); | ||
140 | void btrfs_delayed_inode_exit(void); | ||
141 | #endif | ||
diff --git a/fs/btrfs/dir-item.c b/fs/btrfs/dir-item.c index ab8afed671a..1ddfca78e91 100644 --- a/fs/btrfs/dir-item.c +++ b/fs/btrfs/dir-item.c | |||
@@ -124,8 +124,9 @@ int btrfs_insert_xattr_item(struct btrfs_trans_handle *trans, | |||
124 | * to use for the second index (if one is created). | 124 | * to use for the second index (if one is created). |
125 | */ | 125 | */ |
126 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root | 126 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root |
127 | *root, const char *name, int name_len, u64 dir, | 127 | *root, const char *name, int name_len, |
128 | struct btrfs_key *location, u8 type, u64 index) | 128 | struct inode *dir, struct btrfs_key *location, |
129 | u8 type, u64 index) | ||
129 | { | 130 | { |
130 | int ret = 0; | 131 | int ret = 0; |
131 | int ret2 = 0; | 132 | int ret2 = 0; |
@@ -137,13 +138,17 @@ int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root | |||
137 | struct btrfs_disk_key disk_key; | 138 | struct btrfs_disk_key disk_key; |
138 | u32 data_size; | 139 | u32 data_size; |
139 | 140 | ||
140 | key.objectid = dir; | 141 | key.objectid = btrfs_ino(dir); |
141 | btrfs_set_key_type(&key, BTRFS_DIR_ITEM_KEY); | 142 | btrfs_set_key_type(&key, BTRFS_DIR_ITEM_KEY); |
142 | key.offset = btrfs_name_hash(name, name_len); | 143 | key.offset = btrfs_name_hash(name, name_len); |
143 | 144 | ||
144 | path = btrfs_alloc_path(); | 145 | path = btrfs_alloc_path(); |
146 | if (!path) | ||
147 | return -ENOMEM; | ||
145 | path->leave_spinning = 1; | 148 | path->leave_spinning = 1; |
146 | 149 | ||
150 | btrfs_cpu_key_to_disk(&disk_key, location); | ||
151 | |||
147 | data_size = sizeof(*dir_item) + name_len; | 152 | data_size = sizeof(*dir_item) + name_len; |
148 | dir_item = insert_with_overflow(trans, root, path, &key, data_size, | 153 | dir_item = insert_with_overflow(trans, root, path, &key, data_size, |
149 | name, name_len); | 154 | name, name_len); |
@@ -155,7 +160,6 @@ int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root | |||
155 | } | 160 | } |
156 | 161 | ||
157 | leaf = path->nodes[0]; | 162 | leaf = path->nodes[0]; |
158 | btrfs_cpu_key_to_disk(&disk_key, location); | ||
159 | btrfs_set_dir_item_key(leaf, dir_item, &disk_key); | 163 | btrfs_set_dir_item_key(leaf, dir_item, &disk_key); |
160 | btrfs_set_dir_type(leaf, dir_item, type); | 164 | btrfs_set_dir_type(leaf, dir_item, type); |
161 | btrfs_set_dir_data_len(leaf, dir_item, 0); | 165 | btrfs_set_dir_data_len(leaf, dir_item, 0); |
@@ -174,27 +178,9 @@ second_insert: | |||
174 | } | 178 | } |
175 | btrfs_release_path(path); | 179 | btrfs_release_path(path); |
176 | 180 | ||
177 | btrfs_set_key_type(&key, BTRFS_DIR_INDEX_KEY); | 181 | ret2 = btrfs_insert_delayed_dir_index(trans, root, name, name_len, dir, |
178 | key.offset = index; | 182 | &disk_key, type, index); |
179 | dir_item = insert_with_overflow(trans, root, path, &key, data_size, | ||
180 | name, name_len); | ||
181 | if (IS_ERR(dir_item)) { | ||
182 | ret2 = PTR_ERR(dir_item); | ||
183 | goto out_free; | ||
184 | } | ||
185 | leaf = path->nodes[0]; | ||
186 | btrfs_cpu_key_to_disk(&disk_key, location); | ||
187 | btrfs_set_dir_item_key(leaf, dir_item, &disk_key); | ||
188 | btrfs_set_dir_type(leaf, dir_item, type); | ||
189 | btrfs_set_dir_data_len(leaf, dir_item, 0); | ||
190 | btrfs_set_dir_name_len(leaf, dir_item, name_len); | ||
191 | btrfs_set_dir_transid(leaf, dir_item, trans->transid); | ||
192 | name_ptr = (unsigned long)(dir_item + 1); | ||
193 | write_extent_buffer(leaf, name, name_ptr, name_len); | ||
194 | btrfs_mark_buffer_dirty(leaf); | ||
195 | |||
196 | out_free: | 183 | out_free: |
197 | |||
198 | btrfs_free_path(path); | 184 | btrfs_free_path(path); |
199 | if (ret) | 185 | if (ret) |
200 | return ret; | 186 | return ret; |
diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c index cb9d1b8bfe7..a2eb3a3755d 100644 --- a/fs/btrfs/disk-io.c +++ b/fs/btrfs/disk-io.c | |||
@@ -42,6 +42,7 @@ | |||
42 | #include "locking.h" | 42 | #include "locking.h" |
43 | #include "tree-log.h" | 43 | #include "tree-log.h" |
44 | #include "free-space-cache.h" | 44 | #include "free-space-cache.h" |
45 | #include "inode-map.h" | ||
45 | 46 | ||
46 | static struct extent_io_ops btree_extent_io_ops; | 47 | static struct extent_io_ops btree_extent_io_ops; |
47 | static void end_workqueue_fn(struct btrfs_work *work); | 48 | static void end_workqueue_fn(struct btrfs_work *work); |
@@ -1045,6 +1046,7 @@ static int __setup_root(u32 nodesize, u32 leafsize, u32 sectorsize, | |||
1045 | root->name = NULL; | 1046 | root->name = NULL; |
1046 | root->in_sysfs = 0; | 1047 | root->in_sysfs = 0; |
1047 | root->inode_tree = RB_ROOT; | 1048 | root->inode_tree = RB_ROOT; |
1049 | INIT_RADIX_TREE(&root->delayed_nodes_tree, GFP_ATOMIC); | ||
1048 | root->block_rsv = NULL; | 1050 | root->block_rsv = NULL; |
1049 | root->orphan_block_rsv = NULL; | 1051 | root->orphan_block_rsv = NULL; |
1050 | 1052 | ||
@@ -1298,6 +1300,19 @@ again: | |||
1298 | if (IS_ERR(root)) | 1300 | if (IS_ERR(root)) |
1299 | return root; | 1301 | return root; |
1300 | 1302 | ||
1303 | root->free_ino_ctl = kzalloc(sizeof(*root->free_ino_ctl), GFP_NOFS); | ||
1304 | if (!root->free_ino_ctl) | ||
1305 | goto fail; | ||
1306 | root->free_ino_pinned = kzalloc(sizeof(*root->free_ino_pinned), | ||
1307 | GFP_NOFS); | ||
1308 | if (!root->free_ino_pinned) | ||
1309 | goto fail; | ||
1310 | |||
1311 | btrfs_init_free_ino_ctl(root); | ||
1312 | mutex_init(&root->fs_commit_mutex); | ||
1313 | spin_lock_init(&root->cache_lock); | ||
1314 | init_waitqueue_head(&root->cache_wait); | ||
1315 | |||
1301 | set_anon_super(&root->anon_super, NULL); | 1316 | set_anon_super(&root->anon_super, NULL); |
1302 | 1317 | ||
1303 | if (btrfs_root_refs(&root->root_item) == 0) { | 1318 | if (btrfs_root_refs(&root->root_item) == 0) { |
@@ -1631,6 +1646,13 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1631 | 1646 | ||
1632 | INIT_LIST_HEAD(&fs_info->ordered_extents); | 1647 | INIT_LIST_HEAD(&fs_info->ordered_extents); |
1633 | spin_lock_init(&fs_info->ordered_extent_lock); | 1648 | spin_lock_init(&fs_info->ordered_extent_lock); |
1649 | fs_info->delayed_root = kmalloc(sizeof(struct btrfs_delayed_root), | ||
1650 | GFP_NOFS); | ||
1651 | if (!fs_info->delayed_root) { | ||
1652 | err = -ENOMEM; | ||
1653 | goto fail_iput; | ||
1654 | } | ||
1655 | btrfs_init_delayed_root(fs_info->delayed_root); | ||
1634 | 1656 | ||
1635 | sb->s_blocksize = 4096; | 1657 | sb->s_blocksize = 4096; |
1636 | sb->s_blocksize_bits = blksize_bits(4096); | 1658 | sb->s_blocksize_bits = blksize_bits(4096); |
@@ -1696,7 +1718,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1696 | bh = btrfs_read_dev_super(fs_devices->latest_bdev); | 1718 | bh = btrfs_read_dev_super(fs_devices->latest_bdev); |
1697 | if (!bh) { | 1719 | if (!bh) { |
1698 | err = -EINVAL; | 1720 | err = -EINVAL; |
1699 | goto fail_iput; | 1721 | goto fail_alloc; |
1700 | } | 1722 | } |
1701 | 1723 | ||
1702 | memcpy(&fs_info->super_copy, bh->b_data, sizeof(fs_info->super_copy)); | 1724 | memcpy(&fs_info->super_copy, bh->b_data, sizeof(fs_info->super_copy)); |
@@ -1708,7 +1730,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1708 | 1730 | ||
1709 | disk_super = &fs_info->super_copy; | 1731 | disk_super = &fs_info->super_copy; |
1710 | if (!btrfs_super_root(disk_super)) | 1732 | if (!btrfs_super_root(disk_super)) |
1711 | goto fail_iput; | 1733 | goto fail_alloc; |
1712 | 1734 | ||
1713 | /* check FS state, whether FS is broken. */ | 1735 | /* check FS state, whether FS is broken. */ |
1714 | fs_info->fs_state |= btrfs_super_flags(disk_super); | 1736 | fs_info->fs_state |= btrfs_super_flags(disk_super); |
@@ -1724,7 +1746,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1724 | ret = btrfs_parse_options(tree_root, options); | 1746 | ret = btrfs_parse_options(tree_root, options); |
1725 | if (ret) { | 1747 | if (ret) { |
1726 | err = ret; | 1748 | err = ret; |
1727 | goto fail_iput; | 1749 | goto fail_alloc; |
1728 | } | 1750 | } |
1729 | 1751 | ||
1730 | features = btrfs_super_incompat_flags(disk_super) & | 1752 | features = btrfs_super_incompat_flags(disk_super) & |
@@ -1734,7 +1756,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1734 | "unsupported optional features (%Lx).\n", | 1756 | "unsupported optional features (%Lx).\n", |
1735 | (unsigned long long)features); | 1757 | (unsigned long long)features); |
1736 | err = -EINVAL; | 1758 | err = -EINVAL; |
1737 | goto fail_iput; | 1759 | goto fail_alloc; |
1738 | } | 1760 | } |
1739 | 1761 | ||
1740 | features = btrfs_super_incompat_flags(disk_super); | 1762 | features = btrfs_super_incompat_flags(disk_super); |
@@ -1750,7 +1772,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1750 | "unsupported option features (%Lx).\n", | 1772 | "unsupported option features (%Lx).\n", |
1751 | (unsigned long long)features); | 1773 | (unsigned long long)features); |
1752 | err = -EINVAL; | 1774 | err = -EINVAL; |
1753 | goto fail_iput; | 1775 | goto fail_alloc; |
1754 | } | 1776 | } |
1755 | 1777 | ||
1756 | btrfs_init_workers(&fs_info->generic_worker, | 1778 | btrfs_init_workers(&fs_info->generic_worker, |
@@ -1797,6 +1819,9 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1797 | &fs_info->generic_worker); | 1819 | &fs_info->generic_worker); |
1798 | btrfs_init_workers(&fs_info->endio_freespace_worker, "freespace-write", | 1820 | btrfs_init_workers(&fs_info->endio_freespace_worker, "freespace-write", |
1799 | 1, &fs_info->generic_worker); | 1821 | 1, &fs_info->generic_worker); |
1822 | btrfs_init_workers(&fs_info->delayed_workers, "delayed-meta", | ||
1823 | fs_info->thread_pool_size, | ||
1824 | &fs_info->generic_worker); | ||
1800 | 1825 | ||
1801 | /* | 1826 | /* |
1802 | * endios are largely parallel and should have a very | 1827 | * endios are largely parallel and should have a very |
@@ -1818,6 +1843,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1818 | btrfs_start_workers(&fs_info->endio_meta_write_workers, 1); | 1843 | btrfs_start_workers(&fs_info->endio_meta_write_workers, 1); |
1819 | btrfs_start_workers(&fs_info->endio_write_workers, 1); | 1844 | btrfs_start_workers(&fs_info->endio_write_workers, 1); |
1820 | btrfs_start_workers(&fs_info->endio_freespace_worker, 1); | 1845 | btrfs_start_workers(&fs_info->endio_freespace_worker, 1); |
1846 | btrfs_start_workers(&fs_info->delayed_workers, 1); | ||
1821 | 1847 | ||
1822 | fs_info->bdi.ra_pages *= btrfs_super_num_devices(disk_super); | 1848 | fs_info->bdi.ra_pages *= btrfs_super_num_devices(disk_super); |
1823 | fs_info->bdi.ra_pages = max(fs_info->bdi.ra_pages, | 1849 | fs_info->bdi.ra_pages = max(fs_info->bdi.ra_pages, |
@@ -2074,6 +2100,9 @@ fail_sb_buffer: | |||
2074 | btrfs_stop_workers(&fs_info->endio_write_workers); | 2100 | btrfs_stop_workers(&fs_info->endio_write_workers); |
2075 | btrfs_stop_workers(&fs_info->endio_freespace_worker); | 2101 | btrfs_stop_workers(&fs_info->endio_freespace_worker); |
2076 | btrfs_stop_workers(&fs_info->submit_workers); | 2102 | btrfs_stop_workers(&fs_info->submit_workers); |
2103 | btrfs_stop_workers(&fs_info->delayed_workers); | ||
2104 | fail_alloc: | ||
2105 | kfree(fs_info->delayed_root); | ||
2077 | fail_iput: | 2106 | fail_iput: |
2078 | invalidate_inode_pages2(fs_info->btree_inode->i_mapping); | 2107 | invalidate_inode_pages2(fs_info->btree_inode->i_mapping); |
2079 | iput(fs_info->btree_inode); | 2108 | iput(fs_info->btree_inode); |
@@ -2338,12 +2367,15 @@ int btrfs_free_fs_root(struct btrfs_fs_info *fs_info, struct btrfs_root *root) | |||
2338 | if (btrfs_root_refs(&root->root_item) == 0) | 2367 | if (btrfs_root_refs(&root->root_item) == 0) |
2339 | synchronize_srcu(&fs_info->subvol_srcu); | 2368 | synchronize_srcu(&fs_info->subvol_srcu); |
2340 | 2369 | ||
2370 | __btrfs_remove_free_space_cache(root->free_ino_pinned); | ||
2371 | __btrfs_remove_free_space_cache(root->free_ino_ctl); | ||
2341 | free_fs_root(root); | 2372 | free_fs_root(root); |
2342 | return 0; | 2373 | return 0; |
2343 | } | 2374 | } |
2344 | 2375 | ||
2345 | static void free_fs_root(struct btrfs_root *root) | 2376 | static void free_fs_root(struct btrfs_root *root) |
2346 | { | 2377 | { |
2378 | iput(root->cache_inode); | ||
2347 | WARN_ON(!RB_EMPTY_ROOT(&root->inode_tree)); | 2379 | WARN_ON(!RB_EMPTY_ROOT(&root->inode_tree)); |
2348 | if (root->anon_super.s_dev) { | 2380 | if (root->anon_super.s_dev) { |
2349 | down_write(&root->anon_super.s_umount); | 2381 | down_write(&root->anon_super.s_umount); |
@@ -2351,6 +2383,8 @@ static void free_fs_root(struct btrfs_root *root) | |||
2351 | } | 2383 | } |
2352 | free_extent_buffer(root->node); | 2384 | free_extent_buffer(root->node); |
2353 | free_extent_buffer(root->commit_root); | 2385 | free_extent_buffer(root->commit_root); |
2386 | kfree(root->free_ino_ctl); | ||
2387 | kfree(root->free_ino_pinned); | ||
2354 | kfree(root->name); | 2388 | kfree(root->name); |
2355 | kfree(root); | 2389 | kfree(root); |
2356 | } | 2390 | } |
@@ -2512,6 +2546,7 @@ int close_ctree(struct btrfs_root *root) | |||
2512 | del_fs_roots(fs_info); | 2546 | del_fs_roots(fs_info); |
2513 | 2547 | ||
2514 | iput(fs_info->btree_inode); | 2548 | iput(fs_info->btree_inode); |
2549 | kfree(fs_info->delayed_root); | ||
2515 | 2550 | ||
2516 | btrfs_stop_workers(&fs_info->generic_worker); | 2551 | btrfs_stop_workers(&fs_info->generic_worker); |
2517 | btrfs_stop_workers(&fs_info->fixup_workers); | 2552 | btrfs_stop_workers(&fs_info->fixup_workers); |
@@ -2523,6 +2558,7 @@ int close_ctree(struct btrfs_root *root) | |||
2523 | btrfs_stop_workers(&fs_info->endio_write_workers); | 2558 | btrfs_stop_workers(&fs_info->endio_write_workers); |
2524 | btrfs_stop_workers(&fs_info->endio_freespace_worker); | 2559 | btrfs_stop_workers(&fs_info->endio_freespace_worker); |
2525 | btrfs_stop_workers(&fs_info->submit_workers); | 2560 | btrfs_stop_workers(&fs_info->submit_workers); |
2561 | btrfs_stop_workers(&fs_info->delayed_workers); | ||
2526 | 2562 | ||
2527 | btrfs_close_devices(fs_info->fs_devices); | 2563 | btrfs_close_devices(fs_info->fs_devices); |
2528 | btrfs_mapping_tree_free(&fs_info->mapping_tree); | 2564 | btrfs_mapping_tree_free(&fs_info->mapping_tree); |
@@ -2599,6 +2635,29 @@ void btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr) | |||
2599 | if (current->flags & PF_MEMALLOC) | 2635 | if (current->flags & PF_MEMALLOC) |
2600 | return; | 2636 | return; |
2601 | 2637 | ||
2638 | btrfs_balance_delayed_items(root); | ||
2639 | |||
2640 | num_dirty = root->fs_info->dirty_metadata_bytes; | ||
2641 | |||
2642 | if (num_dirty > thresh) { | ||
2643 | balance_dirty_pages_ratelimited_nr( | ||
2644 | root->fs_info->btree_inode->i_mapping, 1); | ||
2645 | } | ||
2646 | return; | ||
2647 | } | ||
2648 | |||
2649 | void __btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr) | ||
2650 | { | ||
2651 | /* | ||
2652 | * looks as though older kernels can get into trouble with | ||
2653 | * this code, they end up stuck in balance_dirty_pages forever | ||
2654 | */ | ||
2655 | u64 num_dirty; | ||
2656 | unsigned long thresh = 32 * 1024 * 1024; | ||
2657 | |||
2658 | if (current->flags & PF_MEMALLOC) | ||
2659 | return; | ||
2660 | |||
2602 | num_dirty = root->fs_info->dirty_metadata_bytes; | 2661 | num_dirty = root->fs_info->dirty_metadata_bytes; |
2603 | 2662 | ||
2604 | if (num_dirty > thresh) { | 2663 | if (num_dirty > thresh) { |
diff --git a/fs/btrfs/disk-io.h b/fs/btrfs/disk-io.h index 2d75f9e896f..a0b610a67aa 100644 --- a/fs/btrfs/disk-io.h +++ b/fs/btrfs/disk-io.h | |||
@@ -61,6 +61,7 @@ struct btrfs_root *btrfs_read_fs_root_no_name(struct btrfs_fs_info *fs_info, | |||
61 | struct btrfs_key *location); | 61 | struct btrfs_key *location); |
62 | int btrfs_cleanup_fs_roots(struct btrfs_fs_info *fs_info); | 62 | int btrfs_cleanup_fs_roots(struct btrfs_fs_info *fs_info); |
63 | void btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr); | 63 | void btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr); |
64 | void __btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr); | ||
64 | int btrfs_free_fs_root(struct btrfs_fs_info *fs_info, struct btrfs_root *root); | 65 | int btrfs_free_fs_root(struct btrfs_fs_info *fs_info, struct btrfs_root *root); |
65 | void btrfs_mark_buffer_dirty(struct extent_buffer *buf); | 66 | void btrfs_mark_buffer_dirty(struct extent_buffer *buf); |
66 | int btrfs_buffer_uptodate(struct extent_buffer *buf, u64 parent_transid); | 67 | int btrfs_buffer_uptodate(struct extent_buffer *buf, u64 parent_transid); |
diff --git a/fs/btrfs/export.c b/fs/btrfs/export.c index b4ffad859ad..1b8dc33778f 100644 --- a/fs/btrfs/export.c +++ b/fs/btrfs/export.c | |||
@@ -32,7 +32,7 @@ static int btrfs_encode_fh(struct dentry *dentry, u32 *fh, int *max_len, | |||
32 | len = BTRFS_FID_SIZE_NON_CONNECTABLE; | 32 | len = BTRFS_FID_SIZE_NON_CONNECTABLE; |
33 | type = FILEID_BTRFS_WITHOUT_PARENT; | 33 | type = FILEID_BTRFS_WITHOUT_PARENT; |
34 | 34 | ||
35 | fid->objectid = inode->i_ino; | 35 | fid->objectid = btrfs_ino(inode); |
36 | fid->root_objectid = BTRFS_I(inode)->root->objectid; | 36 | fid->root_objectid = BTRFS_I(inode)->root->objectid; |
37 | fid->gen = inode->i_generation; | 37 | fid->gen = inode->i_generation; |
38 | 38 | ||
@@ -178,13 +178,13 @@ static struct dentry *btrfs_get_parent(struct dentry *child) | |||
178 | if (!path) | 178 | if (!path) |
179 | return ERR_PTR(-ENOMEM); | 179 | return ERR_PTR(-ENOMEM); |
180 | 180 | ||
181 | if (dir->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 181 | if (btrfs_ino(dir) == BTRFS_FIRST_FREE_OBJECTID) { |
182 | key.objectid = root->root_key.objectid; | 182 | key.objectid = root->root_key.objectid; |
183 | key.type = BTRFS_ROOT_BACKREF_KEY; | 183 | key.type = BTRFS_ROOT_BACKREF_KEY; |
184 | key.offset = (u64)-1; | 184 | key.offset = (u64)-1; |
185 | root = root->fs_info->tree_root; | 185 | root = root->fs_info->tree_root; |
186 | } else { | 186 | } else { |
187 | key.objectid = dir->i_ino; | 187 | key.objectid = btrfs_ino(dir); |
188 | key.type = BTRFS_INODE_REF_KEY; | 188 | key.type = BTRFS_INODE_REF_KEY; |
189 | key.offset = (u64)-1; | 189 | key.offset = (u64)-1; |
190 | } | 190 | } |
@@ -244,6 +244,7 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
244 | struct btrfs_key key; | 244 | struct btrfs_key key; |
245 | int name_len; | 245 | int name_len; |
246 | int ret; | 246 | int ret; |
247 | u64 ino; | ||
247 | 248 | ||
248 | if (!dir || !inode) | 249 | if (!dir || !inode) |
249 | return -EINVAL; | 250 | return -EINVAL; |
@@ -251,19 +252,21 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
251 | if (!S_ISDIR(dir->i_mode)) | 252 | if (!S_ISDIR(dir->i_mode)) |
252 | return -EINVAL; | 253 | return -EINVAL; |
253 | 254 | ||
255 | ino = btrfs_ino(inode); | ||
256 | |||
254 | path = btrfs_alloc_path(); | 257 | path = btrfs_alloc_path(); |
255 | if (!path) | 258 | if (!path) |
256 | return -ENOMEM; | 259 | return -ENOMEM; |
257 | path->leave_spinning = 1; | 260 | path->leave_spinning = 1; |
258 | 261 | ||
259 | if (inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 262 | if (ino == BTRFS_FIRST_FREE_OBJECTID) { |
260 | key.objectid = BTRFS_I(inode)->root->root_key.objectid; | 263 | key.objectid = BTRFS_I(inode)->root->root_key.objectid; |
261 | key.type = BTRFS_ROOT_BACKREF_KEY; | 264 | key.type = BTRFS_ROOT_BACKREF_KEY; |
262 | key.offset = (u64)-1; | 265 | key.offset = (u64)-1; |
263 | root = root->fs_info->tree_root; | 266 | root = root->fs_info->tree_root; |
264 | } else { | 267 | } else { |
265 | key.objectid = inode->i_ino; | 268 | key.objectid = ino; |
266 | key.offset = dir->i_ino; | 269 | key.offset = btrfs_ino(dir); |
267 | key.type = BTRFS_INODE_REF_KEY; | 270 | key.type = BTRFS_INODE_REF_KEY; |
268 | } | 271 | } |
269 | 272 | ||
@@ -272,7 +275,7 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
272 | btrfs_free_path(path); | 275 | btrfs_free_path(path); |
273 | return ret; | 276 | return ret; |
274 | } else if (ret > 0) { | 277 | } else if (ret > 0) { |
275 | if (inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 278 | if (ino == BTRFS_FIRST_FREE_OBJECTID) { |
276 | path->slots[0]--; | 279 | path->slots[0]--; |
277 | } else { | 280 | } else { |
278 | btrfs_free_path(path); | 281 | btrfs_free_path(path); |
@@ -281,11 +284,11 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
281 | } | 284 | } |
282 | leaf = path->nodes[0]; | 285 | leaf = path->nodes[0]; |
283 | 286 | ||
284 | if (inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 287 | if (ino == BTRFS_FIRST_FREE_OBJECTID) { |
285 | rref = btrfs_item_ptr(leaf, path->slots[0], | 288 | rref = btrfs_item_ptr(leaf, path->slots[0], |
286 | struct btrfs_root_ref); | 289 | struct btrfs_root_ref); |
287 | name_ptr = (unsigned long)(rref + 1); | 290 | name_ptr = (unsigned long)(rref + 1); |
288 | name_len = btrfs_root_ref_name_len(leaf, rref); | 291 | name_len = btrfs_root_ref_name_len(leaf, rref); |
289 | } else { | 292 | } else { |
290 | iref = btrfs_item_ptr(leaf, path->slots[0], | 293 | iref = btrfs_item_ptr(leaf, path->slots[0], |
291 | struct btrfs_inode_ref); | 294 | struct btrfs_inode_ref); |
diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c index b457f195636..98ca149bdbc 100644 --- a/fs/btrfs/extent-tree.c +++ b/fs/btrfs/extent-tree.c | |||
@@ -105,6 +105,7 @@ void btrfs_put_block_group(struct btrfs_block_group_cache *cache) | |||
105 | WARN_ON(cache->pinned > 0); | 105 | WARN_ON(cache->pinned > 0); |
106 | WARN_ON(cache->reserved > 0); | 106 | WARN_ON(cache->reserved > 0); |
107 | WARN_ON(cache->reserved_pinned > 0); | 107 | WARN_ON(cache->reserved_pinned > 0); |
108 | kfree(cache->free_space_ctl); | ||
108 | kfree(cache); | 109 | kfree(cache); |
109 | } | 110 | } |
110 | } | 111 | } |
@@ -3036,7 +3037,8 @@ int btrfs_check_data_free_space(struct inode *inode, u64 bytes) | |||
3036 | /* make sure bytes are sectorsize aligned */ | 3037 | /* make sure bytes are sectorsize aligned */ |
3037 | bytes = (bytes + root->sectorsize - 1) & ~((u64)root->sectorsize - 1); | 3038 | bytes = (bytes + root->sectorsize - 1) & ~((u64)root->sectorsize - 1); |
3038 | 3039 | ||
3039 | if (root == root->fs_info->tree_root) { | 3040 | if (root == root->fs_info->tree_root || |
3041 | BTRFS_I(inode)->location.objectid == BTRFS_FREE_INO_OBJECTID) { | ||
3040 | alloc_chunk = 0; | 3042 | alloc_chunk = 0; |
3041 | committed = 1; | 3043 | committed = 1; |
3042 | } | 3044 | } |
@@ -3834,12 +3836,6 @@ static void release_global_block_rsv(struct btrfs_fs_info *fs_info) | |||
3834 | WARN_ON(fs_info->chunk_block_rsv.reserved > 0); | 3836 | WARN_ON(fs_info->chunk_block_rsv.reserved > 0); |
3835 | } | 3837 | } |
3836 | 3838 | ||
3837 | static u64 calc_trans_metadata_size(struct btrfs_root *root, int num_items) | ||
3838 | { | ||
3839 | return (root->leafsize + root->nodesize * (BTRFS_MAX_LEVEL - 1)) * | ||
3840 | 3 * num_items; | ||
3841 | } | ||
3842 | |||
3843 | int btrfs_trans_reserve_metadata(struct btrfs_trans_handle *trans, | 3839 | int btrfs_trans_reserve_metadata(struct btrfs_trans_handle *trans, |
3844 | struct btrfs_root *root, | 3840 | struct btrfs_root *root, |
3845 | int num_items) | 3841 | int num_items) |
@@ -3850,7 +3846,7 @@ int btrfs_trans_reserve_metadata(struct btrfs_trans_handle *trans, | |||
3850 | if (num_items == 0 || root->fs_info->chunk_root == root) | 3846 | if (num_items == 0 || root->fs_info->chunk_root == root) |
3851 | return 0; | 3847 | return 0; |
3852 | 3848 | ||
3853 | num_bytes = calc_trans_metadata_size(root, num_items); | 3849 | num_bytes = btrfs_calc_trans_metadata_size(root, num_items); |
3854 | ret = btrfs_block_rsv_add(trans, root, &root->fs_info->trans_block_rsv, | 3850 | ret = btrfs_block_rsv_add(trans, root, &root->fs_info->trans_block_rsv, |
3855 | num_bytes); | 3851 | num_bytes); |
3856 | if (!ret) { | 3852 | if (!ret) { |
@@ -3889,14 +3885,14 @@ int btrfs_orphan_reserve_metadata(struct btrfs_trans_handle *trans, | |||
3889 | * If all of the metadata space is used, we can commit | 3885 | * If all of the metadata space is used, we can commit |
3890 | * transaction and use space it freed. | 3886 | * transaction and use space it freed. |
3891 | */ | 3887 | */ |
3892 | u64 num_bytes = calc_trans_metadata_size(root, 4); | 3888 | u64 num_bytes = btrfs_calc_trans_metadata_size(root, 4); |
3893 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); | 3889 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); |
3894 | } | 3890 | } |
3895 | 3891 | ||
3896 | void btrfs_orphan_release_metadata(struct inode *inode) | 3892 | void btrfs_orphan_release_metadata(struct inode *inode) |
3897 | { | 3893 | { |
3898 | struct btrfs_root *root = BTRFS_I(inode)->root; | 3894 | struct btrfs_root *root = BTRFS_I(inode)->root; |
3899 | u64 num_bytes = calc_trans_metadata_size(root, 4); | 3895 | u64 num_bytes = btrfs_calc_trans_metadata_size(root, 4); |
3900 | btrfs_block_rsv_release(root, root->orphan_block_rsv, num_bytes); | 3896 | btrfs_block_rsv_release(root, root->orphan_block_rsv, num_bytes); |
3901 | } | 3897 | } |
3902 | 3898 | ||
@@ -3910,7 +3906,7 @@ int btrfs_snap_reserve_metadata(struct btrfs_trans_handle *trans, | |||
3910 | * two for root back/forward refs, two for directory entries | 3906 | * two for root back/forward refs, two for directory entries |
3911 | * and one for root of the snapshot. | 3907 | * and one for root of the snapshot. |
3912 | */ | 3908 | */ |
3913 | u64 num_bytes = calc_trans_metadata_size(root, 5); | 3909 | u64 num_bytes = btrfs_calc_trans_metadata_size(root, 5); |
3914 | dst_rsv->space_info = src_rsv->space_info; | 3910 | dst_rsv->space_info = src_rsv->space_info; |
3915 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); | 3911 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); |
3916 | } | 3912 | } |
@@ -3939,7 +3935,7 @@ int btrfs_delalloc_reserve_metadata(struct inode *inode, u64 num_bytes) | |||
3939 | 3935 | ||
3940 | if (nr_extents > reserved_extents) { | 3936 | if (nr_extents > reserved_extents) { |
3941 | nr_extents -= reserved_extents; | 3937 | nr_extents -= reserved_extents; |
3942 | to_reserve = calc_trans_metadata_size(root, nr_extents); | 3938 | to_reserve = btrfs_calc_trans_metadata_size(root, nr_extents); |
3943 | } else { | 3939 | } else { |
3944 | nr_extents = 0; | 3940 | nr_extents = 0; |
3945 | to_reserve = 0; | 3941 | to_reserve = 0; |
@@ -3993,7 +3989,7 @@ void btrfs_delalloc_release_metadata(struct inode *inode, u64 num_bytes) | |||
3993 | 3989 | ||
3994 | to_free = calc_csum_metadata_size(inode, num_bytes); | 3990 | to_free = calc_csum_metadata_size(inode, num_bytes); |
3995 | if (nr_extents > 0) | 3991 | if (nr_extents > 0) |
3996 | to_free += calc_trans_metadata_size(root, nr_extents); | 3992 | to_free += btrfs_calc_trans_metadata_size(root, nr_extents); |
3997 | 3993 | ||
3998 | btrfs_block_rsv_release(root, &root->fs_info->delalloc_block_rsv, | 3994 | btrfs_block_rsv_release(root, &root->fs_info->delalloc_block_rsv, |
3999 | to_free); | 3995 | to_free); |
@@ -4754,7 +4750,7 @@ wait_block_group_cache_progress(struct btrfs_block_group_cache *cache, | |||
4754 | return 0; | 4750 | return 0; |
4755 | 4751 | ||
4756 | wait_event(caching_ctl->wait, block_group_cache_done(cache) || | 4752 | wait_event(caching_ctl->wait, block_group_cache_done(cache) || |
4757 | (cache->free_space >= num_bytes)); | 4753 | (cache->free_space_ctl->free_space >= num_bytes)); |
4758 | 4754 | ||
4759 | put_caching_control(caching_ctl); | 4755 | put_caching_control(caching_ctl); |
4760 | return 0; | 4756 | return 0; |
@@ -6908,10 +6904,16 @@ int btrfs_read_block_groups(struct btrfs_root *root) | |||
6908 | ret = -ENOMEM; | 6904 | ret = -ENOMEM; |
6909 | goto error; | 6905 | goto error; |
6910 | } | 6906 | } |
6907 | cache->free_space_ctl = kzalloc(sizeof(*cache->free_space_ctl), | ||
6908 | GFP_NOFS); | ||
6909 | if (!cache->free_space_ctl) { | ||
6910 | kfree(cache); | ||
6911 | ret = -ENOMEM; | ||
6912 | goto error; | ||
6913 | } | ||
6911 | 6914 | ||
6912 | atomic_set(&cache->count, 1); | 6915 | atomic_set(&cache->count, 1); |
6913 | spin_lock_init(&cache->lock); | 6916 | spin_lock_init(&cache->lock); |
6914 | spin_lock_init(&cache->tree_lock); | ||
6915 | cache->fs_info = info; | 6917 | cache->fs_info = info; |
6916 | INIT_LIST_HEAD(&cache->list); | 6918 | INIT_LIST_HEAD(&cache->list); |
6917 | INIT_LIST_HEAD(&cache->cluster_list); | 6919 | INIT_LIST_HEAD(&cache->cluster_list); |
@@ -6919,14 +6921,6 @@ int btrfs_read_block_groups(struct btrfs_root *root) | |||
6919 | if (need_clear) | 6921 | if (need_clear) |
6920 | cache->disk_cache_state = BTRFS_DC_CLEAR; | 6922 | cache->disk_cache_state = BTRFS_DC_CLEAR; |
6921 | 6923 | ||
6922 | /* | ||
6923 | * we only want to have 32k of ram per block group for keeping | ||
6924 | * track of free space, and if we pass 1/2 of that we want to | ||
6925 | * start converting things over to using bitmaps | ||
6926 | */ | ||
6927 | cache->extents_thresh = ((1024 * 32) / 2) / | ||
6928 | sizeof(struct btrfs_free_space); | ||
6929 | |||
6930 | read_extent_buffer(leaf, &cache->item, | 6924 | read_extent_buffer(leaf, &cache->item, |
6931 | btrfs_item_ptr_offset(leaf, path->slots[0]), | 6925 | btrfs_item_ptr_offset(leaf, path->slots[0]), |
6932 | sizeof(cache->item)); | 6926 | sizeof(cache->item)); |
@@ -6937,6 +6931,8 @@ int btrfs_read_block_groups(struct btrfs_root *root) | |||
6937 | cache->flags = btrfs_block_group_flags(&cache->item); | 6931 | cache->flags = btrfs_block_group_flags(&cache->item); |
6938 | cache->sectorsize = root->sectorsize; | 6932 | cache->sectorsize = root->sectorsize; |
6939 | 6933 | ||
6934 | btrfs_init_free_space_ctl(cache); | ||
6935 | |||
6940 | /* | 6936 | /* |
6941 | * We need to exclude the super stripes now so that the space | 6937 | * We need to exclude the super stripes now so that the space |
6942 | * info has super bytes accounted for, otherwise we'll think | 6938 | * info has super bytes accounted for, otherwise we'll think |
@@ -7023,6 +7019,12 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans, | |||
7023 | cache = kzalloc(sizeof(*cache), GFP_NOFS); | 7019 | cache = kzalloc(sizeof(*cache), GFP_NOFS); |
7024 | if (!cache) | 7020 | if (!cache) |
7025 | return -ENOMEM; | 7021 | return -ENOMEM; |
7022 | cache->free_space_ctl = kzalloc(sizeof(*cache->free_space_ctl), | ||
7023 | GFP_NOFS); | ||
7024 | if (!cache->free_space_ctl) { | ||
7025 | kfree(cache); | ||
7026 | return -ENOMEM; | ||
7027 | } | ||
7026 | 7028 | ||
7027 | cache->key.objectid = chunk_offset; | 7029 | cache->key.objectid = chunk_offset; |
7028 | cache->key.offset = size; | 7030 | cache->key.offset = size; |
@@ -7030,19 +7032,13 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans, | |||
7030 | cache->sectorsize = root->sectorsize; | 7032 | cache->sectorsize = root->sectorsize; |
7031 | cache->fs_info = root->fs_info; | 7033 | cache->fs_info = root->fs_info; |
7032 | 7034 | ||
7033 | /* | ||
7034 | * we only want to have 32k of ram per block group for keeping track | ||
7035 | * of free space, and if we pass 1/2 of that we want to start | ||
7036 | * converting things over to using bitmaps | ||
7037 | */ | ||
7038 | cache->extents_thresh = ((1024 * 32) / 2) / | ||
7039 | sizeof(struct btrfs_free_space); | ||
7040 | atomic_set(&cache->count, 1); | 7035 | atomic_set(&cache->count, 1); |
7041 | spin_lock_init(&cache->lock); | 7036 | spin_lock_init(&cache->lock); |
7042 | spin_lock_init(&cache->tree_lock); | ||
7043 | INIT_LIST_HEAD(&cache->list); | 7037 | INIT_LIST_HEAD(&cache->list); |
7044 | INIT_LIST_HEAD(&cache->cluster_list); | 7038 | INIT_LIST_HEAD(&cache->cluster_list); |
7045 | 7039 | ||
7040 | btrfs_init_free_space_ctl(cache); | ||
7041 | |||
7046 | btrfs_set_block_group_used(&cache->item, bytes_used); | 7042 | btrfs_set_block_group_used(&cache->item, bytes_used); |
7047 | btrfs_set_block_group_chunk_objectid(&cache->item, chunk_objectid); | 7043 | btrfs_set_block_group_chunk_objectid(&cache->item, chunk_objectid); |
7048 | cache->flags = type; | 7044 | cache->flags = type; |
@@ -7209,23 +7205,38 @@ out: | |||
7209 | int btrfs_init_space_info(struct btrfs_fs_info *fs_info) | 7205 | int btrfs_init_space_info(struct btrfs_fs_info *fs_info) |
7210 | { | 7206 | { |
7211 | struct btrfs_space_info *space_info; | 7207 | struct btrfs_space_info *space_info; |
7208 | struct btrfs_super_block *disk_super; | ||
7209 | u64 features; | ||
7210 | u64 flags; | ||
7211 | int mixed = 0; | ||
7212 | int ret; | 7212 | int ret; |
7213 | 7213 | ||
7214 | ret = update_space_info(fs_info, BTRFS_BLOCK_GROUP_SYSTEM, 0, 0, | 7214 | disk_super = &fs_info->super_copy; |
7215 | &space_info); | 7215 | if (!btrfs_super_root(disk_super)) |
7216 | if (ret) | 7216 | return 1; |
7217 | return ret; | ||
7218 | 7217 | ||
7219 | ret = update_space_info(fs_info, BTRFS_BLOCK_GROUP_METADATA, 0, 0, | 7218 | features = btrfs_super_incompat_flags(disk_super); |
7220 | &space_info); | 7219 | if (features & BTRFS_FEATURE_INCOMPAT_MIXED_GROUPS) |
7221 | if (ret) | 7220 | mixed = 1; |
7222 | return ret; | ||
7223 | 7221 | ||
7224 | ret = update_space_info(fs_info, BTRFS_BLOCK_GROUP_DATA, 0, 0, | 7222 | flags = BTRFS_BLOCK_GROUP_SYSTEM; |
7225 | &space_info); | 7223 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); |
7226 | if (ret) | 7224 | if (ret) |
7227 | return ret; | 7225 | goto out; |
7228 | 7226 | ||
7227 | if (mixed) { | ||
7228 | flags = BTRFS_BLOCK_GROUP_METADATA | BTRFS_BLOCK_GROUP_DATA; | ||
7229 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); | ||
7230 | } else { | ||
7231 | flags = BTRFS_BLOCK_GROUP_METADATA; | ||
7232 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); | ||
7233 | if (ret) | ||
7234 | goto out; | ||
7235 | |||
7236 | flags = BTRFS_BLOCK_GROUP_DATA; | ||
7237 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); | ||
7238 | } | ||
7239 | out: | ||
7229 | return ret; | 7240 | return ret; |
7230 | } | 7241 | } |
7231 | 7242 | ||
diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c index 91208296ff2..64c8b361b53 100644 --- a/fs/btrfs/extent_io.c +++ b/fs/btrfs/extent_io.c | |||
@@ -2810,7 +2810,7 @@ int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
2810 | * because there might be preallocation past i_size | 2810 | * because there might be preallocation past i_size |
2811 | */ | 2811 | */ |
2812 | ret = btrfs_lookup_file_extent(NULL, BTRFS_I(inode)->root, | 2812 | ret = btrfs_lookup_file_extent(NULL, BTRFS_I(inode)->root, |
2813 | path, inode->i_ino, -1, 0); | 2813 | path, btrfs_ino(inode), -1, 0); |
2814 | if (ret < 0) { | 2814 | if (ret < 0) { |
2815 | btrfs_free_path(path); | 2815 | btrfs_free_path(path); |
2816 | return ret; | 2816 | return ret; |
@@ -2823,7 +2823,7 @@ int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
2823 | found_type = btrfs_key_type(&found_key); | 2823 | found_type = btrfs_key_type(&found_key); |
2824 | 2824 | ||
2825 | /* No extents, but there might be delalloc bits */ | 2825 | /* No extents, but there might be delalloc bits */ |
2826 | if (found_key.objectid != inode->i_ino || | 2826 | if (found_key.objectid != btrfs_ino(inode) || |
2827 | found_type != BTRFS_EXTENT_DATA_KEY) { | 2827 | found_type != BTRFS_EXTENT_DATA_KEY) { |
2828 | /* have to trust i_size as the end */ | 2828 | /* have to trust i_size as the end */ |
2829 | last = (u64)-1; | 2829 | last = (u64)-1; |
diff --git a/fs/btrfs/file-item.c b/fs/btrfs/file-item.c index f47e43d855a..29e014984f6 100644 --- a/fs/btrfs/file-item.c +++ b/fs/btrfs/file-item.c | |||
@@ -208,8 +208,9 @@ static int __btrfs_lookup_bio_sums(struct btrfs_root *root, | |||
208 | EXTENT_NODATASUM, GFP_NOFS); | 208 | EXTENT_NODATASUM, GFP_NOFS); |
209 | } else { | 209 | } else { |
210 | printk(KERN_INFO "btrfs no csum found " | 210 | printk(KERN_INFO "btrfs no csum found " |
211 | "for inode %lu start %llu\n", | 211 | "for inode %llu start %llu\n", |
212 | inode->i_ino, | 212 | (unsigned long long) |
213 | btrfs_ino(inode), | ||
213 | (unsigned long long)offset); | 214 | (unsigned long long)offset); |
214 | } | 215 | } |
215 | item = NULL; | 216 | item = NULL; |
diff --git a/fs/btrfs/file.c b/fs/btrfs/file.c index 566bdf298ea..58ddc444215 100644 --- a/fs/btrfs/file.c +++ b/fs/btrfs/file.c | |||
@@ -298,6 +298,7 @@ int btrfs_drop_extents(struct btrfs_trans_handle *trans, struct inode *inode, | |||
298 | struct btrfs_path *path; | 298 | struct btrfs_path *path; |
299 | struct btrfs_key key; | 299 | struct btrfs_key key; |
300 | struct btrfs_key new_key; | 300 | struct btrfs_key new_key; |
301 | u64 ino = btrfs_ino(inode); | ||
301 | u64 search_start = start; | 302 | u64 search_start = start; |
302 | u64 disk_bytenr = 0; | 303 | u64 disk_bytenr = 0; |
303 | u64 num_bytes = 0; | 304 | u64 num_bytes = 0; |
@@ -318,14 +319,14 @@ int btrfs_drop_extents(struct btrfs_trans_handle *trans, struct inode *inode, | |||
318 | 319 | ||
319 | while (1) { | 320 | while (1) { |
320 | recow = 0; | 321 | recow = 0; |
321 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 322 | ret = btrfs_lookup_file_extent(trans, root, path, ino, |
322 | search_start, -1); | 323 | search_start, -1); |
323 | if (ret < 0) | 324 | if (ret < 0) |
324 | break; | 325 | break; |
325 | if (ret > 0 && path->slots[0] > 0 && search_start == start) { | 326 | if (ret > 0 && path->slots[0] > 0 && search_start == start) { |
326 | leaf = path->nodes[0]; | 327 | leaf = path->nodes[0]; |
327 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0] - 1); | 328 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0] - 1); |
328 | if (key.objectid == inode->i_ino && | 329 | if (key.objectid == ino && |
329 | key.type == BTRFS_EXTENT_DATA_KEY) | 330 | key.type == BTRFS_EXTENT_DATA_KEY) |
330 | path->slots[0]--; | 331 | path->slots[0]--; |
331 | } | 332 | } |
@@ -346,7 +347,7 @@ next_slot: | |||
346 | } | 347 | } |
347 | 348 | ||
348 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); | 349 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); |
349 | if (key.objectid > inode->i_ino || | 350 | if (key.objectid > ino || |
350 | key.type > BTRFS_EXTENT_DATA_KEY || key.offset >= end) | 351 | key.type > BTRFS_EXTENT_DATA_KEY || key.offset >= end) |
351 | break; | 352 | break; |
352 | 353 | ||
@@ -592,6 +593,7 @@ int btrfs_mark_extent_written(struct btrfs_trans_handle *trans, | |||
592 | int del_slot = 0; | 593 | int del_slot = 0; |
593 | int recow; | 594 | int recow; |
594 | int ret; | 595 | int ret; |
596 | u64 ino = btrfs_ino(inode); | ||
595 | 597 | ||
596 | btrfs_drop_extent_cache(inode, start, end - 1, 0); | 598 | btrfs_drop_extent_cache(inode, start, end - 1, 0); |
597 | 599 | ||
@@ -600,7 +602,7 @@ int btrfs_mark_extent_written(struct btrfs_trans_handle *trans, | |||
600 | again: | 602 | again: |
601 | recow = 0; | 603 | recow = 0; |
602 | split = start; | 604 | split = start; |
603 | key.objectid = inode->i_ino; | 605 | key.objectid = ino; |
604 | key.type = BTRFS_EXTENT_DATA_KEY; | 606 | key.type = BTRFS_EXTENT_DATA_KEY; |
605 | key.offset = split; | 607 | key.offset = split; |
606 | 608 | ||
@@ -612,8 +614,7 @@ again: | |||
612 | 614 | ||
613 | leaf = path->nodes[0]; | 615 | leaf = path->nodes[0]; |
614 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); | 616 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); |
615 | BUG_ON(key.objectid != inode->i_ino || | 617 | BUG_ON(key.objectid != ino || key.type != BTRFS_EXTENT_DATA_KEY); |
616 | key.type != BTRFS_EXTENT_DATA_KEY); | ||
617 | fi = btrfs_item_ptr(leaf, path->slots[0], | 618 | fi = btrfs_item_ptr(leaf, path->slots[0], |
618 | struct btrfs_file_extent_item); | 619 | struct btrfs_file_extent_item); |
619 | BUG_ON(btrfs_file_extent_type(leaf, fi) != | 620 | BUG_ON(btrfs_file_extent_type(leaf, fi) != |
@@ -630,7 +631,7 @@ again: | |||
630 | other_start = 0; | 631 | other_start = 0; |
631 | other_end = start; | 632 | other_end = start; |
632 | if (extent_mergeable(leaf, path->slots[0] - 1, | 633 | if (extent_mergeable(leaf, path->slots[0] - 1, |
633 | inode->i_ino, bytenr, orig_offset, | 634 | ino, bytenr, orig_offset, |
634 | &other_start, &other_end)) { | 635 | &other_start, &other_end)) { |
635 | new_key.offset = end; | 636 | new_key.offset = end; |
636 | btrfs_set_item_key_safe(trans, root, path, &new_key); | 637 | btrfs_set_item_key_safe(trans, root, path, &new_key); |
@@ -653,7 +654,7 @@ again: | |||
653 | other_start = end; | 654 | other_start = end; |
654 | other_end = 0; | 655 | other_end = 0; |
655 | if (extent_mergeable(leaf, path->slots[0] + 1, | 656 | if (extent_mergeable(leaf, path->slots[0] + 1, |
656 | inode->i_ino, bytenr, orig_offset, | 657 | ino, bytenr, orig_offset, |
657 | &other_start, &other_end)) { | 658 | &other_start, &other_end)) { |
658 | fi = btrfs_item_ptr(leaf, path->slots[0], | 659 | fi = btrfs_item_ptr(leaf, path->slots[0], |
659 | struct btrfs_file_extent_item); | 660 | struct btrfs_file_extent_item); |
@@ -702,7 +703,7 @@ again: | |||
702 | 703 | ||
703 | ret = btrfs_inc_extent_ref(trans, root, bytenr, num_bytes, 0, | 704 | ret = btrfs_inc_extent_ref(trans, root, bytenr, num_bytes, 0, |
704 | root->root_key.objectid, | 705 | root->root_key.objectid, |
705 | inode->i_ino, orig_offset); | 706 | ino, orig_offset); |
706 | BUG_ON(ret); | 707 | BUG_ON(ret); |
707 | 708 | ||
708 | if (split == start) { | 709 | if (split == start) { |
@@ -718,7 +719,7 @@ again: | |||
718 | other_start = end; | 719 | other_start = end; |
719 | other_end = 0; | 720 | other_end = 0; |
720 | if (extent_mergeable(leaf, path->slots[0] + 1, | 721 | if (extent_mergeable(leaf, path->slots[0] + 1, |
721 | inode->i_ino, bytenr, orig_offset, | 722 | ino, bytenr, orig_offset, |
722 | &other_start, &other_end)) { | 723 | &other_start, &other_end)) { |
723 | if (recow) { | 724 | if (recow) { |
724 | btrfs_release_path(path); | 725 | btrfs_release_path(path); |
@@ -729,13 +730,13 @@ again: | |||
729 | del_nr++; | 730 | del_nr++; |
730 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, | 731 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, |
731 | 0, root->root_key.objectid, | 732 | 0, root->root_key.objectid, |
732 | inode->i_ino, orig_offset); | 733 | ino, orig_offset); |
733 | BUG_ON(ret); | 734 | BUG_ON(ret); |
734 | } | 735 | } |
735 | other_start = 0; | 736 | other_start = 0; |
736 | other_end = start; | 737 | other_end = start; |
737 | if (extent_mergeable(leaf, path->slots[0] - 1, | 738 | if (extent_mergeable(leaf, path->slots[0] - 1, |
738 | inode->i_ino, bytenr, orig_offset, | 739 | ino, bytenr, orig_offset, |
739 | &other_start, &other_end)) { | 740 | &other_start, &other_end)) { |
740 | if (recow) { | 741 | if (recow) { |
741 | btrfs_release_path(path); | 742 | btrfs_release_path(path); |
@@ -746,7 +747,7 @@ again: | |||
746 | del_nr++; | 747 | del_nr++; |
747 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, | 748 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, |
748 | 0, root->root_key.objectid, | 749 | 0, root->root_key.objectid, |
749 | inode->i_ino, orig_offset); | 750 | ino, orig_offset); |
750 | BUG_ON(ret); | 751 | BUG_ON(ret); |
751 | } | 752 | } |
752 | if (del_nr == 0) { | 753 | if (del_nr == 0) { |
diff --git a/fs/btrfs/free-space-cache.c b/fs/btrfs/free-space-cache.c index 0290b0c7b00..70d45795d75 100644 --- a/fs/btrfs/free-space-cache.c +++ b/fs/btrfs/free-space-cache.c | |||
@@ -25,18 +25,17 @@ | |||
25 | #include "transaction.h" | 25 | #include "transaction.h" |
26 | #include "disk-io.h" | 26 | #include "disk-io.h" |
27 | #include "extent_io.h" | 27 | #include "extent_io.h" |
28 | #include "inode-map.h" | ||
28 | 29 | ||
29 | #define BITS_PER_BITMAP (PAGE_CACHE_SIZE * 8) | 30 | #define BITS_PER_BITMAP (PAGE_CACHE_SIZE * 8) |
30 | #define MAX_CACHE_BYTES_PER_GIG (32 * 1024) | 31 | #define MAX_CACHE_BYTES_PER_GIG (32 * 1024) |
31 | 32 | ||
32 | static void recalculate_thresholds(struct btrfs_block_group_cache | 33 | static int link_free_space(struct btrfs_free_space_ctl *ctl, |
33 | *block_group); | ||
34 | static int link_free_space(struct btrfs_block_group_cache *block_group, | ||
35 | struct btrfs_free_space *info); | 34 | struct btrfs_free_space *info); |
36 | 35 | ||
37 | struct inode *lookup_free_space_inode(struct btrfs_root *root, | 36 | static struct inode *__lookup_free_space_inode(struct btrfs_root *root, |
38 | struct btrfs_block_group_cache | 37 | struct btrfs_path *path, |
39 | *block_group, struct btrfs_path *path) | 38 | u64 offset) |
40 | { | 39 | { |
41 | struct btrfs_key key; | 40 | struct btrfs_key key; |
42 | struct btrfs_key location; | 41 | struct btrfs_key location; |
@@ -46,15 +45,8 @@ struct inode *lookup_free_space_inode(struct btrfs_root *root, | |||
46 | struct inode *inode = NULL; | 45 | struct inode *inode = NULL; |
47 | int ret; | 46 | int ret; |
48 | 47 | ||
49 | spin_lock(&block_group->lock); | ||
50 | if (block_group->inode) | ||
51 | inode = igrab(block_group->inode); | ||
52 | spin_unlock(&block_group->lock); | ||
53 | if (inode) | ||
54 | return inode; | ||
55 | |||
56 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 48 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
57 | key.offset = block_group->key.objectid; | 49 | key.offset = offset; |
58 | key.type = 0; | 50 | key.type = 0; |
59 | 51 | ||
60 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | 52 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); |
@@ -84,6 +76,27 @@ struct inode *lookup_free_space_inode(struct btrfs_root *root, | |||
84 | 76 | ||
85 | inode->i_mapping->flags &= ~__GFP_FS; | 77 | inode->i_mapping->flags &= ~__GFP_FS; |
86 | 78 | ||
79 | return inode; | ||
80 | } | ||
81 | |||
82 | struct inode *lookup_free_space_inode(struct btrfs_root *root, | ||
83 | struct btrfs_block_group_cache | ||
84 | *block_group, struct btrfs_path *path) | ||
85 | { | ||
86 | struct inode *inode = NULL; | ||
87 | |||
88 | spin_lock(&block_group->lock); | ||
89 | if (block_group->inode) | ||
90 | inode = igrab(block_group->inode); | ||
91 | spin_unlock(&block_group->lock); | ||
92 | if (inode) | ||
93 | return inode; | ||
94 | |||
95 | inode = __lookup_free_space_inode(root, path, | ||
96 | block_group->key.objectid); | ||
97 | if (IS_ERR(inode)) | ||
98 | return inode; | ||
99 | |||
87 | spin_lock(&block_group->lock); | 100 | spin_lock(&block_group->lock); |
88 | if (!root->fs_info->closing) { | 101 | if (!root->fs_info->closing) { |
89 | block_group->inode = igrab(inode); | 102 | block_group->inode = igrab(inode); |
@@ -94,24 +107,18 @@ struct inode *lookup_free_space_inode(struct btrfs_root *root, | |||
94 | return inode; | 107 | return inode; |
95 | } | 108 | } |
96 | 109 | ||
97 | int create_free_space_inode(struct btrfs_root *root, | 110 | int __create_free_space_inode(struct btrfs_root *root, |
98 | struct btrfs_trans_handle *trans, | 111 | struct btrfs_trans_handle *trans, |
99 | struct btrfs_block_group_cache *block_group, | 112 | struct btrfs_path *path, u64 ino, u64 offset) |
100 | struct btrfs_path *path) | ||
101 | { | 113 | { |
102 | struct btrfs_key key; | 114 | struct btrfs_key key; |
103 | struct btrfs_disk_key disk_key; | 115 | struct btrfs_disk_key disk_key; |
104 | struct btrfs_free_space_header *header; | 116 | struct btrfs_free_space_header *header; |
105 | struct btrfs_inode_item *inode_item; | 117 | struct btrfs_inode_item *inode_item; |
106 | struct extent_buffer *leaf; | 118 | struct extent_buffer *leaf; |
107 | u64 objectid; | ||
108 | int ret; | 119 | int ret; |
109 | 120 | ||
110 | ret = btrfs_find_free_objectid(trans, root, 0, &objectid); | 121 | ret = btrfs_insert_empty_inode(trans, root, path, ino); |
111 | if (ret < 0) | ||
112 | return ret; | ||
113 | |||
114 | ret = btrfs_insert_empty_inode(trans, root, path, objectid); | ||
115 | if (ret) | 122 | if (ret) |
116 | return ret; | 123 | return ret; |
117 | 124 | ||
@@ -131,13 +138,12 @@ int create_free_space_inode(struct btrfs_root *root, | |||
131 | BTRFS_INODE_PREALLOC | BTRFS_INODE_NODATASUM); | 138 | BTRFS_INODE_PREALLOC | BTRFS_INODE_NODATASUM); |
132 | btrfs_set_inode_nlink(leaf, inode_item, 1); | 139 | btrfs_set_inode_nlink(leaf, inode_item, 1); |
133 | btrfs_set_inode_transid(leaf, inode_item, trans->transid); | 140 | btrfs_set_inode_transid(leaf, inode_item, trans->transid); |
134 | btrfs_set_inode_block_group(leaf, inode_item, | 141 | btrfs_set_inode_block_group(leaf, inode_item, offset); |
135 | block_group->key.objectid); | ||
136 | btrfs_mark_buffer_dirty(leaf); | 142 | btrfs_mark_buffer_dirty(leaf); |
137 | btrfs_release_path(path); | 143 | btrfs_release_path(path); |
138 | 144 | ||
139 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 145 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
140 | key.offset = block_group->key.objectid; | 146 | key.offset = offset; |
141 | key.type = 0; | 147 | key.type = 0; |
142 | 148 | ||
143 | ret = btrfs_insert_empty_item(trans, root, path, &key, | 149 | ret = btrfs_insert_empty_item(trans, root, path, &key, |
@@ -157,6 +163,22 @@ int create_free_space_inode(struct btrfs_root *root, | |||
157 | return 0; | 163 | return 0; |
158 | } | 164 | } |
159 | 165 | ||
166 | int create_free_space_inode(struct btrfs_root *root, | ||
167 | struct btrfs_trans_handle *trans, | ||
168 | struct btrfs_block_group_cache *block_group, | ||
169 | struct btrfs_path *path) | ||
170 | { | ||
171 | int ret; | ||
172 | u64 ino; | ||
173 | |||
174 | ret = btrfs_find_free_objectid(root, &ino); | ||
175 | if (ret < 0) | ||
176 | return ret; | ||
177 | |||
178 | return __create_free_space_inode(root, trans, path, ino, | ||
179 | block_group->key.objectid); | ||
180 | } | ||
181 | |||
160 | int btrfs_truncate_free_space_cache(struct btrfs_root *root, | 182 | int btrfs_truncate_free_space_cache(struct btrfs_root *root, |
161 | struct btrfs_trans_handle *trans, | 183 | struct btrfs_trans_handle *trans, |
162 | struct btrfs_path *path, | 184 | struct btrfs_path *path, |
@@ -187,7 +209,8 @@ int btrfs_truncate_free_space_cache(struct btrfs_root *root, | |||
187 | return ret; | 209 | return ret; |
188 | } | 210 | } |
189 | 211 | ||
190 | return btrfs_update_inode(trans, root, inode); | 212 | ret = btrfs_update_inode(trans, root, inode); |
213 | return ret; | ||
191 | } | 214 | } |
192 | 215 | ||
193 | static int readahead_cache(struct inode *inode) | 216 | static int readahead_cache(struct inode *inode) |
@@ -209,15 +232,13 @@ static int readahead_cache(struct inode *inode) | |||
209 | return 0; | 232 | return 0; |
210 | } | 233 | } |
211 | 234 | ||
212 | int load_free_space_cache(struct btrfs_fs_info *fs_info, | 235 | int __load_free_space_cache(struct btrfs_root *root, struct inode *inode, |
213 | struct btrfs_block_group_cache *block_group) | 236 | struct btrfs_free_space_ctl *ctl, |
237 | struct btrfs_path *path, u64 offset) | ||
214 | { | 238 | { |
215 | struct btrfs_root *root = fs_info->tree_root; | ||
216 | struct inode *inode; | ||
217 | struct btrfs_free_space_header *header; | 239 | struct btrfs_free_space_header *header; |
218 | struct extent_buffer *leaf; | 240 | struct extent_buffer *leaf; |
219 | struct page *page; | 241 | struct page *page; |
220 | struct btrfs_path *path; | ||
221 | u32 *checksums = NULL, *crc; | 242 | u32 *checksums = NULL, *crc; |
222 | char *disk_crcs = NULL; | 243 | char *disk_crcs = NULL; |
223 | struct btrfs_key key; | 244 | struct btrfs_key key; |
@@ -225,76 +246,47 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
225 | u64 num_entries; | 246 | u64 num_entries; |
226 | u64 num_bitmaps; | 247 | u64 num_bitmaps; |
227 | u64 generation; | 248 | u64 generation; |
228 | u64 used = btrfs_block_group_used(&block_group->item); | ||
229 | u32 cur_crc = ~(u32)0; | 249 | u32 cur_crc = ~(u32)0; |
230 | pgoff_t index = 0; | 250 | pgoff_t index = 0; |
231 | unsigned long first_page_offset; | 251 | unsigned long first_page_offset; |
232 | int num_checksums; | 252 | int num_checksums; |
233 | int ret = 0; | 253 | int ret = 0, ret2; |
234 | |||
235 | /* | ||
236 | * If we're unmounting then just return, since this does a search on the | ||
237 | * normal root and not the commit root and we could deadlock. | ||
238 | */ | ||
239 | smp_mb(); | ||
240 | if (fs_info->closing) | ||
241 | return 0; | ||
242 | |||
243 | /* | ||
244 | * If this block group has been marked to be cleared for one reason or | ||
245 | * another then we can't trust the on disk cache, so just return. | ||
246 | */ | ||
247 | spin_lock(&block_group->lock); | ||
248 | if (block_group->disk_cache_state != BTRFS_DC_WRITTEN) { | ||
249 | spin_unlock(&block_group->lock); | ||
250 | return 0; | ||
251 | } | ||
252 | spin_unlock(&block_group->lock); | ||
253 | 254 | ||
254 | INIT_LIST_HEAD(&bitmaps); | 255 | INIT_LIST_HEAD(&bitmaps); |
255 | 256 | ||
256 | path = btrfs_alloc_path(); | ||
257 | if (!path) | ||
258 | return 0; | ||
259 | |||
260 | inode = lookup_free_space_inode(root, block_group, path); | ||
261 | if (IS_ERR(inode)) { | ||
262 | btrfs_free_path(path); | ||
263 | return 0; | ||
264 | } | ||
265 | |||
266 | /* Nothing in the space cache, goodbye */ | 257 | /* Nothing in the space cache, goodbye */ |
267 | if (!i_size_read(inode)) { | 258 | if (!i_size_read(inode)) |
268 | btrfs_free_path(path); | ||
269 | goto out; | 259 | goto out; |
270 | } | ||
271 | 260 | ||
272 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 261 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
273 | key.offset = block_group->key.objectid; | 262 | key.offset = offset; |
274 | key.type = 0; | 263 | key.type = 0; |
275 | 264 | ||
276 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | 265 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); |
277 | if (ret) { | 266 | if (ret < 0) |
278 | btrfs_free_path(path); | 267 | goto out; |
268 | else if (ret > 0) { | ||
269 | btrfs_release_path(path); | ||
270 | ret = 0; | ||
279 | goto out; | 271 | goto out; |
280 | } | 272 | } |
281 | 273 | ||
274 | ret = -1; | ||
275 | |||
282 | leaf = path->nodes[0]; | 276 | leaf = path->nodes[0]; |
283 | header = btrfs_item_ptr(leaf, path->slots[0], | 277 | header = btrfs_item_ptr(leaf, path->slots[0], |
284 | struct btrfs_free_space_header); | 278 | struct btrfs_free_space_header); |
285 | num_entries = btrfs_free_space_entries(leaf, header); | 279 | num_entries = btrfs_free_space_entries(leaf, header); |
286 | num_bitmaps = btrfs_free_space_bitmaps(leaf, header); | 280 | num_bitmaps = btrfs_free_space_bitmaps(leaf, header); |
287 | generation = btrfs_free_space_generation(leaf, header); | 281 | generation = btrfs_free_space_generation(leaf, header); |
288 | btrfs_free_path(path); | 282 | btrfs_release_path(path); |
289 | 283 | ||
290 | if (BTRFS_I(inode)->generation != generation) { | 284 | if (BTRFS_I(inode)->generation != generation) { |
291 | printk(KERN_ERR "btrfs: free space inode generation (%llu) did" | 285 | printk(KERN_ERR "btrfs: free space inode generation (%llu) did" |
292 | " not match free space cache generation (%llu) for " | 286 | " not match free space cache generation (%llu)\n", |
293 | "block group %llu\n", | ||
294 | (unsigned long long)BTRFS_I(inode)->generation, | 287 | (unsigned long long)BTRFS_I(inode)->generation, |
295 | (unsigned long long)generation, | 288 | (unsigned long long)generation); |
296 | (unsigned long long)block_group->key.objectid); | 289 | goto out; |
297 | goto free_cache; | ||
298 | } | 290 | } |
299 | 291 | ||
300 | if (!num_entries) | 292 | if (!num_entries) |
@@ -311,10 +303,8 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
311 | goto out; | 303 | goto out; |
312 | 304 | ||
313 | ret = readahead_cache(inode); | 305 | ret = readahead_cache(inode); |
314 | if (ret) { | 306 | if (ret) |
315 | ret = 0; | ||
316 | goto out; | 307 | goto out; |
317 | } | ||
318 | 308 | ||
319 | while (1) { | 309 | while (1) { |
320 | struct btrfs_free_space_entry *entry; | 310 | struct btrfs_free_space_entry *entry; |
@@ -333,10 +323,8 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
333 | } | 323 | } |
334 | 324 | ||
335 | page = grab_cache_page(inode->i_mapping, index); | 325 | page = grab_cache_page(inode->i_mapping, index); |
336 | if (!page) { | 326 | if (!page) |
337 | ret = 0; | ||
338 | goto free_cache; | 327 | goto free_cache; |
339 | } | ||
340 | 328 | ||
341 | if (!PageUptodate(page)) { | 329 | if (!PageUptodate(page)) { |
342 | btrfs_readpage(NULL, page); | 330 | btrfs_readpage(NULL, page); |
@@ -345,9 +333,7 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
345 | unlock_page(page); | 333 | unlock_page(page); |
346 | page_cache_release(page); | 334 | page_cache_release(page); |
347 | printk(KERN_ERR "btrfs: error reading free " | 335 | printk(KERN_ERR "btrfs: error reading free " |
348 | "space cache: %llu\n", | 336 | "space cache\n"); |
349 | (unsigned long long) | ||
350 | block_group->key.objectid); | ||
351 | goto free_cache; | 337 | goto free_cache; |
352 | } | 338 | } |
353 | } | 339 | } |
@@ -360,13 +346,10 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
360 | gen = addr + (sizeof(u32) * num_checksums); | 346 | gen = addr + (sizeof(u32) * num_checksums); |
361 | if (*gen != BTRFS_I(inode)->generation) { | 347 | if (*gen != BTRFS_I(inode)->generation) { |
362 | printk(KERN_ERR "btrfs: space cache generation" | 348 | printk(KERN_ERR "btrfs: space cache generation" |
363 | " (%llu) does not match inode (%llu) " | 349 | " (%llu) does not match inode (%llu)\n", |
364 | "for block group %llu\n", | ||
365 | (unsigned long long)*gen, | 350 | (unsigned long long)*gen, |
366 | (unsigned long long) | 351 | (unsigned long long) |
367 | BTRFS_I(inode)->generation, | 352 | BTRFS_I(inode)->generation); |
368 | (unsigned long long) | ||
369 | block_group->key.objectid); | ||
370 | kunmap(page); | 353 | kunmap(page); |
371 | unlock_page(page); | 354 | unlock_page(page); |
372 | page_cache_release(page); | 355 | page_cache_release(page); |
@@ -382,9 +365,8 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
382 | PAGE_CACHE_SIZE - start_offset); | 365 | PAGE_CACHE_SIZE - start_offset); |
383 | btrfs_csum_final(cur_crc, (char *)&cur_crc); | 366 | btrfs_csum_final(cur_crc, (char *)&cur_crc); |
384 | if (cur_crc != *crc) { | 367 | if (cur_crc != *crc) { |
385 | printk(KERN_ERR "btrfs: crc mismatch for page %lu in " | 368 | printk(KERN_ERR "btrfs: crc mismatch for page %lu\n", |
386 | "block group %llu\n", index, | 369 | index); |
387 | (unsigned long long)block_group->key.objectid); | ||
388 | kunmap(page); | 370 | kunmap(page); |
389 | unlock_page(page); | 371 | unlock_page(page); |
390 | page_cache_release(page); | 372 | page_cache_release(page); |
@@ -417,9 +399,9 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
417 | } | 399 | } |
418 | 400 | ||
419 | if (entry->type == BTRFS_FREE_SPACE_EXTENT) { | 401 | if (entry->type == BTRFS_FREE_SPACE_EXTENT) { |
420 | spin_lock(&block_group->tree_lock); | 402 | spin_lock(&ctl->tree_lock); |
421 | ret = link_free_space(block_group, e); | 403 | ret = link_free_space(ctl, e); |
422 | spin_unlock(&block_group->tree_lock); | 404 | spin_unlock(&ctl->tree_lock); |
423 | BUG_ON(ret); | 405 | BUG_ON(ret); |
424 | } else { | 406 | } else { |
425 | e->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); | 407 | e->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); |
@@ -431,11 +413,11 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
431 | page_cache_release(page); | 413 | page_cache_release(page); |
432 | goto free_cache; | 414 | goto free_cache; |
433 | } | 415 | } |
434 | spin_lock(&block_group->tree_lock); | 416 | spin_lock(&ctl->tree_lock); |
435 | ret = link_free_space(block_group, e); | 417 | ret2 = link_free_space(ctl, e); |
436 | block_group->total_bitmaps++; | 418 | ctl->total_bitmaps++; |
437 | recalculate_thresholds(block_group); | 419 | ctl->op->recalc_thresholds(ctl); |
438 | spin_unlock(&block_group->tree_lock); | 420 | spin_unlock(&ctl->tree_lock); |
439 | list_add_tail(&e->list, &bitmaps); | 421 | list_add_tail(&e->list, &bitmaps); |
440 | } | 422 | } |
441 | 423 | ||
@@ -471,41 +453,97 @@ next: | |||
471 | index++; | 453 | index++; |
472 | } | 454 | } |
473 | 455 | ||
474 | spin_lock(&block_group->tree_lock); | ||
475 | if (block_group->free_space != (block_group->key.offset - used - | ||
476 | block_group->bytes_super)) { | ||
477 | spin_unlock(&block_group->tree_lock); | ||
478 | printk(KERN_ERR "block group %llu has an wrong amount of free " | ||
479 | "space\n", block_group->key.objectid); | ||
480 | ret = 0; | ||
481 | goto free_cache; | ||
482 | } | ||
483 | spin_unlock(&block_group->tree_lock); | ||
484 | |||
485 | ret = 1; | 456 | ret = 1; |
486 | out: | 457 | out: |
487 | kfree(checksums); | 458 | kfree(checksums); |
488 | kfree(disk_crcs); | 459 | kfree(disk_crcs); |
489 | iput(inode); | ||
490 | return ret; | 460 | return ret; |
491 | |||
492 | free_cache: | 461 | free_cache: |
493 | /* This cache is bogus, make sure it gets cleared */ | 462 | __btrfs_remove_free_space_cache(ctl); |
463 | goto out; | ||
464 | } | ||
465 | |||
466 | int load_free_space_cache(struct btrfs_fs_info *fs_info, | ||
467 | struct btrfs_block_group_cache *block_group) | ||
468 | { | ||
469 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
470 | struct btrfs_root *root = fs_info->tree_root; | ||
471 | struct inode *inode; | ||
472 | struct btrfs_path *path; | ||
473 | int ret; | ||
474 | bool matched; | ||
475 | u64 used = btrfs_block_group_used(&block_group->item); | ||
476 | |||
477 | /* | ||
478 | * If we're unmounting then just return, since this does a search on the | ||
479 | * normal root and not the commit root and we could deadlock. | ||
480 | */ | ||
481 | smp_mb(); | ||
482 | if (fs_info->closing) | ||
483 | return 0; | ||
484 | |||
485 | /* | ||
486 | * If this block group has been marked to be cleared for one reason or | ||
487 | * another then we can't trust the on disk cache, so just return. | ||
488 | */ | ||
494 | spin_lock(&block_group->lock); | 489 | spin_lock(&block_group->lock); |
495 | block_group->disk_cache_state = BTRFS_DC_CLEAR; | 490 | if (block_group->disk_cache_state != BTRFS_DC_WRITTEN) { |
491 | spin_unlock(&block_group->lock); | ||
492 | return 0; | ||
493 | } | ||
496 | spin_unlock(&block_group->lock); | 494 | spin_unlock(&block_group->lock); |
497 | btrfs_remove_free_space_cache(block_group); | 495 | |
498 | goto out; | 496 | path = btrfs_alloc_path(); |
497 | if (!path) | ||
498 | return 0; | ||
499 | |||
500 | inode = lookup_free_space_inode(root, block_group, path); | ||
501 | if (IS_ERR(inode)) { | ||
502 | btrfs_free_path(path); | ||
503 | return 0; | ||
504 | } | ||
505 | |||
506 | ret = __load_free_space_cache(fs_info->tree_root, inode, ctl, | ||
507 | path, block_group->key.objectid); | ||
508 | btrfs_free_path(path); | ||
509 | if (ret <= 0) | ||
510 | goto out; | ||
511 | |||
512 | spin_lock(&ctl->tree_lock); | ||
513 | matched = (ctl->free_space == (block_group->key.offset - used - | ||
514 | block_group->bytes_super)); | ||
515 | spin_unlock(&ctl->tree_lock); | ||
516 | |||
517 | if (!matched) { | ||
518 | __btrfs_remove_free_space_cache(ctl); | ||
519 | printk(KERN_ERR "block group %llu has an wrong amount of free " | ||
520 | "space\n", block_group->key.objectid); | ||
521 | ret = -1; | ||
522 | } | ||
523 | out: | ||
524 | if (ret < 0) { | ||
525 | /* This cache is bogus, make sure it gets cleared */ | ||
526 | spin_lock(&block_group->lock); | ||
527 | block_group->disk_cache_state = BTRFS_DC_CLEAR; | ||
528 | spin_unlock(&block_group->lock); | ||
529 | ret = 0; | ||
530 | |||
531 | printk(KERN_ERR "btrfs: failed to load free space cache " | ||
532 | "for block group %llu\n", block_group->key.objectid); | ||
533 | } | ||
534 | |||
535 | iput(inode); | ||
536 | return ret; | ||
499 | } | 537 | } |
500 | 538 | ||
501 | int btrfs_write_out_cache(struct btrfs_root *root, | 539 | int __btrfs_write_out_cache(struct btrfs_root *root, struct inode *inode, |
502 | struct btrfs_trans_handle *trans, | 540 | struct btrfs_free_space_ctl *ctl, |
503 | struct btrfs_block_group_cache *block_group, | 541 | struct btrfs_block_group_cache *block_group, |
504 | struct btrfs_path *path) | 542 | struct btrfs_trans_handle *trans, |
543 | struct btrfs_path *path, u64 offset) | ||
505 | { | 544 | { |
506 | struct btrfs_free_space_header *header; | 545 | struct btrfs_free_space_header *header; |
507 | struct extent_buffer *leaf; | 546 | struct extent_buffer *leaf; |
508 | struct inode *inode; | ||
509 | struct rb_node *node; | 547 | struct rb_node *node; |
510 | struct list_head *pos, *n; | 548 | struct list_head *pos, *n; |
511 | struct page **pages; | 549 | struct page **pages; |
@@ -522,35 +560,18 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
522 | int index = 0, num_pages = 0; | 560 | int index = 0, num_pages = 0; |
523 | int entries = 0; | 561 | int entries = 0; |
524 | int bitmaps = 0; | 562 | int bitmaps = 0; |
525 | int ret = 0; | 563 | int ret = -1; |
526 | bool next_page = false; | 564 | bool next_page = false; |
527 | bool out_of_space = false; | 565 | bool out_of_space = false; |
528 | 566 | ||
529 | root = root->fs_info->tree_root; | ||
530 | |||
531 | INIT_LIST_HEAD(&bitmap_list); | 567 | INIT_LIST_HEAD(&bitmap_list); |
532 | 568 | ||
533 | spin_lock(&block_group->lock); | 569 | node = rb_first(&ctl->free_space_offset); |
534 | if (block_group->disk_cache_state < BTRFS_DC_SETUP) { | 570 | if (!node) |
535 | spin_unlock(&block_group->lock); | ||
536 | return 0; | ||
537 | } | ||
538 | spin_unlock(&block_group->lock); | ||
539 | |||
540 | inode = lookup_free_space_inode(root, block_group, path); | ||
541 | if (IS_ERR(inode)) | ||
542 | return 0; | ||
543 | |||
544 | if (!i_size_read(inode)) { | ||
545 | iput(inode); | ||
546 | return 0; | 571 | return 0; |
547 | } | ||
548 | 572 | ||
549 | node = rb_first(&block_group->free_space_offset); | 573 | if (!i_size_read(inode)) |
550 | if (!node) { | 574 | return -1; |
551 | iput(inode); | ||
552 | return 0; | ||
553 | } | ||
554 | 575 | ||
555 | num_pages = (i_size_read(inode) + PAGE_CACHE_SIZE - 1) >> | 576 | num_pages = (i_size_read(inode) + PAGE_CACHE_SIZE - 1) >> |
556 | PAGE_CACHE_SHIFT; | 577 | PAGE_CACHE_SHIFT; |
@@ -560,16 +581,13 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
560 | 581 | ||
561 | /* We need a checksum per page. */ | 582 | /* We need a checksum per page. */ |
562 | crc = checksums = kzalloc(sizeof(u32) * num_pages, GFP_NOFS); | 583 | crc = checksums = kzalloc(sizeof(u32) * num_pages, GFP_NOFS); |
563 | if (!crc) { | 584 | if (!crc) |
564 | iput(inode); | 585 | return -1; |
565 | return 0; | ||
566 | } | ||
567 | 586 | ||
568 | pages = kzalloc(sizeof(struct page *) * num_pages, GFP_NOFS); | 587 | pages = kzalloc(sizeof(struct page *) * num_pages, GFP_NOFS); |
569 | if (!pages) { | 588 | if (!pages) { |
570 | kfree(crc); | 589 | kfree(crc); |
571 | iput(inode); | 590 | return -1; |
572 | return 0; | ||
573 | } | 591 | } |
574 | 592 | ||
575 | /* Since the first page has all of our checksums and our generation we | 593 | /* Since the first page has all of our checksums and our generation we |
@@ -579,7 +597,7 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
579 | first_page_offset = (sizeof(u32) * num_pages) + sizeof(u64); | 597 | first_page_offset = (sizeof(u32) * num_pages) + sizeof(u64); |
580 | 598 | ||
581 | /* Get the cluster for this block_group if it exists */ | 599 | /* Get the cluster for this block_group if it exists */ |
582 | if (!list_empty(&block_group->cluster_list)) | 600 | if (block_group && !list_empty(&block_group->cluster_list)) |
583 | cluster = list_entry(block_group->cluster_list.next, | 601 | cluster = list_entry(block_group->cluster_list.next, |
584 | struct btrfs_free_cluster, | 602 | struct btrfs_free_cluster, |
585 | block_group_list); | 603 | block_group_list); |
@@ -621,7 +639,8 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
621 | * When searching for pinned extents, we need to start at our start | 639 | * When searching for pinned extents, we need to start at our start |
622 | * offset. | 640 | * offset. |
623 | */ | 641 | */ |
624 | start = block_group->key.objectid; | 642 | if (block_group) |
643 | start = block_group->key.objectid; | ||
625 | 644 | ||
626 | /* Write out the extent entries */ | 645 | /* Write out the extent entries */ |
627 | do { | 646 | do { |
@@ -679,8 +698,9 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
679 | * We want to add any pinned extents to our free space cache | 698 | * We want to add any pinned extents to our free space cache |
680 | * so we don't leak the space | 699 | * so we don't leak the space |
681 | */ | 700 | */ |
682 | while (!next_page && (start < block_group->key.objectid + | 701 | while (block_group && !next_page && |
683 | block_group->key.offset)) { | 702 | (start < block_group->key.objectid + |
703 | block_group->key.offset)) { | ||
684 | ret = find_first_extent_bit(unpin, start, &start, &end, | 704 | ret = find_first_extent_bit(unpin, start, &start, &end, |
685 | EXTENT_DIRTY); | 705 | EXTENT_DIRTY); |
686 | if (ret) { | 706 | if (ret) { |
@@ -798,12 +818,12 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
798 | filemap_write_and_wait(inode->i_mapping); | 818 | filemap_write_and_wait(inode->i_mapping); |
799 | 819 | ||
800 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 820 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
801 | key.offset = block_group->key.objectid; | 821 | key.offset = offset; |
802 | key.type = 0; | 822 | key.type = 0; |
803 | 823 | ||
804 | ret = btrfs_search_slot(trans, root, &key, path, 1, 1); | 824 | ret = btrfs_search_slot(trans, root, &key, path, 1, 1); |
805 | if (ret < 0) { | 825 | if (ret < 0) { |
806 | ret = 0; | 826 | ret = -1; |
807 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, | 827 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, |
808 | EXTENT_DIRTY | EXTENT_DELALLOC | | 828 | EXTENT_DIRTY | EXTENT_DELALLOC | |
809 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, GFP_NOFS); | 829 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, GFP_NOFS); |
@@ -816,8 +836,8 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
816 | path->slots[0]--; | 836 | path->slots[0]--; |
817 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); | 837 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); |
818 | if (found_key.objectid != BTRFS_FREE_SPACE_OBJECTID || | 838 | if (found_key.objectid != BTRFS_FREE_SPACE_OBJECTID || |
819 | found_key.offset != block_group->key.objectid) { | 839 | found_key.offset != offset) { |
820 | ret = 0; | 840 | ret = -1; |
821 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, | 841 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, |
822 | EXTENT_DIRTY | EXTENT_DELALLOC | | 842 | EXTENT_DIRTY | EXTENT_DELALLOC | |
823 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, | 843 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, |
@@ -837,44 +857,78 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
837 | ret = 1; | 857 | ret = 1; |
838 | 858 | ||
839 | out_free: | 859 | out_free: |
840 | if (ret == 0) { | 860 | if (ret != 1) { |
841 | invalidate_inode_pages2_range(inode->i_mapping, 0, index); | 861 | invalidate_inode_pages2_range(inode->i_mapping, 0, index); |
842 | spin_lock(&block_group->lock); | ||
843 | block_group->disk_cache_state = BTRFS_DC_ERROR; | ||
844 | spin_unlock(&block_group->lock); | ||
845 | BTRFS_I(inode)->generation = 0; | 862 | BTRFS_I(inode)->generation = 0; |
846 | } | 863 | } |
847 | kfree(checksums); | 864 | kfree(checksums); |
848 | kfree(pages); | 865 | kfree(pages); |
849 | btrfs_update_inode(trans, root, inode); | 866 | btrfs_update_inode(trans, root, inode); |
867 | return ret; | ||
868 | } | ||
869 | |||
870 | int btrfs_write_out_cache(struct btrfs_root *root, | ||
871 | struct btrfs_trans_handle *trans, | ||
872 | struct btrfs_block_group_cache *block_group, | ||
873 | struct btrfs_path *path) | ||
874 | { | ||
875 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
876 | struct inode *inode; | ||
877 | int ret = 0; | ||
878 | |||
879 | root = root->fs_info->tree_root; | ||
880 | |||
881 | spin_lock(&block_group->lock); | ||
882 | if (block_group->disk_cache_state < BTRFS_DC_SETUP) { | ||
883 | spin_unlock(&block_group->lock); | ||
884 | return 0; | ||
885 | } | ||
886 | spin_unlock(&block_group->lock); | ||
887 | |||
888 | inode = lookup_free_space_inode(root, block_group, path); | ||
889 | if (IS_ERR(inode)) | ||
890 | return 0; | ||
891 | |||
892 | ret = __btrfs_write_out_cache(root, inode, ctl, block_group, trans, | ||
893 | path, block_group->key.objectid); | ||
894 | if (ret < 0) { | ||
895 | spin_lock(&block_group->lock); | ||
896 | block_group->disk_cache_state = BTRFS_DC_ERROR; | ||
897 | spin_unlock(&block_group->lock); | ||
898 | ret = 0; | ||
899 | |||
900 | printk(KERN_ERR "btrfs: failed to write free space cace " | ||
901 | "for block group %llu\n", block_group->key.objectid); | ||
902 | } | ||
903 | |||
850 | iput(inode); | 904 | iput(inode); |
851 | return ret; | 905 | return ret; |
852 | } | 906 | } |
853 | 907 | ||
854 | static inline unsigned long offset_to_bit(u64 bitmap_start, u64 sectorsize, | 908 | static inline unsigned long offset_to_bit(u64 bitmap_start, u32 unit, |
855 | u64 offset) | 909 | u64 offset) |
856 | { | 910 | { |
857 | BUG_ON(offset < bitmap_start); | 911 | BUG_ON(offset < bitmap_start); |
858 | offset -= bitmap_start; | 912 | offset -= bitmap_start; |
859 | return (unsigned long)(div64_u64(offset, sectorsize)); | 913 | return (unsigned long)(div_u64(offset, unit)); |
860 | } | 914 | } |
861 | 915 | ||
862 | static inline unsigned long bytes_to_bits(u64 bytes, u64 sectorsize) | 916 | static inline unsigned long bytes_to_bits(u64 bytes, u32 unit) |
863 | { | 917 | { |
864 | return (unsigned long)(div64_u64(bytes, sectorsize)); | 918 | return (unsigned long)(div_u64(bytes, unit)); |
865 | } | 919 | } |
866 | 920 | ||
867 | static inline u64 offset_to_bitmap(struct btrfs_block_group_cache *block_group, | 921 | static inline u64 offset_to_bitmap(struct btrfs_free_space_ctl *ctl, |
868 | u64 offset) | 922 | u64 offset) |
869 | { | 923 | { |
870 | u64 bitmap_start; | 924 | u64 bitmap_start; |
871 | u64 bytes_per_bitmap; | 925 | u64 bytes_per_bitmap; |
872 | 926 | ||
873 | bytes_per_bitmap = BITS_PER_BITMAP * block_group->sectorsize; | 927 | bytes_per_bitmap = BITS_PER_BITMAP * ctl->unit; |
874 | bitmap_start = offset - block_group->key.objectid; | 928 | bitmap_start = offset - ctl->start; |
875 | bitmap_start = div64_u64(bitmap_start, bytes_per_bitmap); | 929 | bitmap_start = div64_u64(bitmap_start, bytes_per_bitmap); |
876 | bitmap_start *= bytes_per_bitmap; | 930 | bitmap_start *= bytes_per_bitmap; |
877 | bitmap_start += block_group->key.objectid; | 931 | bitmap_start += ctl->start; |
878 | 932 | ||
879 | return bitmap_start; | 933 | return bitmap_start; |
880 | } | 934 | } |
@@ -932,10 +986,10 @@ static int tree_insert_offset(struct rb_root *root, u64 offset, | |||
932 | * offset. | 986 | * offset. |
933 | */ | 987 | */ |
934 | static struct btrfs_free_space * | 988 | static struct btrfs_free_space * |
935 | tree_search_offset(struct btrfs_block_group_cache *block_group, | 989 | tree_search_offset(struct btrfs_free_space_ctl *ctl, |
936 | u64 offset, int bitmap_only, int fuzzy) | 990 | u64 offset, int bitmap_only, int fuzzy) |
937 | { | 991 | { |
938 | struct rb_node *n = block_group->free_space_offset.rb_node; | 992 | struct rb_node *n = ctl->free_space_offset.rb_node; |
939 | struct btrfs_free_space *entry, *prev = NULL; | 993 | struct btrfs_free_space *entry, *prev = NULL; |
940 | 994 | ||
941 | /* find entry that is closest to the 'offset' */ | 995 | /* find entry that is closest to the 'offset' */ |
@@ -1031,8 +1085,7 @@ tree_search_offset(struct btrfs_block_group_cache *block_group, | |||
1031 | break; | 1085 | break; |
1032 | } | 1086 | } |
1033 | } | 1087 | } |
1034 | if (entry->offset + BITS_PER_BITMAP * | 1088 | if (entry->offset + BITS_PER_BITMAP * ctl->unit > offset) |
1035 | block_group->sectorsize > offset) | ||
1036 | return entry; | 1089 | return entry; |
1037 | } else if (entry->offset + entry->bytes > offset) | 1090 | } else if (entry->offset + entry->bytes > offset) |
1038 | return entry; | 1091 | return entry; |
@@ -1043,7 +1096,7 @@ tree_search_offset(struct btrfs_block_group_cache *block_group, | |||
1043 | while (1) { | 1096 | while (1) { |
1044 | if (entry->bitmap) { | 1097 | if (entry->bitmap) { |
1045 | if (entry->offset + BITS_PER_BITMAP * | 1098 | if (entry->offset + BITS_PER_BITMAP * |
1046 | block_group->sectorsize > offset) | 1099 | ctl->unit > offset) |
1047 | break; | 1100 | break; |
1048 | } else { | 1101 | } else { |
1049 | if (entry->offset + entry->bytes > offset) | 1102 | if (entry->offset + entry->bytes > offset) |
@@ -1059,42 +1112,47 @@ tree_search_offset(struct btrfs_block_group_cache *block_group, | |||
1059 | } | 1112 | } |
1060 | 1113 | ||
1061 | static inline void | 1114 | static inline void |
1062 | __unlink_free_space(struct btrfs_block_group_cache *block_group, | 1115 | __unlink_free_space(struct btrfs_free_space_ctl *ctl, |
1063 | struct btrfs_free_space *info) | 1116 | struct btrfs_free_space *info) |
1064 | { | 1117 | { |
1065 | rb_erase(&info->offset_index, &block_group->free_space_offset); | 1118 | rb_erase(&info->offset_index, &ctl->free_space_offset); |
1066 | block_group->free_extents--; | 1119 | ctl->free_extents--; |
1067 | } | 1120 | } |
1068 | 1121 | ||
1069 | static void unlink_free_space(struct btrfs_block_group_cache *block_group, | 1122 | static void unlink_free_space(struct btrfs_free_space_ctl *ctl, |
1070 | struct btrfs_free_space *info) | 1123 | struct btrfs_free_space *info) |
1071 | { | 1124 | { |
1072 | __unlink_free_space(block_group, info); | 1125 | __unlink_free_space(ctl, info); |
1073 | block_group->free_space -= info->bytes; | 1126 | ctl->free_space -= info->bytes; |
1074 | } | 1127 | } |
1075 | 1128 | ||
1076 | static int link_free_space(struct btrfs_block_group_cache *block_group, | 1129 | static int link_free_space(struct btrfs_free_space_ctl *ctl, |
1077 | struct btrfs_free_space *info) | 1130 | struct btrfs_free_space *info) |
1078 | { | 1131 | { |
1079 | int ret = 0; | 1132 | int ret = 0; |
1080 | 1133 | ||
1081 | BUG_ON(!info->bitmap && !info->bytes); | 1134 | BUG_ON(!info->bitmap && !info->bytes); |
1082 | ret = tree_insert_offset(&block_group->free_space_offset, info->offset, | 1135 | ret = tree_insert_offset(&ctl->free_space_offset, info->offset, |
1083 | &info->offset_index, (info->bitmap != NULL)); | 1136 | &info->offset_index, (info->bitmap != NULL)); |
1084 | if (ret) | 1137 | if (ret) |
1085 | return ret; | 1138 | return ret; |
1086 | 1139 | ||
1087 | block_group->free_space += info->bytes; | 1140 | ctl->free_space += info->bytes; |
1088 | block_group->free_extents++; | 1141 | ctl->free_extents++; |
1089 | return ret; | 1142 | return ret; |
1090 | } | 1143 | } |
1091 | 1144 | ||
1092 | static void recalculate_thresholds(struct btrfs_block_group_cache *block_group) | 1145 | static void recalculate_thresholds(struct btrfs_free_space_ctl *ctl) |
1093 | { | 1146 | { |
1147 | struct btrfs_block_group_cache *block_group = ctl->private; | ||
1094 | u64 max_bytes; | 1148 | u64 max_bytes; |
1095 | u64 bitmap_bytes; | 1149 | u64 bitmap_bytes; |
1096 | u64 extent_bytes; | 1150 | u64 extent_bytes; |
1097 | u64 size = block_group->key.offset; | 1151 | u64 size = block_group->key.offset; |
1152 | u64 bytes_per_bg = BITS_PER_BITMAP * block_group->sectorsize; | ||
1153 | int max_bitmaps = div64_u64(size + bytes_per_bg - 1, bytes_per_bg); | ||
1154 | |||
1155 | BUG_ON(ctl->total_bitmaps > max_bitmaps); | ||
1098 | 1156 | ||
1099 | /* | 1157 | /* |
1100 | * The goal is to keep the total amount of memory used per 1gb of space | 1158 | * The goal is to keep the total amount of memory used per 1gb of space |
@@ -1112,10 +1170,10 @@ static void recalculate_thresholds(struct btrfs_block_group_cache *block_group) | |||
1112 | * sure we don't go over our overall goal of MAX_CACHE_BYTES_PER_GIG as | 1170 | * sure we don't go over our overall goal of MAX_CACHE_BYTES_PER_GIG as |
1113 | * we add more bitmaps. | 1171 | * we add more bitmaps. |
1114 | */ | 1172 | */ |
1115 | bitmap_bytes = (block_group->total_bitmaps + 1) * PAGE_CACHE_SIZE; | 1173 | bitmap_bytes = (ctl->total_bitmaps + 1) * PAGE_CACHE_SIZE; |
1116 | 1174 | ||
1117 | if (bitmap_bytes >= max_bytes) { | 1175 | if (bitmap_bytes >= max_bytes) { |
1118 | block_group->extents_thresh = 0; | 1176 | ctl->extents_thresh = 0; |
1119 | return; | 1177 | return; |
1120 | } | 1178 | } |
1121 | 1179 | ||
@@ -1126,47 +1184,43 @@ static void recalculate_thresholds(struct btrfs_block_group_cache *block_group) | |||
1126 | extent_bytes = max_bytes - bitmap_bytes; | 1184 | extent_bytes = max_bytes - bitmap_bytes; |
1127 | extent_bytes = min_t(u64, extent_bytes, div64_u64(max_bytes, 2)); | 1185 | extent_bytes = min_t(u64, extent_bytes, div64_u64(max_bytes, 2)); |
1128 | 1186 | ||
1129 | block_group->extents_thresh = | 1187 | ctl->extents_thresh = |
1130 | div64_u64(extent_bytes, (sizeof(struct btrfs_free_space))); | 1188 | div64_u64(extent_bytes, (sizeof(struct btrfs_free_space))); |
1131 | } | 1189 | } |
1132 | 1190 | ||
1133 | static void bitmap_clear_bits(struct btrfs_block_group_cache *block_group, | 1191 | static void bitmap_clear_bits(struct btrfs_free_space_ctl *ctl, |
1134 | struct btrfs_free_space *info, u64 offset, | 1192 | struct btrfs_free_space *info, u64 offset, |
1135 | u64 bytes) | 1193 | u64 bytes) |
1136 | { | 1194 | { |
1137 | unsigned long start, end; | 1195 | unsigned long start, count; |
1138 | unsigned long i; | ||
1139 | 1196 | ||
1140 | start = offset_to_bit(info->offset, block_group->sectorsize, offset); | 1197 | start = offset_to_bit(info->offset, ctl->unit, offset); |
1141 | end = start + bytes_to_bits(bytes, block_group->sectorsize); | 1198 | count = bytes_to_bits(bytes, ctl->unit); |
1142 | BUG_ON(end > BITS_PER_BITMAP); | 1199 | BUG_ON(start + count > BITS_PER_BITMAP); |
1143 | 1200 | ||
1144 | for (i = start; i < end; i++) | 1201 | bitmap_clear(info->bitmap, start, count); |
1145 | clear_bit(i, info->bitmap); | ||
1146 | 1202 | ||
1147 | info->bytes -= bytes; | 1203 | info->bytes -= bytes; |
1148 | block_group->free_space -= bytes; | 1204 | ctl->free_space -= bytes; |
1149 | } | 1205 | } |
1150 | 1206 | ||
1151 | static void bitmap_set_bits(struct btrfs_block_group_cache *block_group, | 1207 | static void bitmap_set_bits(struct btrfs_free_space_ctl *ctl, |
1152 | struct btrfs_free_space *info, u64 offset, | 1208 | struct btrfs_free_space *info, u64 offset, |
1153 | u64 bytes) | 1209 | u64 bytes) |
1154 | { | 1210 | { |
1155 | unsigned long start, end; | 1211 | unsigned long start, count; |
1156 | unsigned long i; | ||
1157 | 1212 | ||
1158 | start = offset_to_bit(info->offset, block_group->sectorsize, offset); | 1213 | start = offset_to_bit(info->offset, ctl->unit, offset); |
1159 | end = start + bytes_to_bits(bytes, block_group->sectorsize); | 1214 | count = bytes_to_bits(bytes, ctl->unit); |
1160 | BUG_ON(end > BITS_PER_BITMAP); | 1215 | BUG_ON(start + count > BITS_PER_BITMAP); |
1161 | 1216 | ||
1162 | for (i = start; i < end; i++) | 1217 | bitmap_set(info->bitmap, start, count); |
1163 | set_bit(i, info->bitmap); | ||
1164 | 1218 | ||
1165 | info->bytes += bytes; | 1219 | info->bytes += bytes; |
1166 | block_group->free_space += bytes; | 1220 | ctl->free_space += bytes; |
1167 | } | 1221 | } |
1168 | 1222 | ||
1169 | static int search_bitmap(struct btrfs_block_group_cache *block_group, | 1223 | static int search_bitmap(struct btrfs_free_space_ctl *ctl, |
1170 | struct btrfs_free_space *bitmap_info, u64 *offset, | 1224 | struct btrfs_free_space *bitmap_info, u64 *offset, |
1171 | u64 *bytes) | 1225 | u64 *bytes) |
1172 | { | 1226 | { |
@@ -1174,9 +1228,9 @@ static int search_bitmap(struct btrfs_block_group_cache *block_group, | |||
1174 | unsigned long bits, i; | 1228 | unsigned long bits, i; |
1175 | unsigned long next_zero; | 1229 | unsigned long next_zero; |
1176 | 1230 | ||
1177 | i = offset_to_bit(bitmap_info->offset, block_group->sectorsize, | 1231 | i = offset_to_bit(bitmap_info->offset, ctl->unit, |
1178 | max_t(u64, *offset, bitmap_info->offset)); | 1232 | max_t(u64, *offset, bitmap_info->offset)); |
1179 | bits = bytes_to_bits(*bytes, block_group->sectorsize); | 1233 | bits = bytes_to_bits(*bytes, ctl->unit); |
1180 | 1234 | ||
1181 | for (i = find_next_bit(bitmap_info->bitmap, BITS_PER_BITMAP, i); | 1235 | for (i = find_next_bit(bitmap_info->bitmap, BITS_PER_BITMAP, i); |
1182 | i < BITS_PER_BITMAP; | 1236 | i < BITS_PER_BITMAP; |
@@ -1191,29 +1245,25 @@ static int search_bitmap(struct btrfs_block_group_cache *block_group, | |||
1191 | } | 1245 | } |
1192 | 1246 | ||
1193 | if (found_bits) { | 1247 | if (found_bits) { |
1194 | *offset = (u64)(i * block_group->sectorsize) + | 1248 | *offset = (u64)(i * ctl->unit) + bitmap_info->offset; |
1195 | bitmap_info->offset; | 1249 | *bytes = (u64)(found_bits) * ctl->unit; |
1196 | *bytes = (u64)(found_bits) * block_group->sectorsize; | ||
1197 | return 0; | 1250 | return 0; |
1198 | } | 1251 | } |
1199 | 1252 | ||
1200 | return -1; | 1253 | return -1; |
1201 | } | 1254 | } |
1202 | 1255 | ||
1203 | static struct btrfs_free_space *find_free_space(struct btrfs_block_group_cache | 1256 | static struct btrfs_free_space * |
1204 | *block_group, u64 *offset, | 1257 | find_free_space(struct btrfs_free_space_ctl *ctl, u64 *offset, u64 *bytes) |
1205 | u64 *bytes, int debug) | ||
1206 | { | 1258 | { |
1207 | struct btrfs_free_space *entry; | 1259 | struct btrfs_free_space *entry; |
1208 | struct rb_node *node; | 1260 | struct rb_node *node; |
1209 | int ret; | 1261 | int ret; |
1210 | 1262 | ||
1211 | if (!block_group->free_space_offset.rb_node) | 1263 | if (!ctl->free_space_offset.rb_node) |
1212 | return NULL; | 1264 | return NULL; |
1213 | 1265 | ||
1214 | entry = tree_search_offset(block_group, | 1266 | entry = tree_search_offset(ctl, offset_to_bitmap(ctl, *offset), 0, 1); |
1215 | offset_to_bitmap(block_group, *offset), | ||
1216 | 0, 1); | ||
1217 | if (!entry) | 1267 | if (!entry) |
1218 | return NULL; | 1268 | return NULL; |
1219 | 1269 | ||
@@ -1223,7 +1273,7 @@ static struct btrfs_free_space *find_free_space(struct btrfs_block_group_cache | |||
1223 | continue; | 1273 | continue; |
1224 | 1274 | ||
1225 | if (entry->bitmap) { | 1275 | if (entry->bitmap) { |
1226 | ret = search_bitmap(block_group, entry, offset, bytes); | 1276 | ret = search_bitmap(ctl, entry, offset, bytes); |
1227 | if (!ret) | 1277 | if (!ret) |
1228 | return entry; | 1278 | return entry; |
1229 | continue; | 1279 | continue; |
@@ -1237,33 +1287,28 @@ static struct btrfs_free_space *find_free_space(struct btrfs_block_group_cache | |||
1237 | return NULL; | 1287 | return NULL; |
1238 | } | 1288 | } |
1239 | 1289 | ||
1240 | static void add_new_bitmap(struct btrfs_block_group_cache *block_group, | 1290 | static void add_new_bitmap(struct btrfs_free_space_ctl *ctl, |
1241 | struct btrfs_free_space *info, u64 offset) | 1291 | struct btrfs_free_space *info, u64 offset) |
1242 | { | 1292 | { |
1243 | u64 bytes_per_bg = BITS_PER_BITMAP * block_group->sectorsize; | 1293 | info->offset = offset_to_bitmap(ctl, offset); |
1244 | int max_bitmaps = (int)div64_u64(block_group->key.offset + | ||
1245 | bytes_per_bg - 1, bytes_per_bg); | ||
1246 | BUG_ON(block_group->total_bitmaps >= max_bitmaps); | ||
1247 | |||
1248 | info->offset = offset_to_bitmap(block_group, offset); | ||
1249 | info->bytes = 0; | 1294 | info->bytes = 0; |
1250 | link_free_space(block_group, info); | 1295 | link_free_space(ctl, info); |
1251 | block_group->total_bitmaps++; | 1296 | ctl->total_bitmaps++; |
1252 | 1297 | ||
1253 | recalculate_thresholds(block_group); | 1298 | ctl->op->recalc_thresholds(ctl); |
1254 | } | 1299 | } |
1255 | 1300 | ||
1256 | static void free_bitmap(struct btrfs_block_group_cache *block_group, | 1301 | static void free_bitmap(struct btrfs_free_space_ctl *ctl, |
1257 | struct btrfs_free_space *bitmap_info) | 1302 | struct btrfs_free_space *bitmap_info) |
1258 | { | 1303 | { |
1259 | unlink_free_space(block_group, bitmap_info); | 1304 | unlink_free_space(ctl, bitmap_info); |
1260 | kfree(bitmap_info->bitmap); | 1305 | kfree(bitmap_info->bitmap); |
1261 | kmem_cache_free(btrfs_free_space_cachep, bitmap_info); | 1306 | kmem_cache_free(btrfs_free_space_cachep, bitmap_info); |
1262 | block_group->total_bitmaps--; | 1307 | ctl->total_bitmaps--; |
1263 | recalculate_thresholds(block_group); | 1308 | ctl->op->recalc_thresholds(ctl); |
1264 | } | 1309 | } |
1265 | 1310 | ||
1266 | static noinline int remove_from_bitmap(struct btrfs_block_group_cache *block_group, | 1311 | static noinline int remove_from_bitmap(struct btrfs_free_space_ctl *ctl, |
1267 | struct btrfs_free_space *bitmap_info, | 1312 | struct btrfs_free_space *bitmap_info, |
1268 | u64 *offset, u64 *bytes) | 1313 | u64 *offset, u64 *bytes) |
1269 | { | 1314 | { |
@@ -1272,8 +1317,7 @@ static noinline int remove_from_bitmap(struct btrfs_block_group_cache *block_gro | |||
1272 | int ret; | 1317 | int ret; |
1273 | 1318 | ||
1274 | again: | 1319 | again: |
1275 | end = bitmap_info->offset + | 1320 | end = bitmap_info->offset + (u64)(BITS_PER_BITMAP * ctl->unit) - 1; |
1276 | (u64)(BITS_PER_BITMAP * block_group->sectorsize) - 1; | ||
1277 | 1321 | ||
1278 | /* | 1322 | /* |
1279 | * XXX - this can go away after a few releases. | 1323 | * XXX - this can go away after a few releases. |
@@ -1288,24 +1332,22 @@ again: | |||
1288 | search_start = *offset; | 1332 | search_start = *offset; |
1289 | search_bytes = *bytes; | 1333 | search_bytes = *bytes; |
1290 | search_bytes = min(search_bytes, end - search_start + 1); | 1334 | search_bytes = min(search_bytes, end - search_start + 1); |
1291 | ret = search_bitmap(block_group, bitmap_info, &search_start, | 1335 | ret = search_bitmap(ctl, bitmap_info, &search_start, &search_bytes); |
1292 | &search_bytes); | ||
1293 | BUG_ON(ret < 0 || search_start != *offset); | 1336 | BUG_ON(ret < 0 || search_start != *offset); |
1294 | 1337 | ||
1295 | if (*offset > bitmap_info->offset && *offset + *bytes > end) { | 1338 | if (*offset > bitmap_info->offset && *offset + *bytes > end) { |
1296 | bitmap_clear_bits(block_group, bitmap_info, *offset, | 1339 | bitmap_clear_bits(ctl, bitmap_info, *offset, end - *offset + 1); |
1297 | end - *offset + 1); | ||
1298 | *bytes -= end - *offset + 1; | 1340 | *bytes -= end - *offset + 1; |
1299 | *offset = end + 1; | 1341 | *offset = end + 1; |
1300 | } else if (*offset >= bitmap_info->offset && *offset + *bytes <= end) { | 1342 | } else if (*offset >= bitmap_info->offset && *offset + *bytes <= end) { |
1301 | bitmap_clear_bits(block_group, bitmap_info, *offset, *bytes); | 1343 | bitmap_clear_bits(ctl, bitmap_info, *offset, *bytes); |
1302 | *bytes = 0; | 1344 | *bytes = 0; |
1303 | } | 1345 | } |
1304 | 1346 | ||
1305 | if (*bytes) { | 1347 | if (*bytes) { |
1306 | struct rb_node *next = rb_next(&bitmap_info->offset_index); | 1348 | struct rb_node *next = rb_next(&bitmap_info->offset_index); |
1307 | if (!bitmap_info->bytes) | 1349 | if (!bitmap_info->bytes) |
1308 | free_bitmap(block_group, bitmap_info); | 1350 | free_bitmap(ctl, bitmap_info); |
1309 | 1351 | ||
1310 | /* | 1352 | /* |
1311 | * no entry after this bitmap, but we still have bytes to | 1353 | * no entry after this bitmap, but we still have bytes to |
@@ -1332,31 +1374,28 @@ again: | |||
1332 | */ | 1374 | */ |
1333 | search_start = *offset; | 1375 | search_start = *offset; |
1334 | search_bytes = *bytes; | 1376 | search_bytes = *bytes; |
1335 | ret = search_bitmap(block_group, bitmap_info, &search_start, | 1377 | ret = search_bitmap(ctl, bitmap_info, &search_start, |
1336 | &search_bytes); | 1378 | &search_bytes); |
1337 | if (ret < 0 || search_start != *offset) | 1379 | if (ret < 0 || search_start != *offset) |
1338 | return -EAGAIN; | 1380 | return -EAGAIN; |
1339 | 1381 | ||
1340 | goto again; | 1382 | goto again; |
1341 | } else if (!bitmap_info->bytes) | 1383 | } else if (!bitmap_info->bytes) |
1342 | free_bitmap(block_group, bitmap_info); | 1384 | free_bitmap(ctl, bitmap_info); |
1343 | 1385 | ||
1344 | return 0; | 1386 | return 0; |
1345 | } | 1387 | } |
1346 | 1388 | ||
1347 | static int insert_into_bitmap(struct btrfs_block_group_cache *block_group, | 1389 | static bool use_bitmap(struct btrfs_free_space_ctl *ctl, |
1348 | struct btrfs_free_space *info) | 1390 | struct btrfs_free_space *info) |
1349 | { | 1391 | { |
1350 | struct btrfs_free_space *bitmap_info; | 1392 | struct btrfs_block_group_cache *block_group = ctl->private; |
1351 | int added = 0; | ||
1352 | u64 bytes, offset, end; | ||
1353 | int ret; | ||
1354 | 1393 | ||
1355 | /* | 1394 | /* |
1356 | * If we are below the extents threshold then we can add this as an | 1395 | * If we are below the extents threshold then we can add this as an |
1357 | * extent, and don't have to deal with the bitmap | 1396 | * extent, and don't have to deal with the bitmap |
1358 | */ | 1397 | */ |
1359 | if (block_group->free_extents < block_group->extents_thresh) { | 1398 | if (ctl->free_extents < ctl->extents_thresh) { |
1360 | /* | 1399 | /* |
1361 | * If this block group has some small extents we don't want to | 1400 | * If this block group has some small extents we don't want to |
1362 | * use up all of our free slots in the cache with them, we want | 1401 | * use up all of our free slots in the cache with them, we want |
@@ -1365,11 +1404,10 @@ static int insert_into_bitmap(struct btrfs_block_group_cache *block_group, | |||
1365 | * the overhead of a bitmap if we don't have to. | 1404 | * the overhead of a bitmap if we don't have to. |
1366 | */ | 1405 | */ |
1367 | if (info->bytes <= block_group->sectorsize * 4) { | 1406 | if (info->bytes <= block_group->sectorsize * 4) { |
1368 | if (block_group->free_extents * 2 <= | 1407 | if (ctl->free_extents * 2 <= ctl->extents_thresh) |
1369 | block_group->extents_thresh) | 1408 | return false; |
1370 | return 0; | ||
1371 | } else { | 1409 | } else { |
1372 | return 0; | 1410 | return false; |
1373 | } | 1411 | } |
1374 | } | 1412 | } |
1375 | 1413 | ||
@@ -1379,31 +1417,42 @@ static int insert_into_bitmap(struct btrfs_block_group_cache *block_group, | |||
1379 | */ | 1417 | */ |
1380 | if (BITS_PER_BITMAP * block_group->sectorsize > | 1418 | if (BITS_PER_BITMAP * block_group->sectorsize > |
1381 | block_group->key.offset) | 1419 | block_group->key.offset) |
1382 | return 0; | 1420 | return false; |
1421 | |||
1422 | return true; | ||
1423 | } | ||
1424 | |||
1425 | static int insert_into_bitmap(struct btrfs_free_space_ctl *ctl, | ||
1426 | struct btrfs_free_space *info) | ||
1427 | { | ||
1428 | struct btrfs_free_space *bitmap_info; | ||
1429 | int added = 0; | ||
1430 | u64 bytes, offset, end; | ||
1431 | int ret; | ||
1383 | 1432 | ||
1384 | bytes = info->bytes; | 1433 | bytes = info->bytes; |
1385 | offset = info->offset; | 1434 | offset = info->offset; |
1386 | 1435 | ||
1436 | if (!ctl->op->use_bitmap(ctl, info)) | ||
1437 | return 0; | ||
1438 | |||
1387 | again: | 1439 | again: |
1388 | bitmap_info = tree_search_offset(block_group, | 1440 | bitmap_info = tree_search_offset(ctl, offset_to_bitmap(ctl, offset), |
1389 | offset_to_bitmap(block_group, offset), | ||
1390 | 1, 0); | 1441 | 1, 0); |
1391 | if (!bitmap_info) { | 1442 | if (!bitmap_info) { |
1392 | BUG_ON(added); | 1443 | BUG_ON(added); |
1393 | goto new_bitmap; | 1444 | goto new_bitmap; |
1394 | } | 1445 | } |
1395 | 1446 | ||
1396 | end = bitmap_info->offset + | 1447 | end = bitmap_info->offset + (u64)(BITS_PER_BITMAP * ctl->unit); |
1397 | (u64)(BITS_PER_BITMAP * block_group->sectorsize); | ||
1398 | 1448 | ||
1399 | if (offset >= bitmap_info->offset && offset + bytes > end) { | 1449 | if (offset >= bitmap_info->offset && offset + bytes > end) { |
1400 | bitmap_set_bits(block_group, bitmap_info, offset, | 1450 | bitmap_set_bits(ctl, bitmap_info, offset, end - offset); |
1401 | end - offset); | ||
1402 | bytes -= end - offset; | 1451 | bytes -= end - offset; |
1403 | offset = end; | 1452 | offset = end; |
1404 | added = 0; | 1453 | added = 0; |
1405 | } else if (offset >= bitmap_info->offset && offset + bytes <= end) { | 1454 | } else if (offset >= bitmap_info->offset && offset + bytes <= end) { |
1406 | bitmap_set_bits(block_group, bitmap_info, offset, bytes); | 1455 | bitmap_set_bits(ctl, bitmap_info, offset, bytes); |
1407 | bytes = 0; | 1456 | bytes = 0; |
1408 | } else { | 1457 | } else { |
1409 | BUG(); | 1458 | BUG(); |
@@ -1417,19 +1466,19 @@ again: | |||
1417 | 1466 | ||
1418 | new_bitmap: | 1467 | new_bitmap: |
1419 | if (info && info->bitmap) { | 1468 | if (info && info->bitmap) { |
1420 | add_new_bitmap(block_group, info, offset); | 1469 | add_new_bitmap(ctl, info, offset); |
1421 | added = 1; | 1470 | added = 1; |
1422 | info = NULL; | 1471 | info = NULL; |
1423 | goto again; | 1472 | goto again; |
1424 | } else { | 1473 | } else { |
1425 | spin_unlock(&block_group->tree_lock); | 1474 | spin_unlock(&ctl->tree_lock); |
1426 | 1475 | ||
1427 | /* no pre-allocated info, allocate a new one */ | 1476 | /* no pre-allocated info, allocate a new one */ |
1428 | if (!info) { | 1477 | if (!info) { |
1429 | info = kmem_cache_zalloc(btrfs_free_space_cachep, | 1478 | info = kmem_cache_zalloc(btrfs_free_space_cachep, |
1430 | GFP_NOFS); | 1479 | GFP_NOFS); |
1431 | if (!info) { | 1480 | if (!info) { |
1432 | spin_lock(&block_group->tree_lock); | 1481 | spin_lock(&ctl->tree_lock); |
1433 | ret = -ENOMEM; | 1482 | ret = -ENOMEM; |
1434 | goto out; | 1483 | goto out; |
1435 | } | 1484 | } |
@@ -1437,7 +1486,7 @@ new_bitmap: | |||
1437 | 1486 | ||
1438 | /* allocate the bitmap */ | 1487 | /* allocate the bitmap */ |
1439 | info->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); | 1488 | info->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); |
1440 | spin_lock(&block_group->tree_lock); | 1489 | spin_lock(&ctl->tree_lock); |
1441 | if (!info->bitmap) { | 1490 | if (!info->bitmap) { |
1442 | ret = -ENOMEM; | 1491 | ret = -ENOMEM; |
1443 | goto out; | 1492 | goto out; |
@@ -1455,8 +1504,8 @@ out: | |||
1455 | return ret; | 1504 | return ret; |
1456 | } | 1505 | } |
1457 | 1506 | ||
1458 | static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | 1507 | static bool try_merge_free_space(struct btrfs_free_space_ctl *ctl, |
1459 | struct btrfs_free_space *info, bool update_stat) | 1508 | struct btrfs_free_space *info, bool update_stat) |
1460 | { | 1509 | { |
1461 | struct btrfs_free_space *left_info; | 1510 | struct btrfs_free_space *left_info; |
1462 | struct btrfs_free_space *right_info; | 1511 | struct btrfs_free_space *right_info; |
@@ -1469,18 +1518,18 @@ static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | |||
1469 | * are adding, if there is remove that struct and add a new one to | 1518 | * are adding, if there is remove that struct and add a new one to |
1470 | * cover the entire range | 1519 | * cover the entire range |
1471 | */ | 1520 | */ |
1472 | right_info = tree_search_offset(block_group, offset + bytes, 0, 0); | 1521 | right_info = tree_search_offset(ctl, offset + bytes, 0, 0); |
1473 | if (right_info && rb_prev(&right_info->offset_index)) | 1522 | if (right_info && rb_prev(&right_info->offset_index)) |
1474 | left_info = rb_entry(rb_prev(&right_info->offset_index), | 1523 | left_info = rb_entry(rb_prev(&right_info->offset_index), |
1475 | struct btrfs_free_space, offset_index); | 1524 | struct btrfs_free_space, offset_index); |
1476 | else | 1525 | else |
1477 | left_info = tree_search_offset(block_group, offset - 1, 0, 0); | 1526 | left_info = tree_search_offset(ctl, offset - 1, 0, 0); |
1478 | 1527 | ||
1479 | if (right_info && !right_info->bitmap) { | 1528 | if (right_info && !right_info->bitmap) { |
1480 | if (update_stat) | 1529 | if (update_stat) |
1481 | unlink_free_space(block_group, right_info); | 1530 | unlink_free_space(ctl, right_info); |
1482 | else | 1531 | else |
1483 | __unlink_free_space(block_group, right_info); | 1532 | __unlink_free_space(ctl, right_info); |
1484 | info->bytes += right_info->bytes; | 1533 | info->bytes += right_info->bytes; |
1485 | kmem_cache_free(btrfs_free_space_cachep, right_info); | 1534 | kmem_cache_free(btrfs_free_space_cachep, right_info); |
1486 | merged = true; | 1535 | merged = true; |
@@ -1489,9 +1538,9 @@ static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | |||
1489 | if (left_info && !left_info->bitmap && | 1538 | if (left_info && !left_info->bitmap && |
1490 | left_info->offset + left_info->bytes == offset) { | 1539 | left_info->offset + left_info->bytes == offset) { |
1491 | if (update_stat) | 1540 | if (update_stat) |
1492 | unlink_free_space(block_group, left_info); | 1541 | unlink_free_space(ctl, left_info); |
1493 | else | 1542 | else |
1494 | __unlink_free_space(block_group, left_info); | 1543 | __unlink_free_space(ctl, left_info); |
1495 | info->offset = left_info->offset; | 1544 | info->offset = left_info->offset; |
1496 | info->bytes += left_info->bytes; | 1545 | info->bytes += left_info->bytes; |
1497 | kmem_cache_free(btrfs_free_space_cachep, left_info); | 1546 | kmem_cache_free(btrfs_free_space_cachep, left_info); |
@@ -1501,8 +1550,8 @@ static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | |||
1501 | return merged; | 1550 | return merged; |
1502 | } | 1551 | } |
1503 | 1552 | ||
1504 | int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | 1553 | int __btrfs_add_free_space(struct btrfs_free_space_ctl *ctl, |
1505 | u64 offset, u64 bytes) | 1554 | u64 offset, u64 bytes) |
1506 | { | 1555 | { |
1507 | struct btrfs_free_space *info; | 1556 | struct btrfs_free_space *info; |
1508 | int ret = 0; | 1557 | int ret = 0; |
@@ -1514,9 +1563,9 @@ int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | |||
1514 | info->offset = offset; | 1563 | info->offset = offset; |
1515 | info->bytes = bytes; | 1564 | info->bytes = bytes; |
1516 | 1565 | ||
1517 | spin_lock(&block_group->tree_lock); | 1566 | spin_lock(&ctl->tree_lock); |
1518 | 1567 | ||
1519 | if (try_merge_free_space(block_group, info, true)) | 1568 | if (try_merge_free_space(ctl, info, true)) |
1520 | goto link; | 1569 | goto link; |
1521 | 1570 | ||
1522 | /* | 1571 | /* |
@@ -1524,7 +1573,7 @@ int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | |||
1524 | * extent then we know we're going to have to allocate a new extent, so | 1573 | * extent then we know we're going to have to allocate a new extent, so |
1525 | * before we do that see if we need to drop this into a bitmap | 1574 | * before we do that see if we need to drop this into a bitmap |
1526 | */ | 1575 | */ |
1527 | ret = insert_into_bitmap(block_group, info); | 1576 | ret = insert_into_bitmap(ctl, info); |
1528 | if (ret < 0) { | 1577 | if (ret < 0) { |
1529 | goto out; | 1578 | goto out; |
1530 | } else if (ret) { | 1579 | } else if (ret) { |
@@ -1532,11 +1581,11 @@ int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | |||
1532 | goto out; | 1581 | goto out; |
1533 | } | 1582 | } |
1534 | link: | 1583 | link: |
1535 | ret = link_free_space(block_group, info); | 1584 | ret = link_free_space(ctl, info); |
1536 | if (ret) | 1585 | if (ret) |
1537 | kmem_cache_free(btrfs_free_space_cachep, info); | 1586 | kmem_cache_free(btrfs_free_space_cachep, info); |
1538 | out: | 1587 | out: |
1539 | spin_unlock(&block_group->tree_lock); | 1588 | spin_unlock(&ctl->tree_lock); |
1540 | 1589 | ||
1541 | if (ret) { | 1590 | if (ret) { |
1542 | printk(KERN_CRIT "btrfs: unable to add free space :%d\n", ret); | 1591 | printk(KERN_CRIT "btrfs: unable to add free space :%d\n", ret); |
@@ -1549,21 +1598,21 @@ out: | |||
1549 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, | 1598 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, |
1550 | u64 offset, u64 bytes) | 1599 | u64 offset, u64 bytes) |
1551 | { | 1600 | { |
1601 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1552 | struct btrfs_free_space *info; | 1602 | struct btrfs_free_space *info; |
1553 | struct btrfs_free_space *next_info = NULL; | 1603 | struct btrfs_free_space *next_info = NULL; |
1554 | int ret = 0; | 1604 | int ret = 0; |
1555 | 1605 | ||
1556 | spin_lock(&block_group->tree_lock); | 1606 | spin_lock(&ctl->tree_lock); |
1557 | 1607 | ||
1558 | again: | 1608 | again: |
1559 | info = tree_search_offset(block_group, offset, 0, 0); | 1609 | info = tree_search_offset(ctl, offset, 0, 0); |
1560 | if (!info) { | 1610 | if (!info) { |
1561 | /* | 1611 | /* |
1562 | * oops didn't find an extent that matched the space we wanted | 1612 | * oops didn't find an extent that matched the space we wanted |
1563 | * to remove, look for a bitmap instead | 1613 | * to remove, look for a bitmap instead |
1564 | */ | 1614 | */ |
1565 | info = tree_search_offset(block_group, | 1615 | info = tree_search_offset(ctl, offset_to_bitmap(ctl, offset), |
1566 | offset_to_bitmap(block_group, offset), | ||
1567 | 1, 0); | 1616 | 1, 0); |
1568 | if (!info) { | 1617 | if (!info) { |
1569 | WARN_ON(1); | 1618 | WARN_ON(1); |
@@ -1578,8 +1627,8 @@ again: | |||
1578 | offset_index); | 1627 | offset_index); |
1579 | 1628 | ||
1580 | if (next_info->bitmap) | 1629 | if (next_info->bitmap) |
1581 | end = next_info->offset + BITS_PER_BITMAP * | 1630 | end = next_info->offset + |
1582 | block_group->sectorsize - 1; | 1631 | BITS_PER_BITMAP * ctl->unit - 1; |
1583 | else | 1632 | else |
1584 | end = next_info->offset + next_info->bytes; | 1633 | end = next_info->offset + next_info->bytes; |
1585 | 1634 | ||
@@ -1599,20 +1648,20 @@ again: | |||
1599 | } | 1648 | } |
1600 | 1649 | ||
1601 | if (info->bytes == bytes) { | 1650 | if (info->bytes == bytes) { |
1602 | unlink_free_space(block_group, info); | 1651 | unlink_free_space(ctl, info); |
1603 | if (info->bitmap) { | 1652 | if (info->bitmap) { |
1604 | kfree(info->bitmap); | 1653 | kfree(info->bitmap); |
1605 | block_group->total_bitmaps--; | 1654 | ctl->total_bitmaps--; |
1606 | } | 1655 | } |
1607 | kmem_cache_free(btrfs_free_space_cachep, info); | 1656 | kmem_cache_free(btrfs_free_space_cachep, info); |
1608 | goto out_lock; | 1657 | goto out_lock; |
1609 | } | 1658 | } |
1610 | 1659 | ||
1611 | if (!info->bitmap && info->offset == offset) { | 1660 | if (!info->bitmap && info->offset == offset) { |
1612 | unlink_free_space(block_group, info); | 1661 | unlink_free_space(ctl, info); |
1613 | info->offset += bytes; | 1662 | info->offset += bytes; |
1614 | info->bytes -= bytes; | 1663 | info->bytes -= bytes; |
1615 | link_free_space(block_group, info); | 1664 | link_free_space(ctl, info); |
1616 | goto out_lock; | 1665 | goto out_lock; |
1617 | } | 1666 | } |
1618 | 1667 | ||
@@ -1626,13 +1675,13 @@ again: | |||
1626 | * first unlink the old info and then | 1675 | * first unlink the old info and then |
1627 | * insert it again after the hole we're creating | 1676 | * insert it again after the hole we're creating |
1628 | */ | 1677 | */ |
1629 | unlink_free_space(block_group, info); | 1678 | unlink_free_space(ctl, info); |
1630 | if (offset + bytes < info->offset + info->bytes) { | 1679 | if (offset + bytes < info->offset + info->bytes) { |
1631 | u64 old_end = info->offset + info->bytes; | 1680 | u64 old_end = info->offset + info->bytes; |
1632 | 1681 | ||
1633 | info->offset = offset + bytes; | 1682 | info->offset = offset + bytes; |
1634 | info->bytes = old_end - info->offset; | 1683 | info->bytes = old_end - info->offset; |
1635 | ret = link_free_space(block_group, info); | 1684 | ret = link_free_space(ctl, info); |
1636 | WARN_ON(ret); | 1685 | WARN_ON(ret); |
1637 | if (ret) | 1686 | if (ret) |
1638 | goto out_lock; | 1687 | goto out_lock; |
@@ -1642,7 +1691,7 @@ again: | |||
1642 | */ | 1691 | */ |
1643 | kmem_cache_free(btrfs_free_space_cachep, info); | 1692 | kmem_cache_free(btrfs_free_space_cachep, info); |
1644 | } | 1693 | } |
1645 | spin_unlock(&block_group->tree_lock); | 1694 | spin_unlock(&ctl->tree_lock); |
1646 | 1695 | ||
1647 | /* step two, insert a new info struct to cover | 1696 | /* step two, insert a new info struct to cover |
1648 | * anything before the hole | 1697 | * anything before the hole |
@@ -1653,12 +1702,12 @@ again: | |||
1653 | goto out; | 1702 | goto out; |
1654 | } | 1703 | } |
1655 | 1704 | ||
1656 | ret = remove_from_bitmap(block_group, info, &offset, &bytes); | 1705 | ret = remove_from_bitmap(ctl, info, &offset, &bytes); |
1657 | if (ret == -EAGAIN) | 1706 | if (ret == -EAGAIN) |
1658 | goto again; | 1707 | goto again; |
1659 | BUG_ON(ret); | 1708 | BUG_ON(ret); |
1660 | out_lock: | 1709 | out_lock: |
1661 | spin_unlock(&block_group->tree_lock); | 1710 | spin_unlock(&ctl->tree_lock); |
1662 | out: | 1711 | out: |
1663 | return ret; | 1712 | return ret; |
1664 | } | 1713 | } |
@@ -1666,11 +1715,12 @@ out: | |||
1666 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, | 1715 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, |
1667 | u64 bytes) | 1716 | u64 bytes) |
1668 | { | 1717 | { |
1718 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1669 | struct btrfs_free_space *info; | 1719 | struct btrfs_free_space *info; |
1670 | struct rb_node *n; | 1720 | struct rb_node *n; |
1671 | int count = 0; | 1721 | int count = 0; |
1672 | 1722 | ||
1673 | for (n = rb_first(&block_group->free_space_offset); n; n = rb_next(n)) { | 1723 | for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) { |
1674 | info = rb_entry(n, struct btrfs_free_space, offset_index); | 1724 | info = rb_entry(n, struct btrfs_free_space, offset_index); |
1675 | if (info->bytes >= bytes) | 1725 | if (info->bytes >= bytes) |
1676 | count++; | 1726 | count++; |
@@ -1685,6 +1735,30 @@ void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, | |||
1685 | "\n", count); | 1735 | "\n", count); |
1686 | } | 1736 | } |
1687 | 1737 | ||
1738 | static struct btrfs_free_space_op free_space_op = { | ||
1739 | .recalc_thresholds = recalculate_thresholds, | ||
1740 | .use_bitmap = use_bitmap, | ||
1741 | }; | ||
1742 | |||
1743 | void btrfs_init_free_space_ctl(struct btrfs_block_group_cache *block_group) | ||
1744 | { | ||
1745 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1746 | |||
1747 | spin_lock_init(&ctl->tree_lock); | ||
1748 | ctl->unit = block_group->sectorsize; | ||
1749 | ctl->start = block_group->key.objectid; | ||
1750 | ctl->private = block_group; | ||
1751 | ctl->op = &free_space_op; | ||
1752 | |||
1753 | /* | ||
1754 | * we only want to have 32k of ram per block group for keeping | ||
1755 | * track of free space, and if we pass 1/2 of that we want to | ||
1756 | * start converting things over to using bitmaps | ||
1757 | */ | ||
1758 | ctl->extents_thresh = ((1024 * 32) / 2) / | ||
1759 | sizeof(struct btrfs_free_space); | ||
1760 | } | ||
1761 | |||
1688 | /* | 1762 | /* |
1689 | * for a given cluster, put all of its extents back into the free | 1763 | * for a given cluster, put all of its extents back into the free |
1690 | * space cache. If the block group passed doesn't match the block group | 1764 | * space cache. If the block group passed doesn't match the block group |
@@ -1696,6 +1770,7 @@ __btrfs_return_cluster_to_free_space( | |||
1696 | struct btrfs_block_group_cache *block_group, | 1770 | struct btrfs_block_group_cache *block_group, |
1697 | struct btrfs_free_cluster *cluster) | 1771 | struct btrfs_free_cluster *cluster) |
1698 | { | 1772 | { |
1773 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1699 | struct btrfs_free_space *entry; | 1774 | struct btrfs_free_space *entry; |
1700 | struct rb_node *node; | 1775 | struct rb_node *node; |
1701 | 1776 | ||
@@ -1717,8 +1792,8 @@ __btrfs_return_cluster_to_free_space( | |||
1717 | 1792 | ||
1718 | bitmap = (entry->bitmap != NULL); | 1793 | bitmap = (entry->bitmap != NULL); |
1719 | if (!bitmap) | 1794 | if (!bitmap) |
1720 | try_merge_free_space(block_group, entry, false); | 1795 | try_merge_free_space(ctl, entry, false); |
1721 | tree_insert_offset(&block_group->free_space_offset, | 1796 | tree_insert_offset(&ctl->free_space_offset, |
1722 | entry->offset, &entry->offset_index, bitmap); | 1797 | entry->offset, &entry->offset_index, bitmap); |
1723 | } | 1798 | } |
1724 | cluster->root = RB_ROOT; | 1799 | cluster->root = RB_ROOT; |
@@ -1729,14 +1804,38 @@ out: | |||
1729 | return 0; | 1804 | return 0; |
1730 | } | 1805 | } |
1731 | 1806 | ||
1732 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache *block_group) | 1807 | void __btrfs_remove_free_space_cache_locked(struct btrfs_free_space_ctl *ctl) |
1733 | { | 1808 | { |
1734 | struct btrfs_free_space *info; | 1809 | struct btrfs_free_space *info; |
1735 | struct rb_node *node; | 1810 | struct rb_node *node; |
1811 | |||
1812 | while ((node = rb_last(&ctl->free_space_offset)) != NULL) { | ||
1813 | info = rb_entry(node, struct btrfs_free_space, offset_index); | ||
1814 | unlink_free_space(ctl, info); | ||
1815 | kfree(info->bitmap); | ||
1816 | kmem_cache_free(btrfs_free_space_cachep, info); | ||
1817 | if (need_resched()) { | ||
1818 | spin_unlock(&ctl->tree_lock); | ||
1819 | cond_resched(); | ||
1820 | spin_lock(&ctl->tree_lock); | ||
1821 | } | ||
1822 | } | ||
1823 | } | ||
1824 | |||
1825 | void __btrfs_remove_free_space_cache(struct btrfs_free_space_ctl *ctl) | ||
1826 | { | ||
1827 | spin_lock(&ctl->tree_lock); | ||
1828 | __btrfs_remove_free_space_cache_locked(ctl); | ||
1829 | spin_unlock(&ctl->tree_lock); | ||
1830 | } | ||
1831 | |||
1832 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache *block_group) | ||
1833 | { | ||
1834 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1736 | struct btrfs_free_cluster *cluster; | 1835 | struct btrfs_free_cluster *cluster; |
1737 | struct list_head *head; | 1836 | struct list_head *head; |
1738 | 1837 | ||
1739 | spin_lock(&block_group->tree_lock); | 1838 | spin_lock(&ctl->tree_lock); |
1740 | while ((head = block_group->cluster_list.next) != | 1839 | while ((head = block_group->cluster_list.next) != |
1741 | &block_group->cluster_list) { | 1840 | &block_group->cluster_list) { |
1742 | cluster = list_entry(head, struct btrfs_free_cluster, | 1841 | cluster = list_entry(head, struct btrfs_free_cluster, |
@@ -1745,60 +1844,46 @@ void btrfs_remove_free_space_cache(struct btrfs_block_group_cache *block_group) | |||
1745 | WARN_ON(cluster->block_group != block_group); | 1844 | WARN_ON(cluster->block_group != block_group); |
1746 | __btrfs_return_cluster_to_free_space(block_group, cluster); | 1845 | __btrfs_return_cluster_to_free_space(block_group, cluster); |
1747 | if (need_resched()) { | 1846 | if (need_resched()) { |
1748 | spin_unlock(&block_group->tree_lock); | 1847 | spin_unlock(&ctl->tree_lock); |
1749 | cond_resched(); | 1848 | cond_resched(); |
1750 | spin_lock(&block_group->tree_lock); | 1849 | spin_lock(&ctl->tree_lock); |
1751 | } | 1850 | } |
1752 | } | 1851 | } |
1852 | __btrfs_remove_free_space_cache_locked(ctl); | ||
1853 | spin_unlock(&ctl->tree_lock); | ||
1753 | 1854 | ||
1754 | while ((node = rb_last(&block_group->free_space_offset)) != NULL) { | ||
1755 | info = rb_entry(node, struct btrfs_free_space, offset_index); | ||
1756 | if (!info->bitmap) { | ||
1757 | unlink_free_space(block_group, info); | ||
1758 | kmem_cache_free(btrfs_free_space_cachep, info); | ||
1759 | } else { | ||
1760 | free_bitmap(block_group, info); | ||
1761 | } | ||
1762 | |||
1763 | if (need_resched()) { | ||
1764 | spin_unlock(&block_group->tree_lock); | ||
1765 | cond_resched(); | ||
1766 | spin_lock(&block_group->tree_lock); | ||
1767 | } | ||
1768 | } | ||
1769 | |||
1770 | spin_unlock(&block_group->tree_lock); | ||
1771 | } | 1855 | } |
1772 | 1856 | ||
1773 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, | 1857 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, |
1774 | u64 offset, u64 bytes, u64 empty_size) | 1858 | u64 offset, u64 bytes, u64 empty_size) |
1775 | { | 1859 | { |
1860 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1776 | struct btrfs_free_space *entry = NULL; | 1861 | struct btrfs_free_space *entry = NULL; |
1777 | u64 bytes_search = bytes + empty_size; | 1862 | u64 bytes_search = bytes + empty_size; |
1778 | u64 ret = 0; | 1863 | u64 ret = 0; |
1779 | 1864 | ||
1780 | spin_lock(&block_group->tree_lock); | 1865 | spin_lock(&ctl->tree_lock); |
1781 | entry = find_free_space(block_group, &offset, &bytes_search, 0); | 1866 | entry = find_free_space(ctl, &offset, &bytes_search); |
1782 | if (!entry) | 1867 | if (!entry) |
1783 | goto out; | 1868 | goto out; |
1784 | 1869 | ||
1785 | ret = offset; | 1870 | ret = offset; |
1786 | if (entry->bitmap) { | 1871 | if (entry->bitmap) { |
1787 | bitmap_clear_bits(block_group, entry, offset, bytes); | 1872 | bitmap_clear_bits(ctl, entry, offset, bytes); |
1788 | if (!entry->bytes) | 1873 | if (!entry->bytes) |
1789 | free_bitmap(block_group, entry); | 1874 | free_bitmap(ctl, entry); |
1790 | } else { | 1875 | } else { |
1791 | unlink_free_space(block_group, entry); | 1876 | unlink_free_space(ctl, entry); |
1792 | entry->offset += bytes; | 1877 | entry->offset += bytes; |
1793 | entry->bytes -= bytes; | 1878 | entry->bytes -= bytes; |
1794 | if (!entry->bytes) | 1879 | if (!entry->bytes) |
1795 | kmem_cache_free(btrfs_free_space_cachep, entry); | 1880 | kmem_cache_free(btrfs_free_space_cachep, entry); |
1796 | else | 1881 | else |
1797 | link_free_space(block_group, entry); | 1882 | link_free_space(ctl, entry); |
1798 | } | 1883 | } |
1799 | 1884 | ||
1800 | out: | 1885 | out: |
1801 | spin_unlock(&block_group->tree_lock); | 1886 | spin_unlock(&ctl->tree_lock); |
1802 | 1887 | ||
1803 | return ret; | 1888 | return ret; |
1804 | } | 1889 | } |
@@ -1815,6 +1900,7 @@ int btrfs_return_cluster_to_free_space( | |||
1815 | struct btrfs_block_group_cache *block_group, | 1900 | struct btrfs_block_group_cache *block_group, |
1816 | struct btrfs_free_cluster *cluster) | 1901 | struct btrfs_free_cluster *cluster) |
1817 | { | 1902 | { |
1903 | struct btrfs_free_space_ctl *ctl; | ||
1818 | int ret; | 1904 | int ret; |
1819 | 1905 | ||
1820 | /* first, get a safe pointer to the block group */ | 1906 | /* first, get a safe pointer to the block group */ |
@@ -1833,10 +1919,12 @@ int btrfs_return_cluster_to_free_space( | |||
1833 | atomic_inc(&block_group->count); | 1919 | atomic_inc(&block_group->count); |
1834 | spin_unlock(&cluster->lock); | 1920 | spin_unlock(&cluster->lock); |
1835 | 1921 | ||
1922 | ctl = block_group->free_space_ctl; | ||
1923 | |||
1836 | /* now return any extents the cluster had on it */ | 1924 | /* now return any extents the cluster had on it */ |
1837 | spin_lock(&block_group->tree_lock); | 1925 | spin_lock(&ctl->tree_lock); |
1838 | ret = __btrfs_return_cluster_to_free_space(block_group, cluster); | 1926 | ret = __btrfs_return_cluster_to_free_space(block_group, cluster); |
1839 | spin_unlock(&block_group->tree_lock); | 1927 | spin_unlock(&ctl->tree_lock); |
1840 | 1928 | ||
1841 | /* finally drop our ref */ | 1929 | /* finally drop our ref */ |
1842 | btrfs_put_block_group(block_group); | 1930 | btrfs_put_block_group(block_group); |
@@ -1848,6 +1936,7 @@ static u64 btrfs_alloc_from_bitmap(struct btrfs_block_group_cache *block_group, | |||
1848 | struct btrfs_free_space *entry, | 1936 | struct btrfs_free_space *entry, |
1849 | u64 bytes, u64 min_start) | 1937 | u64 bytes, u64 min_start) |
1850 | { | 1938 | { |
1939 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1851 | int err; | 1940 | int err; |
1852 | u64 search_start = cluster->window_start; | 1941 | u64 search_start = cluster->window_start; |
1853 | u64 search_bytes = bytes; | 1942 | u64 search_bytes = bytes; |
@@ -1856,13 +1945,12 @@ static u64 btrfs_alloc_from_bitmap(struct btrfs_block_group_cache *block_group, | |||
1856 | search_start = min_start; | 1945 | search_start = min_start; |
1857 | search_bytes = bytes; | 1946 | search_bytes = bytes; |
1858 | 1947 | ||
1859 | err = search_bitmap(block_group, entry, &search_start, | 1948 | err = search_bitmap(ctl, entry, &search_start, &search_bytes); |
1860 | &search_bytes); | ||
1861 | if (err) | 1949 | if (err) |
1862 | return 0; | 1950 | return 0; |
1863 | 1951 | ||
1864 | ret = search_start; | 1952 | ret = search_start; |
1865 | bitmap_clear_bits(block_group, entry, ret, bytes); | 1953 | bitmap_clear_bits(ctl, entry, ret, bytes); |
1866 | 1954 | ||
1867 | return ret; | 1955 | return ret; |
1868 | } | 1956 | } |
@@ -1876,6 +1964,7 @@ u64 btrfs_alloc_from_cluster(struct btrfs_block_group_cache *block_group, | |||
1876 | struct btrfs_free_cluster *cluster, u64 bytes, | 1964 | struct btrfs_free_cluster *cluster, u64 bytes, |
1877 | u64 min_start) | 1965 | u64 min_start) |
1878 | { | 1966 | { |
1967 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1879 | struct btrfs_free_space *entry = NULL; | 1968 | struct btrfs_free_space *entry = NULL; |
1880 | struct rb_node *node; | 1969 | struct rb_node *node; |
1881 | u64 ret = 0; | 1970 | u64 ret = 0; |
@@ -1933,20 +2022,20 @@ out: | |||
1933 | if (!ret) | 2022 | if (!ret) |
1934 | return 0; | 2023 | return 0; |
1935 | 2024 | ||
1936 | spin_lock(&block_group->tree_lock); | 2025 | spin_lock(&ctl->tree_lock); |
1937 | 2026 | ||
1938 | block_group->free_space -= bytes; | 2027 | ctl->free_space -= bytes; |
1939 | if (entry->bytes == 0) { | 2028 | if (entry->bytes == 0) { |
1940 | block_group->free_extents--; | 2029 | ctl->free_extents--; |
1941 | if (entry->bitmap) { | 2030 | if (entry->bitmap) { |
1942 | kfree(entry->bitmap); | 2031 | kfree(entry->bitmap); |
1943 | block_group->total_bitmaps--; | 2032 | ctl->total_bitmaps--; |
1944 | recalculate_thresholds(block_group); | 2033 | ctl->op->recalc_thresholds(ctl); |
1945 | } | 2034 | } |
1946 | kmem_cache_free(btrfs_free_space_cachep, entry); | 2035 | kmem_cache_free(btrfs_free_space_cachep, entry); |
1947 | } | 2036 | } |
1948 | 2037 | ||
1949 | spin_unlock(&block_group->tree_lock); | 2038 | spin_unlock(&ctl->tree_lock); |
1950 | 2039 | ||
1951 | return ret; | 2040 | return ret; |
1952 | } | 2041 | } |
@@ -1956,6 +2045,7 @@ static int btrfs_bitmap_cluster(struct btrfs_block_group_cache *block_group, | |||
1956 | struct btrfs_free_cluster *cluster, | 2045 | struct btrfs_free_cluster *cluster, |
1957 | u64 offset, u64 bytes, u64 min_bytes) | 2046 | u64 offset, u64 bytes, u64 min_bytes) |
1958 | { | 2047 | { |
2048 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1959 | unsigned long next_zero; | 2049 | unsigned long next_zero; |
1960 | unsigned long i; | 2050 | unsigned long i; |
1961 | unsigned long search_bits; | 2051 | unsigned long search_bits; |
@@ -2010,7 +2100,7 @@ again: | |||
2010 | 2100 | ||
2011 | cluster->window_start = start * block_group->sectorsize + | 2101 | cluster->window_start = start * block_group->sectorsize + |
2012 | entry->offset; | 2102 | entry->offset; |
2013 | rb_erase(&entry->offset_index, &block_group->free_space_offset); | 2103 | rb_erase(&entry->offset_index, &ctl->free_space_offset); |
2014 | ret = tree_insert_offset(&cluster->root, entry->offset, | 2104 | ret = tree_insert_offset(&cluster->root, entry->offset, |
2015 | &entry->offset_index, 1); | 2105 | &entry->offset_index, 1); |
2016 | BUG_ON(ret); | 2106 | BUG_ON(ret); |
@@ -2025,6 +2115,7 @@ static int setup_cluster_no_bitmap(struct btrfs_block_group_cache *block_group, | |||
2025 | struct btrfs_free_cluster *cluster, | 2115 | struct btrfs_free_cluster *cluster, |
2026 | u64 offset, u64 bytes, u64 min_bytes) | 2116 | u64 offset, u64 bytes, u64 min_bytes) |
2027 | { | 2117 | { |
2118 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2028 | struct btrfs_free_space *first = NULL; | 2119 | struct btrfs_free_space *first = NULL; |
2029 | struct btrfs_free_space *entry = NULL; | 2120 | struct btrfs_free_space *entry = NULL; |
2030 | struct btrfs_free_space *prev = NULL; | 2121 | struct btrfs_free_space *prev = NULL; |
@@ -2035,7 +2126,7 @@ static int setup_cluster_no_bitmap(struct btrfs_block_group_cache *block_group, | |||
2035 | u64 max_extent; | 2126 | u64 max_extent; |
2036 | u64 max_gap = 128 * 1024; | 2127 | u64 max_gap = 128 * 1024; |
2037 | 2128 | ||
2038 | entry = tree_search_offset(block_group, offset, 0, 1); | 2129 | entry = tree_search_offset(ctl, offset, 0, 1); |
2039 | if (!entry) | 2130 | if (!entry) |
2040 | return -ENOSPC; | 2131 | return -ENOSPC; |
2041 | 2132 | ||
@@ -2101,7 +2192,7 @@ static int setup_cluster_no_bitmap(struct btrfs_block_group_cache *block_group, | |||
2101 | if (entry->bitmap) | 2192 | if (entry->bitmap) |
2102 | continue; | 2193 | continue; |
2103 | 2194 | ||
2104 | rb_erase(&entry->offset_index, &block_group->free_space_offset); | 2195 | rb_erase(&entry->offset_index, &ctl->free_space_offset); |
2105 | ret = tree_insert_offset(&cluster->root, entry->offset, | 2196 | ret = tree_insert_offset(&cluster->root, entry->offset, |
2106 | &entry->offset_index, 0); | 2197 | &entry->offset_index, 0); |
2107 | BUG_ON(ret); | 2198 | BUG_ON(ret); |
@@ -2120,16 +2211,15 @@ static int setup_cluster_bitmap(struct btrfs_block_group_cache *block_group, | |||
2120 | struct btrfs_free_cluster *cluster, | 2211 | struct btrfs_free_cluster *cluster, |
2121 | u64 offset, u64 bytes, u64 min_bytes) | 2212 | u64 offset, u64 bytes, u64 min_bytes) |
2122 | { | 2213 | { |
2214 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2123 | struct btrfs_free_space *entry; | 2215 | struct btrfs_free_space *entry; |
2124 | struct rb_node *node; | 2216 | struct rb_node *node; |
2125 | int ret = -ENOSPC; | 2217 | int ret = -ENOSPC; |
2126 | 2218 | ||
2127 | if (block_group->total_bitmaps == 0) | 2219 | if (ctl->total_bitmaps == 0) |
2128 | return -ENOSPC; | 2220 | return -ENOSPC; |
2129 | 2221 | ||
2130 | entry = tree_search_offset(block_group, | 2222 | entry = tree_search_offset(ctl, offset_to_bitmap(ctl, offset), 0, 1); |
2131 | offset_to_bitmap(block_group, offset), | ||
2132 | 0, 1); | ||
2133 | if (!entry) | 2223 | if (!entry) |
2134 | return -ENOSPC; | 2224 | return -ENOSPC; |
2135 | 2225 | ||
@@ -2162,6 +2252,7 @@ int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | |||
2162 | struct btrfs_free_cluster *cluster, | 2252 | struct btrfs_free_cluster *cluster, |
2163 | u64 offset, u64 bytes, u64 empty_size) | 2253 | u64 offset, u64 bytes, u64 empty_size) |
2164 | { | 2254 | { |
2255 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2165 | u64 min_bytes; | 2256 | u64 min_bytes; |
2166 | int ret; | 2257 | int ret; |
2167 | 2258 | ||
@@ -2181,14 +2272,14 @@ int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | |||
2181 | } else | 2272 | } else |
2182 | min_bytes = max(bytes, (bytes + empty_size) >> 2); | 2273 | min_bytes = max(bytes, (bytes + empty_size) >> 2); |
2183 | 2274 | ||
2184 | spin_lock(&block_group->tree_lock); | 2275 | spin_lock(&ctl->tree_lock); |
2185 | 2276 | ||
2186 | /* | 2277 | /* |
2187 | * If we know we don't have enough space to make a cluster don't even | 2278 | * If we know we don't have enough space to make a cluster don't even |
2188 | * bother doing all the work to try and find one. | 2279 | * bother doing all the work to try and find one. |
2189 | */ | 2280 | */ |
2190 | if (block_group->free_space < min_bytes) { | 2281 | if (ctl->free_space < min_bytes) { |
2191 | spin_unlock(&block_group->tree_lock); | 2282 | spin_unlock(&ctl->tree_lock); |
2192 | return -ENOSPC; | 2283 | return -ENOSPC; |
2193 | } | 2284 | } |
2194 | 2285 | ||
@@ -2214,7 +2305,7 @@ int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | |||
2214 | } | 2305 | } |
2215 | out: | 2306 | out: |
2216 | spin_unlock(&cluster->lock); | 2307 | spin_unlock(&cluster->lock); |
2217 | spin_unlock(&block_group->tree_lock); | 2308 | spin_unlock(&ctl->tree_lock); |
2218 | 2309 | ||
2219 | return ret; | 2310 | return ret; |
2220 | } | 2311 | } |
@@ -2235,6 +2326,7 @@ void btrfs_init_free_cluster(struct btrfs_free_cluster *cluster) | |||
2235 | int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | 2326 | int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, |
2236 | u64 *trimmed, u64 start, u64 end, u64 minlen) | 2327 | u64 *trimmed, u64 start, u64 end, u64 minlen) |
2237 | { | 2328 | { |
2329 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2238 | struct btrfs_free_space *entry = NULL; | 2330 | struct btrfs_free_space *entry = NULL; |
2239 | struct btrfs_fs_info *fs_info = block_group->fs_info; | 2331 | struct btrfs_fs_info *fs_info = block_group->fs_info; |
2240 | u64 bytes = 0; | 2332 | u64 bytes = 0; |
@@ -2244,52 +2336,50 @@ int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | |||
2244 | *trimmed = 0; | 2336 | *trimmed = 0; |
2245 | 2337 | ||
2246 | while (start < end) { | 2338 | while (start < end) { |
2247 | spin_lock(&block_group->tree_lock); | 2339 | spin_lock(&ctl->tree_lock); |
2248 | 2340 | ||
2249 | if (block_group->free_space < minlen) { | 2341 | if (ctl->free_space < minlen) { |
2250 | spin_unlock(&block_group->tree_lock); | 2342 | spin_unlock(&ctl->tree_lock); |
2251 | break; | 2343 | break; |
2252 | } | 2344 | } |
2253 | 2345 | ||
2254 | entry = tree_search_offset(block_group, start, 0, 1); | 2346 | entry = tree_search_offset(ctl, start, 0, 1); |
2255 | if (!entry) | 2347 | if (!entry) |
2256 | entry = tree_search_offset(block_group, | 2348 | entry = tree_search_offset(ctl, |
2257 | offset_to_bitmap(block_group, | 2349 | offset_to_bitmap(ctl, start), |
2258 | start), | ||
2259 | 1, 1); | 2350 | 1, 1); |
2260 | 2351 | ||
2261 | if (!entry || entry->offset >= end) { | 2352 | if (!entry || entry->offset >= end) { |
2262 | spin_unlock(&block_group->tree_lock); | 2353 | spin_unlock(&ctl->tree_lock); |
2263 | break; | 2354 | break; |
2264 | } | 2355 | } |
2265 | 2356 | ||
2266 | if (entry->bitmap) { | 2357 | if (entry->bitmap) { |
2267 | ret = search_bitmap(block_group, entry, &start, &bytes); | 2358 | ret = search_bitmap(ctl, entry, &start, &bytes); |
2268 | if (!ret) { | 2359 | if (!ret) { |
2269 | if (start >= end) { | 2360 | if (start >= end) { |
2270 | spin_unlock(&block_group->tree_lock); | 2361 | spin_unlock(&ctl->tree_lock); |
2271 | break; | 2362 | break; |
2272 | } | 2363 | } |
2273 | bytes = min(bytes, end - start); | 2364 | bytes = min(bytes, end - start); |
2274 | bitmap_clear_bits(block_group, entry, | 2365 | bitmap_clear_bits(ctl, entry, start, bytes); |
2275 | start, bytes); | ||
2276 | if (entry->bytes == 0) | 2366 | if (entry->bytes == 0) |
2277 | free_bitmap(block_group, entry); | 2367 | free_bitmap(ctl, entry); |
2278 | } else { | 2368 | } else { |
2279 | start = entry->offset + BITS_PER_BITMAP * | 2369 | start = entry->offset + BITS_PER_BITMAP * |
2280 | block_group->sectorsize; | 2370 | block_group->sectorsize; |
2281 | spin_unlock(&block_group->tree_lock); | 2371 | spin_unlock(&ctl->tree_lock); |
2282 | ret = 0; | 2372 | ret = 0; |
2283 | continue; | 2373 | continue; |
2284 | } | 2374 | } |
2285 | } else { | 2375 | } else { |
2286 | start = entry->offset; | 2376 | start = entry->offset; |
2287 | bytes = min(entry->bytes, end - start); | 2377 | bytes = min(entry->bytes, end - start); |
2288 | unlink_free_space(block_group, entry); | 2378 | unlink_free_space(ctl, entry); |
2289 | kmem_cache_free(btrfs_free_space_cachep, entry); | 2379 | kmem_cache_free(btrfs_free_space_cachep, entry); |
2290 | } | 2380 | } |
2291 | 2381 | ||
2292 | spin_unlock(&block_group->tree_lock); | 2382 | spin_unlock(&ctl->tree_lock); |
2293 | 2383 | ||
2294 | if (bytes >= minlen) { | 2384 | if (bytes >= minlen) { |
2295 | int update_ret; | 2385 | int update_ret; |
@@ -2301,8 +2391,7 @@ int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | |||
2301 | bytes, | 2391 | bytes, |
2302 | &actually_trimmed); | 2392 | &actually_trimmed); |
2303 | 2393 | ||
2304 | btrfs_add_free_space(block_group, | 2394 | btrfs_add_free_space(block_group, start, bytes); |
2305 | start, bytes); | ||
2306 | if (!update_ret) | 2395 | if (!update_ret) |
2307 | btrfs_update_reserved_bytes(block_group, | 2396 | btrfs_update_reserved_bytes(block_group, |
2308 | bytes, 0, 1); | 2397 | bytes, 0, 1); |
@@ -2324,3 +2413,145 @@ int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | |||
2324 | 2413 | ||
2325 | return ret; | 2414 | return ret; |
2326 | } | 2415 | } |
2416 | |||
2417 | /* | ||
2418 | * Find the left-most item in the cache tree, and then return the | ||
2419 | * smallest inode number in the item. | ||
2420 | * | ||
2421 | * Note: the returned inode number may not be the smallest one in | ||
2422 | * the tree, if the left-most item is a bitmap. | ||
2423 | */ | ||
2424 | u64 btrfs_find_ino_for_alloc(struct btrfs_root *fs_root) | ||
2425 | { | ||
2426 | struct btrfs_free_space_ctl *ctl = fs_root->free_ino_ctl; | ||
2427 | struct btrfs_free_space *entry = NULL; | ||
2428 | u64 ino = 0; | ||
2429 | |||
2430 | spin_lock(&ctl->tree_lock); | ||
2431 | |||
2432 | if (RB_EMPTY_ROOT(&ctl->free_space_offset)) | ||
2433 | goto out; | ||
2434 | |||
2435 | entry = rb_entry(rb_first(&ctl->free_space_offset), | ||
2436 | struct btrfs_free_space, offset_index); | ||
2437 | |||
2438 | if (!entry->bitmap) { | ||
2439 | ino = entry->offset; | ||
2440 | |||
2441 | unlink_free_space(ctl, entry); | ||
2442 | entry->offset++; | ||
2443 | entry->bytes--; | ||
2444 | if (!entry->bytes) | ||
2445 | kmem_cache_free(btrfs_free_space_cachep, entry); | ||
2446 | else | ||
2447 | link_free_space(ctl, entry); | ||
2448 | } else { | ||
2449 | u64 offset = 0; | ||
2450 | u64 count = 1; | ||
2451 | int ret; | ||
2452 | |||
2453 | ret = search_bitmap(ctl, entry, &offset, &count); | ||
2454 | BUG_ON(ret); | ||
2455 | |||
2456 | ino = offset; | ||
2457 | bitmap_clear_bits(ctl, entry, offset, 1); | ||
2458 | if (entry->bytes == 0) | ||
2459 | free_bitmap(ctl, entry); | ||
2460 | } | ||
2461 | out: | ||
2462 | spin_unlock(&ctl->tree_lock); | ||
2463 | |||
2464 | return ino; | ||
2465 | } | ||
2466 | |||
2467 | struct inode *lookup_free_ino_inode(struct btrfs_root *root, | ||
2468 | struct btrfs_path *path) | ||
2469 | { | ||
2470 | struct inode *inode = NULL; | ||
2471 | |||
2472 | spin_lock(&root->cache_lock); | ||
2473 | if (root->cache_inode) | ||
2474 | inode = igrab(root->cache_inode); | ||
2475 | spin_unlock(&root->cache_lock); | ||
2476 | if (inode) | ||
2477 | return inode; | ||
2478 | |||
2479 | inode = __lookup_free_space_inode(root, path, 0); | ||
2480 | if (IS_ERR(inode)) | ||
2481 | return inode; | ||
2482 | |||
2483 | spin_lock(&root->cache_lock); | ||
2484 | if (!root->fs_info->closing) | ||
2485 | root->cache_inode = igrab(inode); | ||
2486 | spin_unlock(&root->cache_lock); | ||
2487 | |||
2488 | return inode; | ||
2489 | } | ||
2490 | |||
2491 | int create_free_ino_inode(struct btrfs_root *root, | ||
2492 | struct btrfs_trans_handle *trans, | ||
2493 | struct btrfs_path *path) | ||
2494 | { | ||
2495 | return __create_free_space_inode(root, trans, path, | ||
2496 | BTRFS_FREE_INO_OBJECTID, 0); | ||
2497 | } | ||
2498 | |||
2499 | int load_free_ino_cache(struct btrfs_fs_info *fs_info, struct btrfs_root *root) | ||
2500 | { | ||
2501 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
2502 | struct btrfs_path *path; | ||
2503 | struct inode *inode; | ||
2504 | int ret = 0; | ||
2505 | u64 root_gen = btrfs_root_generation(&root->root_item); | ||
2506 | |||
2507 | /* | ||
2508 | * If we're unmounting then just return, since this does a search on the | ||
2509 | * normal root and not the commit root and we could deadlock. | ||
2510 | */ | ||
2511 | smp_mb(); | ||
2512 | if (fs_info->closing) | ||
2513 | return 0; | ||
2514 | |||
2515 | path = btrfs_alloc_path(); | ||
2516 | if (!path) | ||
2517 | return 0; | ||
2518 | |||
2519 | inode = lookup_free_ino_inode(root, path); | ||
2520 | if (IS_ERR(inode)) | ||
2521 | goto out; | ||
2522 | |||
2523 | if (root_gen != BTRFS_I(inode)->generation) | ||
2524 | goto out_put; | ||
2525 | |||
2526 | ret = __load_free_space_cache(root, inode, ctl, path, 0); | ||
2527 | |||
2528 | if (ret < 0) | ||
2529 | printk(KERN_ERR "btrfs: failed to load free ino cache for " | ||
2530 | "root %llu\n", root->root_key.objectid); | ||
2531 | out_put: | ||
2532 | iput(inode); | ||
2533 | out: | ||
2534 | btrfs_free_path(path); | ||
2535 | return ret; | ||
2536 | } | ||
2537 | |||
2538 | int btrfs_write_out_ino_cache(struct btrfs_root *root, | ||
2539 | struct btrfs_trans_handle *trans, | ||
2540 | struct btrfs_path *path) | ||
2541 | { | ||
2542 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
2543 | struct inode *inode; | ||
2544 | int ret; | ||
2545 | |||
2546 | inode = lookup_free_ino_inode(root, path); | ||
2547 | if (IS_ERR(inode)) | ||
2548 | return 0; | ||
2549 | |||
2550 | ret = __btrfs_write_out_cache(root, inode, ctl, NULL, trans, path, 0); | ||
2551 | if (ret < 0) | ||
2552 | printk(KERN_ERR "btrfs: failed to write free ino cache " | ||
2553 | "for root %llu\n", root->root_key.objectid); | ||
2554 | |||
2555 | iput(inode); | ||
2556 | return ret; | ||
2557 | } | ||
diff --git a/fs/btrfs/free-space-cache.h b/fs/btrfs/free-space-cache.h index 12b2b5165f8..8f2613f779e 100644 --- a/fs/btrfs/free-space-cache.h +++ b/fs/btrfs/free-space-cache.h | |||
@@ -27,6 +27,25 @@ struct btrfs_free_space { | |||
27 | struct list_head list; | 27 | struct list_head list; |
28 | }; | 28 | }; |
29 | 29 | ||
30 | struct btrfs_free_space_ctl { | ||
31 | spinlock_t tree_lock; | ||
32 | struct rb_root free_space_offset; | ||
33 | u64 free_space; | ||
34 | int extents_thresh; | ||
35 | int free_extents; | ||
36 | int total_bitmaps; | ||
37 | int unit; | ||
38 | u64 start; | ||
39 | struct btrfs_free_space_op *op; | ||
40 | void *private; | ||
41 | }; | ||
42 | |||
43 | struct btrfs_free_space_op { | ||
44 | void (*recalc_thresholds)(struct btrfs_free_space_ctl *ctl); | ||
45 | bool (*use_bitmap)(struct btrfs_free_space_ctl *ctl, | ||
46 | struct btrfs_free_space *info); | ||
47 | }; | ||
48 | |||
30 | struct inode *lookup_free_space_inode(struct btrfs_root *root, | 49 | struct inode *lookup_free_space_inode(struct btrfs_root *root, |
31 | struct btrfs_block_group_cache | 50 | struct btrfs_block_group_cache |
32 | *block_group, struct btrfs_path *path); | 51 | *block_group, struct btrfs_path *path); |
@@ -45,14 +64,36 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
45 | struct btrfs_trans_handle *trans, | 64 | struct btrfs_trans_handle *trans, |
46 | struct btrfs_block_group_cache *block_group, | 65 | struct btrfs_block_group_cache *block_group, |
47 | struct btrfs_path *path); | 66 | struct btrfs_path *path); |
48 | int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | 67 | |
49 | u64 bytenr, u64 size); | 68 | struct inode *lookup_free_ino_inode(struct btrfs_root *root, |
69 | struct btrfs_path *path); | ||
70 | int create_free_ino_inode(struct btrfs_root *root, | ||
71 | struct btrfs_trans_handle *trans, | ||
72 | struct btrfs_path *path); | ||
73 | int load_free_ino_cache(struct btrfs_fs_info *fs_info, | ||
74 | struct btrfs_root *root); | ||
75 | int btrfs_write_out_ino_cache(struct btrfs_root *root, | ||
76 | struct btrfs_trans_handle *trans, | ||
77 | struct btrfs_path *path); | ||
78 | |||
79 | void btrfs_init_free_space_ctl(struct btrfs_block_group_cache *block_group); | ||
80 | int __btrfs_add_free_space(struct btrfs_free_space_ctl *ctl, | ||
81 | u64 bytenr, u64 size); | ||
82 | static inline int | ||
83 | btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | ||
84 | u64 bytenr, u64 size) | ||
85 | { | ||
86 | return __btrfs_add_free_space(block_group->free_space_ctl, | ||
87 | bytenr, size); | ||
88 | } | ||
50 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, | 89 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, |
51 | u64 bytenr, u64 size); | 90 | u64 bytenr, u64 size); |
91 | void __btrfs_remove_free_space_cache(struct btrfs_free_space_ctl *ctl); | ||
52 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache | 92 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache |
53 | *block_group); | 93 | *block_group); |
54 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, | 94 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, |
55 | u64 offset, u64 bytes, u64 empty_size); | 95 | u64 offset, u64 bytes, u64 empty_size); |
96 | u64 btrfs_find_ino_for_alloc(struct btrfs_root *fs_root); | ||
56 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, | 97 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, |
57 | u64 bytes); | 98 | u64 bytes); |
58 | int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | 99 | int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, |
diff --git a/fs/btrfs/inode-map.c b/fs/btrfs/inode-map.c index c05a08f4c41..00097051262 100644 --- a/fs/btrfs/inode-map.c +++ b/fs/btrfs/inode-map.c | |||
@@ -16,11 +16,430 @@ | |||
16 | * Boston, MA 021110-1307, USA. | 16 | * Boston, MA 021110-1307, USA. |
17 | */ | 17 | */ |
18 | 18 | ||
19 | #include <linux/delay.h> | ||
20 | #include <linux/kthread.h> | ||
21 | #include <linux/pagemap.h> | ||
22 | |||
19 | #include "ctree.h" | 23 | #include "ctree.h" |
20 | #include "disk-io.h" | 24 | #include "disk-io.h" |
25 | #include "free-space-cache.h" | ||
26 | #include "inode-map.h" | ||
21 | #include "transaction.h" | 27 | #include "transaction.h" |
22 | 28 | ||
23 | int btrfs_find_highest_inode(struct btrfs_root *root, u64 *objectid) | 29 | static int caching_kthread(void *data) |
30 | { | ||
31 | struct btrfs_root *root = data; | ||
32 | struct btrfs_fs_info *fs_info = root->fs_info; | ||
33 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
34 | struct btrfs_key key; | ||
35 | struct btrfs_path *path; | ||
36 | struct extent_buffer *leaf; | ||
37 | u64 last = (u64)-1; | ||
38 | int slot; | ||
39 | int ret; | ||
40 | |||
41 | path = btrfs_alloc_path(); | ||
42 | if (!path) | ||
43 | return -ENOMEM; | ||
44 | |||
45 | /* Since the commit root is read-only, we can safely skip locking. */ | ||
46 | path->skip_locking = 1; | ||
47 | path->search_commit_root = 1; | ||
48 | path->reada = 2; | ||
49 | |||
50 | key.objectid = BTRFS_FIRST_FREE_OBJECTID; | ||
51 | key.offset = 0; | ||
52 | key.type = BTRFS_INODE_ITEM_KEY; | ||
53 | again: | ||
54 | /* need to make sure the commit_root doesn't disappear */ | ||
55 | mutex_lock(&root->fs_commit_mutex); | ||
56 | |||
57 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | ||
58 | if (ret < 0) | ||
59 | goto out; | ||
60 | |||
61 | while (1) { | ||
62 | smp_mb(); | ||
63 | if (fs_info->closing > 1) | ||
64 | goto out; | ||
65 | |||
66 | leaf = path->nodes[0]; | ||
67 | slot = path->slots[0]; | ||
68 | if (path->slots[0] >= btrfs_header_nritems(leaf)) { | ||
69 | ret = btrfs_next_leaf(root, path); | ||
70 | if (ret < 0) | ||
71 | goto out; | ||
72 | else if (ret > 0) | ||
73 | break; | ||
74 | |||
75 | if (need_resched() || | ||
76 | btrfs_transaction_in_commit(fs_info)) { | ||
77 | leaf = path->nodes[0]; | ||
78 | |||
79 | if (btrfs_header_nritems(leaf) == 0) { | ||
80 | WARN_ON(1); | ||
81 | break; | ||
82 | } | ||
83 | |||
84 | /* | ||
85 | * Save the key so we can advances forward | ||
86 | * in the next search. | ||
87 | */ | ||
88 | btrfs_item_key_to_cpu(leaf, &key, 0); | ||
89 | btrfs_release_path(path); | ||
90 | root->cache_progress = last; | ||
91 | mutex_unlock(&root->fs_commit_mutex); | ||
92 | schedule_timeout(1); | ||
93 | goto again; | ||
94 | } else | ||
95 | continue; | ||
96 | } | ||
97 | |||
98 | btrfs_item_key_to_cpu(leaf, &key, slot); | ||
99 | |||
100 | if (key.type != BTRFS_INODE_ITEM_KEY) | ||
101 | goto next; | ||
102 | |||
103 | if (key.objectid >= BTRFS_LAST_FREE_OBJECTID) | ||
104 | break; | ||
105 | |||
106 | if (last != (u64)-1 && last + 1 != key.objectid) { | ||
107 | __btrfs_add_free_space(ctl, last + 1, | ||
108 | key.objectid - last - 1); | ||
109 | wake_up(&root->cache_wait); | ||
110 | } | ||
111 | |||
112 | last = key.objectid; | ||
113 | next: | ||
114 | path->slots[0]++; | ||
115 | } | ||
116 | |||
117 | if (last < BTRFS_LAST_FREE_OBJECTID - 1) { | ||
118 | __btrfs_add_free_space(ctl, last + 1, | ||
119 | BTRFS_LAST_FREE_OBJECTID - last - 1); | ||
120 | } | ||
121 | |||
122 | spin_lock(&root->cache_lock); | ||
123 | root->cached = BTRFS_CACHE_FINISHED; | ||
124 | spin_unlock(&root->cache_lock); | ||
125 | |||
126 | root->cache_progress = (u64)-1; | ||
127 | btrfs_unpin_free_ino(root); | ||
128 | out: | ||
129 | wake_up(&root->cache_wait); | ||
130 | mutex_unlock(&root->fs_commit_mutex); | ||
131 | |||
132 | btrfs_free_path(path); | ||
133 | |||
134 | return ret; | ||
135 | } | ||
136 | |||
137 | static void start_caching(struct btrfs_root *root) | ||
138 | { | ||
139 | struct task_struct *tsk; | ||
140 | int ret; | ||
141 | |||
142 | spin_lock(&root->cache_lock); | ||
143 | if (root->cached != BTRFS_CACHE_NO) { | ||
144 | spin_unlock(&root->cache_lock); | ||
145 | return; | ||
146 | } | ||
147 | |||
148 | root->cached = BTRFS_CACHE_STARTED; | ||
149 | spin_unlock(&root->cache_lock); | ||
150 | |||
151 | ret = load_free_ino_cache(root->fs_info, root); | ||
152 | if (ret == 1) { | ||
153 | spin_lock(&root->cache_lock); | ||
154 | root->cached = BTRFS_CACHE_FINISHED; | ||
155 | spin_unlock(&root->cache_lock); | ||
156 | return; | ||
157 | } | ||
158 | |||
159 | tsk = kthread_run(caching_kthread, root, "btrfs-ino-cache-%llu\n", | ||
160 | root->root_key.objectid); | ||
161 | BUG_ON(IS_ERR(tsk)); | ||
162 | } | ||
163 | |||
164 | int btrfs_find_free_ino(struct btrfs_root *root, u64 *objectid) | ||
165 | { | ||
166 | again: | ||
167 | *objectid = btrfs_find_ino_for_alloc(root); | ||
168 | |||
169 | if (*objectid != 0) | ||
170 | return 0; | ||
171 | |||
172 | start_caching(root); | ||
173 | |||
174 | wait_event(root->cache_wait, | ||
175 | root->cached == BTRFS_CACHE_FINISHED || | ||
176 | root->free_ino_ctl->free_space > 0); | ||
177 | |||
178 | if (root->cached == BTRFS_CACHE_FINISHED && | ||
179 | root->free_ino_ctl->free_space == 0) | ||
180 | return -ENOSPC; | ||
181 | else | ||
182 | goto again; | ||
183 | } | ||
184 | |||
185 | void btrfs_return_ino(struct btrfs_root *root, u64 objectid) | ||
186 | { | ||
187 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
188 | struct btrfs_free_space_ctl *pinned = root->free_ino_pinned; | ||
189 | again: | ||
190 | if (root->cached == BTRFS_CACHE_FINISHED) { | ||
191 | __btrfs_add_free_space(ctl, objectid, 1); | ||
192 | } else { | ||
193 | /* | ||
194 | * If we are in the process of caching free ino chunks, | ||
195 | * to avoid adding the same inode number to the free_ino | ||
196 | * tree twice due to cross transaction, we'll leave it | ||
197 | * in the pinned tree until a transaction is committed | ||
198 | * or the caching work is done. | ||
199 | */ | ||
200 | |||
201 | mutex_lock(&root->fs_commit_mutex); | ||
202 | spin_lock(&root->cache_lock); | ||
203 | if (root->cached == BTRFS_CACHE_FINISHED) { | ||
204 | spin_unlock(&root->cache_lock); | ||
205 | mutex_unlock(&root->fs_commit_mutex); | ||
206 | goto again; | ||
207 | } | ||
208 | spin_unlock(&root->cache_lock); | ||
209 | |||
210 | start_caching(root); | ||
211 | |||
212 | if (objectid <= root->cache_progress) | ||
213 | __btrfs_add_free_space(ctl, objectid, 1); | ||
214 | else | ||
215 | __btrfs_add_free_space(pinned, objectid, 1); | ||
216 | |||
217 | mutex_unlock(&root->fs_commit_mutex); | ||
218 | } | ||
219 | } | ||
220 | |||
221 | /* | ||
222 | * When a transaction is committed, we'll move those inode numbers which | ||
223 | * are smaller than root->cache_progress from pinned tree to free_ino tree, | ||
224 | * and others will just be dropped, because the commit root we were | ||
225 | * searching has changed. | ||
226 | * | ||
227 | * Must be called with root->fs_commit_mutex held | ||
228 | */ | ||
229 | void btrfs_unpin_free_ino(struct btrfs_root *root) | ||
230 | { | ||
231 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
232 | struct rb_root *rbroot = &root->free_ino_pinned->free_space_offset; | ||
233 | struct btrfs_free_space *info; | ||
234 | struct rb_node *n; | ||
235 | u64 count; | ||
236 | |||
237 | while (1) { | ||
238 | n = rb_first(rbroot); | ||
239 | if (!n) | ||
240 | break; | ||
241 | |||
242 | info = rb_entry(n, struct btrfs_free_space, offset_index); | ||
243 | BUG_ON(info->bitmap); | ||
244 | |||
245 | if (info->offset > root->cache_progress) | ||
246 | goto free; | ||
247 | else if (info->offset + info->bytes > root->cache_progress) | ||
248 | count = root->cache_progress - info->offset + 1; | ||
249 | else | ||
250 | count = info->bytes; | ||
251 | |||
252 | __btrfs_add_free_space(ctl, info->offset, count); | ||
253 | free: | ||
254 | rb_erase(&info->offset_index, rbroot); | ||
255 | kfree(info); | ||
256 | } | ||
257 | } | ||
258 | |||
259 | #define INIT_THRESHOLD (((1024 * 32) / 2) / sizeof(struct btrfs_free_space)) | ||
260 | #define INODES_PER_BITMAP (PAGE_CACHE_SIZE * 8) | ||
261 | |||
262 | /* | ||
263 | * The goal is to keep the memory used by the free_ino tree won't | ||
264 | * exceed the memory if we use bitmaps only. | ||
265 | */ | ||
266 | static void recalculate_thresholds(struct btrfs_free_space_ctl *ctl) | ||
267 | { | ||
268 | struct btrfs_free_space *info; | ||
269 | struct rb_node *n; | ||
270 | int max_ino; | ||
271 | int max_bitmaps; | ||
272 | |||
273 | n = rb_last(&ctl->free_space_offset); | ||
274 | if (!n) { | ||
275 | ctl->extents_thresh = INIT_THRESHOLD; | ||
276 | return; | ||
277 | } | ||
278 | info = rb_entry(n, struct btrfs_free_space, offset_index); | ||
279 | |||
280 | /* | ||
281 | * Find the maximum inode number in the filesystem. Note we | ||
282 | * ignore the fact that this can be a bitmap, because we are | ||
283 | * not doing precise calculation. | ||
284 | */ | ||
285 | max_ino = info->bytes - 1; | ||
286 | |||
287 | max_bitmaps = ALIGN(max_ino, INODES_PER_BITMAP) / INODES_PER_BITMAP; | ||
288 | if (max_bitmaps <= ctl->total_bitmaps) { | ||
289 | ctl->extents_thresh = 0; | ||
290 | return; | ||
291 | } | ||
292 | |||
293 | ctl->extents_thresh = (max_bitmaps - ctl->total_bitmaps) * | ||
294 | PAGE_CACHE_SIZE / sizeof(*info); | ||
295 | } | ||
296 | |||
297 | /* | ||
298 | * We don't fall back to bitmap, if we are below the extents threshold | ||
299 | * or this chunk of inode numbers is a big one. | ||
300 | */ | ||
301 | static bool use_bitmap(struct btrfs_free_space_ctl *ctl, | ||
302 | struct btrfs_free_space *info) | ||
303 | { | ||
304 | if (ctl->free_extents < ctl->extents_thresh || | ||
305 | info->bytes > INODES_PER_BITMAP / 10) | ||
306 | return false; | ||
307 | |||
308 | return true; | ||
309 | } | ||
310 | |||
311 | static struct btrfs_free_space_op free_ino_op = { | ||
312 | .recalc_thresholds = recalculate_thresholds, | ||
313 | .use_bitmap = use_bitmap, | ||
314 | }; | ||
315 | |||
316 | static void pinned_recalc_thresholds(struct btrfs_free_space_ctl *ctl) | ||
317 | { | ||
318 | } | ||
319 | |||
320 | static bool pinned_use_bitmap(struct btrfs_free_space_ctl *ctl, | ||
321 | struct btrfs_free_space *info) | ||
322 | { | ||
323 | /* | ||
324 | * We always use extents for two reasons: | ||
325 | * | ||
326 | * - The pinned tree is only used during the process of caching | ||
327 | * work. | ||
328 | * - Make code simpler. See btrfs_unpin_free_ino(). | ||
329 | */ | ||
330 | return false; | ||
331 | } | ||
332 | |||
333 | static struct btrfs_free_space_op pinned_free_ino_op = { | ||
334 | .recalc_thresholds = pinned_recalc_thresholds, | ||
335 | .use_bitmap = pinned_use_bitmap, | ||
336 | }; | ||
337 | |||
338 | void btrfs_init_free_ino_ctl(struct btrfs_root *root) | ||
339 | { | ||
340 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
341 | struct btrfs_free_space_ctl *pinned = root->free_ino_pinned; | ||
342 | |||
343 | spin_lock_init(&ctl->tree_lock); | ||
344 | ctl->unit = 1; | ||
345 | ctl->start = 0; | ||
346 | ctl->private = NULL; | ||
347 | ctl->op = &free_ino_op; | ||
348 | |||
349 | /* | ||
350 | * Initially we allow to use 16K of ram to cache chunks of | ||
351 | * inode numbers before we resort to bitmaps. This is somewhat | ||
352 | * arbitrary, but it will be adjusted in runtime. | ||
353 | */ | ||
354 | ctl->extents_thresh = INIT_THRESHOLD; | ||
355 | |||
356 | spin_lock_init(&pinned->tree_lock); | ||
357 | pinned->unit = 1; | ||
358 | pinned->start = 0; | ||
359 | pinned->private = NULL; | ||
360 | pinned->extents_thresh = 0; | ||
361 | pinned->op = &pinned_free_ino_op; | ||
362 | } | ||
363 | |||
364 | int btrfs_save_ino_cache(struct btrfs_root *root, | ||
365 | struct btrfs_trans_handle *trans) | ||
366 | { | ||
367 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
368 | struct btrfs_path *path; | ||
369 | struct inode *inode; | ||
370 | u64 alloc_hint = 0; | ||
371 | int ret; | ||
372 | int prealloc; | ||
373 | bool retry = false; | ||
374 | |||
375 | path = btrfs_alloc_path(); | ||
376 | if (!path) | ||
377 | return -ENOMEM; | ||
378 | again: | ||
379 | inode = lookup_free_ino_inode(root, path); | ||
380 | if (IS_ERR(inode) && PTR_ERR(inode) != -ENOENT) { | ||
381 | ret = PTR_ERR(inode); | ||
382 | goto out; | ||
383 | } | ||
384 | |||
385 | if (IS_ERR(inode)) { | ||
386 | BUG_ON(retry); | ||
387 | retry = true; | ||
388 | |||
389 | ret = create_free_ino_inode(root, trans, path); | ||
390 | if (ret) | ||
391 | goto out; | ||
392 | goto again; | ||
393 | } | ||
394 | |||
395 | BTRFS_I(inode)->generation = 0; | ||
396 | ret = btrfs_update_inode(trans, root, inode); | ||
397 | WARN_ON(ret); | ||
398 | |||
399 | if (i_size_read(inode) > 0) { | ||
400 | ret = btrfs_truncate_free_space_cache(root, trans, path, inode); | ||
401 | if (ret) | ||
402 | goto out_put; | ||
403 | } | ||
404 | |||
405 | spin_lock(&root->cache_lock); | ||
406 | if (root->cached != BTRFS_CACHE_FINISHED) { | ||
407 | ret = -1; | ||
408 | spin_unlock(&root->cache_lock); | ||
409 | goto out_put; | ||
410 | } | ||
411 | spin_unlock(&root->cache_lock); | ||
412 | |||
413 | spin_lock(&ctl->tree_lock); | ||
414 | prealloc = sizeof(struct btrfs_free_space) * ctl->free_extents; | ||
415 | prealloc = ALIGN(prealloc, PAGE_CACHE_SIZE); | ||
416 | prealloc += ctl->total_bitmaps * PAGE_CACHE_SIZE; | ||
417 | spin_unlock(&ctl->tree_lock); | ||
418 | |||
419 | /* Just to make sure we have enough space */ | ||
420 | prealloc += 8 * PAGE_CACHE_SIZE; | ||
421 | |||
422 | ret = btrfs_check_data_free_space(inode, prealloc); | ||
423 | if (ret) | ||
424 | goto out_put; | ||
425 | |||
426 | ret = btrfs_prealloc_file_range_trans(inode, trans, 0, 0, prealloc, | ||
427 | prealloc, prealloc, &alloc_hint); | ||
428 | if (ret) | ||
429 | goto out_put; | ||
430 | btrfs_free_reserved_data_space(inode, prealloc); | ||
431 | |||
432 | out_put: | ||
433 | iput(inode); | ||
434 | out: | ||
435 | if (ret == 0) | ||
436 | ret = btrfs_write_out_ino_cache(root, trans, path); | ||
437 | |||
438 | btrfs_free_path(path); | ||
439 | return ret; | ||
440 | } | ||
441 | |||
442 | static int btrfs_find_highest_objectid(struct btrfs_root *root, u64 *objectid) | ||
24 | { | 443 | { |
25 | struct btrfs_path *path; | 444 | struct btrfs_path *path; |
26 | int ret; | 445 | int ret; |
@@ -55,15 +474,14 @@ error: | |||
55 | return ret; | 474 | return ret; |
56 | } | 475 | } |
57 | 476 | ||
58 | int btrfs_find_free_objectid(struct btrfs_trans_handle *trans, | 477 | int btrfs_find_free_objectid(struct btrfs_root *root, u64 *objectid) |
59 | struct btrfs_root *root, | ||
60 | u64 dirid, u64 *objectid) | ||
61 | { | 478 | { |
62 | int ret; | 479 | int ret; |
63 | mutex_lock(&root->objectid_mutex); | 480 | mutex_lock(&root->objectid_mutex); |
64 | 481 | ||
65 | if (unlikely(root->highest_objectid < BTRFS_FIRST_FREE_OBJECTID)) { | 482 | if (unlikely(root->highest_objectid < BTRFS_FIRST_FREE_OBJECTID)) { |
66 | ret = btrfs_find_highest_inode(root, &root->highest_objectid); | 483 | ret = btrfs_find_highest_objectid(root, |
484 | &root->highest_objectid); | ||
67 | if (ret) | 485 | if (ret) |
68 | goto out; | 486 | goto out; |
69 | } | 487 | } |
diff --git a/fs/btrfs/inode-map.h b/fs/btrfs/inode-map.h new file mode 100644 index 00000000000..ddb347bfee2 --- /dev/null +++ b/fs/btrfs/inode-map.h | |||
@@ -0,0 +1,13 @@ | |||
1 | #ifndef __BTRFS_INODE_MAP | ||
2 | #define __BTRFS_INODE_MAP | ||
3 | |||
4 | void btrfs_init_free_ino_ctl(struct btrfs_root *root); | ||
5 | void btrfs_unpin_free_ino(struct btrfs_root *root); | ||
6 | void btrfs_return_ino(struct btrfs_root *root, u64 objectid); | ||
7 | int btrfs_find_free_ino(struct btrfs_root *root, u64 *objectid); | ||
8 | int btrfs_save_ino_cache(struct btrfs_root *root, | ||
9 | struct btrfs_trans_handle *trans); | ||
10 | |||
11 | int btrfs_find_free_objectid(struct btrfs_root *root, u64 *objectid); | ||
12 | |||
13 | #endif | ||
diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c index 1d1017f9155..8ae72c3eedb 100644 --- a/fs/btrfs/inode.c +++ b/fs/btrfs/inode.c | |||
@@ -52,6 +52,7 @@ | |||
52 | #include "compression.h" | 52 | #include "compression.h" |
53 | #include "locking.h" | 53 | #include "locking.h" |
54 | #include "free-space-cache.h" | 54 | #include "free-space-cache.h" |
55 | #include "inode-map.h" | ||
55 | 56 | ||
56 | struct btrfs_iget_args { | 57 | struct btrfs_iget_args { |
57 | u64 ino; | 58 | u64 ino; |
@@ -139,7 +140,7 @@ static noinline int insert_inline_extent(struct btrfs_trans_handle *trans, | |||
139 | path->leave_spinning = 1; | 140 | path->leave_spinning = 1; |
140 | btrfs_set_trans_block_group(trans, inode); | 141 | btrfs_set_trans_block_group(trans, inode); |
141 | 142 | ||
142 | key.objectid = inode->i_ino; | 143 | key.objectid = btrfs_ino(inode); |
143 | key.offset = start; | 144 | key.offset = start; |
144 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); | 145 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); |
145 | datasize = btrfs_file_extent_calc_inline_size(cur_size); | 146 | datasize = btrfs_file_extent_calc_inline_size(cur_size); |
@@ -746,6 +747,15 @@ static u64 get_extent_allocation_hint(struct inode *inode, u64 start, | |||
746 | return alloc_hint; | 747 | return alloc_hint; |
747 | } | 748 | } |
748 | 749 | ||
750 | static inline bool is_free_space_inode(struct btrfs_root *root, | ||
751 | struct inode *inode) | ||
752 | { | ||
753 | if (root == root->fs_info->tree_root || | ||
754 | BTRFS_I(inode)->location.objectid == BTRFS_FREE_INO_OBJECTID) | ||
755 | return true; | ||
756 | return false; | ||
757 | } | ||
758 | |||
749 | /* | 759 | /* |
750 | * when extent_io.c finds a delayed allocation range in the file, | 760 | * when extent_io.c finds a delayed allocation range in the file, |
751 | * the call backs end up in this code. The basic idea is to | 761 | * the call backs end up in this code. The basic idea is to |
@@ -778,7 +788,7 @@ static noinline int cow_file_range(struct inode *inode, | |||
778 | struct extent_map_tree *em_tree = &BTRFS_I(inode)->extent_tree; | 788 | struct extent_map_tree *em_tree = &BTRFS_I(inode)->extent_tree; |
779 | int ret = 0; | 789 | int ret = 0; |
780 | 790 | ||
781 | BUG_ON(root == root->fs_info->tree_root); | 791 | BUG_ON(is_free_space_inode(root, inode)); |
782 | trans = btrfs_join_transaction(root, 1); | 792 | trans = btrfs_join_transaction(root, 1); |
783 | BUG_ON(IS_ERR(trans)); | 793 | BUG_ON(IS_ERR(trans)); |
784 | btrfs_set_trans_block_group(trans, inode); | 794 | btrfs_set_trans_block_group(trans, inode); |
@@ -1050,29 +1060,31 @@ static noinline int run_delalloc_nocow(struct inode *inode, | |||
1050 | int type; | 1060 | int type; |
1051 | int nocow; | 1061 | int nocow; |
1052 | int check_prev = 1; | 1062 | int check_prev = 1; |
1053 | bool nolock = false; | 1063 | bool nolock; |
1064 | u64 ino = btrfs_ino(inode); | ||
1054 | 1065 | ||
1055 | path = btrfs_alloc_path(); | 1066 | path = btrfs_alloc_path(); |
1056 | BUG_ON(!path); | 1067 | BUG_ON(!path); |
1057 | if (root == root->fs_info->tree_root) { | 1068 | |
1058 | nolock = true; | 1069 | nolock = is_free_space_inode(root, inode); |
1070 | |||
1071 | if (nolock) | ||
1059 | trans = btrfs_join_transaction_nolock(root, 1); | 1072 | trans = btrfs_join_transaction_nolock(root, 1); |
1060 | } else { | 1073 | else |
1061 | trans = btrfs_join_transaction(root, 1); | 1074 | trans = btrfs_join_transaction(root, 1); |
1062 | } | ||
1063 | BUG_ON(IS_ERR(trans)); | 1075 | BUG_ON(IS_ERR(trans)); |
1064 | 1076 | ||
1065 | cow_start = (u64)-1; | 1077 | cow_start = (u64)-1; |
1066 | cur_offset = start; | 1078 | cur_offset = start; |
1067 | while (1) { | 1079 | while (1) { |
1068 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 1080 | ret = btrfs_lookup_file_extent(trans, root, path, ino, |
1069 | cur_offset, 0); | 1081 | cur_offset, 0); |
1070 | BUG_ON(ret < 0); | 1082 | BUG_ON(ret < 0); |
1071 | if (ret > 0 && path->slots[0] > 0 && check_prev) { | 1083 | if (ret > 0 && path->slots[0] > 0 && check_prev) { |
1072 | leaf = path->nodes[0]; | 1084 | leaf = path->nodes[0]; |
1073 | btrfs_item_key_to_cpu(leaf, &found_key, | 1085 | btrfs_item_key_to_cpu(leaf, &found_key, |
1074 | path->slots[0] - 1); | 1086 | path->slots[0] - 1); |
1075 | if (found_key.objectid == inode->i_ino && | 1087 | if (found_key.objectid == ino && |
1076 | found_key.type == BTRFS_EXTENT_DATA_KEY) | 1088 | found_key.type == BTRFS_EXTENT_DATA_KEY) |
1077 | path->slots[0]--; | 1089 | path->slots[0]--; |
1078 | } | 1090 | } |
@@ -1093,7 +1105,7 @@ next_slot: | |||
1093 | num_bytes = 0; | 1105 | num_bytes = 0; |
1094 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); | 1106 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); |
1095 | 1107 | ||
1096 | if (found_key.objectid > inode->i_ino || | 1108 | if (found_key.objectid > ino || |
1097 | found_key.type > BTRFS_EXTENT_DATA_KEY || | 1109 | found_key.type > BTRFS_EXTENT_DATA_KEY || |
1098 | found_key.offset > end) | 1110 | found_key.offset > end) |
1099 | break; | 1111 | break; |
@@ -1128,7 +1140,7 @@ next_slot: | |||
1128 | goto out_check; | 1140 | goto out_check; |
1129 | if (btrfs_extent_readonly(root, disk_bytenr)) | 1141 | if (btrfs_extent_readonly(root, disk_bytenr)) |
1130 | goto out_check; | 1142 | goto out_check; |
1131 | if (btrfs_cross_ref_exist(trans, root, inode->i_ino, | 1143 | if (btrfs_cross_ref_exist(trans, root, ino, |
1132 | found_key.offset - | 1144 | found_key.offset - |
1133 | extent_offset, disk_bytenr)) | 1145 | extent_offset, disk_bytenr)) |
1134 | goto out_check; | 1146 | goto out_check; |
@@ -1317,8 +1329,7 @@ static int btrfs_set_bit_hook(struct inode *inode, | |||
1317 | if (!(state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { | 1329 | if (!(state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { |
1318 | struct btrfs_root *root = BTRFS_I(inode)->root; | 1330 | struct btrfs_root *root = BTRFS_I(inode)->root; |
1319 | u64 len = state->end + 1 - state->start; | 1331 | u64 len = state->end + 1 - state->start; |
1320 | int do_list = (root->root_key.objectid != | 1332 | bool do_list = !is_free_space_inode(root, inode); |
1321 | BTRFS_ROOT_TREE_OBJECTID); | ||
1322 | 1333 | ||
1323 | if (*bits & EXTENT_FIRST_DELALLOC) | 1334 | if (*bits & EXTENT_FIRST_DELALLOC) |
1324 | *bits &= ~EXTENT_FIRST_DELALLOC; | 1335 | *bits &= ~EXTENT_FIRST_DELALLOC; |
@@ -1351,8 +1362,7 @@ static int btrfs_clear_bit_hook(struct inode *inode, | |||
1351 | if ((state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { | 1362 | if ((state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { |
1352 | struct btrfs_root *root = BTRFS_I(inode)->root; | 1363 | struct btrfs_root *root = BTRFS_I(inode)->root; |
1353 | u64 len = state->end + 1 - state->start; | 1364 | u64 len = state->end + 1 - state->start; |
1354 | int do_list = (root->root_key.objectid != | 1365 | bool do_list = !is_free_space_inode(root, inode); |
1355 | BTRFS_ROOT_TREE_OBJECTID); | ||
1356 | 1366 | ||
1357 | if (*bits & EXTENT_FIRST_DELALLOC) | 1367 | if (*bits & EXTENT_FIRST_DELALLOC) |
1358 | *bits &= ~EXTENT_FIRST_DELALLOC; | 1368 | *bits &= ~EXTENT_FIRST_DELALLOC; |
@@ -1459,7 +1469,7 @@ static int btrfs_submit_bio_hook(struct inode *inode, int rw, struct bio *bio, | |||
1459 | 1469 | ||
1460 | skip_sum = BTRFS_I(inode)->flags & BTRFS_INODE_NODATASUM; | 1470 | skip_sum = BTRFS_I(inode)->flags & BTRFS_INODE_NODATASUM; |
1461 | 1471 | ||
1462 | if (root == root->fs_info->tree_root) | 1472 | if (is_free_space_inode(root, inode)) |
1463 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 2); | 1473 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 2); |
1464 | else | 1474 | else |
1465 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 0); | 1475 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 0); |
@@ -1645,7 +1655,7 @@ static int insert_reserved_file_extent(struct btrfs_trans_handle *trans, | |||
1645 | &hint, 0); | 1655 | &hint, 0); |
1646 | BUG_ON(ret); | 1656 | BUG_ON(ret); |
1647 | 1657 | ||
1648 | ins.objectid = inode->i_ino; | 1658 | ins.objectid = btrfs_ino(inode); |
1649 | ins.offset = file_pos; | 1659 | ins.offset = file_pos; |
1650 | ins.type = BTRFS_EXTENT_DATA_KEY; | 1660 | ins.type = BTRFS_EXTENT_DATA_KEY; |
1651 | ret = btrfs_insert_empty_item(trans, root, path, &ins, sizeof(*fi)); | 1661 | ret = btrfs_insert_empty_item(trans, root, path, &ins, sizeof(*fi)); |
@@ -1676,7 +1686,7 @@ static int insert_reserved_file_extent(struct btrfs_trans_handle *trans, | |||
1676 | ins.type = BTRFS_EXTENT_ITEM_KEY; | 1686 | ins.type = BTRFS_EXTENT_ITEM_KEY; |
1677 | ret = btrfs_alloc_reserved_file_extent(trans, root, | 1687 | ret = btrfs_alloc_reserved_file_extent(trans, root, |
1678 | root->root_key.objectid, | 1688 | root->root_key.objectid, |
1679 | inode->i_ino, file_pos, &ins); | 1689 | btrfs_ino(inode), file_pos, &ins); |
1680 | BUG_ON(ret); | 1690 | BUG_ON(ret); |
1681 | btrfs_free_path(path); | 1691 | btrfs_free_path(path); |
1682 | 1692 | ||
@@ -1702,7 +1712,7 @@ static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end) | |||
1702 | struct extent_state *cached_state = NULL; | 1712 | struct extent_state *cached_state = NULL; |
1703 | int compress_type = 0; | 1713 | int compress_type = 0; |
1704 | int ret; | 1714 | int ret; |
1705 | bool nolock = false; | 1715 | bool nolock; |
1706 | 1716 | ||
1707 | ret = btrfs_dec_test_ordered_pending(inode, &ordered_extent, start, | 1717 | ret = btrfs_dec_test_ordered_pending(inode, &ordered_extent, start, |
1708 | end - start + 1); | 1718 | end - start + 1); |
@@ -1710,7 +1720,7 @@ static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end) | |||
1710 | return 0; | 1720 | return 0; |
1711 | BUG_ON(!ordered_extent); | 1721 | BUG_ON(!ordered_extent); |
1712 | 1722 | ||
1713 | nolock = (root == root->fs_info->tree_root); | 1723 | nolock = is_free_space_inode(root, inode); |
1714 | 1724 | ||
1715 | if (test_bit(BTRFS_ORDERED_NOCOW, &ordered_extent->flags)) { | 1725 | if (test_bit(BTRFS_ORDERED_NOCOW, &ordered_extent->flags)) { |
1716 | BUG_ON(!list_empty(&ordered_extent->list)); | 1726 | BUG_ON(!list_empty(&ordered_extent->list)); |
@@ -2005,8 +2015,9 @@ good: | |||
2005 | return 0; | 2015 | return 0; |
2006 | 2016 | ||
2007 | zeroit: | 2017 | zeroit: |
2008 | printk_ratelimited(KERN_INFO "btrfs csum failed ino %lu off %llu csum %u " | 2018 | printk_ratelimited(KERN_INFO "btrfs csum failed ino %llu off %llu csum %u " |
2009 | "private %llu\n", page->mapping->host->i_ino, | 2019 | "private %llu\n", |
2020 | (unsigned long long)btrfs_ino(page->mapping->host), | ||
2010 | (unsigned long long)start, csum, | 2021 | (unsigned long long)start, csum, |
2011 | (unsigned long long)private); | 2022 | (unsigned long long)private); |
2012 | memset(kaddr + offset, 1, end - start + 1); | 2023 | memset(kaddr + offset, 1, end - start + 1); |
@@ -2243,7 +2254,7 @@ int btrfs_orphan_add(struct btrfs_trans_handle *trans, struct inode *inode) | |||
2243 | 2254 | ||
2244 | /* insert an orphan item to track this unlinked/truncated file */ | 2255 | /* insert an orphan item to track this unlinked/truncated file */ |
2245 | if (insert >= 1) { | 2256 | if (insert >= 1) { |
2246 | ret = btrfs_insert_orphan_item(trans, root, inode->i_ino); | 2257 | ret = btrfs_insert_orphan_item(trans, root, btrfs_ino(inode)); |
2247 | BUG_ON(ret); | 2258 | BUG_ON(ret); |
2248 | } | 2259 | } |
2249 | 2260 | ||
@@ -2280,7 +2291,7 @@ int btrfs_orphan_del(struct btrfs_trans_handle *trans, struct inode *inode) | |||
2280 | spin_unlock(&root->orphan_lock); | 2291 | spin_unlock(&root->orphan_lock); |
2281 | 2292 | ||
2282 | if (trans && delete_item) { | 2293 | if (trans && delete_item) { |
2283 | ret = btrfs_del_orphan_item(trans, root, inode->i_ino); | 2294 | ret = btrfs_del_orphan_item(trans, root, btrfs_ino(inode)); |
2284 | BUG_ON(ret); | 2295 | BUG_ON(ret); |
2285 | } | 2296 | } |
2286 | 2297 | ||
@@ -2542,7 +2553,8 @@ static void btrfs_read_locked_inode(struct inode *inode) | |||
2542 | * try to precache a NULL acl entry for files that don't have | 2553 | * try to precache a NULL acl entry for files that don't have |
2543 | * any xattrs or acls | 2554 | * any xattrs or acls |
2544 | */ | 2555 | */ |
2545 | maybe_acls = acls_after_inode_item(leaf, path->slots[0], inode->i_ino); | 2556 | maybe_acls = acls_after_inode_item(leaf, path->slots[0], |
2557 | btrfs_ino(inode)); | ||
2546 | if (!maybe_acls) | 2558 | if (!maybe_acls) |
2547 | cache_no_acl(inode); | 2559 | cache_no_acl(inode); |
2548 | 2560 | ||
@@ -2646,11 +2658,26 @@ noinline int btrfs_update_inode(struct btrfs_trans_handle *trans, | |||
2646 | struct extent_buffer *leaf; | 2658 | struct extent_buffer *leaf; |
2647 | int ret; | 2659 | int ret; |
2648 | 2660 | ||
2661 | /* | ||
2662 | * If root is tree root, it means this inode is used to | ||
2663 | * store free space information. And these inodes are updated | ||
2664 | * when committing the transaction, so they needn't delaye to | ||
2665 | * be updated, or deadlock will occured. | ||
2666 | */ | ||
2667 | if (!is_free_space_inode(root, inode)) { | ||
2668 | ret = btrfs_delayed_update_inode(trans, root, inode); | ||
2669 | if (!ret) | ||
2670 | btrfs_set_inode_last_trans(trans, inode); | ||
2671 | return ret; | ||
2672 | } | ||
2673 | |||
2649 | path = btrfs_alloc_path(); | 2674 | path = btrfs_alloc_path(); |
2650 | BUG_ON(!path); | 2675 | if (!path) |
2676 | return -ENOMEM; | ||
2677 | |||
2651 | path->leave_spinning = 1; | 2678 | path->leave_spinning = 1; |
2652 | ret = btrfs_lookup_inode(trans, root, path, | 2679 | ret = btrfs_lookup_inode(trans, root, path, &BTRFS_I(inode)->location, |
2653 | &BTRFS_I(inode)->location, 1); | 2680 | 1); |
2654 | if (ret) { | 2681 | if (ret) { |
2655 | if (ret > 0) | 2682 | if (ret > 0) |
2656 | ret = -ENOENT; | 2683 | ret = -ENOENT; |
@@ -2660,7 +2687,7 @@ noinline int btrfs_update_inode(struct btrfs_trans_handle *trans, | |||
2660 | btrfs_unlock_up_safe(path, 1); | 2687 | btrfs_unlock_up_safe(path, 1); |
2661 | leaf = path->nodes[0]; | 2688 | leaf = path->nodes[0]; |
2662 | inode_item = btrfs_item_ptr(leaf, path->slots[0], | 2689 | inode_item = btrfs_item_ptr(leaf, path->slots[0], |
2663 | struct btrfs_inode_item); | 2690 | struct btrfs_inode_item); |
2664 | 2691 | ||
2665 | fill_inode_item(trans, leaf, inode_item, inode); | 2692 | fill_inode_item(trans, leaf, inode_item, inode); |
2666 | btrfs_mark_buffer_dirty(leaf); | 2693 | btrfs_mark_buffer_dirty(leaf); |
@@ -2671,7 +2698,6 @@ failed: | |||
2671 | return ret; | 2698 | return ret; |
2672 | } | 2699 | } |
2673 | 2700 | ||
2674 | |||
2675 | /* | 2701 | /* |
2676 | * unlink helper that gets used here in inode.c and in the tree logging | 2702 | * unlink helper that gets used here in inode.c and in the tree logging |
2677 | * recovery code. It remove a link in a directory with a given name, and | 2703 | * recovery code. It remove a link in a directory with a given name, and |
@@ -2688,6 +2714,8 @@ static int __btrfs_unlink_inode(struct btrfs_trans_handle *trans, | |||
2688 | struct btrfs_dir_item *di; | 2714 | struct btrfs_dir_item *di; |
2689 | struct btrfs_key key; | 2715 | struct btrfs_key key; |
2690 | u64 index; | 2716 | u64 index; |
2717 | u64 ino = btrfs_ino(inode); | ||
2718 | u64 dir_ino = btrfs_ino(dir); | ||
2691 | 2719 | ||
2692 | path = btrfs_alloc_path(); | 2720 | path = btrfs_alloc_path(); |
2693 | if (!path) { | 2721 | if (!path) { |
@@ -2696,7 +2724,7 @@ static int __btrfs_unlink_inode(struct btrfs_trans_handle *trans, | |||
2696 | } | 2724 | } |
2697 | 2725 | ||
2698 | path->leave_spinning = 1; | 2726 | path->leave_spinning = 1; |
2699 | di = btrfs_lookup_dir_item(trans, root, path, dir->i_ino, | 2727 | di = btrfs_lookup_dir_item(trans, root, path, dir_ino, |
2700 | name, name_len, -1); | 2728 | name, name_len, -1); |
2701 | if (IS_ERR(di)) { | 2729 | if (IS_ERR(di)) { |
2702 | ret = PTR_ERR(di); | 2730 | ret = PTR_ERR(di); |
@@ -2713,31 +2741,21 @@ static int __btrfs_unlink_inode(struct btrfs_trans_handle *trans, | |||
2713 | goto err; | 2741 | goto err; |
2714 | btrfs_release_path(path); | 2742 | btrfs_release_path(path); |
2715 | 2743 | ||
2716 | ret = btrfs_del_inode_ref(trans, root, name, name_len, | 2744 | ret = btrfs_del_inode_ref(trans, root, name, name_len, ino, |
2717 | inode->i_ino, | 2745 | dir_ino, &index); |
2718 | dir->i_ino, &index); | ||
2719 | if (ret) { | 2746 | if (ret) { |
2720 | printk(KERN_INFO "btrfs failed to delete reference to %.*s, " | 2747 | printk(KERN_INFO "btrfs failed to delete reference to %.*s, " |
2721 | "inode %lu parent %lu\n", name_len, name, | 2748 | "inode %llu parent %llu\n", name_len, name, |
2722 | inode->i_ino, dir->i_ino); | 2749 | (unsigned long long)ino, (unsigned long long)dir_ino); |
2723 | goto err; | 2750 | goto err; |
2724 | } | 2751 | } |
2725 | 2752 | ||
2726 | di = btrfs_lookup_dir_index_item(trans, root, path, dir->i_ino, | 2753 | ret = btrfs_delete_delayed_dir_index(trans, root, dir, index); |
2727 | index, name, name_len, -1); | 2754 | if (ret) |
2728 | if (IS_ERR(di)) { | ||
2729 | ret = PTR_ERR(di); | ||
2730 | goto err; | ||
2731 | } | ||
2732 | if (!di) { | ||
2733 | ret = -ENOENT; | ||
2734 | goto err; | 2755 | goto err; |
2735 | } | ||
2736 | ret = btrfs_delete_one_dir_name(trans, root, path, di); | ||
2737 | btrfs_release_path(path); | ||
2738 | 2756 | ||
2739 | ret = btrfs_del_inode_ref_in_log(trans, root, name, name_len, | 2757 | ret = btrfs_del_inode_ref_in_log(trans, root, name, name_len, |
2740 | inode, dir->i_ino); | 2758 | inode, dir_ino); |
2741 | BUG_ON(ret != 0 && ret != -ENOENT); | 2759 | BUG_ON(ret != 0 && ret != -ENOENT); |
2742 | 2760 | ||
2743 | ret = btrfs_del_dir_entries_in_log(trans, root, name, name_len, | 2761 | ret = btrfs_del_dir_entries_in_log(trans, root, name, name_len, |
@@ -2815,12 +2833,14 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2815 | int check_link = 1; | 2833 | int check_link = 1; |
2816 | int err = -ENOSPC; | 2834 | int err = -ENOSPC; |
2817 | int ret; | 2835 | int ret; |
2836 | u64 ino = btrfs_ino(inode); | ||
2837 | u64 dir_ino = btrfs_ino(dir); | ||
2818 | 2838 | ||
2819 | trans = btrfs_start_transaction(root, 10); | 2839 | trans = btrfs_start_transaction(root, 10); |
2820 | if (!IS_ERR(trans) || PTR_ERR(trans) != -ENOSPC) | 2840 | if (!IS_ERR(trans) || PTR_ERR(trans) != -ENOSPC) |
2821 | return trans; | 2841 | return trans; |
2822 | 2842 | ||
2823 | if (inode->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) | 2843 | if (ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) |
2824 | return ERR_PTR(-ENOSPC); | 2844 | return ERR_PTR(-ENOSPC); |
2825 | 2845 | ||
2826 | /* check if there is someone else holds reference */ | 2846 | /* check if there is someone else holds reference */ |
@@ -2879,7 +2899,7 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2879 | 2899 | ||
2880 | if (ret == 0 && S_ISREG(inode->i_mode)) { | 2900 | if (ret == 0 && S_ISREG(inode->i_mode)) { |
2881 | ret = btrfs_lookup_file_extent(trans, root, path, | 2901 | ret = btrfs_lookup_file_extent(trans, root, path, |
2882 | inode->i_ino, (u64)-1, 0); | 2902 | ino, (u64)-1, 0); |
2883 | if (ret < 0) { | 2903 | if (ret < 0) { |
2884 | err = ret; | 2904 | err = ret; |
2885 | goto out; | 2905 | goto out; |
@@ -2895,7 +2915,7 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2895 | goto out; | 2915 | goto out; |
2896 | } | 2916 | } |
2897 | 2917 | ||
2898 | di = btrfs_lookup_dir_item(trans, root, path, dir->i_ino, | 2918 | di = btrfs_lookup_dir_item(trans, root, path, dir_ino, |
2899 | dentry->d_name.name, dentry->d_name.len, 0); | 2919 | dentry->d_name.name, dentry->d_name.len, 0); |
2900 | if (IS_ERR(di)) { | 2920 | if (IS_ERR(di)) { |
2901 | err = PTR_ERR(di); | 2921 | err = PTR_ERR(di); |
@@ -2912,7 +2932,7 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2912 | 2932 | ||
2913 | ref = btrfs_lookup_inode_ref(trans, root, path, | 2933 | ref = btrfs_lookup_inode_ref(trans, root, path, |
2914 | dentry->d_name.name, dentry->d_name.len, | 2934 | dentry->d_name.name, dentry->d_name.len, |
2915 | inode->i_ino, dir->i_ino, 0); | 2935 | ino, dir_ino, 0); |
2916 | if (IS_ERR(ref)) { | 2936 | if (IS_ERR(ref)) { |
2917 | err = PTR_ERR(ref); | 2937 | err = PTR_ERR(ref); |
2918 | goto out; | 2938 | goto out; |
@@ -2923,7 +2943,15 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2923 | index = btrfs_inode_ref_index(path->nodes[0], ref); | 2943 | index = btrfs_inode_ref_index(path->nodes[0], ref); |
2924 | btrfs_release_path(path); | 2944 | btrfs_release_path(path); |
2925 | 2945 | ||
2926 | di = btrfs_lookup_dir_index_item(trans, root, path, dir->i_ino, index, | 2946 | /* |
2947 | * This is a commit root search, if we can lookup inode item and other | ||
2948 | * relative items in the commit root, it means the transaction of | ||
2949 | * dir/file creation has been committed, and the dir index item that we | ||
2950 | * delay to insert has also been inserted into the commit root. So | ||
2951 | * we needn't worry about the delayed insertion of the dir index item | ||
2952 | * here. | ||
2953 | */ | ||
2954 | di = btrfs_lookup_dir_index_item(trans, root, path, dir_ino, index, | ||
2927 | dentry->d_name.name, dentry->d_name.len, 0); | 2955 | dentry->d_name.name, dentry->d_name.len, 0); |
2928 | if (IS_ERR(di)) { | 2956 | if (IS_ERR(di)) { |
2929 | err = PTR_ERR(di); | 2957 | err = PTR_ERR(di); |
@@ -2998,12 +3026,13 @@ int btrfs_unlink_subvol(struct btrfs_trans_handle *trans, | |||
2998 | struct btrfs_key key; | 3026 | struct btrfs_key key; |
2999 | u64 index; | 3027 | u64 index; |
3000 | int ret; | 3028 | int ret; |
3029 | u64 dir_ino = btrfs_ino(dir); | ||
3001 | 3030 | ||
3002 | path = btrfs_alloc_path(); | 3031 | path = btrfs_alloc_path(); |
3003 | if (!path) | 3032 | if (!path) |
3004 | return -ENOMEM; | 3033 | return -ENOMEM; |
3005 | 3034 | ||
3006 | di = btrfs_lookup_dir_item(trans, root, path, dir->i_ino, | 3035 | di = btrfs_lookup_dir_item(trans, root, path, dir_ino, |
3007 | name, name_len, -1); | 3036 | name, name_len, -1); |
3008 | BUG_ON(IS_ERR_OR_NULL(di)); | 3037 | BUG_ON(IS_ERR_OR_NULL(di)); |
3009 | 3038 | ||
@@ -3016,10 +3045,10 @@ int btrfs_unlink_subvol(struct btrfs_trans_handle *trans, | |||
3016 | 3045 | ||
3017 | ret = btrfs_del_root_ref(trans, root->fs_info->tree_root, | 3046 | ret = btrfs_del_root_ref(trans, root->fs_info->tree_root, |
3018 | objectid, root->root_key.objectid, | 3047 | objectid, root->root_key.objectid, |
3019 | dir->i_ino, &index, name, name_len); | 3048 | dir_ino, &index, name, name_len); |
3020 | if (ret < 0) { | 3049 | if (ret < 0) { |
3021 | BUG_ON(ret != -ENOENT); | 3050 | BUG_ON(ret != -ENOENT); |
3022 | di = btrfs_search_dir_index_item(root, path, dir->i_ino, | 3051 | di = btrfs_search_dir_index_item(root, path, dir_ino, |
3023 | name, name_len); | 3052 | name, name_len); |
3024 | BUG_ON(IS_ERR_OR_NULL(di)); | 3053 | BUG_ON(IS_ERR_OR_NULL(di)); |
3025 | 3054 | ||
@@ -3028,24 +3057,16 @@ int btrfs_unlink_subvol(struct btrfs_trans_handle *trans, | |||
3028 | btrfs_release_path(path); | 3057 | btrfs_release_path(path); |
3029 | index = key.offset; | 3058 | index = key.offset; |
3030 | } | 3059 | } |
3060 | btrfs_release_path(path); | ||
3031 | 3061 | ||
3032 | di = btrfs_lookup_dir_index_item(trans, root, path, dir->i_ino, | 3062 | ret = btrfs_delete_delayed_dir_index(trans, root, dir, index); |
3033 | index, name, name_len, -1); | ||
3034 | BUG_ON(IS_ERR_OR_NULL(di)); | ||
3035 | |||
3036 | leaf = path->nodes[0]; | ||
3037 | btrfs_dir_item_key_to_cpu(leaf, di, &key); | ||
3038 | WARN_ON(key.type != BTRFS_ROOT_ITEM_KEY || key.objectid != objectid); | ||
3039 | ret = btrfs_delete_one_dir_name(trans, root, path, di); | ||
3040 | BUG_ON(ret); | 3063 | BUG_ON(ret); |
3041 | btrfs_release_path(path); | ||
3042 | 3064 | ||
3043 | btrfs_i_size_write(dir, dir->i_size - name_len * 2); | 3065 | btrfs_i_size_write(dir, dir->i_size - name_len * 2); |
3044 | dir->i_mtime = dir->i_ctime = CURRENT_TIME; | 3066 | dir->i_mtime = dir->i_ctime = CURRENT_TIME; |
3045 | ret = btrfs_update_inode(trans, root, dir); | 3067 | ret = btrfs_update_inode(trans, root, dir); |
3046 | BUG_ON(ret); | 3068 | BUG_ON(ret); |
3047 | 3069 | ||
3048 | btrfs_free_path(path); | ||
3049 | return 0; | 3070 | return 0; |
3050 | } | 3071 | } |
3051 | 3072 | ||
@@ -3058,7 +3079,7 @@ static int btrfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
3058 | unsigned long nr = 0; | 3079 | unsigned long nr = 0; |
3059 | 3080 | ||
3060 | if (inode->i_size > BTRFS_EMPTY_DIR_SIZE || | 3081 | if (inode->i_size > BTRFS_EMPTY_DIR_SIZE || |
3061 | inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) | 3082 | btrfs_ino(inode) == BTRFS_FIRST_FREE_OBJECTID) |
3062 | return -ENOTEMPTY; | 3083 | return -ENOTEMPTY; |
3063 | 3084 | ||
3064 | trans = __unlink_start_trans(dir, dentry); | 3085 | trans = __unlink_start_trans(dir, dentry); |
@@ -3067,7 +3088,7 @@ static int btrfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
3067 | 3088 | ||
3068 | btrfs_set_trans_block_group(trans, dir); | 3089 | btrfs_set_trans_block_group(trans, dir); |
3069 | 3090 | ||
3070 | if (unlikely(inode->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { | 3091 | if (unlikely(btrfs_ino(inode) == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { |
3071 | err = btrfs_unlink_subvol(trans, root, dir, | 3092 | err = btrfs_unlink_subvol(trans, root, dir, |
3072 | BTRFS_I(inode)->location.objectid, | 3093 | BTRFS_I(inode)->location.objectid, |
3073 | dentry->d_name.name, | 3094 | dentry->d_name.name, |
@@ -3127,17 +3148,27 @@ int btrfs_truncate_inode_items(struct btrfs_trans_handle *trans, | |||
3127 | int encoding; | 3148 | int encoding; |
3128 | int ret; | 3149 | int ret; |
3129 | int err = 0; | 3150 | int err = 0; |
3151 | u64 ino = btrfs_ino(inode); | ||
3130 | 3152 | ||
3131 | BUG_ON(new_size > 0 && min_type != BTRFS_EXTENT_DATA_KEY); | 3153 | BUG_ON(new_size > 0 && min_type != BTRFS_EXTENT_DATA_KEY); |
3132 | 3154 | ||
3133 | if (root->ref_cows || root == root->fs_info->tree_root) | 3155 | if (root->ref_cows || root == root->fs_info->tree_root) |
3134 | btrfs_drop_extent_cache(inode, new_size & (~mask), (u64)-1, 0); | 3156 | btrfs_drop_extent_cache(inode, new_size & (~mask), (u64)-1, 0); |
3135 | 3157 | ||
3158 | /* | ||
3159 | * This function is also used to drop the items in the log tree before | ||
3160 | * we relog the inode, so if root != BTRFS_I(inode)->root, it means | ||
3161 | * it is used to drop the loged items. So we shouldn't kill the delayed | ||
3162 | * items. | ||
3163 | */ | ||
3164 | if (min_type == 0 && root == BTRFS_I(inode)->root) | ||
3165 | btrfs_kill_delayed_inode_items(inode); | ||
3166 | |||
3136 | path = btrfs_alloc_path(); | 3167 | path = btrfs_alloc_path(); |
3137 | BUG_ON(!path); | 3168 | BUG_ON(!path); |
3138 | path->reada = -1; | 3169 | path->reada = -1; |
3139 | 3170 | ||
3140 | key.objectid = inode->i_ino; | 3171 | key.objectid = ino; |
3141 | key.offset = (u64)-1; | 3172 | key.offset = (u64)-1; |
3142 | key.type = (u8)-1; | 3173 | key.type = (u8)-1; |
3143 | 3174 | ||
@@ -3165,7 +3196,7 @@ search_again: | |||
3165 | found_type = btrfs_key_type(&found_key); | 3196 | found_type = btrfs_key_type(&found_key); |
3166 | encoding = 0; | 3197 | encoding = 0; |
3167 | 3198 | ||
3168 | if (found_key.objectid != inode->i_ino) | 3199 | if (found_key.objectid != ino) |
3169 | break; | 3200 | break; |
3170 | 3201 | ||
3171 | if (found_type < min_type) | 3202 | if (found_type < min_type) |
@@ -3284,7 +3315,7 @@ delete: | |||
3284 | ret = btrfs_free_extent(trans, root, extent_start, | 3315 | ret = btrfs_free_extent(trans, root, extent_start, |
3285 | extent_num_bytes, 0, | 3316 | extent_num_bytes, 0, |
3286 | btrfs_header_owner(leaf), | 3317 | btrfs_header_owner(leaf), |
3287 | inode->i_ino, extent_offset); | 3318 | ino, extent_offset); |
3288 | BUG_ON(ret); | 3319 | BUG_ON(ret); |
3289 | } | 3320 | } |
3290 | 3321 | ||
@@ -3293,7 +3324,9 @@ delete: | |||
3293 | 3324 | ||
3294 | if (path->slots[0] == 0 || | 3325 | if (path->slots[0] == 0 || |
3295 | path->slots[0] != pending_del_slot) { | 3326 | path->slots[0] != pending_del_slot) { |
3296 | if (root->ref_cows) { | 3327 | if (root->ref_cows && |
3328 | BTRFS_I(inode)->location.objectid != | ||
3329 | BTRFS_FREE_INO_OBJECTID) { | ||
3297 | err = -EAGAIN; | 3330 | err = -EAGAIN; |
3298 | goto out; | 3331 | goto out; |
3299 | } | 3332 | } |
@@ -3483,7 +3516,7 @@ int btrfs_cont_expand(struct inode *inode, loff_t oldsize, loff_t size) | |||
3483 | break; | 3516 | break; |
3484 | 3517 | ||
3485 | err = btrfs_insert_file_extent(trans, root, | 3518 | err = btrfs_insert_file_extent(trans, root, |
3486 | inode->i_ino, cur_offset, 0, | 3519 | btrfs_ino(inode), cur_offset, 0, |
3487 | 0, hole_size, 0, hole_size, | 3520 | 0, hole_size, 0, hole_size, |
3488 | 0, 0, 0); | 3521 | 0, 0, 0); |
3489 | if (err) | 3522 | if (err) |
@@ -3585,7 +3618,7 @@ void btrfs_evict_inode(struct inode *inode) | |||
3585 | 3618 | ||
3586 | truncate_inode_pages(&inode->i_data, 0); | 3619 | truncate_inode_pages(&inode->i_data, 0); |
3587 | if (inode->i_nlink && (btrfs_root_refs(&root->root_item) != 0 || | 3620 | if (inode->i_nlink && (btrfs_root_refs(&root->root_item) != 0 || |
3588 | root == root->fs_info->tree_root)) | 3621 | is_free_space_inode(root, inode))) |
3589 | goto no_delete; | 3622 | goto no_delete; |
3590 | 3623 | ||
3591 | if (is_bad_inode(inode)) { | 3624 | if (is_bad_inode(inode)) { |
@@ -3638,6 +3671,10 @@ void btrfs_evict_inode(struct inode *inode) | |||
3638 | BUG_ON(ret); | 3671 | BUG_ON(ret); |
3639 | } | 3672 | } |
3640 | 3673 | ||
3674 | if (!(root == root->fs_info->tree_root || | ||
3675 | root->root_key.objectid == BTRFS_TREE_RELOC_OBJECTID)) | ||
3676 | btrfs_return_ino(root, btrfs_ino(inode)); | ||
3677 | |||
3641 | nr = trans->blocks_used; | 3678 | nr = trans->blocks_used; |
3642 | btrfs_end_transaction(trans, root); | 3679 | btrfs_end_transaction(trans, root); |
3643 | btrfs_btree_balance_dirty(root, nr); | 3680 | btrfs_btree_balance_dirty(root, nr); |
@@ -3663,7 +3700,7 @@ static int btrfs_inode_by_name(struct inode *dir, struct dentry *dentry, | |||
3663 | path = btrfs_alloc_path(); | 3700 | path = btrfs_alloc_path(); |
3664 | BUG_ON(!path); | 3701 | BUG_ON(!path); |
3665 | 3702 | ||
3666 | di = btrfs_lookup_dir_item(NULL, root, path, dir->i_ino, name, | 3703 | di = btrfs_lookup_dir_item(NULL, root, path, btrfs_ino(dir), name, |
3667 | namelen, 0); | 3704 | namelen, 0); |
3668 | if (IS_ERR(di)) | 3705 | if (IS_ERR(di)) |
3669 | ret = PTR_ERR(di); | 3706 | ret = PTR_ERR(di); |
@@ -3716,7 +3753,7 @@ static int fixup_tree_root_location(struct btrfs_root *root, | |||
3716 | 3753 | ||
3717 | leaf = path->nodes[0]; | 3754 | leaf = path->nodes[0]; |
3718 | ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref); | 3755 | ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref); |
3719 | if (btrfs_root_ref_dirid(leaf, ref) != dir->i_ino || | 3756 | if (btrfs_root_ref_dirid(leaf, ref) != btrfs_ino(dir) || |
3720 | btrfs_root_ref_name_len(leaf, ref) != dentry->d_name.len) | 3757 | btrfs_root_ref_name_len(leaf, ref) != dentry->d_name.len) |
3721 | goto out; | 3758 | goto out; |
3722 | 3759 | ||
@@ -3755,6 +3792,7 @@ static void inode_tree_add(struct inode *inode) | |||
3755 | struct btrfs_inode *entry; | 3792 | struct btrfs_inode *entry; |
3756 | struct rb_node **p; | 3793 | struct rb_node **p; |
3757 | struct rb_node *parent; | 3794 | struct rb_node *parent; |
3795 | u64 ino = btrfs_ino(inode); | ||
3758 | again: | 3796 | again: |
3759 | p = &root->inode_tree.rb_node; | 3797 | p = &root->inode_tree.rb_node; |
3760 | parent = NULL; | 3798 | parent = NULL; |
@@ -3767,9 +3805,9 @@ again: | |||
3767 | parent = *p; | 3805 | parent = *p; |
3768 | entry = rb_entry(parent, struct btrfs_inode, rb_node); | 3806 | entry = rb_entry(parent, struct btrfs_inode, rb_node); |
3769 | 3807 | ||
3770 | if (inode->i_ino < entry->vfs_inode.i_ino) | 3808 | if (ino < btrfs_ino(&entry->vfs_inode)) |
3771 | p = &parent->rb_left; | 3809 | p = &parent->rb_left; |
3772 | else if (inode->i_ino > entry->vfs_inode.i_ino) | 3810 | else if (ino > btrfs_ino(&entry->vfs_inode)) |
3773 | p = &parent->rb_right; | 3811 | p = &parent->rb_right; |
3774 | else { | 3812 | else { |
3775 | WARN_ON(!(entry->vfs_inode.i_state & | 3813 | WARN_ON(!(entry->vfs_inode.i_state & |
@@ -3833,9 +3871,9 @@ again: | |||
3833 | prev = node; | 3871 | prev = node; |
3834 | entry = rb_entry(node, struct btrfs_inode, rb_node); | 3872 | entry = rb_entry(node, struct btrfs_inode, rb_node); |
3835 | 3873 | ||
3836 | if (objectid < entry->vfs_inode.i_ino) | 3874 | if (objectid < btrfs_ino(&entry->vfs_inode)) |
3837 | node = node->rb_left; | 3875 | node = node->rb_left; |
3838 | else if (objectid > entry->vfs_inode.i_ino) | 3876 | else if (objectid > btrfs_ino(&entry->vfs_inode)) |
3839 | node = node->rb_right; | 3877 | node = node->rb_right; |
3840 | else | 3878 | else |
3841 | break; | 3879 | break; |
@@ -3843,7 +3881,7 @@ again: | |||
3843 | if (!node) { | 3881 | if (!node) { |
3844 | while (prev) { | 3882 | while (prev) { |
3845 | entry = rb_entry(prev, struct btrfs_inode, rb_node); | 3883 | entry = rb_entry(prev, struct btrfs_inode, rb_node); |
3846 | if (objectid <= entry->vfs_inode.i_ino) { | 3884 | if (objectid <= btrfs_ino(&entry->vfs_inode)) { |
3847 | node = prev; | 3885 | node = prev; |
3848 | break; | 3886 | break; |
3849 | } | 3887 | } |
@@ -3852,7 +3890,7 @@ again: | |||
3852 | } | 3890 | } |
3853 | while (node) { | 3891 | while (node) { |
3854 | entry = rb_entry(node, struct btrfs_inode, rb_node); | 3892 | entry = rb_entry(node, struct btrfs_inode, rb_node); |
3855 | objectid = entry->vfs_inode.i_ino + 1; | 3893 | objectid = btrfs_ino(&entry->vfs_inode) + 1; |
3856 | inode = igrab(&entry->vfs_inode); | 3894 | inode = igrab(&entry->vfs_inode); |
3857 | if (inode) { | 3895 | if (inode) { |
3858 | spin_unlock(&root->inode_lock); | 3896 | spin_unlock(&root->inode_lock); |
@@ -3890,7 +3928,7 @@ static int btrfs_init_locked_inode(struct inode *inode, void *p) | |||
3890 | static int btrfs_find_actor(struct inode *inode, void *opaque) | 3928 | static int btrfs_find_actor(struct inode *inode, void *opaque) |
3891 | { | 3929 | { |
3892 | struct btrfs_iget_args *args = opaque; | 3930 | struct btrfs_iget_args *args = opaque; |
3893 | return args->ino == inode->i_ino && | 3931 | return args->ino == btrfs_ino(inode) && |
3894 | args->root == BTRFS_I(inode)->root; | 3932 | args->root == BTRFS_I(inode)->root; |
3895 | } | 3933 | } |
3896 | 3934 | ||
@@ -4035,7 +4073,7 @@ static struct dentry *btrfs_lookup(struct inode *dir, struct dentry *dentry, | |||
4035 | return d_splice_alias(inode, dentry); | 4073 | return d_splice_alias(inode, dentry); |
4036 | } | 4074 | } |
4037 | 4075 | ||
4038 | static unsigned char btrfs_filetype_table[] = { | 4076 | unsigned char btrfs_filetype_table[] = { |
4039 | DT_UNKNOWN, DT_REG, DT_DIR, DT_CHR, DT_BLK, DT_FIFO, DT_SOCK, DT_LNK | 4077 | DT_UNKNOWN, DT_REG, DT_DIR, DT_CHR, DT_BLK, DT_FIFO, DT_SOCK, DT_LNK |
4040 | }; | 4078 | }; |
4041 | 4079 | ||
@@ -4049,6 +4087,8 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4049 | struct btrfs_key key; | 4087 | struct btrfs_key key; |
4050 | struct btrfs_key found_key; | 4088 | struct btrfs_key found_key; |
4051 | struct btrfs_path *path; | 4089 | struct btrfs_path *path; |
4090 | struct list_head ins_list; | ||
4091 | struct list_head del_list; | ||
4052 | int ret; | 4092 | int ret; |
4053 | struct extent_buffer *leaf; | 4093 | struct extent_buffer *leaf; |
4054 | int slot; | 4094 | int slot; |
@@ -4061,6 +4101,7 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4061 | char tmp_name[32]; | 4101 | char tmp_name[32]; |
4062 | char *name_ptr; | 4102 | char *name_ptr; |
4063 | int name_len; | 4103 | int name_len; |
4104 | int is_curr = 0; /* filp->f_pos points to the current index? */ | ||
4064 | 4105 | ||
4065 | /* FIXME, use a real flag for deciding about the key type */ | 4106 | /* FIXME, use a real flag for deciding about the key type */ |
4066 | if (root->fs_info->tree_root == root) | 4107 | if (root->fs_info->tree_root == root) |
@@ -4068,9 +4109,7 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4068 | 4109 | ||
4069 | /* special case for "." */ | 4110 | /* special case for "." */ |
4070 | if (filp->f_pos == 0) { | 4111 | if (filp->f_pos == 0) { |
4071 | over = filldir(dirent, ".", 1, | 4112 | over = filldir(dirent, ".", 1, 1, btrfs_ino(inode), DT_DIR); |
4072 | 1, inode->i_ino, | ||
4073 | DT_DIR); | ||
4074 | if (over) | 4113 | if (over) |
4075 | return 0; | 4114 | return 0; |
4076 | filp->f_pos = 1; | 4115 | filp->f_pos = 1; |
@@ -4085,11 +4124,19 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4085 | filp->f_pos = 2; | 4124 | filp->f_pos = 2; |
4086 | } | 4125 | } |
4087 | path = btrfs_alloc_path(); | 4126 | path = btrfs_alloc_path(); |
4127 | if (!path) | ||
4128 | return -ENOMEM; | ||
4088 | path->reada = 2; | 4129 | path->reada = 2; |
4089 | 4130 | ||
4131 | if (key_type == BTRFS_DIR_INDEX_KEY) { | ||
4132 | INIT_LIST_HEAD(&ins_list); | ||
4133 | INIT_LIST_HEAD(&del_list); | ||
4134 | btrfs_get_delayed_items(inode, &ins_list, &del_list); | ||
4135 | } | ||
4136 | |||
4090 | btrfs_set_key_type(&key, key_type); | 4137 | btrfs_set_key_type(&key, key_type); |
4091 | key.offset = filp->f_pos; | 4138 | key.offset = filp->f_pos; |
4092 | key.objectid = inode->i_ino; | 4139 | key.objectid = btrfs_ino(inode); |
4093 | 4140 | ||
4094 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | 4141 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); |
4095 | if (ret < 0) | 4142 | if (ret < 0) |
@@ -4116,8 +4163,13 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4116 | break; | 4163 | break; |
4117 | if (found_key.offset < filp->f_pos) | 4164 | if (found_key.offset < filp->f_pos) |
4118 | goto next; | 4165 | goto next; |
4166 | if (key_type == BTRFS_DIR_INDEX_KEY && | ||
4167 | btrfs_should_delete_dir_index(&del_list, | ||
4168 | found_key.offset)) | ||
4169 | goto next; | ||
4119 | 4170 | ||
4120 | filp->f_pos = found_key.offset; | 4171 | filp->f_pos = found_key.offset; |
4172 | is_curr = 1; | ||
4121 | 4173 | ||
4122 | di = btrfs_item_ptr(leaf, slot, struct btrfs_dir_item); | 4174 | di = btrfs_item_ptr(leaf, slot, struct btrfs_dir_item); |
4123 | di_cur = 0; | 4175 | di_cur = 0; |
@@ -4172,6 +4224,15 @@ next: | |||
4172 | path->slots[0]++; | 4224 | path->slots[0]++; |
4173 | } | 4225 | } |
4174 | 4226 | ||
4227 | if (key_type == BTRFS_DIR_INDEX_KEY) { | ||
4228 | if (is_curr) | ||
4229 | filp->f_pos++; | ||
4230 | ret = btrfs_readdir_delayed_dir_index(filp, dirent, filldir, | ||
4231 | &ins_list); | ||
4232 | if (ret) | ||
4233 | goto nopos; | ||
4234 | } | ||
4235 | |||
4175 | /* Reached end of directory/root. Bump pos past the last item. */ | 4236 | /* Reached end of directory/root. Bump pos past the last item. */ |
4176 | if (key_type == BTRFS_DIR_INDEX_KEY) | 4237 | if (key_type == BTRFS_DIR_INDEX_KEY) |
4177 | /* | 4238 | /* |
@@ -4184,6 +4245,8 @@ next: | |||
4184 | nopos: | 4245 | nopos: |
4185 | ret = 0; | 4246 | ret = 0; |
4186 | err: | 4247 | err: |
4248 | if (key_type == BTRFS_DIR_INDEX_KEY) | ||
4249 | btrfs_put_delayed_items(&ins_list, &del_list); | ||
4187 | btrfs_free_path(path); | 4250 | btrfs_free_path(path); |
4188 | return ret; | 4251 | return ret; |
4189 | } | 4252 | } |
@@ -4199,7 +4262,8 @@ int btrfs_write_inode(struct inode *inode, struct writeback_control *wbc) | |||
4199 | return 0; | 4262 | return 0; |
4200 | 4263 | ||
4201 | smp_mb(); | 4264 | smp_mb(); |
4202 | nolock = (root->fs_info->closing && root == root->fs_info->tree_root); | 4265 | if (root->fs_info->closing && is_free_space_inode(root, inode)) |
4266 | nolock = true; | ||
4203 | 4267 | ||
4204 | if (wbc->sync_mode == WB_SYNC_ALL) { | 4268 | if (wbc->sync_mode == WB_SYNC_ALL) { |
4205 | if (nolock) | 4269 | if (nolock) |
@@ -4243,8 +4307,9 @@ void btrfs_dirty_inode(struct inode *inode) | |||
4243 | trans = btrfs_start_transaction(root, 1); | 4307 | trans = btrfs_start_transaction(root, 1); |
4244 | if (IS_ERR(trans)) { | 4308 | if (IS_ERR(trans)) { |
4245 | printk_ratelimited(KERN_ERR "btrfs: fail to " | 4309 | printk_ratelimited(KERN_ERR "btrfs: fail to " |
4246 | "dirty inode %lu error %ld\n", | 4310 | "dirty inode %llu error %ld\n", |
4247 | inode->i_ino, PTR_ERR(trans)); | 4311 | (unsigned long long)btrfs_ino(inode), |
4312 | PTR_ERR(trans)); | ||
4248 | return; | 4313 | return; |
4249 | } | 4314 | } |
4250 | btrfs_set_trans_block_group(trans, inode); | 4315 | btrfs_set_trans_block_group(trans, inode); |
@@ -4252,11 +4317,14 @@ void btrfs_dirty_inode(struct inode *inode) | |||
4252 | ret = btrfs_update_inode(trans, root, inode); | 4317 | ret = btrfs_update_inode(trans, root, inode); |
4253 | if (ret) { | 4318 | if (ret) { |
4254 | printk_ratelimited(KERN_ERR "btrfs: fail to " | 4319 | printk_ratelimited(KERN_ERR "btrfs: fail to " |
4255 | "dirty inode %lu error %d\n", | 4320 | "dirty inode %llu error %d\n", |
4256 | inode->i_ino, ret); | 4321 | (unsigned long long)btrfs_ino(inode), |
4322 | ret); | ||
4257 | } | 4323 | } |
4258 | } | 4324 | } |
4259 | btrfs_end_transaction(trans, root); | 4325 | btrfs_end_transaction(trans, root); |
4326 | if (BTRFS_I(inode)->delayed_node) | ||
4327 | btrfs_balance_delayed_items(root); | ||
4260 | } | 4328 | } |
4261 | 4329 | ||
4262 | /* | 4330 | /* |
@@ -4272,7 +4340,7 @@ static int btrfs_set_inode_index_count(struct inode *inode) | |||
4272 | struct extent_buffer *leaf; | 4340 | struct extent_buffer *leaf; |
4273 | int ret; | 4341 | int ret; |
4274 | 4342 | ||
4275 | key.objectid = inode->i_ino; | 4343 | key.objectid = btrfs_ino(inode); |
4276 | btrfs_set_key_type(&key, BTRFS_DIR_INDEX_KEY); | 4344 | btrfs_set_key_type(&key, BTRFS_DIR_INDEX_KEY); |
4277 | key.offset = (u64)-1; | 4345 | key.offset = (u64)-1; |
4278 | 4346 | ||
@@ -4304,7 +4372,7 @@ static int btrfs_set_inode_index_count(struct inode *inode) | |||
4304 | leaf = path->nodes[0]; | 4372 | leaf = path->nodes[0]; |
4305 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); | 4373 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); |
4306 | 4374 | ||
4307 | if (found_key.objectid != inode->i_ino || | 4375 | if (found_key.objectid != btrfs_ino(inode) || |
4308 | btrfs_key_type(&found_key) != BTRFS_DIR_INDEX_KEY) { | 4376 | btrfs_key_type(&found_key) != BTRFS_DIR_INDEX_KEY) { |
4309 | BTRFS_I(inode)->index_cnt = 2; | 4377 | BTRFS_I(inode)->index_cnt = 2; |
4310 | goto out; | 4378 | goto out; |
@@ -4325,9 +4393,12 @@ int btrfs_set_inode_index(struct inode *dir, u64 *index) | |||
4325 | int ret = 0; | 4393 | int ret = 0; |
4326 | 4394 | ||
4327 | if (BTRFS_I(dir)->index_cnt == (u64)-1) { | 4395 | if (BTRFS_I(dir)->index_cnt == (u64)-1) { |
4328 | ret = btrfs_set_inode_index_count(dir); | 4396 | ret = btrfs_inode_delayed_dir_index_count(dir); |
4329 | if (ret) | 4397 | if (ret) { |
4330 | return ret; | 4398 | ret = btrfs_set_inode_index_count(dir); |
4399 | if (ret) | ||
4400 | return ret; | ||
4401 | } | ||
4331 | } | 4402 | } |
4332 | 4403 | ||
4333 | *index = BTRFS_I(dir)->index_cnt; | 4404 | *index = BTRFS_I(dir)->index_cnt; |
@@ -4363,6 +4434,12 @@ static struct inode *btrfs_new_inode(struct btrfs_trans_handle *trans, | |||
4363 | return ERR_PTR(-ENOMEM); | 4434 | return ERR_PTR(-ENOMEM); |
4364 | } | 4435 | } |
4365 | 4436 | ||
4437 | /* | ||
4438 | * we have to initialize this early, so we can reclaim the inode | ||
4439 | * number if we fail afterwards in this function. | ||
4440 | */ | ||
4441 | inode->i_ino = objectid; | ||
4442 | |||
4366 | if (dir) { | 4443 | if (dir) { |
4367 | trace_btrfs_inode_request(dir); | 4444 | trace_btrfs_inode_request(dir); |
4368 | 4445 | ||
@@ -4408,7 +4485,6 @@ static struct inode *btrfs_new_inode(struct btrfs_trans_handle *trans, | |||
4408 | goto fail; | 4485 | goto fail; |
4409 | 4486 | ||
4410 | inode_init_owner(inode, dir, mode); | 4487 | inode_init_owner(inode, dir, mode); |
4411 | inode->i_ino = objectid; | ||
4412 | inode_set_bytes(inode, 0); | 4488 | inode_set_bytes(inode, 0); |
4413 | inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME; | 4489 | inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME; |
4414 | inode_item = btrfs_item_ptr(path->nodes[0], path->slots[0], | 4490 | inode_item = btrfs_item_ptr(path->nodes[0], path->slots[0], |
@@ -4472,29 +4548,29 @@ int btrfs_add_link(struct btrfs_trans_handle *trans, | |||
4472 | int ret = 0; | 4548 | int ret = 0; |
4473 | struct btrfs_key key; | 4549 | struct btrfs_key key; |
4474 | struct btrfs_root *root = BTRFS_I(parent_inode)->root; | 4550 | struct btrfs_root *root = BTRFS_I(parent_inode)->root; |
4551 | u64 ino = btrfs_ino(inode); | ||
4552 | u64 parent_ino = btrfs_ino(parent_inode); | ||
4475 | 4553 | ||
4476 | if (unlikely(inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 4554 | if (unlikely(ino == BTRFS_FIRST_FREE_OBJECTID)) { |
4477 | memcpy(&key, &BTRFS_I(inode)->root->root_key, sizeof(key)); | 4555 | memcpy(&key, &BTRFS_I(inode)->root->root_key, sizeof(key)); |
4478 | } else { | 4556 | } else { |
4479 | key.objectid = inode->i_ino; | 4557 | key.objectid = ino; |
4480 | btrfs_set_key_type(&key, BTRFS_INODE_ITEM_KEY); | 4558 | btrfs_set_key_type(&key, BTRFS_INODE_ITEM_KEY); |
4481 | key.offset = 0; | 4559 | key.offset = 0; |
4482 | } | 4560 | } |
4483 | 4561 | ||
4484 | if (unlikely(inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 4562 | if (unlikely(ino == BTRFS_FIRST_FREE_OBJECTID)) { |
4485 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, | 4563 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, |
4486 | key.objectid, root->root_key.objectid, | 4564 | key.objectid, root->root_key.objectid, |
4487 | parent_inode->i_ino, | 4565 | parent_ino, index, name, name_len); |
4488 | index, name, name_len); | ||
4489 | } else if (add_backref) { | 4566 | } else if (add_backref) { |
4490 | ret = btrfs_insert_inode_ref(trans, root, | 4567 | ret = btrfs_insert_inode_ref(trans, root, name, name_len, ino, |
4491 | name, name_len, inode->i_ino, | 4568 | parent_ino, index); |
4492 | parent_inode->i_ino, index); | ||
4493 | } | 4569 | } |
4494 | 4570 | ||
4495 | if (ret == 0) { | 4571 | if (ret == 0) { |
4496 | ret = btrfs_insert_dir_item(trans, root, name, name_len, | 4572 | ret = btrfs_insert_dir_item(trans, root, name, name_len, |
4497 | parent_inode->i_ino, &key, | 4573 | parent_inode, &key, |
4498 | btrfs_inode_type(inode), index); | 4574 | btrfs_inode_type(inode), index); |
4499 | BUG_ON(ret); | 4575 | BUG_ON(ret); |
4500 | 4576 | ||
@@ -4537,10 +4613,6 @@ static int btrfs_mknod(struct inode *dir, struct dentry *dentry, | |||
4537 | if (!new_valid_dev(rdev)) | 4613 | if (!new_valid_dev(rdev)) |
4538 | return -EINVAL; | 4614 | return -EINVAL; |
4539 | 4615 | ||
4540 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
4541 | if (err) | ||
4542 | return err; | ||
4543 | |||
4544 | /* | 4616 | /* |
4545 | * 2 for inode item and ref | 4617 | * 2 for inode item and ref |
4546 | * 2 for dir items | 4618 | * 2 for dir items |
@@ -4552,8 +4624,12 @@ static int btrfs_mknod(struct inode *dir, struct dentry *dentry, | |||
4552 | 4624 | ||
4553 | btrfs_set_trans_block_group(trans, dir); | 4625 | btrfs_set_trans_block_group(trans, dir); |
4554 | 4626 | ||
4627 | err = btrfs_find_free_ino(root, &objectid); | ||
4628 | if (err) | ||
4629 | goto out_unlock; | ||
4630 | |||
4555 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 4631 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
4556 | dentry->d_name.len, dir->i_ino, objectid, | 4632 | dentry->d_name.len, btrfs_ino(dir), objectid, |
4557 | BTRFS_I(dir)->block_group, mode, &index); | 4633 | BTRFS_I(dir)->block_group, mode, &index); |
4558 | if (IS_ERR(inode)) { | 4634 | if (IS_ERR(inode)) { |
4559 | err = PTR_ERR(inode); | 4635 | err = PTR_ERR(inode); |
@@ -4600,9 +4676,6 @@ static int btrfs_create(struct inode *dir, struct dentry *dentry, | |||
4600 | u64 objectid; | 4676 | u64 objectid; |
4601 | u64 index = 0; | 4677 | u64 index = 0; |
4602 | 4678 | ||
4603 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
4604 | if (err) | ||
4605 | return err; | ||
4606 | /* | 4679 | /* |
4607 | * 2 for inode item and ref | 4680 | * 2 for inode item and ref |
4608 | * 2 for dir items | 4681 | * 2 for dir items |
@@ -4614,8 +4687,12 @@ static int btrfs_create(struct inode *dir, struct dentry *dentry, | |||
4614 | 4687 | ||
4615 | btrfs_set_trans_block_group(trans, dir); | 4688 | btrfs_set_trans_block_group(trans, dir); |
4616 | 4689 | ||
4690 | err = btrfs_find_free_ino(root, &objectid); | ||
4691 | if (err) | ||
4692 | goto out_unlock; | ||
4693 | |||
4617 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 4694 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
4618 | dentry->d_name.len, dir->i_ino, objectid, | 4695 | dentry->d_name.len, btrfs_ino(dir), objectid, |
4619 | BTRFS_I(dir)->block_group, mode, &index); | 4696 | BTRFS_I(dir)->block_group, mode, &index); |
4620 | if (IS_ERR(inode)) { | 4697 | if (IS_ERR(inode)) { |
4621 | err = PTR_ERR(inode); | 4698 | err = PTR_ERR(inode); |
@@ -4726,10 +4803,6 @@ static int btrfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
4726 | u64 index = 0; | 4803 | u64 index = 0; |
4727 | unsigned long nr = 1; | 4804 | unsigned long nr = 1; |
4728 | 4805 | ||
4729 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
4730 | if (err) | ||
4731 | return err; | ||
4732 | |||
4733 | /* | 4806 | /* |
4734 | * 2 items for inode and ref | 4807 | * 2 items for inode and ref |
4735 | * 2 items for dir items | 4808 | * 2 items for dir items |
@@ -4740,8 +4813,12 @@ static int btrfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
4740 | return PTR_ERR(trans); | 4813 | return PTR_ERR(trans); |
4741 | btrfs_set_trans_block_group(trans, dir); | 4814 | btrfs_set_trans_block_group(trans, dir); |
4742 | 4815 | ||
4816 | err = btrfs_find_free_ino(root, &objectid); | ||
4817 | if (err) | ||
4818 | goto out_fail; | ||
4819 | |||
4743 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 4820 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
4744 | dentry->d_name.len, dir->i_ino, objectid, | 4821 | dentry->d_name.len, btrfs_ino(dir), objectid, |
4745 | BTRFS_I(dir)->block_group, S_IFDIR | mode, | 4822 | BTRFS_I(dir)->block_group, S_IFDIR | mode, |
4746 | &index); | 4823 | &index); |
4747 | if (IS_ERR(inode)) { | 4824 | if (IS_ERR(inode)) { |
@@ -4864,7 +4941,7 @@ struct extent_map *btrfs_get_extent(struct inode *inode, struct page *page, | |||
4864 | u64 bytenr; | 4941 | u64 bytenr; |
4865 | u64 extent_start = 0; | 4942 | u64 extent_start = 0; |
4866 | u64 extent_end = 0; | 4943 | u64 extent_end = 0; |
4867 | u64 objectid = inode->i_ino; | 4944 | u64 objectid = btrfs_ino(inode); |
4868 | u32 found_type; | 4945 | u32 found_type; |
4869 | struct btrfs_path *path = NULL; | 4946 | struct btrfs_path *path = NULL; |
4870 | struct btrfs_root *root = BTRFS_I(inode)->root; | 4947 | struct btrfs_root *root = BTRFS_I(inode)->root; |
@@ -5372,7 +5449,7 @@ static noinline int can_nocow_odirect(struct btrfs_trans_handle *trans, | |||
5372 | if (!path) | 5449 | if (!path) |
5373 | return -ENOMEM; | 5450 | return -ENOMEM; |
5374 | 5451 | ||
5375 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 5452 | ret = btrfs_lookup_file_extent(trans, root, path, btrfs_ino(inode), |
5376 | offset, 0); | 5453 | offset, 0); |
5377 | if (ret < 0) | 5454 | if (ret < 0) |
5378 | goto out; | 5455 | goto out; |
@@ -5389,7 +5466,7 @@ static noinline int can_nocow_odirect(struct btrfs_trans_handle *trans, | |||
5389 | ret = 0; | 5466 | ret = 0; |
5390 | leaf = path->nodes[0]; | 5467 | leaf = path->nodes[0]; |
5391 | btrfs_item_key_to_cpu(leaf, &key, slot); | 5468 | btrfs_item_key_to_cpu(leaf, &key, slot); |
5392 | if (key.objectid != inode->i_ino || | 5469 | if (key.objectid != btrfs_ino(inode) || |
5393 | key.type != BTRFS_EXTENT_DATA_KEY) { | 5470 | key.type != BTRFS_EXTENT_DATA_KEY) { |
5394 | /* not our file or wrong item type, must cow */ | 5471 | /* not our file or wrong item type, must cow */ |
5395 | goto out; | 5472 | goto out; |
@@ -5423,7 +5500,7 @@ static noinline int can_nocow_odirect(struct btrfs_trans_handle *trans, | |||
5423 | * look for other files referencing this extent, if we | 5500 | * look for other files referencing this extent, if we |
5424 | * find any we must cow | 5501 | * find any we must cow |
5425 | */ | 5502 | */ |
5426 | if (btrfs_cross_ref_exist(trans, root, inode->i_ino, | 5503 | if (btrfs_cross_ref_exist(trans, root, btrfs_ino(inode), |
5427 | key.offset - backref_offset, disk_bytenr)) | 5504 | key.offset - backref_offset, disk_bytenr)) |
5428 | goto out; | 5505 | goto out; |
5429 | 5506 | ||
@@ -5613,9 +5690,10 @@ static void btrfs_endio_direct_read(struct bio *bio, int err) | |||
5613 | 5690 | ||
5614 | flush_dcache_page(bvec->bv_page); | 5691 | flush_dcache_page(bvec->bv_page); |
5615 | if (csum != *private) { | 5692 | if (csum != *private) { |
5616 | printk(KERN_ERR "btrfs csum failed ino %lu off" | 5693 | printk(KERN_ERR "btrfs csum failed ino %llu off" |
5617 | " %llu csum %u private %u\n", | 5694 | " %llu csum %u private %u\n", |
5618 | inode->i_ino, (unsigned long long)start, | 5695 | (unsigned long long)btrfs_ino(inode), |
5696 | (unsigned long long)start, | ||
5619 | csum, *private); | 5697 | csum, *private); |
5620 | err = -EIO; | 5698 | err = -EIO; |
5621 | } | 5699 | } |
@@ -5762,9 +5840,9 @@ static void btrfs_end_dio_bio(struct bio *bio, int err) | |||
5762 | struct btrfs_dio_private *dip = bio->bi_private; | 5840 | struct btrfs_dio_private *dip = bio->bi_private; |
5763 | 5841 | ||
5764 | if (err) { | 5842 | if (err) { |
5765 | printk(KERN_ERR "btrfs direct IO failed ino %lu rw %lu " | 5843 | printk(KERN_ERR "btrfs direct IO failed ino %llu rw %lu " |
5766 | "sector %#Lx len %u err no %d\n", | 5844 | "sector %#Lx len %u err no %d\n", |
5767 | dip->inode->i_ino, bio->bi_rw, | 5845 | (unsigned long long)btrfs_ino(dip->inode), bio->bi_rw, |
5768 | (unsigned long long)bio->bi_sector, bio->bi_size, err); | 5846 | (unsigned long long)bio->bi_sector, bio->bi_size, err); |
5769 | dip->errors = 1; | 5847 | dip->errors = 1; |
5770 | 5848 | ||
@@ -6607,6 +6685,8 @@ struct inode *btrfs_alloc_inode(struct super_block *sb) | |||
6607 | ei->dummy_inode = 0; | 6685 | ei->dummy_inode = 0; |
6608 | ei->force_compress = BTRFS_COMPRESS_NONE; | 6686 | ei->force_compress = BTRFS_COMPRESS_NONE; |
6609 | 6687 | ||
6688 | ei->delayed_node = NULL; | ||
6689 | |||
6610 | inode = &ei->vfs_inode; | 6690 | inode = &ei->vfs_inode; |
6611 | extent_map_tree_init(&ei->extent_tree); | 6691 | extent_map_tree_init(&ei->extent_tree); |
6612 | extent_io_tree_init(&ei->io_tree, &inode->i_data); | 6692 | extent_io_tree_init(&ei->io_tree, &inode->i_data); |
@@ -6674,8 +6754,8 @@ void btrfs_destroy_inode(struct inode *inode) | |||
6674 | 6754 | ||
6675 | spin_lock(&root->orphan_lock); | 6755 | spin_lock(&root->orphan_lock); |
6676 | if (!list_empty(&BTRFS_I(inode)->i_orphan)) { | 6756 | if (!list_empty(&BTRFS_I(inode)->i_orphan)) { |
6677 | printk(KERN_INFO "BTRFS: inode %lu still on the orphan list\n", | 6757 | printk(KERN_INFO "BTRFS: inode %llu still on the orphan list\n", |
6678 | inode->i_ino); | 6758 | (unsigned long long)btrfs_ino(inode)); |
6679 | list_del_init(&BTRFS_I(inode)->i_orphan); | 6759 | list_del_init(&BTRFS_I(inode)->i_orphan); |
6680 | } | 6760 | } |
6681 | spin_unlock(&root->orphan_lock); | 6761 | spin_unlock(&root->orphan_lock); |
@@ -6697,6 +6777,7 @@ void btrfs_destroy_inode(struct inode *inode) | |||
6697 | inode_tree_del(inode); | 6777 | inode_tree_del(inode); |
6698 | btrfs_drop_extent_cache(inode, 0, (u64)-1, 0); | 6778 | btrfs_drop_extent_cache(inode, 0, (u64)-1, 0); |
6699 | free: | 6779 | free: |
6780 | btrfs_remove_delayed_node(inode); | ||
6700 | call_rcu(&inode->i_rcu, btrfs_i_callback); | 6781 | call_rcu(&inode->i_rcu, btrfs_i_callback); |
6701 | } | 6782 | } |
6702 | 6783 | ||
@@ -6705,7 +6786,7 @@ int btrfs_drop_inode(struct inode *inode) | |||
6705 | struct btrfs_root *root = BTRFS_I(inode)->root; | 6786 | struct btrfs_root *root = BTRFS_I(inode)->root; |
6706 | 6787 | ||
6707 | if (btrfs_root_refs(&root->root_item) == 0 && | 6788 | if (btrfs_root_refs(&root->root_item) == 0 && |
6708 | root != root->fs_info->tree_root) | 6789 | !is_free_space_inode(root, inode)) |
6709 | return 1; | 6790 | return 1; |
6710 | else | 6791 | else |
6711 | return generic_drop_inode(inode); | 6792 | return generic_drop_inode(inode); |
@@ -6808,38 +6889,39 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6808 | struct btrfs_trans_handle *trans; | 6889 | struct btrfs_trans_handle *trans; |
6809 | struct btrfs_root *root = BTRFS_I(old_dir)->root; | 6890 | struct btrfs_root *root = BTRFS_I(old_dir)->root; |
6810 | struct btrfs_root *dest = BTRFS_I(new_dir)->root; | 6891 | struct btrfs_root *dest = BTRFS_I(new_dir)->root; |
6811 | struct inode *newinode = new_dentry->d_inode; | 6892 | struct inode *new_inode = new_dentry->d_inode; |
6812 | struct inode *old_inode = old_dentry->d_inode; | 6893 | struct inode *old_inode = old_dentry->d_inode; |
6813 | struct timespec ctime = CURRENT_TIME; | 6894 | struct timespec ctime = CURRENT_TIME; |
6814 | u64 index = 0; | 6895 | u64 index = 0; |
6815 | u64 root_objectid; | 6896 | u64 root_objectid; |
6816 | int ret; | 6897 | int ret; |
6898 | u64 old_ino = btrfs_ino(old_inode); | ||
6817 | 6899 | ||
6818 | if (new_dir->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) | 6900 | if (btrfs_ino(new_dir) == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) |
6819 | return -EPERM; | 6901 | return -EPERM; |
6820 | 6902 | ||
6821 | /* we only allow rename subvolume link between subvolumes */ | 6903 | /* we only allow rename subvolume link between subvolumes */ |
6822 | if (old_inode->i_ino != BTRFS_FIRST_FREE_OBJECTID && root != dest) | 6904 | if (old_ino != BTRFS_FIRST_FREE_OBJECTID && root != dest) |
6823 | return -EXDEV; | 6905 | return -EXDEV; |
6824 | 6906 | ||
6825 | if (old_inode->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID || | 6907 | if (old_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID || |
6826 | (newinode && newinode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) | 6908 | (new_inode && btrfs_ino(new_inode) == BTRFS_FIRST_FREE_OBJECTID)) |
6827 | return -ENOTEMPTY; | 6909 | return -ENOTEMPTY; |
6828 | 6910 | ||
6829 | if (S_ISDIR(old_inode->i_mode) && newinode && | 6911 | if (S_ISDIR(old_inode->i_mode) && new_inode && |
6830 | newinode->i_size > BTRFS_EMPTY_DIR_SIZE) | 6912 | new_inode->i_size > BTRFS_EMPTY_DIR_SIZE) |
6831 | return -ENOTEMPTY; | 6913 | return -ENOTEMPTY; |
6832 | /* | 6914 | /* |
6833 | * we're using rename to replace one file with another. | 6915 | * we're using rename to replace one file with another. |
6834 | * and the replacement file is large. Start IO on it now so | 6916 | * and the replacement file is large. Start IO on it now so |
6835 | * we don't add too much work to the end of the transaction | 6917 | * we don't add too much work to the end of the transaction |
6836 | */ | 6918 | */ |
6837 | if (newinode && S_ISREG(old_inode->i_mode) && newinode->i_size && | 6919 | if (new_inode && S_ISREG(old_inode->i_mode) && new_inode->i_size && |
6838 | old_inode->i_size > BTRFS_ORDERED_OPERATIONS_FLUSH_LIMIT) | 6920 | old_inode->i_size > BTRFS_ORDERED_OPERATIONS_FLUSH_LIMIT) |
6839 | filemap_flush(old_inode->i_mapping); | 6921 | filemap_flush(old_inode->i_mapping); |
6840 | 6922 | ||
6841 | /* close the racy window with snapshot create/destroy ioctl */ | 6923 | /* close the racy window with snapshot create/destroy ioctl */ |
6842 | if (old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) | 6924 | if (old_ino == BTRFS_FIRST_FREE_OBJECTID) |
6843 | down_read(&root->fs_info->subvol_sem); | 6925 | down_read(&root->fs_info->subvol_sem); |
6844 | /* | 6926 | /* |
6845 | * We want to reserve the absolute worst case amount of items. So if | 6927 | * We want to reserve the absolute worst case amount of items. So if |
@@ -6864,15 +6946,15 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6864 | if (ret) | 6946 | if (ret) |
6865 | goto out_fail; | 6947 | goto out_fail; |
6866 | 6948 | ||
6867 | if (unlikely(old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 6949 | if (unlikely(old_ino == BTRFS_FIRST_FREE_OBJECTID)) { |
6868 | /* force full log commit if subvolume involved. */ | 6950 | /* force full log commit if subvolume involved. */ |
6869 | root->fs_info->last_trans_log_full_commit = trans->transid; | 6951 | root->fs_info->last_trans_log_full_commit = trans->transid; |
6870 | } else { | 6952 | } else { |
6871 | ret = btrfs_insert_inode_ref(trans, dest, | 6953 | ret = btrfs_insert_inode_ref(trans, dest, |
6872 | new_dentry->d_name.name, | 6954 | new_dentry->d_name.name, |
6873 | new_dentry->d_name.len, | 6955 | new_dentry->d_name.len, |
6874 | old_inode->i_ino, | 6956 | old_ino, |
6875 | new_dir->i_ino, index); | 6957 | btrfs_ino(new_dir), index); |
6876 | if (ret) | 6958 | if (ret) |
6877 | goto out_fail; | 6959 | goto out_fail; |
6878 | /* | 6960 | /* |
@@ -6888,10 +6970,8 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6888 | * make sure the inode gets flushed if it is replacing | 6970 | * make sure the inode gets flushed if it is replacing |
6889 | * something. | 6971 | * something. |
6890 | */ | 6972 | */ |
6891 | if (newinode && newinode->i_size && | 6973 | if (new_inode && new_inode->i_size && S_ISREG(old_inode->i_mode)) |
6892 | old_inode && S_ISREG(old_inode->i_mode)) { | ||
6893 | btrfs_add_ordered_operation(trans, root, old_inode); | 6974 | btrfs_add_ordered_operation(trans, root, old_inode); |
6894 | } | ||
6895 | 6975 | ||
6896 | old_dir->i_ctime = old_dir->i_mtime = ctime; | 6976 | old_dir->i_ctime = old_dir->i_mtime = ctime; |
6897 | new_dir->i_ctime = new_dir->i_mtime = ctime; | 6977 | new_dir->i_ctime = new_dir->i_mtime = ctime; |
@@ -6900,7 +6980,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6900 | if (old_dentry->d_parent != new_dentry->d_parent) | 6980 | if (old_dentry->d_parent != new_dentry->d_parent) |
6901 | btrfs_record_unlink_dir(trans, old_dir, old_inode, 1); | 6981 | btrfs_record_unlink_dir(trans, old_dir, old_inode, 1); |
6902 | 6982 | ||
6903 | if (unlikely(old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 6983 | if (unlikely(old_ino == BTRFS_FIRST_FREE_OBJECTID)) { |
6904 | root_objectid = BTRFS_I(old_inode)->root->root_key.objectid; | 6984 | root_objectid = BTRFS_I(old_inode)->root->root_key.objectid; |
6905 | ret = btrfs_unlink_subvol(trans, root, old_dir, root_objectid, | 6985 | ret = btrfs_unlink_subvol(trans, root, old_dir, root_objectid, |
6906 | old_dentry->d_name.name, | 6986 | old_dentry->d_name.name, |
@@ -6915,16 +6995,16 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6915 | } | 6995 | } |
6916 | BUG_ON(ret); | 6996 | BUG_ON(ret); |
6917 | 6997 | ||
6918 | if (newinode) { | 6998 | if (new_inode) { |
6919 | newinode->i_ctime = CURRENT_TIME; | 6999 | new_inode->i_ctime = CURRENT_TIME; |
6920 | if (unlikely(newinode->i_ino == | 7000 | if (unlikely(btrfs_ino(new_inode) == |
6921 | BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { | 7001 | BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { |
6922 | root_objectid = BTRFS_I(newinode)->location.objectid; | 7002 | root_objectid = BTRFS_I(new_inode)->location.objectid; |
6923 | ret = btrfs_unlink_subvol(trans, dest, new_dir, | 7003 | ret = btrfs_unlink_subvol(trans, dest, new_dir, |
6924 | root_objectid, | 7004 | root_objectid, |
6925 | new_dentry->d_name.name, | 7005 | new_dentry->d_name.name, |
6926 | new_dentry->d_name.len); | 7006 | new_dentry->d_name.len); |
6927 | BUG_ON(newinode->i_nlink == 0); | 7007 | BUG_ON(new_inode->i_nlink == 0); |
6928 | } else { | 7008 | } else { |
6929 | ret = btrfs_unlink_inode(trans, dest, new_dir, | 7009 | ret = btrfs_unlink_inode(trans, dest, new_dir, |
6930 | new_dentry->d_inode, | 7010 | new_dentry->d_inode, |
@@ -6932,7 +7012,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6932 | new_dentry->d_name.len); | 7012 | new_dentry->d_name.len); |
6933 | } | 7013 | } |
6934 | BUG_ON(ret); | 7014 | BUG_ON(ret); |
6935 | if (newinode->i_nlink == 0) { | 7015 | if (new_inode->i_nlink == 0) { |
6936 | ret = btrfs_orphan_add(trans, new_dentry->d_inode); | 7016 | ret = btrfs_orphan_add(trans, new_dentry->d_inode); |
6937 | BUG_ON(ret); | 7017 | BUG_ON(ret); |
6938 | } | 7018 | } |
@@ -6945,7 +7025,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6945 | new_dentry->d_name.len, 0, index); | 7025 | new_dentry->d_name.len, 0, index); |
6946 | BUG_ON(ret); | 7026 | BUG_ON(ret); |
6947 | 7027 | ||
6948 | if (old_inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) { | 7028 | if (old_ino != BTRFS_FIRST_FREE_OBJECTID) { |
6949 | struct dentry *parent = dget_parent(new_dentry); | 7029 | struct dentry *parent = dget_parent(new_dentry); |
6950 | btrfs_log_new_name(trans, old_inode, old_dir, parent); | 7030 | btrfs_log_new_name(trans, old_inode, old_dir, parent); |
6951 | dput(parent); | 7031 | dput(parent); |
@@ -6954,7 +7034,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6954 | out_fail: | 7034 | out_fail: |
6955 | btrfs_end_transaction_throttle(trans, root); | 7035 | btrfs_end_transaction_throttle(trans, root); |
6956 | out_notrans: | 7036 | out_notrans: |
6957 | if (old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) | 7037 | if (old_ino == BTRFS_FIRST_FREE_OBJECTID) |
6958 | up_read(&root->fs_info->subvol_sem); | 7038 | up_read(&root->fs_info->subvol_sem); |
6959 | 7039 | ||
6960 | return ret; | 7040 | return ret; |
@@ -7031,9 +7111,6 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
7031 | if (name_len > BTRFS_MAX_INLINE_DATA_SIZE(root)) | 7111 | if (name_len > BTRFS_MAX_INLINE_DATA_SIZE(root)) |
7032 | return -ENAMETOOLONG; | 7112 | return -ENAMETOOLONG; |
7033 | 7113 | ||
7034 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
7035 | if (err) | ||
7036 | return err; | ||
7037 | /* | 7114 | /* |
7038 | * 2 items for inode item and ref | 7115 | * 2 items for inode item and ref |
7039 | * 2 items for dir items | 7116 | * 2 items for dir items |
@@ -7045,8 +7122,12 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
7045 | 7122 | ||
7046 | btrfs_set_trans_block_group(trans, dir); | 7123 | btrfs_set_trans_block_group(trans, dir); |
7047 | 7124 | ||
7125 | err = btrfs_find_free_ino(root, &objectid); | ||
7126 | if (err) | ||
7127 | goto out_unlock; | ||
7128 | |||
7048 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 7129 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
7049 | dentry->d_name.len, dir->i_ino, objectid, | 7130 | dentry->d_name.len, btrfs_ino(dir), objectid, |
7050 | BTRFS_I(dir)->block_group, S_IFLNK|S_IRWXUGO, | 7131 | BTRFS_I(dir)->block_group, S_IFLNK|S_IRWXUGO, |
7051 | &index); | 7132 | &index); |
7052 | if (IS_ERR(inode)) { | 7133 | if (IS_ERR(inode)) { |
@@ -7078,7 +7159,7 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
7078 | 7159 | ||
7079 | path = btrfs_alloc_path(); | 7160 | path = btrfs_alloc_path(); |
7080 | BUG_ON(!path); | 7161 | BUG_ON(!path); |
7081 | key.objectid = inode->i_ino; | 7162 | key.objectid = btrfs_ino(inode); |
7082 | key.offset = 0; | 7163 | key.offset = 0; |
7083 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); | 7164 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); |
7084 | datasize = btrfs_file_extent_calc_inline_size(name_len); | 7165 | datasize = btrfs_file_extent_calc_inline_size(name_len); |
diff --git a/fs/btrfs/ioctl.c b/fs/btrfs/ioctl.c index d11fc6548e1..ed8c055ab70 100644 --- a/fs/btrfs/ioctl.c +++ b/fs/btrfs/ioctl.c | |||
@@ -50,6 +50,7 @@ | |||
50 | #include "print-tree.h" | 50 | #include "print-tree.h" |
51 | #include "volumes.h" | 51 | #include "volumes.h" |
52 | #include "locking.h" | 52 | #include "locking.h" |
53 | #include "inode-map.h" | ||
53 | 54 | ||
54 | /* Mask out flags that are inappropriate for the given type of inode. */ | 55 | /* Mask out flags that are inappropriate for the given type of inode. */ |
55 | static inline __u32 btrfs_mask_flags(umode_t mode, __u32 flags) | 56 | static inline __u32 btrfs_mask_flags(umode_t mode, __u32 flags) |
@@ -81,6 +82,13 @@ static unsigned int btrfs_flags_to_ioctl(unsigned int flags) | |||
81 | iflags |= FS_NOATIME_FL; | 82 | iflags |= FS_NOATIME_FL; |
82 | if (flags & BTRFS_INODE_DIRSYNC) | 83 | if (flags & BTRFS_INODE_DIRSYNC) |
83 | iflags |= FS_DIRSYNC_FL; | 84 | iflags |= FS_DIRSYNC_FL; |
85 | if (flags & BTRFS_INODE_NODATACOW) | ||
86 | iflags |= FS_NOCOW_FL; | ||
87 | |||
88 | if ((flags & BTRFS_INODE_COMPRESS) && !(flags & BTRFS_INODE_NOCOMPRESS)) | ||
89 | iflags |= FS_COMPR_FL; | ||
90 | else if (flags & BTRFS_INODE_NOCOMPRESS) | ||
91 | iflags |= FS_NOCOMP_FL; | ||
84 | 92 | ||
85 | return iflags; | 93 | return iflags; |
86 | } | 94 | } |
@@ -144,16 +152,13 @@ static int check_flags(unsigned int flags) | |||
144 | if (flags & ~(FS_IMMUTABLE_FL | FS_APPEND_FL | \ | 152 | if (flags & ~(FS_IMMUTABLE_FL | FS_APPEND_FL | \ |
145 | FS_NOATIME_FL | FS_NODUMP_FL | \ | 153 | FS_NOATIME_FL | FS_NODUMP_FL | \ |
146 | FS_SYNC_FL | FS_DIRSYNC_FL | \ | 154 | FS_SYNC_FL | FS_DIRSYNC_FL | \ |
147 | FS_NOCOMP_FL | FS_COMPR_FL | \ | 155 | FS_NOCOMP_FL | FS_COMPR_FL | |
148 | FS_NOCOW_FL | FS_COW_FL)) | 156 | FS_NOCOW_FL)) |
149 | return -EOPNOTSUPP; | 157 | return -EOPNOTSUPP; |
150 | 158 | ||
151 | if ((flags & FS_NOCOMP_FL) && (flags & FS_COMPR_FL)) | 159 | if ((flags & FS_NOCOMP_FL) && (flags & FS_COMPR_FL)) |
152 | return -EINVAL; | 160 | return -EINVAL; |
153 | 161 | ||
154 | if ((flags & FS_NOCOW_FL) && (flags & FS_COW_FL)) | ||
155 | return -EINVAL; | ||
156 | |||
157 | return 0; | 162 | return 0; |
158 | } | 163 | } |
159 | 164 | ||
@@ -218,6 +223,10 @@ static int btrfs_ioctl_setflags(struct file *file, void __user *arg) | |||
218 | ip->flags |= BTRFS_INODE_DIRSYNC; | 223 | ip->flags |= BTRFS_INODE_DIRSYNC; |
219 | else | 224 | else |
220 | ip->flags &= ~BTRFS_INODE_DIRSYNC; | 225 | ip->flags &= ~BTRFS_INODE_DIRSYNC; |
226 | if (flags & FS_NOCOW_FL) | ||
227 | ip->flags |= BTRFS_INODE_NODATACOW; | ||
228 | else | ||
229 | ip->flags &= ~BTRFS_INODE_NODATACOW; | ||
221 | 230 | ||
222 | /* | 231 | /* |
223 | * The COMPRESS flag can only be changed by users, while the NOCOMPRESS | 232 | * The COMPRESS flag can only be changed by users, while the NOCOMPRESS |
@@ -230,11 +239,9 @@ static int btrfs_ioctl_setflags(struct file *file, void __user *arg) | |||
230 | } else if (flags & FS_COMPR_FL) { | 239 | } else if (flags & FS_COMPR_FL) { |
231 | ip->flags |= BTRFS_INODE_COMPRESS; | 240 | ip->flags |= BTRFS_INODE_COMPRESS; |
232 | ip->flags &= ~BTRFS_INODE_NOCOMPRESS; | 241 | ip->flags &= ~BTRFS_INODE_NOCOMPRESS; |
242 | } else { | ||
243 | ip->flags &= ~(BTRFS_INODE_COMPRESS | BTRFS_INODE_NOCOMPRESS); | ||
233 | } | 244 | } |
234 | if (flags & FS_NOCOW_FL) | ||
235 | ip->flags |= BTRFS_INODE_NODATACOW; | ||
236 | else if (flags & FS_COW_FL) | ||
237 | ip->flags &= ~BTRFS_INODE_NODATACOW; | ||
238 | 245 | ||
239 | trans = btrfs_join_transaction(root, 1); | 246 | trans = btrfs_join_transaction(root, 1); |
240 | BUG_ON(IS_ERR(trans)); | 247 | BUG_ON(IS_ERR(trans)); |
@@ -323,8 +330,7 @@ static noinline int create_subvol(struct btrfs_root *root, | |||
323 | u64 new_dirid = BTRFS_FIRST_FREE_OBJECTID; | 330 | u64 new_dirid = BTRFS_FIRST_FREE_OBJECTID; |
324 | u64 index = 0; | 331 | u64 index = 0; |
325 | 332 | ||
326 | ret = btrfs_find_free_objectid(NULL, root->fs_info->tree_root, | 333 | ret = btrfs_find_free_objectid(root->fs_info->tree_root, &objectid); |
327 | 0, &objectid); | ||
328 | if (ret) { | 334 | if (ret) { |
329 | dput(parent); | 335 | dput(parent); |
330 | return ret; | 336 | return ret; |
@@ -416,7 +422,7 @@ static noinline int create_subvol(struct btrfs_root *root, | |||
416 | BUG_ON(ret); | 422 | BUG_ON(ret); |
417 | 423 | ||
418 | ret = btrfs_insert_dir_item(trans, root, | 424 | ret = btrfs_insert_dir_item(trans, root, |
419 | name, namelen, dir->i_ino, &key, | 425 | name, namelen, dir, &key, |
420 | BTRFS_FT_DIR, index); | 426 | BTRFS_FT_DIR, index); |
421 | if (ret) | 427 | if (ret) |
422 | goto fail; | 428 | goto fail; |
@@ -427,7 +433,7 @@ static noinline int create_subvol(struct btrfs_root *root, | |||
427 | 433 | ||
428 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, | 434 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, |
429 | objectid, root->root_key.objectid, | 435 | objectid, root->root_key.objectid, |
430 | dir->i_ino, index, name, namelen); | 436 | btrfs_ino(dir), index, name, namelen); |
431 | 437 | ||
432 | BUG_ON(ret); | 438 | BUG_ON(ret); |
433 | 439 | ||
@@ -1123,7 +1129,7 @@ static noinline int btrfs_ioctl_subvol_getflags(struct file *file, | |||
1123 | int ret = 0; | 1129 | int ret = 0; |
1124 | u64 flags = 0; | 1130 | u64 flags = 0; |
1125 | 1131 | ||
1126 | if (inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) | 1132 | if (btrfs_ino(inode) != BTRFS_FIRST_FREE_OBJECTID) |
1127 | return -EINVAL; | 1133 | return -EINVAL; |
1128 | 1134 | ||
1129 | down_read(&root->fs_info->subvol_sem); | 1135 | down_read(&root->fs_info->subvol_sem); |
@@ -1150,7 +1156,7 @@ static noinline int btrfs_ioctl_subvol_setflags(struct file *file, | |||
1150 | if (root->fs_info->sb->s_flags & MS_RDONLY) | 1156 | if (root->fs_info->sb->s_flags & MS_RDONLY) |
1151 | return -EROFS; | 1157 | return -EROFS; |
1152 | 1158 | ||
1153 | if (inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) | 1159 | if (btrfs_ino(inode) != BTRFS_FIRST_FREE_OBJECTID) |
1154 | return -EINVAL; | 1160 | return -EINVAL; |
1155 | 1161 | ||
1156 | if (copy_from_user(&flags, arg, sizeof(flags))) | 1162 | if (copy_from_user(&flags, arg, sizeof(flags))) |
@@ -1633,7 +1639,7 @@ static noinline int btrfs_ioctl_snap_destroy(struct file *file, | |||
1633 | goto out_dput; | 1639 | goto out_dput; |
1634 | } | 1640 | } |
1635 | 1641 | ||
1636 | if (inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) { | 1642 | if (btrfs_ino(inode) != BTRFS_FIRST_FREE_OBJECTID) { |
1637 | err = -EINVAL; | 1643 | err = -EINVAL; |
1638 | goto out_dput; | 1644 | goto out_dput; |
1639 | } | 1645 | } |
@@ -1919,7 +1925,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
1919 | } | 1925 | } |
1920 | 1926 | ||
1921 | /* clone data */ | 1927 | /* clone data */ |
1922 | key.objectid = src->i_ino; | 1928 | key.objectid = btrfs_ino(src); |
1923 | key.type = BTRFS_EXTENT_DATA_KEY; | 1929 | key.type = BTRFS_EXTENT_DATA_KEY; |
1924 | key.offset = 0; | 1930 | key.offset = 0; |
1925 | 1931 | ||
@@ -1946,7 +1952,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
1946 | 1952 | ||
1947 | btrfs_item_key_to_cpu(leaf, &key, slot); | 1953 | btrfs_item_key_to_cpu(leaf, &key, slot); |
1948 | if (btrfs_key_type(&key) > BTRFS_EXTENT_DATA_KEY || | 1954 | if (btrfs_key_type(&key) > BTRFS_EXTENT_DATA_KEY || |
1949 | key.objectid != src->i_ino) | 1955 | key.objectid != btrfs_ino(src)) |
1950 | break; | 1956 | break; |
1951 | 1957 | ||
1952 | if (btrfs_key_type(&key) == BTRFS_EXTENT_DATA_KEY) { | 1958 | if (btrfs_key_type(&key) == BTRFS_EXTENT_DATA_KEY) { |
@@ -1989,7 +1995,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
1989 | goto next; | 1995 | goto next; |
1990 | 1996 | ||
1991 | memcpy(&new_key, &key, sizeof(new_key)); | 1997 | memcpy(&new_key, &key, sizeof(new_key)); |
1992 | new_key.objectid = inode->i_ino; | 1998 | new_key.objectid = btrfs_ino(inode); |
1993 | if (off <= key.offset) | 1999 | if (off <= key.offset) |
1994 | new_key.offset = key.offset + destoff - off; | 2000 | new_key.offset = key.offset + destoff - off; |
1995 | else | 2001 | else |
@@ -2043,7 +2049,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
2043 | ret = btrfs_inc_extent_ref(trans, root, | 2049 | ret = btrfs_inc_extent_ref(trans, root, |
2044 | disko, diskl, 0, | 2050 | disko, diskl, 0, |
2045 | root->root_key.objectid, | 2051 | root->root_key.objectid, |
2046 | inode->i_ino, | 2052 | btrfs_ino(inode), |
2047 | new_key.offset - datao); | 2053 | new_key.offset - datao); |
2048 | BUG_ON(ret); | 2054 | BUG_ON(ret); |
2049 | } | 2055 | } |
diff --git a/fs/btrfs/relocation.c b/fs/btrfs/relocation.c index f726e72dd36..051992c7fcc 100644 --- a/fs/btrfs/relocation.c +++ b/fs/btrfs/relocation.c | |||
@@ -30,6 +30,7 @@ | |||
30 | #include "btrfs_inode.h" | 30 | #include "btrfs_inode.h" |
31 | #include "async-thread.h" | 31 | #include "async-thread.h" |
32 | #include "free-space-cache.h" | 32 | #include "free-space-cache.h" |
33 | #include "inode-map.h" | ||
33 | 34 | ||
34 | /* | 35 | /* |
35 | * backref_node, mapping_node and tree_block start with this | 36 | * backref_node, mapping_node and tree_block start with this |
@@ -1409,9 +1410,9 @@ again: | |||
1409 | prev = node; | 1410 | prev = node; |
1410 | entry = rb_entry(node, struct btrfs_inode, rb_node); | 1411 | entry = rb_entry(node, struct btrfs_inode, rb_node); |
1411 | 1412 | ||
1412 | if (objectid < entry->vfs_inode.i_ino) | 1413 | if (objectid < btrfs_ino(&entry->vfs_inode)) |
1413 | node = node->rb_left; | 1414 | node = node->rb_left; |
1414 | else if (objectid > entry->vfs_inode.i_ino) | 1415 | else if (objectid > btrfs_ino(&entry->vfs_inode)) |
1415 | node = node->rb_right; | 1416 | node = node->rb_right; |
1416 | else | 1417 | else |
1417 | break; | 1418 | break; |
@@ -1419,7 +1420,7 @@ again: | |||
1419 | if (!node) { | 1420 | if (!node) { |
1420 | while (prev) { | 1421 | while (prev) { |
1421 | entry = rb_entry(prev, struct btrfs_inode, rb_node); | 1422 | entry = rb_entry(prev, struct btrfs_inode, rb_node); |
1422 | if (objectid <= entry->vfs_inode.i_ino) { | 1423 | if (objectid <= btrfs_ino(&entry->vfs_inode)) { |
1423 | node = prev; | 1424 | node = prev; |
1424 | break; | 1425 | break; |
1425 | } | 1426 | } |
@@ -1434,7 +1435,7 @@ again: | |||
1434 | return inode; | 1435 | return inode; |
1435 | } | 1436 | } |
1436 | 1437 | ||
1437 | objectid = entry->vfs_inode.i_ino + 1; | 1438 | objectid = btrfs_ino(&entry->vfs_inode) + 1; |
1438 | if (cond_resched_lock(&root->inode_lock)) | 1439 | if (cond_resched_lock(&root->inode_lock)) |
1439 | goto again; | 1440 | goto again; |
1440 | 1441 | ||
@@ -1470,7 +1471,7 @@ static int get_new_location(struct inode *reloc_inode, u64 *new_bytenr, | |||
1470 | return -ENOMEM; | 1471 | return -ENOMEM; |
1471 | 1472 | ||
1472 | bytenr -= BTRFS_I(reloc_inode)->index_cnt; | 1473 | bytenr -= BTRFS_I(reloc_inode)->index_cnt; |
1473 | ret = btrfs_lookup_file_extent(NULL, root, path, reloc_inode->i_ino, | 1474 | ret = btrfs_lookup_file_extent(NULL, root, path, btrfs_ino(reloc_inode), |
1474 | bytenr, 0); | 1475 | bytenr, 0); |
1475 | if (ret < 0) | 1476 | if (ret < 0) |
1476 | goto out; | 1477 | goto out; |
@@ -1558,11 +1559,11 @@ int replace_file_extents(struct btrfs_trans_handle *trans, | |||
1558 | if (first) { | 1559 | if (first) { |
1559 | inode = find_next_inode(root, key.objectid); | 1560 | inode = find_next_inode(root, key.objectid); |
1560 | first = 0; | 1561 | first = 0; |
1561 | } else if (inode && inode->i_ino < key.objectid) { | 1562 | } else if (inode && btrfs_ino(inode) < key.objectid) { |
1562 | btrfs_add_delayed_iput(inode); | 1563 | btrfs_add_delayed_iput(inode); |
1563 | inode = find_next_inode(root, key.objectid); | 1564 | inode = find_next_inode(root, key.objectid); |
1564 | } | 1565 | } |
1565 | if (inode && inode->i_ino == key.objectid) { | 1566 | if (inode && btrfs_ino(inode) == key.objectid) { |
1566 | end = key.offset + | 1567 | end = key.offset + |
1567 | btrfs_file_extent_num_bytes(leaf, fi); | 1568 | btrfs_file_extent_num_bytes(leaf, fi); |
1568 | WARN_ON(!IS_ALIGNED(key.offset, | 1569 | WARN_ON(!IS_ALIGNED(key.offset, |
@@ -1893,6 +1894,7 @@ static int invalidate_extent_cache(struct btrfs_root *root, | |||
1893 | struct inode *inode = NULL; | 1894 | struct inode *inode = NULL; |
1894 | u64 objectid; | 1895 | u64 objectid; |
1895 | u64 start, end; | 1896 | u64 start, end; |
1897 | u64 ino; | ||
1896 | 1898 | ||
1897 | objectid = min_key->objectid; | 1899 | objectid = min_key->objectid; |
1898 | while (1) { | 1900 | while (1) { |
@@ -1905,17 +1907,18 @@ static int invalidate_extent_cache(struct btrfs_root *root, | |||
1905 | inode = find_next_inode(root, objectid); | 1907 | inode = find_next_inode(root, objectid); |
1906 | if (!inode) | 1908 | if (!inode) |
1907 | break; | 1909 | break; |
1910 | ino = btrfs_ino(inode); | ||
1908 | 1911 | ||
1909 | if (inode->i_ino > max_key->objectid) { | 1912 | if (ino > max_key->objectid) { |
1910 | iput(inode); | 1913 | iput(inode); |
1911 | break; | 1914 | break; |
1912 | } | 1915 | } |
1913 | 1916 | ||
1914 | objectid = inode->i_ino + 1; | 1917 | objectid = ino + 1; |
1915 | if (!S_ISREG(inode->i_mode)) | 1918 | if (!S_ISREG(inode->i_mode)) |
1916 | continue; | 1919 | continue; |
1917 | 1920 | ||
1918 | if (unlikely(min_key->objectid == inode->i_ino)) { | 1921 | if (unlikely(min_key->objectid == ino)) { |
1919 | if (min_key->type > BTRFS_EXTENT_DATA_KEY) | 1922 | if (min_key->type > BTRFS_EXTENT_DATA_KEY) |
1920 | continue; | 1923 | continue; |
1921 | if (min_key->type < BTRFS_EXTENT_DATA_KEY) | 1924 | if (min_key->type < BTRFS_EXTENT_DATA_KEY) |
@@ -1928,7 +1931,7 @@ static int invalidate_extent_cache(struct btrfs_root *root, | |||
1928 | start = 0; | 1931 | start = 0; |
1929 | } | 1932 | } |
1930 | 1933 | ||
1931 | if (unlikely(max_key->objectid == inode->i_ino)) { | 1934 | if (unlikely(max_key->objectid == ino)) { |
1932 | if (max_key->type < BTRFS_EXTENT_DATA_KEY) | 1935 | if (max_key->type < BTRFS_EXTENT_DATA_KEY) |
1933 | continue; | 1936 | continue; |
1934 | if (max_key->type > BTRFS_EXTENT_DATA_KEY) { | 1937 | if (max_key->type > BTRFS_EXTENT_DATA_KEY) { |
@@ -3897,7 +3900,7 @@ struct inode *create_reloc_inode(struct btrfs_fs_info *fs_info, | |||
3897 | if (IS_ERR(trans)) | 3900 | if (IS_ERR(trans)) |
3898 | return ERR_CAST(trans); | 3901 | return ERR_CAST(trans); |
3899 | 3902 | ||
3900 | err = btrfs_find_free_objectid(trans, root, objectid, &objectid); | 3903 | err = btrfs_find_free_objectid(root, &objectid); |
3901 | if (err) | 3904 | if (err) |
3902 | goto out; | 3905 | goto out; |
3903 | 3906 | ||
diff --git a/fs/btrfs/super.c b/fs/btrfs/super.c index 3e28521643f..fb72e2bea88 100644 --- a/fs/btrfs/super.c +++ b/fs/btrfs/super.c | |||
@@ -40,6 +40,7 @@ | |||
40 | #include <linux/magic.h> | 40 | #include <linux/magic.h> |
41 | #include <linux/slab.h> | 41 | #include <linux/slab.h> |
42 | #include "compat.h" | 42 | #include "compat.h" |
43 | #include "delayed-inode.h" | ||
43 | #include "ctree.h" | 44 | #include "ctree.h" |
44 | #include "disk-io.h" | 45 | #include "disk-io.h" |
45 | #include "transaction.h" | 46 | #include "transaction.h" |
@@ -1206,10 +1207,14 @@ static int __init init_btrfs_fs(void) | |||
1206 | if (err) | 1207 | if (err) |
1207 | goto free_extent_io; | 1208 | goto free_extent_io; |
1208 | 1209 | ||
1209 | err = btrfs_interface_init(); | 1210 | err = btrfs_delayed_inode_init(); |
1210 | if (err) | 1211 | if (err) |
1211 | goto free_extent_map; | 1212 | goto free_extent_map; |
1212 | 1213 | ||
1214 | err = btrfs_interface_init(); | ||
1215 | if (err) | ||
1216 | goto free_delayed_inode; | ||
1217 | |||
1213 | err = register_filesystem(&btrfs_fs_type); | 1218 | err = register_filesystem(&btrfs_fs_type); |
1214 | if (err) | 1219 | if (err) |
1215 | goto unregister_ioctl; | 1220 | goto unregister_ioctl; |
@@ -1219,6 +1224,8 @@ static int __init init_btrfs_fs(void) | |||
1219 | 1224 | ||
1220 | unregister_ioctl: | 1225 | unregister_ioctl: |
1221 | btrfs_interface_exit(); | 1226 | btrfs_interface_exit(); |
1227 | free_delayed_inode: | ||
1228 | btrfs_delayed_inode_exit(); | ||
1222 | free_extent_map: | 1229 | free_extent_map: |
1223 | extent_map_exit(); | 1230 | extent_map_exit(); |
1224 | free_extent_io: | 1231 | free_extent_io: |
@@ -1235,6 +1242,7 @@ free_sysfs: | |||
1235 | static void __exit exit_btrfs_fs(void) | 1242 | static void __exit exit_btrfs_fs(void) |
1236 | { | 1243 | { |
1237 | btrfs_destroy_cachep(); | 1244 | btrfs_destroy_cachep(); |
1245 | btrfs_delayed_inode_exit(); | ||
1238 | extent_map_exit(); | 1246 | extent_map_exit(); |
1239 | extent_io_exit(); | 1247 | extent_io_exit(); |
1240 | btrfs_interface_exit(); | 1248 | btrfs_interface_exit(); |
diff --git a/fs/btrfs/sysfs.c b/fs/btrfs/sysfs.c index ab9633fd72a..c3c223ae669 100644 --- a/fs/btrfs/sysfs.c +++ b/fs/btrfs/sysfs.c | |||
@@ -174,18 +174,6 @@ static const struct sysfs_ops btrfs_root_attr_ops = { | |||
174 | .store = btrfs_root_attr_store, | 174 | .store = btrfs_root_attr_store, |
175 | }; | 175 | }; |
176 | 176 | ||
177 | static struct kobj_type btrfs_root_ktype = { | ||
178 | .default_attrs = btrfs_root_attrs, | ||
179 | .sysfs_ops = &btrfs_root_attr_ops, | ||
180 | .release = btrfs_root_release, | ||
181 | }; | ||
182 | |||
183 | static struct kobj_type btrfs_super_ktype = { | ||
184 | .default_attrs = btrfs_super_attrs, | ||
185 | .sysfs_ops = &btrfs_super_attr_ops, | ||
186 | .release = btrfs_super_release, | ||
187 | }; | ||
188 | |||
189 | /* /sys/fs/btrfs/ entry */ | 177 | /* /sys/fs/btrfs/ entry */ |
190 | static struct kset *btrfs_kset; | 178 | static struct kset *btrfs_kset; |
191 | 179 | ||
diff --git a/fs/btrfs/transaction.c b/fs/btrfs/transaction.c index 211aceeb9ea..33679fc710c 100644 --- a/fs/btrfs/transaction.c +++ b/fs/btrfs/transaction.c | |||
@@ -27,6 +27,7 @@ | |||
27 | #include "transaction.h" | 27 | #include "transaction.h" |
28 | #include "locking.h" | 28 | #include "locking.h" |
29 | #include "tree-log.h" | 29 | #include "tree-log.h" |
30 | #include "inode-map.h" | ||
30 | 31 | ||
31 | #define BTRFS_ROOT_TRANS_TAG 0 | 32 | #define BTRFS_ROOT_TRANS_TAG 0 |
32 | 33 | ||
@@ -443,19 +444,40 @@ static int __btrfs_end_transaction(struct btrfs_trans_handle *trans, | |||
443 | int btrfs_end_transaction(struct btrfs_trans_handle *trans, | 444 | int btrfs_end_transaction(struct btrfs_trans_handle *trans, |
444 | struct btrfs_root *root) | 445 | struct btrfs_root *root) |
445 | { | 446 | { |
446 | return __btrfs_end_transaction(trans, root, 0, 1); | 447 | int ret; |
448 | |||
449 | ret = __btrfs_end_transaction(trans, root, 0, 1); | ||
450 | if (ret) | ||
451 | return ret; | ||
452 | return 0; | ||
447 | } | 453 | } |
448 | 454 | ||
449 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, | 455 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, |
450 | struct btrfs_root *root) | 456 | struct btrfs_root *root) |
451 | { | 457 | { |
452 | return __btrfs_end_transaction(trans, root, 1, 1); | 458 | int ret; |
459 | |||
460 | ret = __btrfs_end_transaction(trans, root, 1, 1); | ||
461 | if (ret) | ||
462 | return ret; | ||
463 | return 0; | ||
453 | } | 464 | } |
454 | 465 | ||
455 | int btrfs_end_transaction_nolock(struct btrfs_trans_handle *trans, | 466 | int btrfs_end_transaction_nolock(struct btrfs_trans_handle *trans, |
456 | struct btrfs_root *root) | 467 | struct btrfs_root *root) |
457 | { | 468 | { |
458 | return __btrfs_end_transaction(trans, root, 0, 0); | 469 | int ret; |
470 | |||
471 | ret = __btrfs_end_transaction(trans, root, 0, 0); | ||
472 | if (ret) | ||
473 | return ret; | ||
474 | return 0; | ||
475 | } | ||
476 | |||
477 | int btrfs_end_transaction_dmeta(struct btrfs_trans_handle *trans, | ||
478 | struct btrfs_root *root) | ||
479 | { | ||
480 | return __btrfs_end_transaction(trans, root, 1, 1); | ||
459 | } | 481 | } |
460 | 482 | ||
461 | /* | 483 | /* |
@@ -716,8 +738,14 @@ static noinline int commit_fs_roots(struct btrfs_trans_handle *trans, | |||
716 | btrfs_update_reloc_root(trans, root); | 738 | btrfs_update_reloc_root(trans, root); |
717 | btrfs_orphan_commit_root(trans, root); | 739 | btrfs_orphan_commit_root(trans, root); |
718 | 740 | ||
741 | btrfs_save_ino_cache(root, trans); | ||
742 | |||
719 | if (root->commit_root != root->node) { | 743 | if (root->commit_root != root->node) { |
744 | mutex_lock(&root->fs_commit_mutex); | ||
720 | switch_commit_root(root); | 745 | switch_commit_root(root); |
746 | btrfs_unpin_free_ino(root); | ||
747 | mutex_unlock(&root->fs_commit_mutex); | ||
748 | |||
721 | btrfs_set_root_node(&root->root_item, | 749 | btrfs_set_root_node(&root->root_item, |
722 | root->node); | 750 | root->node); |
723 | } | 751 | } |
@@ -795,7 +823,7 @@ static noinline int create_pending_snapshot(struct btrfs_trans_handle *trans, | |||
795 | goto fail; | 823 | goto fail; |
796 | } | 824 | } |
797 | 825 | ||
798 | ret = btrfs_find_free_objectid(trans, tree_root, 0, &objectid); | 826 | ret = btrfs_find_free_objectid(tree_root, &objectid); |
799 | if (ret) { | 827 | if (ret) { |
800 | pending->error = ret; | 828 | pending->error = ret; |
801 | goto fail; | 829 | goto fail; |
@@ -832,7 +860,7 @@ static noinline int create_pending_snapshot(struct btrfs_trans_handle *trans, | |||
832 | BUG_ON(ret); | 860 | BUG_ON(ret); |
833 | ret = btrfs_insert_dir_item(trans, parent_root, | 861 | ret = btrfs_insert_dir_item(trans, parent_root, |
834 | dentry->d_name.name, dentry->d_name.len, | 862 | dentry->d_name.name, dentry->d_name.len, |
835 | parent_inode->i_ino, &key, | 863 | parent_inode, &key, |
836 | BTRFS_FT_DIR, index); | 864 | BTRFS_FT_DIR, index); |
837 | BUG_ON(ret); | 865 | BUG_ON(ret); |
838 | 866 | ||
@@ -874,7 +902,7 @@ static noinline int create_pending_snapshot(struct btrfs_trans_handle *trans, | |||
874 | */ | 902 | */ |
875 | ret = btrfs_add_root_ref(trans, tree_root, objectid, | 903 | ret = btrfs_add_root_ref(trans, tree_root, objectid, |
876 | parent_root->root_key.objectid, | 904 | parent_root->root_key.objectid, |
877 | parent_inode->i_ino, index, | 905 | btrfs_ino(parent_inode), index, |
878 | dentry->d_name.name, dentry->d_name.len); | 906 | dentry->d_name.name, dentry->d_name.len); |
879 | BUG_ON(ret); | 907 | BUG_ON(ret); |
880 | dput(parent); | 908 | dput(parent); |
@@ -902,6 +930,14 @@ static noinline int create_pending_snapshots(struct btrfs_trans_handle *trans, | |||
902 | int ret; | 930 | int ret; |
903 | 931 | ||
904 | list_for_each_entry(pending, head, list) { | 932 | list_for_each_entry(pending, head, list) { |
933 | /* | ||
934 | * We must deal with the delayed items before creating | ||
935 | * snapshots, or we will create a snapthot with inconsistent | ||
936 | * information. | ||
937 | */ | ||
938 | ret = btrfs_run_delayed_items(trans, fs_info->fs_root); | ||
939 | BUG_ON(ret); | ||
940 | |||
905 | ret = create_pending_snapshot(trans, fs_info, pending); | 941 | ret = create_pending_snapshot(trans, fs_info, pending); |
906 | BUG_ON(ret); | 942 | BUG_ON(ret); |
907 | } | 943 | } |
@@ -1155,6 +1191,9 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, | |||
1155 | BUG_ON(ret); | 1191 | BUG_ON(ret); |
1156 | } | 1192 | } |
1157 | 1193 | ||
1194 | ret = btrfs_run_delayed_items(trans, root); | ||
1195 | BUG_ON(ret); | ||
1196 | |||
1158 | /* | 1197 | /* |
1159 | * rename don't use btrfs_join_transaction, so, once we | 1198 | * rename don't use btrfs_join_transaction, so, once we |
1160 | * set the transaction to blocked above, we aren't going | 1199 | * set the transaction to blocked above, we aren't going |
@@ -1181,6 +1220,9 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, | |||
1181 | ret = create_pending_snapshots(trans, root->fs_info); | 1220 | ret = create_pending_snapshots(trans, root->fs_info); |
1182 | BUG_ON(ret); | 1221 | BUG_ON(ret); |
1183 | 1222 | ||
1223 | ret = btrfs_run_delayed_items(trans, root); | ||
1224 | BUG_ON(ret); | ||
1225 | |||
1184 | ret = btrfs_run_delayed_refs(trans, root, (unsigned long)-1); | 1226 | ret = btrfs_run_delayed_refs(trans, root, (unsigned long)-1); |
1185 | BUG_ON(ret); | 1227 | BUG_ON(ret); |
1186 | 1228 | ||
@@ -1297,6 +1339,8 @@ int btrfs_clean_old_snapshots(struct btrfs_root *root) | |||
1297 | root = list_entry(list.next, struct btrfs_root, root_list); | 1339 | root = list_entry(list.next, struct btrfs_root, root_list); |
1298 | list_del(&root->root_list); | 1340 | list_del(&root->root_list); |
1299 | 1341 | ||
1342 | btrfs_kill_all_delayed_nodes(root); | ||
1343 | |||
1300 | if (btrfs_header_backref_rev(root->node) < | 1344 | if (btrfs_header_backref_rev(root->node) < |
1301 | BTRFS_MIXED_BACKREF_REV) | 1345 | BTRFS_MIXED_BACKREF_REV) |
1302 | btrfs_drop_snapshot(root, NULL, 0); | 1346 | btrfs_drop_snapshot(root, NULL, 0); |
diff --git a/fs/btrfs/transaction.h b/fs/btrfs/transaction.h index 000a41008c3..804c88639e5 100644 --- a/fs/btrfs/transaction.h +++ b/fs/btrfs/transaction.h | |||
@@ -112,6 +112,8 @@ int btrfs_commit_transaction_async(struct btrfs_trans_handle *trans, | |||
112 | int wait_for_unblock); | 112 | int wait_for_unblock); |
113 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, | 113 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, |
114 | struct btrfs_root *root); | 114 | struct btrfs_root *root); |
115 | int btrfs_end_transaction_dmeta(struct btrfs_trans_handle *trans, | ||
116 | struct btrfs_root *root); | ||
115 | int btrfs_should_end_transaction(struct btrfs_trans_handle *trans, | 117 | int btrfs_should_end_transaction(struct btrfs_trans_handle *trans, |
116 | struct btrfs_root *root); | 118 | struct btrfs_root *root); |
117 | void btrfs_throttle(struct btrfs_root *root); | 119 | void btrfs_throttle(struct btrfs_root *root); |
diff --git a/fs/btrfs/tree-log.c b/fs/btrfs/tree-log.c index c599e8c2a53..a794b9f6013 100644 --- a/fs/btrfs/tree-log.c +++ b/fs/btrfs/tree-log.c | |||
@@ -519,7 +519,7 @@ static noinline int replay_one_extent(struct btrfs_trans_handle *trans, | |||
519 | * file. This must be done before the btrfs_drop_extents run | 519 | * file. This must be done before the btrfs_drop_extents run |
520 | * so we don't try to drop this extent. | 520 | * so we don't try to drop this extent. |
521 | */ | 521 | */ |
522 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 522 | ret = btrfs_lookup_file_extent(trans, root, path, btrfs_ino(inode), |
523 | start, 0); | 523 | start, 0); |
524 | 524 | ||
525 | if (ret == 0 && | 525 | if (ret == 0 && |
@@ -832,7 +832,7 @@ again: | |||
832 | read_extent_buffer(eb, name, (unsigned long)(ref + 1), namelen); | 832 | read_extent_buffer(eb, name, (unsigned long)(ref + 1), namelen); |
833 | 833 | ||
834 | /* if we already have a perfect match, we're done */ | 834 | /* if we already have a perfect match, we're done */ |
835 | if (inode_in_dir(root, path, dir->i_ino, inode->i_ino, | 835 | if (inode_in_dir(root, path, btrfs_ino(dir), btrfs_ino(inode), |
836 | btrfs_inode_ref_index(eb, ref), | 836 | btrfs_inode_ref_index(eb, ref), |
837 | name, namelen)) { | 837 | name, namelen)) { |
838 | goto out; | 838 | goto out; |
@@ -960,8 +960,9 @@ static noinline int fixup_inode_link_count(struct btrfs_trans_handle *trans, | |||
960 | unsigned long ptr; | 960 | unsigned long ptr; |
961 | unsigned long ptr_end; | 961 | unsigned long ptr_end; |
962 | int name_len; | 962 | int name_len; |
963 | u64 ino = btrfs_ino(inode); | ||
963 | 964 | ||
964 | key.objectid = inode->i_ino; | 965 | key.objectid = ino; |
965 | key.type = BTRFS_INODE_REF_KEY; | 966 | key.type = BTRFS_INODE_REF_KEY; |
966 | key.offset = (u64)-1; | 967 | key.offset = (u64)-1; |
967 | 968 | ||
@@ -980,7 +981,7 @@ static noinline int fixup_inode_link_count(struct btrfs_trans_handle *trans, | |||
980 | } | 981 | } |
981 | btrfs_item_key_to_cpu(path->nodes[0], &key, | 982 | btrfs_item_key_to_cpu(path->nodes[0], &key, |
982 | path->slots[0]); | 983 | path->slots[0]); |
983 | if (key.objectid != inode->i_ino || | 984 | if (key.objectid != ino || |
984 | key.type != BTRFS_INODE_REF_KEY) | 985 | key.type != BTRFS_INODE_REF_KEY) |
985 | break; | 986 | break; |
986 | ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); | 987 | ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); |
@@ -1011,10 +1012,10 @@ static noinline int fixup_inode_link_count(struct btrfs_trans_handle *trans, | |||
1011 | if (inode->i_nlink == 0) { | 1012 | if (inode->i_nlink == 0) { |
1012 | if (S_ISDIR(inode->i_mode)) { | 1013 | if (S_ISDIR(inode->i_mode)) { |
1013 | ret = replay_dir_deletes(trans, root, NULL, path, | 1014 | ret = replay_dir_deletes(trans, root, NULL, path, |
1014 | inode->i_ino, 1); | 1015 | ino, 1); |
1015 | BUG_ON(ret); | 1016 | BUG_ON(ret); |
1016 | } | 1017 | } |
1017 | ret = insert_orphan_item(trans, root, inode->i_ino); | 1018 | ret = insert_orphan_item(trans, root, ino); |
1018 | BUG_ON(ret); | 1019 | BUG_ON(ret); |
1019 | } | 1020 | } |
1020 | btrfs_free_path(path); | 1021 | btrfs_free_path(path); |
@@ -2197,6 +2198,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2197 | int ret; | 2198 | int ret; |
2198 | int err = 0; | 2199 | int err = 0; |
2199 | int bytes_del = 0; | 2200 | int bytes_del = 0; |
2201 | u64 dir_ino = btrfs_ino(dir); | ||
2200 | 2202 | ||
2201 | if (BTRFS_I(dir)->logged_trans < trans->transid) | 2203 | if (BTRFS_I(dir)->logged_trans < trans->transid) |
2202 | return 0; | 2204 | return 0; |
@@ -2214,7 +2216,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2214 | goto out_unlock; | 2216 | goto out_unlock; |
2215 | } | 2217 | } |
2216 | 2218 | ||
2217 | di = btrfs_lookup_dir_item(trans, log, path, dir->i_ino, | 2219 | di = btrfs_lookup_dir_item(trans, log, path, dir_ino, |
2218 | name, name_len, -1); | 2220 | name, name_len, -1); |
2219 | if (IS_ERR(di)) { | 2221 | if (IS_ERR(di)) { |
2220 | err = PTR_ERR(di); | 2222 | err = PTR_ERR(di); |
@@ -2226,7 +2228,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2226 | BUG_ON(ret); | 2228 | BUG_ON(ret); |
2227 | } | 2229 | } |
2228 | btrfs_release_path(path); | 2230 | btrfs_release_path(path); |
2229 | di = btrfs_lookup_dir_index_item(trans, log, path, dir->i_ino, | 2231 | di = btrfs_lookup_dir_index_item(trans, log, path, dir_ino, |
2230 | index, name, name_len, -1); | 2232 | index, name, name_len, -1); |
2231 | if (IS_ERR(di)) { | 2233 | if (IS_ERR(di)) { |
2232 | err = PTR_ERR(di); | 2234 | err = PTR_ERR(di); |
@@ -2244,7 +2246,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2244 | if (bytes_del) { | 2246 | if (bytes_del) { |
2245 | struct btrfs_key key; | 2247 | struct btrfs_key key; |
2246 | 2248 | ||
2247 | key.objectid = dir->i_ino; | 2249 | key.objectid = dir_ino; |
2248 | key.offset = 0; | 2250 | key.offset = 0; |
2249 | key.type = BTRFS_INODE_ITEM_KEY; | 2251 | key.type = BTRFS_INODE_ITEM_KEY; |
2250 | btrfs_release_path(path); | 2252 | btrfs_release_path(path); |
@@ -2303,7 +2305,7 @@ int btrfs_del_inode_ref_in_log(struct btrfs_trans_handle *trans, | |||
2303 | log = root->log_root; | 2305 | log = root->log_root; |
2304 | mutex_lock(&BTRFS_I(inode)->log_mutex); | 2306 | mutex_lock(&BTRFS_I(inode)->log_mutex); |
2305 | 2307 | ||
2306 | ret = btrfs_del_inode_ref(trans, log, name, name_len, inode->i_ino, | 2308 | ret = btrfs_del_inode_ref(trans, log, name, name_len, btrfs_ino(inode), |
2307 | dirid, &index); | 2309 | dirid, &index); |
2308 | mutex_unlock(&BTRFS_I(inode)->log_mutex); | 2310 | mutex_unlock(&BTRFS_I(inode)->log_mutex); |
2309 | if (ret == -ENOSPC) { | 2311 | if (ret == -ENOSPC) { |
@@ -2369,13 +2371,14 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2369 | int nritems; | 2371 | int nritems; |
2370 | u64 first_offset = min_offset; | 2372 | u64 first_offset = min_offset; |
2371 | u64 last_offset = (u64)-1; | 2373 | u64 last_offset = (u64)-1; |
2374 | u64 ino = btrfs_ino(inode); | ||
2372 | 2375 | ||
2373 | log = root->log_root; | 2376 | log = root->log_root; |
2374 | max_key.objectid = inode->i_ino; | 2377 | max_key.objectid = ino; |
2375 | max_key.offset = (u64)-1; | 2378 | max_key.offset = (u64)-1; |
2376 | max_key.type = key_type; | 2379 | max_key.type = key_type; |
2377 | 2380 | ||
2378 | min_key.objectid = inode->i_ino; | 2381 | min_key.objectid = ino; |
2379 | min_key.type = key_type; | 2382 | min_key.type = key_type; |
2380 | min_key.offset = min_offset; | 2383 | min_key.offset = min_offset; |
2381 | 2384 | ||
@@ -2388,9 +2391,8 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2388 | * we didn't find anything from this transaction, see if there | 2391 | * we didn't find anything from this transaction, see if there |
2389 | * is anything at all | 2392 | * is anything at all |
2390 | */ | 2393 | */ |
2391 | if (ret != 0 || min_key.objectid != inode->i_ino || | 2394 | if (ret != 0 || min_key.objectid != ino || min_key.type != key_type) { |
2392 | min_key.type != key_type) { | 2395 | min_key.objectid = ino; |
2393 | min_key.objectid = inode->i_ino; | ||
2394 | min_key.type = key_type; | 2396 | min_key.type = key_type; |
2395 | min_key.offset = (u64)-1; | 2397 | min_key.offset = (u64)-1; |
2396 | btrfs_release_path(path); | 2398 | btrfs_release_path(path); |
@@ -2399,7 +2401,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2399 | btrfs_release_path(path); | 2401 | btrfs_release_path(path); |
2400 | return ret; | 2402 | return ret; |
2401 | } | 2403 | } |
2402 | ret = btrfs_previous_item(root, path, inode->i_ino, key_type); | 2404 | ret = btrfs_previous_item(root, path, ino, key_type); |
2403 | 2405 | ||
2404 | /* if ret == 0 there are items for this type, | 2406 | /* if ret == 0 there are items for this type, |
2405 | * create a range to tell us the last key of this type. | 2407 | * create a range to tell us the last key of this type. |
@@ -2417,7 +2419,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2417 | } | 2419 | } |
2418 | 2420 | ||
2419 | /* go backward to find any previous key */ | 2421 | /* go backward to find any previous key */ |
2420 | ret = btrfs_previous_item(root, path, inode->i_ino, key_type); | 2422 | ret = btrfs_previous_item(root, path, ino, key_type); |
2421 | if (ret == 0) { | 2423 | if (ret == 0) { |
2422 | struct btrfs_key tmp; | 2424 | struct btrfs_key tmp; |
2423 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); | 2425 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); |
@@ -2452,8 +2454,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2452 | for (i = path->slots[0]; i < nritems; i++) { | 2454 | for (i = path->slots[0]; i < nritems; i++) { |
2453 | btrfs_item_key_to_cpu(src, &min_key, i); | 2455 | btrfs_item_key_to_cpu(src, &min_key, i); |
2454 | 2456 | ||
2455 | if (min_key.objectid != inode->i_ino || | 2457 | if (min_key.objectid != ino || min_key.type != key_type) |
2456 | min_key.type != key_type) | ||
2457 | goto done; | 2458 | goto done; |
2458 | ret = overwrite_item(trans, log, dst_path, src, i, | 2459 | ret = overwrite_item(trans, log, dst_path, src, i, |
2459 | &min_key); | 2460 | &min_key); |
@@ -2474,7 +2475,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2474 | goto done; | 2475 | goto done; |
2475 | } | 2476 | } |
2476 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); | 2477 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); |
2477 | if (tmp.objectid != inode->i_ino || tmp.type != key_type) { | 2478 | if (tmp.objectid != ino || tmp.type != key_type) { |
2478 | last_offset = (u64)-1; | 2479 | last_offset = (u64)-1; |
2479 | goto done; | 2480 | goto done; |
2480 | } | 2481 | } |
@@ -2500,8 +2501,7 @@ done: | |||
2500 | * is valid | 2501 | * is valid |
2501 | */ | 2502 | */ |
2502 | ret = insert_dir_log_key(trans, log, path, key_type, | 2503 | ret = insert_dir_log_key(trans, log, path, key_type, |
2503 | inode->i_ino, first_offset, | 2504 | ino, first_offset, last_offset); |
2504 | last_offset); | ||
2505 | if (ret) | 2505 | if (ret) |
2506 | err = ret; | 2506 | err = ret; |
2507 | } | 2507 | } |
@@ -2745,6 +2745,7 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2745 | int nritems; | 2745 | int nritems; |
2746 | int ins_start_slot = 0; | 2746 | int ins_start_slot = 0; |
2747 | int ins_nr; | 2747 | int ins_nr; |
2748 | u64 ino = btrfs_ino(inode); | ||
2748 | 2749 | ||
2749 | log = root->log_root; | 2750 | log = root->log_root; |
2750 | 2751 | ||
@@ -2757,11 +2758,11 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2757 | return -ENOMEM; | 2758 | return -ENOMEM; |
2758 | } | 2759 | } |
2759 | 2760 | ||
2760 | min_key.objectid = inode->i_ino; | 2761 | min_key.objectid = ino; |
2761 | min_key.type = BTRFS_INODE_ITEM_KEY; | 2762 | min_key.type = BTRFS_INODE_ITEM_KEY; |
2762 | min_key.offset = 0; | 2763 | min_key.offset = 0; |
2763 | 2764 | ||
2764 | max_key.objectid = inode->i_ino; | 2765 | max_key.objectid = ino; |
2765 | 2766 | ||
2766 | /* today the code can only do partial logging of directories */ | 2767 | /* today the code can only do partial logging of directories */ |
2767 | if (!S_ISDIR(inode->i_mode)) | 2768 | if (!S_ISDIR(inode->i_mode)) |
@@ -2773,6 +2774,13 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2773 | max_key.type = (u8)-1; | 2774 | max_key.type = (u8)-1; |
2774 | max_key.offset = (u64)-1; | 2775 | max_key.offset = (u64)-1; |
2775 | 2776 | ||
2777 | ret = btrfs_commit_inode_delayed_items(trans, inode); | ||
2778 | if (ret) { | ||
2779 | btrfs_free_path(path); | ||
2780 | btrfs_free_path(dst_path); | ||
2781 | return ret; | ||
2782 | } | ||
2783 | |||
2776 | mutex_lock(&BTRFS_I(inode)->log_mutex); | 2784 | mutex_lock(&BTRFS_I(inode)->log_mutex); |
2777 | 2785 | ||
2778 | /* | 2786 | /* |
@@ -2784,8 +2792,7 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2784 | 2792 | ||
2785 | if (inode_only == LOG_INODE_EXISTS) | 2793 | if (inode_only == LOG_INODE_EXISTS) |
2786 | max_key_type = BTRFS_XATTR_ITEM_KEY; | 2794 | max_key_type = BTRFS_XATTR_ITEM_KEY; |
2787 | ret = drop_objectid_items(trans, log, path, | 2795 | ret = drop_objectid_items(trans, log, path, ino, max_key_type); |
2788 | inode->i_ino, max_key_type); | ||
2789 | } else { | 2796 | } else { |
2790 | ret = btrfs_truncate_inode_items(trans, log, inode, 0, 0); | 2797 | ret = btrfs_truncate_inode_items(trans, log, inode, 0, 0); |
2791 | } | 2798 | } |
@@ -2803,7 +2810,7 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2803 | break; | 2810 | break; |
2804 | again: | 2811 | again: |
2805 | /* note, ins_nr might be > 0 here, cleanup outside the loop */ | 2812 | /* note, ins_nr might be > 0 here, cleanup outside the loop */ |
2806 | if (min_key.objectid != inode->i_ino) | 2813 | if (min_key.objectid != ino) |
2807 | break; | 2814 | break; |
2808 | if (min_key.type > max_key.type) | 2815 | if (min_key.type > max_key.type) |
2809 | break; | 2816 | break; |
diff --git a/fs/btrfs/xattr.c b/fs/btrfs/xattr.c index 4ca88d1e18e..f3107e4b4d5 100644 --- a/fs/btrfs/xattr.c +++ b/fs/btrfs/xattr.c | |||
@@ -44,7 +44,7 @@ ssize_t __btrfs_getxattr(struct inode *inode, const char *name, | |||
44 | return -ENOMEM; | 44 | return -ENOMEM; |
45 | 45 | ||
46 | /* lookup the xattr by name */ | 46 | /* lookup the xattr by name */ |
47 | di = btrfs_lookup_xattr(NULL, root, path, inode->i_ino, name, | 47 | di = btrfs_lookup_xattr(NULL, root, path, btrfs_ino(inode), name, |
48 | strlen(name), 0); | 48 | strlen(name), 0); |
49 | if (!di) { | 49 | if (!di) { |
50 | ret = -ENODATA; | 50 | ret = -ENODATA; |
@@ -103,7 +103,7 @@ static int do_setxattr(struct btrfs_trans_handle *trans, | |||
103 | return -ENOMEM; | 103 | return -ENOMEM; |
104 | 104 | ||
105 | /* first lets see if we already have this xattr */ | 105 | /* first lets see if we already have this xattr */ |
106 | di = btrfs_lookup_xattr(trans, root, path, inode->i_ino, name, | 106 | di = btrfs_lookup_xattr(trans, root, path, btrfs_ino(inode), name, |
107 | strlen(name), -1); | 107 | strlen(name), -1); |
108 | if (IS_ERR(di)) { | 108 | if (IS_ERR(di)) { |
109 | ret = PTR_ERR(di); | 109 | ret = PTR_ERR(di); |
@@ -136,7 +136,7 @@ static int do_setxattr(struct btrfs_trans_handle *trans, | |||
136 | } | 136 | } |
137 | 137 | ||
138 | /* ok we have to create a completely new xattr */ | 138 | /* ok we have to create a completely new xattr */ |
139 | ret = btrfs_insert_xattr_item(trans, root, path, inode->i_ino, | 139 | ret = btrfs_insert_xattr_item(trans, root, path, btrfs_ino(inode), |
140 | name, name_len, value, size); | 140 | name, name_len, value, size); |
141 | BUG_ON(ret); | 141 | BUG_ON(ret); |
142 | out: | 142 | out: |
@@ -190,7 +190,7 @@ ssize_t btrfs_listxattr(struct dentry *dentry, char *buffer, size_t size) | |||
190 | * NOTE: we set key.offset = 0; because we want to start with the | 190 | * NOTE: we set key.offset = 0; because we want to start with the |
191 | * first xattr that we find and walk forward | 191 | * first xattr that we find and walk forward |
192 | */ | 192 | */ |
193 | key.objectid = inode->i_ino; | 193 | key.objectid = btrfs_ino(inode); |
194 | btrfs_set_key_type(&key, BTRFS_XATTR_ITEM_KEY); | 194 | btrfs_set_key_type(&key, BTRFS_XATTR_ITEM_KEY); |
195 | key.offset = 0; | 195 | key.offset = 0; |
196 | 196 | ||