diff options
Diffstat (limited to 'fs')
82 files changed, 4633 insertions, 1803 deletions
diff --git a/fs/block_dev.c b/fs/block_dev.c index 5147bdd3b8e1..257b00e98428 100644 --- a/fs/block_dev.c +++ b/fs/block_dev.c | |||
@@ -1102,6 +1102,7 @@ static int __blkdev_get(struct block_device *bdev, fmode_t mode, int for_part) | |||
1102 | if (!bdev->bd_part) | 1102 | if (!bdev->bd_part) |
1103 | goto out_clear; | 1103 | goto out_clear; |
1104 | 1104 | ||
1105 | ret = 0; | ||
1105 | if (disk->fops->open) { | 1106 | if (disk->fops->open) { |
1106 | ret = disk->fops->open(bdev, mode); | 1107 | ret = disk->fops->open(bdev, mode); |
1107 | if (ret == -ERESTARTSYS) { | 1108 | if (ret == -ERESTARTSYS) { |
@@ -1118,9 +1119,18 @@ static int __blkdev_get(struct block_device *bdev, fmode_t mode, int for_part) | |||
1118 | put_disk(disk); | 1119 | put_disk(disk); |
1119 | goto restart; | 1120 | goto restart; |
1120 | } | 1121 | } |
1121 | if (ret) | ||
1122 | goto out_clear; | ||
1123 | } | 1122 | } |
1123 | /* | ||
1124 | * If the device is invalidated, rescan partition | ||
1125 | * if open succeeded or failed with -ENOMEDIUM. | ||
1126 | * The latter is necessary to prevent ghost | ||
1127 | * partitions on a removed medium. | ||
1128 | */ | ||
1129 | if (bdev->bd_invalidated && (!ret || ret == -ENOMEDIUM)) | ||
1130 | rescan_partitions(disk, bdev); | ||
1131 | if (ret) | ||
1132 | goto out_clear; | ||
1133 | |||
1124 | if (!bdev->bd_openers) { | 1134 | if (!bdev->bd_openers) { |
1125 | bd_set_size(bdev,(loff_t)get_capacity(disk)<<9); | 1135 | bd_set_size(bdev,(loff_t)get_capacity(disk)<<9); |
1126 | bdi = blk_get_backing_dev_info(bdev); | 1136 | bdi = blk_get_backing_dev_info(bdev); |
@@ -1128,8 +1138,6 @@ static int __blkdev_get(struct block_device *bdev, fmode_t mode, int for_part) | |||
1128 | bdi = &default_backing_dev_info; | 1138 | bdi = &default_backing_dev_info; |
1129 | bdev_inode_switch_bdi(bdev->bd_inode, bdi); | 1139 | bdev_inode_switch_bdi(bdev->bd_inode, bdi); |
1130 | } | 1140 | } |
1131 | if (bdev->bd_invalidated) | ||
1132 | rescan_partitions(disk, bdev); | ||
1133 | } else { | 1141 | } else { |
1134 | struct block_device *whole; | 1142 | struct block_device *whole; |
1135 | whole = bdget_disk(disk, 0); | 1143 | whole = bdget_disk(disk, 0); |
@@ -1153,13 +1161,14 @@ static int __blkdev_get(struct block_device *bdev, fmode_t mode, int for_part) | |||
1153 | } | 1161 | } |
1154 | } else { | 1162 | } else { |
1155 | if (bdev->bd_contains == bdev) { | 1163 | if (bdev->bd_contains == bdev) { |
1156 | if (bdev->bd_disk->fops->open) { | 1164 | ret = 0; |
1165 | if (bdev->bd_disk->fops->open) | ||
1157 | ret = bdev->bd_disk->fops->open(bdev, mode); | 1166 | ret = bdev->bd_disk->fops->open(bdev, mode); |
1158 | if (ret) | 1167 | /* the same as first opener case, read comment there */ |
1159 | goto out_unlock_bdev; | 1168 | if (bdev->bd_invalidated && (!ret || ret == -ENOMEDIUM)) |
1160 | } | ||
1161 | if (bdev->bd_invalidated) | ||
1162 | rescan_partitions(bdev->bd_disk, bdev); | 1169 | rescan_partitions(bdev->bd_disk, bdev); |
1170 | if (ret) | ||
1171 | goto out_unlock_bdev; | ||
1163 | } | 1172 | } |
1164 | /* only one opener holds refs to the module and disk */ | 1173 | /* only one opener holds refs to the module and disk */ |
1165 | module_put(disk->fops->owner); | 1174 | module_put(disk->fops->owner); |
diff --git a/fs/btrfs/Makefile b/fs/btrfs/Makefile index 31610ea73aec..a8411c22313d 100644 --- a/fs/btrfs/Makefile +++ b/fs/btrfs/Makefile | |||
@@ -7,4 +7,4 @@ btrfs-y += super.o ctree.o extent-tree.o print-tree.o root-tree.o dir-item.o \ | |||
7 | extent_map.o sysfs.o struct-funcs.o xattr.o ordered-data.o \ | 7 | extent_map.o sysfs.o struct-funcs.o xattr.o ordered-data.o \ |
8 | extent_io.o volumes.o async-thread.o ioctl.o locking.o orphan.o \ | 8 | extent_io.o volumes.o async-thread.o ioctl.o locking.o orphan.o \ |
9 | export.o tree-log.o acl.o free-space-cache.o zlib.o lzo.o \ | 9 | export.o tree-log.o acl.o free-space-cache.o zlib.o lzo.o \ |
10 | compression.o delayed-ref.o relocation.o | 10 | compression.o delayed-ref.o relocation.o delayed-inode.o |
diff --git a/fs/btrfs/acl.c b/fs/btrfs/acl.c index 1a21c99a91b8..f66fc9959733 100644 --- a/fs/btrfs/acl.c +++ b/fs/btrfs/acl.c | |||
@@ -178,12 +178,13 @@ static int btrfs_xattr_acl_set(struct dentry *dentry, const char *name, | |||
178 | 178 | ||
179 | if (value) { | 179 | if (value) { |
180 | acl = posix_acl_from_xattr(value, size); | 180 | acl = posix_acl_from_xattr(value, size); |
181 | if (IS_ERR(acl)) | ||
182 | return PTR_ERR(acl); | ||
183 | |||
181 | if (acl) { | 184 | if (acl) { |
182 | ret = posix_acl_valid(acl); | 185 | ret = posix_acl_valid(acl); |
183 | if (ret) | 186 | if (ret) |
184 | goto out; | 187 | goto out; |
185 | } else if (IS_ERR(acl)) { | ||
186 | return PTR_ERR(acl); | ||
187 | } | 188 | } |
188 | } | 189 | } |
189 | 190 | ||
diff --git a/fs/btrfs/btrfs_inode.h b/fs/btrfs/btrfs_inode.h index 57c3bb2884ce..d0b0e43a6a8b 100644 --- a/fs/btrfs/btrfs_inode.h +++ b/fs/btrfs/btrfs_inode.h | |||
@@ -22,6 +22,7 @@ | |||
22 | #include "extent_map.h" | 22 | #include "extent_map.h" |
23 | #include "extent_io.h" | 23 | #include "extent_io.h" |
24 | #include "ordered-data.h" | 24 | #include "ordered-data.h" |
25 | #include "delayed-inode.h" | ||
25 | 26 | ||
26 | /* in memory btrfs inode */ | 27 | /* in memory btrfs inode */ |
27 | struct btrfs_inode { | 28 | struct btrfs_inode { |
@@ -158,14 +159,27 @@ struct btrfs_inode { | |||
158 | */ | 159 | */ |
159 | unsigned force_compress:4; | 160 | unsigned force_compress:4; |
160 | 161 | ||
162 | struct btrfs_delayed_node *delayed_node; | ||
163 | |||
161 | struct inode vfs_inode; | 164 | struct inode vfs_inode; |
162 | }; | 165 | }; |
163 | 166 | ||
167 | extern unsigned char btrfs_filetype_table[]; | ||
168 | |||
164 | static inline struct btrfs_inode *BTRFS_I(struct inode *inode) | 169 | static inline struct btrfs_inode *BTRFS_I(struct inode *inode) |
165 | { | 170 | { |
166 | return container_of(inode, struct btrfs_inode, vfs_inode); | 171 | return container_of(inode, struct btrfs_inode, vfs_inode); |
167 | } | 172 | } |
168 | 173 | ||
174 | static inline u64 btrfs_ino(struct inode *inode) | ||
175 | { | ||
176 | u64 ino = BTRFS_I(inode)->location.objectid; | ||
177 | |||
178 | if (ino <= BTRFS_FIRST_FREE_OBJECTID) | ||
179 | ino = inode->i_ino; | ||
180 | return ino; | ||
181 | } | ||
182 | |||
169 | static inline void btrfs_i_size_write(struct inode *inode, u64 size) | 183 | static inline void btrfs_i_size_write(struct inode *inode, u64 size) |
170 | { | 184 | { |
171 | i_size_write(inode, size); | 185 | i_size_write(inode, size); |
diff --git a/fs/btrfs/compression.c b/fs/btrfs/compression.c index d4cd0f0cd695..bfe42b03eaf9 100644 --- a/fs/btrfs/compression.c +++ b/fs/btrfs/compression.c | |||
@@ -125,9 +125,10 @@ static int check_compressed_csum(struct inode *inode, | |||
125 | kunmap_atomic(kaddr, KM_USER0); | 125 | kunmap_atomic(kaddr, KM_USER0); |
126 | 126 | ||
127 | if (csum != *cb_sum) { | 127 | if (csum != *cb_sum) { |
128 | printk(KERN_INFO "btrfs csum failed ino %lu " | 128 | printk(KERN_INFO "btrfs csum failed ino %llu " |
129 | "extent %llu csum %u " | 129 | "extent %llu csum %u " |
130 | "wanted %u mirror %d\n", inode->i_ino, | 130 | "wanted %u mirror %d\n", |
131 | (unsigned long long)btrfs_ino(inode), | ||
131 | (unsigned long long)disk_start, | 132 | (unsigned long long)disk_start, |
132 | csum, *cb_sum, cb->mirror_num); | 133 | csum, *cb_sum, cb->mirror_num); |
133 | ret = -EIO; | 134 | ret = -EIO; |
diff --git a/fs/btrfs/ctree.c b/fs/btrfs/ctree.c index fad8f23d70f0..b6cbeed226b1 100644 --- a/fs/btrfs/ctree.c +++ b/fs/btrfs/ctree.c | |||
@@ -38,11 +38,6 @@ static int balance_node_right(struct btrfs_trans_handle *trans, | |||
38 | struct extent_buffer *src_buf); | 38 | struct extent_buffer *src_buf); |
39 | static int del_ptr(struct btrfs_trans_handle *trans, struct btrfs_root *root, | 39 | static int del_ptr(struct btrfs_trans_handle *trans, struct btrfs_root *root, |
40 | struct btrfs_path *path, int level, int slot); | 40 | struct btrfs_path *path, int level, int slot); |
41 | static int setup_items_for_insert(struct btrfs_trans_handle *trans, | ||
42 | struct btrfs_root *root, struct btrfs_path *path, | ||
43 | struct btrfs_key *cpu_key, u32 *data_size, | ||
44 | u32 total_data, u32 total_size, int nr); | ||
45 | |||
46 | 41 | ||
47 | struct btrfs_path *btrfs_alloc_path(void) | 42 | struct btrfs_path *btrfs_alloc_path(void) |
48 | { | 43 | { |
@@ -74,8 +69,8 @@ noinline void btrfs_set_path_blocking(struct btrfs_path *p) | |||
74 | * retake all the spinlocks in the path. You can safely use NULL | 69 | * retake all the spinlocks in the path. You can safely use NULL |
75 | * for held | 70 | * for held |
76 | */ | 71 | */ |
77 | static noinline void btrfs_clear_path_blocking(struct btrfs_path *p, | 72 | noinline void btrfs_clear_path_blocking(struct btrfs_path *p, |
78 | struct extent_buffer *held) | 73 | struct extent_buffer *held) |
79 | { | 74 | { |
80 | int i; | 75 | int i; |
81 | 76 | ||
@@ -3559,11 +3554,10 @@ out: | |||
3559 | * to save stack depth by doing the bulk of the work in a function | 3554 | * to save stack depth by doing the bulk of the work in a function |
3560 | * that doesn't call btrfs_search_slot | 3555 | * that doesn't call btrfs_search_slot |
3561 | */ | 3556 | */ |
3562 | static noinline_for_stack int | 3557 | int setup_items_for_insert(struct btrfs_trans_handle *trans, |
3563 | setup_items_for_insert(struct btrfs_trans_handle *trans, | 3558 | struct btrfs_root *root, struct btrfs_path *path, |
3564 | struct btrfs_root *root, struct btrfs_path *path, | 3559 | struct btrfs_key *cpu_key, u32 *data_size, |
3565 | struct btrfs_key *cpu_key, u32 *data_size, | 3560 | u32 total_data, u32 total_size, int nr) |
3566 | u32 total_data, u32 total_size, int nr) | ||
3567 | { | 3561 | { |
3568 | struct btrfs_item *item; | 3562 | struct btrfs_item *item; |
3569 | int i; | 3563 | int i; |
diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h index 343304dec6d1..e7d40791ec9f 100644 --- a/fs/btrfs/ctree.h +++ b/fs/btrfs/ctree.h | |||
@@ -105,6 +105,12 @@ struct btrfs_ordered_sum; | |||
105 | /* For storing free space cache */ | 105 | /* For storing free space cache */ |
106 | #define BTRFS_FREE_SPACE_OBJECTID -11ULL | 106 | #define BTRFS_FREE_SPACE_OBJECTID -11ULL |
107 | 107 | ||
108 | /* | ||
109 | * The inode number assigned to the special inode for sotring | ||
110 | * free ino cache | ||
111 | */ | ||
112 | #define BTRFS_FREE_INO_OBJECTID -12ULL | ||
113 | |||
108 | /* dummy objectid represents multiple objectids */ | 114 | /* dummy objectid represents multiple objectids */ |
109 | #define BTRFS_MULTIPLE_OBJECTIDS -255ULL | 115 | #define BTRFS_MULTIPLE_OBJECTIDS -255ULL |
110 | 116 | ||
@@ -830,9 +836,6 @@ struct btrfs_block_group_cache { | |||
830 | u64 bytes_super; | 836 | u64 bytes_super; |
831 | u64 flags; | 837 | u64 flags; |
832 | u64 sectorsize; | 838 | u64 sectorsize; |
833 | int extents_thresh; | ||
834 | int free_extents; | ||
835 | int total_bitmaps; | ||
836 | unsigned int ro:1; | 839 | unsigned int ro:1; |
837 | unsigned int dirty:1; | 840 | unsigned int dirty:1; |
838 | unsigned int iref:1; | 841 | unsigned int iref:1; |
@@ -847,9 +850,7 @@ struct btrfs_block_group_cache { | |||
847 | struct btrfs_space_info *space_info; | 850 | struct btrfs_space_info *space_info; |
848 | 851 | ||
849 | /* free space cache stuff */ | 852 | /* free space cache stuff */ |
850 | spinlock_t tree_lock; | 853 | struct btrfs_free_space_ctl *free_space_ctl; |
851 | struct rb_root free_space_offset; | ||
852 | u64 free_space; | ||
853 | 854 | ||
854 | /* block group cache stuff */ | 855 | /* block group cache stuff */ |
855 | struct rb_node cache_node; | 856 | struct rb_node cache_node; |
@@ -869,6 +870,7 @@ struct btrfs_block_group_cache { | |||
869 | struct reloc_control; | 870 | struct reloc_control; |
870 | struct btrfs_device; | 871 | struct btrfs_device; |
871 | struct btrfs_fs_devices; | 872 | struct btrfs_fs_devices; |
873 | struct btrfs_delayed_root; | ||
872 | struct btrfs_fs_info { | 874 | struct btrfs_fs_info { |
873 | u8 fsid[BTRFS_FSID_SIZE]; | 875 | u8 fsid[BTRFS_FSID_SIZE]; |
874 | u8 chunk_tree_uuid[BTRFS_UUID_SIZE]; | 876 | u8 chunk_tree_uuid[BTRFS_UUID_SIZE]; |
@@ -895,7 +897,10 @@ struct btrfs_fs_info { | |||
895 | /* logical->physical extent mapping */ | 897 | /* logical->physical extent mapping */ |
896 | struct btrfs_mapping_tree mapping_tree; | 898 | struct btrfs_mapping_tree mapping_tree; |
897 | 899 | ||
898 | /* block reservation for extent, checksum and root tree */ | 900 | /* |
901 | * block reservation for extent, checksum, root tree and | ||
902 | * delayed dir index item | ||
903 | */ | ||
899 | struct btrfs_block_rsv global_block_rsv; | 904 | struct btrfs_block_rsv global_block_rsv; |
900 | /* block reservation for delay allocation */ | 905 | /* block reservation for delay allocation */ |
901 | struct btrfs_block_rsv delalloc_block_rsv; | 906 | struct btrfs_block_rsv delalloc_block_rsv; |
@@ -1022,6 +1027,7 @@ struct btrfs_fs_info { | |||
1022 | * for the sys_munmap function call path | 1027 | * for the sys_munmap function call path |
1023 | */ | 1028 | */ |
1024 | struct btrfs_workers fixup_workers; | 1029 | struct btrfs_workers fixup_workers; |
1030 | struct btrfs_workers delayed_workers; | ||
1025 | struct task_struct *transaction_kthread; | 1031 | struct task_struct *transaction_kthread; |
1026 | struct task_struct *cleaner_kthread; | 1032 | struct task_struct *cleaner_kthread; |
1027 | int thread_pool_size; | 1033 | int thread_pool_size; |
@@ -1079,6 +1085,8 @@ struct btrfs_fs_info { | |||
1079 | 1085 | ||
1080 | /* filesystem state */ | 1086 | /* filesystem state */ |
1081 | u64 fs_state; | 1087 | u64 fs_state; |
1088 | |||
1089 | struct btrfs_delayed_root *delayed_root; | ||
1082 | }; | 1090 | }; |
1083 | 1091 | ||
1084 | /* | 1092 | /* |
@@ -1107,6 +1115,16 @@ struct btrfs_root { | |||
1107 | spinlock_t accounting_lock; | 1115 | spinlock_t accounting_lock; |
1108 | struct btrfs_block_rsv *block_rsv; | 1116 | struct btrfs_block_rsv *block_rsv; |
1109 | 1117 | ||
1118 | /* free ino cache stuff */ | ||
1119 | struct mutex fs_commit_mutex; | ||
1120 | struct btrfs_free_space_ctl *free_ino_ctl; | ||
1121 | enum btrfs_caching_type cached; | ||
1122 | spinlock_t cache_lock; | ||
1123 | wait_queue_head_t cache_wait; | ||
1124 | struct btrfs_free_space_ctl *free_ino_pinned; | ||
1125 | u64 cache_progress; | ||
1126 | struct inode *cache_inode; | ||
1127 | |||
1110 | struct mutex log_mutex; | 1128 | struct mutex log_mutex; |
1111 | wait_queue_head_t log_writer_wait; | 1129 | wait_queue_head_t log_writer_wait; |
1112 | wait_queue_head_t log_commit_wait[2]; | 1130 | wait_queue_head_t log_commit_wait[2]; |
@@ -1162,6 +1180,11 @@ struct btrfs_root { | |||
1162 | struct rb_root inode_tree; | 1180 | struct rb_root inode_tree; |
1163 | 1181 | ||
1164 | /* | 1182 | /* |
1183 | * radix tree that keeps track of delayed nodes of every inode, | ||
1184 | * protected by inode_lock | ||
1185 | */ | ||
1186 | struct radix_tree_root delayed_nodes_tree; | ||
1187 | /* | ||
1165 | * right now this just gets used so that a root has its own devid | 1188 | * right now this just gets used so that a root has its own devid |
1166 | * for stat. It may be used for more later | 1189 | * for stat. It may be used for more later |
1167 | */ | 1190 | */ |
@@ -2034,6 +2057,13 @@ static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info) | |||
2034 | } | 2057 | } |
2035 | 2058 | ||
2036 | /* extent-tree.c */ | 2059 | /* extent-tree.c */ |
2060 | static inline u64 btrfs_calc_trans_metadata_size(struct btrfs_root *root, | ||
2061 | int num_items) | ||
2062 | { | ||
2063 | return (root->leafsize + root->nodesize * (BTRFS_MAX_LEVEL - 1)) * | ||
2064 | 3 * num_items; | ||
2065 | } | ||
2066 | |||
2037 | void btrfs_put_block_group(struct btrfs_block_group_cache *cache); | 2067 | void btrfs_put_block_group(struct btrfs_block_group_cache *cache); |
2038 | int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans, | 2068 | int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans, |
2039 | struct btrfs_root *root, unsigned long count); | 2069 | struct btrfs_root *root, unsigned long count); |
@@ -2226,6 +2256,8 @@ void btrfs_release_path(struct btrfs_path *p); | |||
2226 | struct btrfs_path *btrfs_alloc_path(void); | 2256 | struct btrfs_path *btrfs_alloc_path(void); |
2227 | void btrfs_free_path(struct btrfs_path *p); | 2257 | void btrfs_free_path(struct btrfs_path *p); |
2228 | void btrfs_set_path_blocking(struct btrfs_path *p); | 2258 | void btrfs_set_path_blocking(struct btrfs_path *p); |
2259 | void btrfs_clear_path_blocking(struct btrfs_path *p, | ||
2260 | struct extent_buffer *held); | ||
2229 | void btrfs_unlock_up_safe(struct btrfs_path *p, int level); | 2261 | void btrfs_unlock_up_safe(struct btrfs_path *p, int level); |
2230 | 2262 | ||
2231 | int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root, | 2263 | int btrfs_del_items(struct btrfs_trans_handle *trans, struct btrfs_root *root, |
@@ -2237,6 +2269,10 @@ static inline int btrfs_del_item(struct btrfs_trans_handle *trans, | |||
2237 | return btrfs_del_items(trans, root, path, path->slots[0], 1); | 2269 | return btrfs_del_items(trans, root, path, path->slots[0], 1); |
2238 | } | 2270 | } |
2239 | 2271 | ||
2272 | int setup_items_for_insert(struct btrfs_trans_handle *trans, | ||
2273 | struct btrfs_root *root, struct btrfs_path *path, | ||
2274 | struct btrfs_key *cpu_key, u32 *data_size, | ||
2275 | u32 total_data, u32 total_size, int nr); | ||
2240 | int btrfs_insert_item(struct btrfs_trans_handle *trans, struct btrfs_root | 2276 | int btrfs_insert_item(struct btrfs_trans_handle *trans, struct btrfs_root |
2241 | *root, struct btrfs_key *key, void *data, u32 data_size); | 2277 | *root, struct btrfs_key *key, void *data, u32 data_size); |
2242 | int btrfs_insert_empty_items(struct btrfs_trans_handle *trans, | 2278 | int btrfs_insert_empty_items(struct btrfs_trans_handle *trans, |
@@ -2293,7 +2329,7 @@ void btrfs_check_and_init_root_item(struct btrfs_root_item *item); | |||
2293 | /* dir-item.c */ | 2329 | /* dir-item.c */ |
2294 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, | 2330 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, |
2295 | struct btrfs_root *root, const char *name, | 2331 | struct btrfs_root *root, const char *name, |
2296 | int name_len, u64 dir, | 2332 | int name_len, struct inode *dir, |
2297 | struct btrfs_key *location, u8 type, u64 index); | 2333 | struct btrfs_key *location, u8 type, u64 index); |
2298 | struct btrfs_dir_item *btrfs_lookup_dir_item(struct btrfs_trans_handle *trans, | 2334 | struct btrfs_dir_item *btrfs_lookup_dir_item(struct btrfs_trans_handle *trans, |
2299 | struct btrfs_root *root, | 2335 | struct btrfs_root *root, |
@@ -2338,12 +2374,6 @@ int btrfs_del_orphan_item(struct btrfs_trans_handle *trans, | |||
2338 | struct btrfs_root *root, u64 offset); | 2374 | struct btrfs_root *root, u64 offset); |
2339 | int btrfs_find_orphan_item(struct btrfs_root *root, u64 offset); | 2375 | int btrfs_find_orphan_item(struct btrfs_root *root, u64 offset); |
2340 | 2376 | ||
2341 | /* inode-map.c */ | ||
2342 | int btrfs_find_free_objectid(struct btrfs_trans_handle *trans, | ||
2343 | struct btrfs_root *fs_root, | ||
2344 | u64 dirid, u64 *objectid); | ||
2345 | int btrfs_find_highest_inode(struct btrfs_root *fs_root, u64 *objectid); | ||
2346 | |||
2347 | /* inode-item.c */ | 2377 | /* inode-item.c */ |
2348 | int btrfs_insert_inode_ref(struct btrfs_trans_handle *trans, | 2378 | int btrfs_insert_inode_ref(struct btrfs_trans_handle *trans, |
2349 | struct btrfs_root *root, | 2379 | struct btrfs_root *root, |
diff --git a/fs/btrfs/delayed-inode.c b/fs/btrfs/delayed-inode.c new file mode 100644 index 000000000000..01e29503a54b --- /dev/null +++ b/fs/btrfs/delayed-inode.c | |||
@@ -0,0 +1,1695 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2011 Fujitsu. All rights reserved. | ||
3 | * Written by Miao Xie <miaox@cn.fujitsu.com> | ||
4 | * | ||
5 | * This program is free software; you can redistribute it and/or | ||
6 | * modify it under the terms of the GNU General Public | ||
7 | * License v2 as published by the Free Software Foundation. | ||
8 | * | ||
9 | * This program is distributed in the hope that it will be useful, | ||
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | ||
12 | * General Public License for more details. | ||
13 | * | ||
14 | * You should have received a copy of the GNU General Public | ||
15 | * License along with this program; if not, write to the | ||
16 | * Free Software Foundation, Inc., 59 Temple Place - Suite 330, | ||
17 | * Boston, MA 021110-1307, USA. | ||
18 | */ | ||
19 | |||
20 | #include <linux/slab.h> | ||
21 | #include "delayed-inode.h" | ||
22 | #include "disk-io.h" | ||
23 | #include "transaction.h" | ||
24 | |||
25 | #define BTRFS_DELAYED_WRITEBACK 400 | ||
26 | #define BTRFS_DELAYED_BACKGROUND 100 | ||
27 | |||
28 | static struct kmem_cache *delayed_node_cache; | ||
29 | |||
30 | int __init btrfs_delayed_inode_init(void) | ||
31 | { | ||
32 | delayed_node_cache = kmem_cache_create("delayed_node", | ||
33 | sizeof(struct btrfs_delayed_node), | ||
34 | 0, | ||
35 | SLAB_RECLAIM_ACCOUNT | SLAB_MEM_SPREAD, | ||
36 | NULL); | ||
37 | if (!delayed_node_cache) | ||
38 | return -ENOMEM; | ||
39 | return 0; | ||
40 | } | ||
41 | |||
42 | void btrfs_delayed_inode_exit(void) | ||
43 | { | ||
44 | if (delayed_node_cache) | ||
45 | kmem_cache_destroy(delayed_node_cache); | ||
46 | } | ||
47 | |||
48 | static inline void btrfs_init_delayed_node( | ||
49 | struct btrfs_delayed_node *delayed_node, | ||
50 | struct btrfs_root *root, u64 inode_id) | ||
51 | { | ||
52 | delayed_node->root = root; | ||
53 | delayed_node->inode_id = inode_id; | ||
54 | atomic_set(&delayed_node->refs, 0); | ||
55 | delayed_node->count = 0; | ||
56 | delayed_node->in_list = 0; | ||
57 | delayed_node->inode_dirty = 0; | ||
58 | delayed_node->ins_root = RB_ROOT; | ||
59 | delayed_node->del_root = RB_ROOT; | ||
60 | mutex_init(&delayed_node->mutex); | ||
61 | delayed_node->index_cnt = 0; | ||
62 | INIT_LIST_HEAD(&delayed_node->n_list); | ||
63 | INIT_LIST_HEAD(&delayed_node->p_list); | ||
64 | delayed_node->bytes_reserved = 0; | ||
65 | } | ||
66 | |||
67 | static inline int btrfs_is_continuous_delayed_item( | ||
68 | struct btrfs_delayed_item *item1, | ||
69 | struct btrfs_delayed_item *item2) | ||
70 | { | ||
71 | if (item1->key.type == BTRFS_DIR_INDEX_KEY && | ||
72 | item1->key.objectid == item2->key.objectid && | ||
73 | item1->key.type == item2->key.type && | ||
74 | item1->key.offset + 1 == item2->key.offset) | ||
75 | return 1; | ||
76 | return 0; | ||
77 | } | ||
78 | |||
79 | static inline struct btrfs_delayed_root *btrfs_get_delayed_root( | ||
80 | struct btrfs_root *root) | ||
81 | { | ||
82 | return root->fs_info->delayed_root; | ||
83 | } | ||
84 | |||
85 | static struct btrfs_delayed_node *btrfs_get_or_create_delayed_node( | ||
86 | struct inode *inode) | ||
87 | { | ||
88 | struct btrfs_delayed_node *node; | ||
89 | struct btrfs_inode *btrfs_inode = BTRFS_I(inode); | ||
90 | struct btrfs_root *root = btrfs_inode->root; | ||
91 | u64 ino = btrfs_ino(inode); | ||
92 | int ret; | ||
93 | |||
94 | again: | ||
95 | node = ACCESS_ONCE(btrfs_inode->delayed_node); | ||
96 | if (node) { | ||
97 | atomic_inc(&node->refs); /* can be accessed */ | ||
98 | return node; | ||
99 | } | ||
100 | |||
101 | spin_lock(&root->inode_lock); | ||
102 | node = radix_tree_lookup(&root->delayed_nodes_tree, ino); | ||
103 | if (node) { | ||
104 | if (btrfs_inode->delayed_node) { | ||
105 | spin_unlock(&root->inode_lock); | ||
106 | goto again; | ||
107 | } | ||
108 | btrfs_inode->delayed_node = node; | ||
109 | atomic_inc(&node->refs); /* can be accessed */ | ||
110 | atomic_inc(&node->refs); /* cached in the inode */ | ||
111 | spin_unlock(&root->inode_lock); | ||
112 | return node; | ||
113 | } | ||
114 | spin_unlock(&root->inode_lock); | ||
115 | |||
116 | node = kmem_cache_alloc(delayed_node_cache, GFP_NOFS); | ||
117 | if (!node) | ||
118 | return ERR_PTR(-ENOMEM); | ||
119 | btrfs_init_delayed_node(node, root, ino); | ||
120 | |||
121 | atomic_inc(&node->refs); /* cached in the btrfs inode */ | ||
122 | atomic_inc(&node->refs); /* can be accessed */ | ||
123 | |||
124 | ret = radix_tree_preload(GFP_NOFS & ~__GFP_HIGHMEM); | ||
125 | if (ret) { | ||
126 | kmem_cache_free(delayed_node_cache, node); | ||
127 | return ERR_PTR(ret); | ||
128 | } | ||
129 | |||
130 | spin_lock(&root->inode_lock); | ||
131 | ret = radix_tree_insert(&root->delayed_nodes_tree, ino, node); | ||
132 | if (ret == -EEXIST) { | ||
133 | kmem_cache_free(delayed_node_cache, node); | ||
134 | spin_unlock(&root->inode_lock); | ||
135 | radix_tree_preload_end(); | ||
136 | goto again; | ||
137 | } | ||
138 | btrfs_inode->delayed_node = node; | ||
139 | spin_unlock(&root->inode_lock); | ||
140 | radix_tree_preload_end(); | ||
141 | |||
142 | return node; | ||
143 | } | ||
144 | |||
145 | /* | ||
146 | * Call it when holding delayed_node->mutex | ||
147 | * | ||
148 | * If mod = 1, add this node into the prepared list. | ||
149 | */ | ||
150 | static void btrfs_queue_delayed_node(struct btrfs_delayed_root *root, | ||
151 | struct btrfs_delayed_node *node, | ||
152 | int mod) | ||
153 | { | ||
154 | spin_lock(&root->lock); | ||
155 | if (node->in_list) { | ||
156 | if (!list_empty(&node->p_list)) | ||
157 | list_move_tail(&node->p_list, &root->prepare_list); | ||
158 | else if (mod) | ||
159 | list_add_tail(&node->p_list, &root->prepare_list); | ||
160 | } else { | ||
161 | list_add_tail(&node->n_list, &root->node_list); | ||
162 | list_add_tail(&node->p_list, &root->prepare_list); | ||
163 | atomic_inc(&node->refs); /* inserted into list */ | ||
164 | root->nodes++; | ||
165 | node->in_list = 1; | ||
166 | } | ||
167 | spin_unlock(&root->lock); | ||
168 | } | ||
169 | |||
170 | /* Call it when holding delayed_node->mutex */ | ||
171 | static void btrfs_dequeue_delayed_node(struct btrfs_delayed_root *root, | ||
172 | struct btrfs_delayed_node *node) | ||
173 | { | ||
174 | spin_lock(&root->lock); | ||
175 | if (node->in_list) { | ||
176 | root->nodes--; | ||
177 | atomic_dec(&node->refs); /* not in the list */ | ||
178 | list_del_init(&node->n_list); | ||
179 | if (!list_empty(&node->p_list)) | ||
180 | list_del_init(&node->p_list); | ||
181 | node->in_list = 0; | ||
182 | } | ||
183 | spin_unlock(&root->lock); | ||
184 | } | ||
185 | |||
186 | struct btrfs_delayed_node *btrfs_first_delayed_node( | ||
187 | struct btrfs_delayed_root *delayed_root) | ||
188 | { | ||
189 | struct list_head *p; | ||
190 | struct btrfs_delayed_node *node = NULL; | ||
191 | |||
192 | spin_lock(&delayed_root->lock); | ||
193 | if (list_empty(&delayed_root->node_list)) | ||
194 | goto out; | ||
195 | |||
196 | p = delayed_root->node_list.next; | ||
197 | node = list_entry(p, struct btrfs_delayed_node, n_list); | ||
198 | atomic_inc(&node->refs); | ||
199 | out: | ||
200 | spin_unlock(&delayed_root->lock); | ||
201 | |||
202 | return node; | ||
203 | } | ||
204 | |||
205 | struct btrfs_delayed_node *btrfs_next_delayed_node( | ||
206 | struct btrfs_delayed_node *node) | ||
207 | { | ||
208 | struct btrfs_delayed_root *delayed_root; | ||
209 | struct list_head *p; | ||
210 | struct btrfs_delayed_node *next = NULL; | ||
211 | |||
212 | delayed_root = node->root->fs_info->delayed_root; | ||
213 | spin_lock(&delayed_root->lock); | ||
214 | if (!node->in_list) { /* not in the list */ | ||
215 | if (list_empty(&delayed_root->node_list)) | ||
216 | goto out; | ||
217 | p = delayed_root->node_list.next; | ||
218 | } else if (list_is_last(&node->n_list, &delayed_root->node_list)) | ||
219 | goto out; | ||
220 | else | ||
221 | p = node->n_list.next; | ||
222 | |||
223 | next = list_entry(p, struct btrfs_delayed_node, n_list); | ||
224 | atomic_inc(&next->refs); | ||
225 | out: | ||
226 | spin_unlock(&delayed_root->lock); | ||
227 | |||
228 | return next; | ||
229 | } | ||
230 | |||
231 | static void __btrfs_release_delayed_node( | ||
232 | struct btrfs_delayed_node *delayed_node, | ||
233 | int mod) | ||
234 | { | ||
235 | struct btrfs_delayed_root *delayed_root; | ||
236 | |||
237 | if (!delayed_node) | ||
238 | return; | ||
239 | |||
240 | delayed_root = delayed_node->root->fs_info->delayed_root; | ||
241 | |||
242 | mutex_lock(&delayed_node->mutex); | ||
243 | if (delayed_node->count) | ||
244 | btrfs_queue_delayed_node(delayed_root, delayed_node, mod); | ||
245 | else | ||
246 | btrfs_dequeue_delayed_node(delayed_root, delayed_node); | ||
247 | mutex_unlock(&delayed_node->mutex); | ||
248 | |||
249 | if (atomic_dec_and_test(&delayed_node->refs)) { | ||
250 | struct btrfs_root *root = delayed_node->root; | ||
251 | spin_lock(&root->inode_lock); | ||
252 | if (atomic_read(&delayed_node->refs) == 0) { | ||
253 | radix_tree_delete(&root->delayed_nodes_tree, | ||
254 | delayed_node->inode_id); | ||
255 | kmem_cache_free(delayed_node_cache, delayed_node); | ||
256 | } | ||
257 | spin_unlock(&root->inode_lock); | ||
258 | } | ||
259 | } | ||
260 | |||
261 | static inline void btrfs_release_delayed_node(struct btrfs_delayed_node *node) | ||
262 | { | ||
263 | __btrfs_release_delayed_node(node, 0); | ||
264 | } | ||
265 | |||
266 | struct btrfs_delayed_node *btrfs_first_prepared_delayed_node( | ||
267 | struct btrfs_delayed_root *delayed_root) | ||
268 | { | ||
269 | struct list_head *p; | ||
270 | struct btrfs_delayed_node *node = NULL; | ||
271 | |||
272 | spin_lock(&delayed_root->lock); | ||
273 | if (list_empty(&delayed_root->prepare_list)) | ||
274 | goto out; | ||
275 | |||
276 | p = delayed_root->prepare_list.next; | ||
277 | list_del_init(p); | ||
278 | node = list_entry(p, struct btrfs_delayed_node, p_list); | ||
279 | atomic_inc(&node->refs); | ||
280 | out: | ||
281 | spin_unlock(&delayed_root->lock); | ||
282 | |||
283 | return node; | ||
284 | } | ||
285 | |||
286 | static inline void btrfs_release_prepared_delayed_node( | ||
287 | struct btrfs_delayed_node *node) | ||
288 | { | ||
289 | __btrfs_release_delayed_node(node, 1); | ||
290 | } | ||
291 | |||
292 | struct btrfs_delayed_item *btrfs_alloc_delayed_item(u32 data_len) | ||
293 | { | ||
294 | struct btrfs_delayed_item *item; | ||
295 | item = kmalloc(sizeof(*item) + data_len, GFP_NOFS); | ||
296 | if (item) { | ||
297 | item->data_len = data_len; | ||
298 | item->ins_or_del = 0; | ||
299 | item->bytes_reserved = 0; | ||
300 | item->block_rsv = NULL; | ||
301 | item->delayed_node = NULL; | ||
302 | atomic_set(&item->refs, 1); | ||
303 | } | ||
304 | return item; | ||
305 | } | ||
306 | |||
307 | /* | ||
308 | * __btrfs_lookup_delayed_item - look up the delayed item by key | ||
309 | * @delayed_node: pointer to the delayed node | ||
310 | * @key: the key to look up | ||
311 | * @prev: used to store the prev item if the right item isn't found | ||
312 | * @next: used to store the next item if the right item isn't found | ||
313 | * | ||
314 | * Note: if we don't find the right item, we will return the prev item and | ||
315 | * the next item. | ||
316 | */ | ||
317 | static struct btrfs_delayed_item *__btrfs_lookup_delayed_item( | ||
318 | struct rb_root *root, | ||
319 | struct btrfs_key *key, | ||
320 | struct btrfs_delayed_item **prev, | ||
321 | struct btrfs_delayed_item **next) | ||
322 | { | ||
323 | struct rb_node *node, *prev_node = NULL; | ||
324 | struct btrfs_delayed_item *delayed_item = NULL; | ||
325 | int ret = 0; | ||
326 | |||
327 | node = root->rb_node; | ||
328 | |||
329 | while (node) { | ||
330 | delayed_item = rb_entry(node, struct btrfs_delayed_item, | ||
331 | rb_node); | ||
332 | prev_node = node; | ||
333 | ret = btrfs_comp_cpu_keys(&delayed_item->key, key); | ||
334 | if (ret < 0) | ||
335 | node = node->rb_right; | ||
336 | else if (ret > 0) | ||
337 | node = node->rb_left; | ||
338 | else | ||
339 | return delayed_item; | ||
340 | } | ||
341 | |||
342 | if (prev) { | ||
343 | if (!prev_node) | ||
344 | *prev = NULL; | ||
345 | else if (ret < 0) | ||
346 | *prev = delayed_item; | ||
347 | else if ((node = rb_prev(prev_node)) != NULL) { | ||
348 | *prev = rb_entry(node, struct btrfs_delayed_item, | ||
349 | rb_node); | ||
350 | } else | ||
351 | *prev = NULL; | ||
352 | } | ||
353 | |||
354 | if (next) { | ||
355 | if (!prev_node) | ||
356 | *next = NULL; | ||
357 | else if (ret > 0) | ||
358 | *next = delayed_item; | ||
359 | else if ((node = rb_next(prev_node)) != NULL) { | ||
360 | *next = rb_entry(node, struct btrfs_delayed_item, | ||
361 | rb_node); | ||
362 | } else | ||
363 | *next = NULL; | ||
364 | } | ||
365 | return NULL; | ||
366 | } | ||
367 | |||
368 | struct btrfs_delayed_item *__btrfs_lookup_delayed_insertion_item( | ||
369 | struct btrfs_delayed_node *delayed_node, | ||
370 | struct btrfs_key *key) | ||
371 | { | ||
372 | struct btrfs_delayed_item *item; | ||
373 | |||
374 | item = __btrfs_lookup_delayed_item(&delayed_node->ins_root, key, | ||
375 | NULL, NULL); | ||
376 | return item; | ||
377 | } | ||
378 | |||
379 | struct btrfs_delayed_item *__btrfs_lookup_delayed_deletion_item( | ||
380 | struct btrfs_delayed_node *delayed_node, | ||
381 | struct btrfs_key *key) | ||
382 | { | ||
383 | struct btrfs_delayed_item *item; | ||
384 | |||
385 | item = __btrfs_lookup_delayed_item(&delayed_node->del_root, key, | ||
386 | NULL, NULL); | ||
387 | return item; | ||
388 | } | ||
389 | |||
390 | struct btrfs_delayed_item *__btrfs_search_delayed_insertion_item( | ||
391 | struct btrfs_delayed_node *delayed_node, | ||
392 | struct btrfs_key *key) | ||
393 | { | ||
394 | struct btrfs_delayed_item *item, *next; | ||
395 | |||
396 | item = __btrfs_lookup_delayed_item(&delayed_node->ins_root, key, | ||
397 | NULL, &next); | ||
398 | if (!item) | ||
399 | item = next; | ||
400 | |||
401 | return item; | ||
402 | } | ||
403 | |||
404 | struct btrfs_delayed_item *__btrfs_search_delayed_deletion_item( | ||
405 | struct btrfs_delayed_node *delayed_node, | ||
406 | struct btrfs_key *key) | ||
407 | { | ||
408 | struct btrfs_delayed_item *item, *next; | ||
409 | |||
410 | item = __btrfs_lookup_delayed_item(&delayed_node->del_root, key, | ||
411 | NULL, &next); | ||
412 | if (!item) | ||
413 | item = next; | ||
414 | |||
415 | return item; | ||
416 | } | ||
417 | |||
418 | static int __btrfs_add_delayed_item(struct btrfs_delayed_node *delayed_node, | ||
419 | struct btrfs_delayed_item *ins, | ||
420 | int action) | ||
421 | { | ||
422 | struct rb_node **p, *node; | ||
423 | struct rb_node *parent_node = NULL; | ||
424 | struct rb_root *root; | ||
425 | struct btrfs_delayed_item *item; | ||
426 | int cmp; | ||
427 | |||
428 | if (action == BTRFS_DELAYED_INSERTION_ITEM) | ||
429 | root = &delayed_node->ins_root; | ||
430 | else if (action == BTRFS_DELAYED_DELETION_ITEM) | ||
431 | root = &delayed_node->del_root; | ||
432 | else | ||
433 | BUG(); | ||
434 | p = &root->rb_node; | ||
435 | node = &ins->rb_node; | ||
436 | |||
437 | while (*p) { | ||
438 | parent_node = *p; | ||
439 | item = rb_entry(parent_node, struct btrfs_delayed_item, | ||
440 | rb_node); | ||
441 | |||
442 | cmp = btrfs_comp_cpu_keys(&item->key, &ins->key); | ||
443 | if (cmp < 0) | ||
444 | p = &(*p)->rb_right; | ||
445 | else if (cmp > 0) | ||
446 | p = &(*p)->rb_left; | ||
447 | else | ||
448 | return -EEXIST; | ||
449 | } | ||
450 | |||
451 | rb_link_node(node, parent_node, p); | ||
452 | rb_insert_color(node, root); | ||
453 | ins->delayed_node = delayed_node; | ||
454 | ins->ins_or_del = action; | ||
455 | |||
456 | if (ins->key.type == BTRFS_DIR_INDEX_KEY && | ||
457 | action == BTRFS_DELAYED_INSERTION_ITEM && | ||
458 | ins->key.offset >= delayed_node->index_cnt) | ||
459 | delayed_node->index_cnt = ins->key.offset + 1; | ||
460 | |||
461 | delayed_node->count++; | ||
462 | atomic_inc(&delayed_node->root->fs_info->delayed_root->items); | ||
463 | return 0; | ||
464 | } | ||
465 | |||
466 | static int __btrfs_add_delayed_insertion_item(struct btrfs_delayed_node *node, | ||
467 | struct btrfs_delayed_item *item) | ||
468 | { | ||
469 | return __btrfs_add_delayed_item(node, item, | ||
470 | BTRFS_DELAYED_INSERTION_ITEM); | ||
471 | } | ||
472 | |||
473 | static int __btrfs_add_delayed_deletion_item(struct btrfs_delayed_node *node, | ||
474 | struct btrfs_delayed_item *item) | ||
475 | { | ||
476 | return __btrfs_add_delayed_item(node, item, | ||
477 | BTRFS_DELAYED_DELETION_ITEM); | ||
478 | } | ||
479 | |||
480 | static void __btrfs_remove_delayed_item(struct btrfs_delayed_item *delayed_item) | ||
481 | { | ||
482 | struct rb_root *root; | ||
483 | struct btrfs_delayed_root *delayed_root; | ||
484 | |||
485 | delayed_root = delayed_item->delayed_node->root->fs_info->delayed_root; | ||
486 | |||
487 | BUG_ON(!delayed_root); | ||
488 | BUG_ON(delayed_item->ins_or_del != BTRFS_DELAYED_DELETION_ITEM && | ||
489 | delayed_item->ins_or_del != BTRFS_DELAYED_INSERTION_ITEM); | ||
490 | |||
491 | if (delayed_item->ins_or_del == BTRFS_DELAYED_INSERTION_ITEM) | ||
492 | root = &delayed_item->delayed_node->ins_root; | ||
493 | else | ||
494 | root = &delayed_item->delayed_node->del_root; | ||
495 | |||
496 | rb_erase(&delayed_item->rb_node, root); | ||
497 | delayed_item->delayed_node->count--; | ||
498 | atomic_dec(&delayed_root->items); | ||
499 | if (atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND && | ||
500 | waitqueue_active(&delayed_root->wait)) | ||
501 | wake_up(&delayed_root->wait); | ||
502 | } | ||
503 | |||
504 | static void btrfs_release_delayed_item(struct btrfs_delayed_item *item) | ||
505 | { | ||
506 | if (item) { | ||
507 | __btrfs_remove_delayed_item(item); | ||
508 | if (atomic_dec_and_test(&item->refs)) | ||
509 | kfree(item); | ||
510 | } | ||
511 | } | ||
512 | |||
513 | struct btrfs_delayed_item *__btrfs_first_delayed_insertion_item( | ||
514 | struct btrfs_delayed_node *delayed_node) | ||
515 | { | ||
516 | struct rb_node *p; | ||
517 | struct btrfs_delayed_item *item = NULL; | ||
518 | |||
519 | p = rb_first(&delayed_node->ins_root); | ||
520 | if (p) | ||
521 | item = rb_entry(p, struct btrfs_delayed_item, rb_node); | ||
522 | |||
523 | return item; | ||
524 | } | ||
525 | |||
526 | struct btrfs_delayed_item *__btrfs_first_delayed_deletion_item( | ||
527 | struct btrfs_delayed_node *delayed_node) | ||
528 | { | ||
529 | struct rb_node *p; | ||
530 | struct btrfs_delayed_item *item = NULL; | ||
531 | |||
532 | p = rb_first(&delayed_node->del_root); | ||
533 | if (p) | ||
534 | item = rb_entry(p, struct btrfs_delayed_item, rb_node); | ||
535 | |||
536 | return item; | ||
537 | } | ||
538 | |||
539 | struct btrfs_delayed_item *__btrfs_next_delayed_item( | ||
540 | struct btrfs_delayed_item *item) | ||
541 | { | ||
542 | struct rb_node *p; | ||
543 | struct btrfs_delayed_item *next = NULL; | ||
544 | |||
545 | p = rb_next(&item->rb_node); | ||
546 | if (p) | ||
547 | next = rb_entry(p, struct btrfs_delayed_item, rb_node); | ||
548 | |||
549 | return next; | ||
550 | } | ||
551 | |||
552 | static inline struct btrfs_delayed_node *btrfs_get_delayed_node( | ||
553 | struct inode *inode) | ||
554 | { | ||
555 | struct btrfs_inode *btrfs_inode = BTRFS_I(inode); | ||
556 | struct btrfs_delayed_node *delayed_node; | ||
557 | |||
558 | delayed_node = btrfs_inode->delayed_node; | ||
559 | if (delayed_node) | ||
560 | atomic_inc(&delayed_node->refs); | ||
561 | |||
562 | return delayed_node; | ||
563 | } | ||
564 | |||
565 | static inline struct btrfs_root *btrfs_get_fs_root(struct btrfs_root *root, | ||
566 | u64 root_id) | ||
567 | { | ||
568 | struct btrfs_key root_key; | ||
569 | |||
570 | if (root->objectid == root_id) | ||
571 | return root; | ||
572 | |||
573 | root_key.objectid = root_id; | ||
574 | root_key.type = BTRFS_ROOT_ITEM_KEY; | ||
575 | root_key.offset = (u64)-1; | ||
576 | return btrfs_read_fs_root_no_name(root->fs_info, &root_key); | ||
577 | } | ||
578 | |||
579 | static int btrfs_delayed_item_reserve_metadata(struct btrfs_trans_handle *trans, | ||
580 | struct btrfs_root *root, | ||
581 | struct btrfs_delayed_item *item) | ||
582 | { | ||
583 | struct btrfs_block_rsv *src_rsv; | ||
584 | struct btrfs_block_rsv *dst_rsv; | ||
585 | u64 num_bytes; | ||
586 | int ret; | ||
587 | |||
588 | if (!trans->bytes_reserved) | ||
589 | return 0; | ||
590 | |||
591 | src_rsv = trans->block_rsv; | ||
592 | dst_rsv = &root->fs_info->global_block_rsv; | ||
593 | |||
594 | num_bytes = btrfs_calc_trans_metadata_size(root, 1); | ||
595 | ret = btrfs_block_rsv_migrate(src_rsv, dst_rsv, num_bytes); | ||
596 | if (!ret) { | ||
597 | item->bytes_reserved = num_bytes; | ||
598 | item->block_rsv = dst_rsv; | ||
599 | } | ||
600 | |||
601 | return ret; | ||
602 | } | ||
603 | |||
604 | static void btrfs_delayed_item_release_metadata(struct btrfs_root *root, | ||
605 | struct btrfs_delayed_item *item) | ||
606 | { | ||
607 | if (!item->bytes_reserved) | ||
608 | return; | ||
609 | |||
610 | btrfs_block_rsv_release(root, item->block_rsv, | ||
611 | item->bytes_reserved); | ||
612 | } | ||
613 | |||
614 | static int btrfs_delayed_inode_reserve_metadata( | ||
615 | struct btrfs_trans_handle *trans, | ||
616 | struct btrfs_root *root, | ||
617 | struct btrfs_delayed_node *node) | ||
618 | { | ||
619 | struct btrfs_block_rsv *src_rsv; | ||
620 | struct btrfs_block_rsv *dst_rsv; | ||
621 | u64 num_bytes; | ||
622 | int ret; | ||
623 | |||
624 | if (!trans->bytes_reserved) | ||
625 | return 0; | ||
626 | |||
627 | src_rsv = trans->block_rsv; | ||
628 | dst_rsv = &root->fs_info->global_block_rsv; | ||
629 | |||
630 | num_bytes = btrfs_calc_trans_metadata_size(root, 1); | ||
631 | ret = btrfs_block_rsv_migrate(src_rsv, dst_rsv, num_bytes); | ||
632 | if (!ret) | ||
633 | node->bytes_reserved = num_bytes; | ||
634 | |||
635 | return ret; | ||
636 | } | ||
637 | |||
638 | static void btrfs_delayed_inode_release_metadata(struct btrfs_root *root, | ||
639 | struct btrfs_delayed_node *node) | ||
640 | { | ||
641 | struct btrfs_block_rsv *rsv; | ||
642 | |||
643 | if (!node->bytes_reserved) | ||
644 | return; | ||
645 | |||
646 | rsv = &root->fs_info->global_block_rsv; | ||
647 | btrfs_block_rsv_release(root, rsv, | ||
648 | node->bytes_reserved); | ||
649 | node->bytes_reserved = 0; | ||
650 | } | ||
651 | |||
652 | /* | ||
653 | * This helper will insert some continuous items into the same leaf according | ||
654 | * to the free space of the leaf. | ||
655 | */ | ||
656 | static int btrfs_batch_insert_items(struct btrfs_trans_handle *trans, | ||
657 | struct btrfs_root *root, | ||
658 | struct btrfs_path *path, | ||
659 | struct btrfs_delayed_item *item) | ||
660 | { | ||
661 | struct btrfs_delayed_item *curr, *next; | ||
662 | int free_space; | ||
663 | int total_data_size = 0, total_size = 0; | ||
664 | struct extent_buffer *leaf; | ||
665 | char *data_ptr; | ||
666 | struct btrfs_key *keys; | ||
667 | u32 *data_size; | ||
668 | struct list_head head; | ||
669 | int slot; | ||
670 | int nitems; | ||
671 | int i; | ||
672 | int ret = 0; | ||
673 | |||
674 | BUG_ON(!path->nodes[0]); | ||
675 | |||
676 | leaf = path->nodes[0]; | ||
677 | free_space = btrfs_leaf_free_space(root, leaf); | ||
678 | INIT_LIST_HEAD(&head); | ||
679 | |||
680 | next = item; | ||
681 | |||
682 | /* | ||
683 | * count the number of the continuous items that we can insert in batch | ||
684 | */ | ||
685 | while (total_size + next->data_len + sizeof(struct btrfs_item) <= | ||
686 | free_space) { | ||
687 | total_data_size += next->data_len; | ||
688 | total_size += next->data_len + sizeof(struct btrfs_item); | ||
689 | list_add_tail(&next->tree_list, &head); | ||
690 | nitems++; | ||
691 | |||
692 | curr = next; | ||
693 | next = __btrfs_next_delayed_item(curr); | ||
694 | if (!next) | ||
695 | break; | ||
696 | |||
697 | if (!btrfs_is_continuous_delayed_item(curr, next)) | ||
698 | break; | ||
699 | } | ||
700 | |||
701 | if (!nitems) { | ||
702 | ret = 0; | ||
703 | goto out; | ||
704 | } | ||
705 | |||
706 | /* | ||
707 | * we need allocate some memory space, but it might cause the task | ||
708 | * to sleep, so we set all locked nodes in the path to blocking locks | ||
709 | * first. | ||
710 | */ | ||
711 | btrfs_set_path_blocking(path); | ||
712 | |||
713 | keys = kmalloc(sizeof(struct btrfs_key) * nitems, GFP_NOFS); | ||
714 | if (!keys) { | ||
715 | ret = -ENOMEM; | ||
716 | goto out; | ||
717 | } | ||
718 | |||
719 | data_size = kmalloc(sizeof(u32) * nitems, GFP_NOFS); | ||
720 | if (!data_size) { | ||
721 | ret = -ENOMEM; | ||
722 | goto error; | ||
723 | } | ||
724 | |||
725 | /* get keys of all the delayed items */ | ||
726 | i = 0; | ||
727 | list_for_each_entry(next, &head, tree_list) { | ||
728 | keys[i] = next->key; | ||
729 | data_size[i] = next->data_len; | ||
730 | i++; | ||
731 | } | ||
732 | |||
733 | /* reset all the locked nodes in the patch to spinning locks. */ | ||
734 | btrfs_clear_path_blocking(path, NULL); | ||
735 | |||
736 | /* insert the keys of the items */ | ||
737 | ret = setup_items_for_insert(trans, root, path, keys, data_size, | ||
738 | total_data_size, total_size, nitems); | ||
739 | if (ret) | ||
740 | goto error; | ||
741 | |||
742 | /* insert the dir index items */ | ||
743 | slot = path->slots[0]; | ||
744 | list_for_each_entry_safe(curr, next, &head, tree_list) { | ||
745 | data_ptr = btrfs_item_ptr(leaf, slot, char); | ||
746 | write_extent_buffer(leaf, &curr->data, | ||
747 | (unsigned long)data_ptr, | ||
748 | curr->data_len); | ||
749 | slot++; | ||
750 | |||
751 | btrfs_delayed_item_release_metadata(root, curr); | ||
752 | |||
753 | list_del(&curr->tree_list); | ||
754 | btrfs_release_delayed_item(curr); | ||
755 | } | ||
756 | |||
757 | error: | ||
758 | kfree(data_size); | ||
759 | kfree(keys); | ||
760 | out: | ||
761 | return ret; | ||
762 | } | ||
763 | |||
764 | /* | ||
765 | * This helper can just do simple insertion that needn't extend item for new | ||
766 | * data, such as directory name index insertion, inode insertion. | ||
767 | */ | ||
768 | static int btrfs_insert_delayed_item(struct btrfs_trans_handle *trans, | ||
769 | struct btrfs_root *root, | ||
770 | struct btrfs_path *path, | ||
771 | struct btrfs_delayed_item *delayed_item) | ||
772 | { | ||
773 | struct extent_buffer *leaf; | ||
774 | struct btrfs_item *item; | ||
775 | char *ptr; | ||
776 | int ret; | ||
777 | |||
778 | ret = btrfs_insert_empty_item(trans, root, path, &delayed_item->key, | ||
779 | delayed_item->data_len); | ||
780 | if (ret < 0 && ret != -EEXIST) | ||
781 | return ret; | ||
782 | |||
783 | leaf = path->nodes[0]; | ||
784 | |||
785 | item = btrfs_item_nr(leaf, path->slots[0]); | ||
786 | ptr = btrfs_item_ptr(leaf, path->slots[0], char); | ||
787 | |||
788 | write_extent_buffer(leaf, delayed_item->data, (unsigned long)ptr, | ||
789 | delayed_item->data_len); | ||
790 | btrfs_mark_buffer_dirty(leaf); | ||
791 | |||
792 | btrfs_delayed_item_release_metadata(root, delayed_item); | ||
793 | return 0; | ||
794 | } | ||
795 | |||
796 | /* | ||
797 | * we insert an item first, then if there are some continuous items, we try | ||
798 | * to insert those items into the same leaf. | ||
799 | */ | ||
800 | static int btrfs_insert_delayed_items(struct btrfs_trans_handle *trans, | ||
801 | struct btrfs_path *path, | ||
802 | struct btrfs_root *root, | ||
803 | struct btrfs_delayed_node *node) | ||
804 | { | ||
805 | struct btrfs_delayed_item *curr, *prev; | ||
806 | int ret = 0; | ||
807 | |||
808 | do_again: | ||
809 | mutex_lock(&node->mutex); | ||
810 | curr = __btrfs_first_delayed_insertion_item(node); | ||
811 | if (!curr) | ||
812 | goto insert_end; | ||
813 | |||
814 | ret = btrfs_insert_delayed_item(trans, root, path, curr); | ||
815 | if (ret < 0) { | ||
816 | btrfs_release_path(path); | ||
817 | goto insert_end; | ||
818 | } | ||
819 | |||
820 | prev = curr; | ||
821 | curr = __btrfs_next_delayed_item(prev); | ||
822 | if (curr && btrfs_is_continuous_delayed_item(prev, curr)) { | ||
823 | /* insert the continuous items into the same leaf */ | ||
824 | path->slots[0]++; | ||
825 | btrfs_batch_insert_items(trans, root, path, curr); | ||
826 | } | ||
827 | btrfs_release_delayed_item(prev); | ||
828 | btrfs_mark_buffer_dirty(path->nodes[0]); | ||
829 | |||
830 | btrfs_release_path(path); | ||
831 | mutex_unlock(&node->mutex); | ||
832 | goto do_again; | ||
833 | |||
834 | insert_end: | ||
835 | mutex_unlock(&node->mutex); | ||
836 | return ret; | ||
837 | } | ||
838 | |||
839 | static int btrfs_batch_delete_items(struct btrfs_trans_handle *trans, | ||
840 | struct btrfs_root *root, | ||
841 | struct btrfs_path *path, | ||
842 | struct btrfs_delayed_item *item) | ||
843 | { | ||
844 | struct btrfs_delayed_item *curr, *next; | ||
845 | struct extent_buffer *leaf; | ||
846 | struct btrfs_key key; | ||
847 | struct list_head head; | ||
848 | int nitems, i, last_item; | ||
849 | int ret = 0; | ||
850 | |||
851 | BUG_ON(!path->nodes[0]); | ||
852 | |||
853 | leaf = path->nodes[0]; | ||
854 | |||
855 | i = path->slots[0]; | ||
856 | last_item = btrfs_header_nritems(leaf) - 1; | ||
857 | if (i > last_item) | ||
858 | return -ENOENT; /* FIXME: Is errno suitable? */ | ||
859 | |||
860 | next = item; | ||
861 | INIT_LIST_HEAD(&head); | ||
862 | btrfs_item_key_to_cpu(leaf, &key, i); | ||
863 | nitems = 0; | ||
864 | /* | ||
865 | * count the number of the dir index items that we can delete in batch | ||
866 | */ | ||
867 | while (btrfs_comp_cpu_keys(&next->key, &key) == 0) { | ||
868 | list_add_tail(&next->tree_list, &head); | ||
869 | nitems++; | ||
870 | |||
871 | curr = next; | ||
872 | next = __btrfs_next_delayed_item(curr); | ||
873 | if (!next) | ||
874 | break; | ||
875 | |||
876 | if (!btrfs_is_continuous_delayed_item(curr, next)) | ||
877 | break; | ||
878 | |||
879 | i++; | ||
880 | if (i > last_item) | ||
881 | break; | ||
882 | btrfs_item_key_to_cpu(leaf, &key, i); | ||
883 | } | ||
884 | |||
885 | if (!nitems) | ||
886 | return 0; | ||
887 | |||
888 | ret = btrfs_del_items(trans, root, path, path->slots[0], nitems); | ||
889 | if (ret) | ||
890 | goto out; | ||
891 | |||
892 | list_for_each_entry_safe(curr, next, &head, tree_list) { | ||
893 | btrfs_delayed_item_release_metadata(root, curr); | ||
894 | list_del(&curr->tree_list); | ||
895 | btrfs_release_delayed_item(curr); | ||
896 | } | ||
897 | |||
898 | out: | ||
899 | return ret; | ||
900 | } | ||
901 | |||
902 | static int btrfs_delete_delayed_items(struct btrfs_trans_handle *trans, | ||
903 | struct btrfs_path *path, | ||
904 | struct btrfs_root *root, | ||
905 | struct btrfs_delayed_node *node) | ||
906 | { | ||
907 | struct btrfs_delayed_item *curr, *prev; | ||
908 | int ret = 0; | ||
909 | |||
910 | do_again: | ||
911 | mutex_lock(&node->mutex); | ||
912 | curr = __btrfs_first_delayed_deletion_item(node); | ||
913 | if (!curr) | ||
914 | goto delete_fail; | ||
915 | |||
916 | ret = btrfs_search_slot(trans, root, &curr->key, path, -1, 1); | ||
917 | if (ret < 0) | ||
918 | goto delete_fail; | ||
919 | else if (ret > 0) { | ||
920 | /* | ||
921 | * can't find the item which the node points to, so this node | ||
922 | * is invalid, just drop it. | ||
923 | */ | ||
924 | prev = curr; | ||
925 | curr = __btrfs_next_delayed_item(prev); | ||
926 | btrfs_release_delayed_item(prev); | ||
927 | ret = 0; | ||
928 | btrfs_release_path(path); | ||
929 | if (curr) | ||
930 | goto do_again; | ||
931 | else | ||
932 | goto delete_fail; | ||
933 | } | ||
934 | |||
935 | btrfs_batch_delete_items(trans, root, path, curr); | ||
936 | btrfs_release_path(path); | ||
937 | mutex_unlock(&node->mutex); | ||
938 | goto do_again; | ||
939 | |||
940 | delete_fail: | ||
941 | btrfs_release_path(path); | ||
942 | mutex_unlock(&node->mutex); | ||
943 | return ret; | ||
944 | } | ||
945 | |||
946 | static void btrfs_release_delayed_inode(struct btrfs_delayed_node *delayed_node) | ||
947 | { | ||
948 | struct btrfs_delayed_root *delayed_root; | ||
949 | |||
950 | if (delayed_node && delayed_node->inode_dirty) { | ||
951 | BUG_ON(!delayed_node->root); | ||
952 | delayed_node->inode_dirty = 0; | ||
953 | delayed_node->count--; | ||
954 | |||
955 | delayed_root = delayed_node->root->fs_info->delayed_root; | ||
956 | atomic_dec(&delayed_root->items); | ||
957 | if (atomic_read(&delayed_root->items) < | ||
958 | BTRFS_DELAYED_BACKGROUND && | ||
959 | waitqueue_active(&delayed_root->wait)) | ||
960 | wake_up(&delayed_root->wait); | ||
961 | } | ||
962 | } | ||
963 | |||
964 | static int btrfs_update_delayed_inode(struct btrfs_trans_handle *trans, | ||
965 | struct btrfs_root *root, | ||
966 | struct btrfs_path *path, | ||
967 | struct btrfs_delayed_node *node) | ||
968 | { | ||
969 | struct btrfs_key key; | ||
970 | struct btrfs_inode_item *inode_item; | ||
971 | struct extent_buffer *leaf; | ||
972 | int ret; | ||
973 | |||
974 | mutex_lock(&node->mutex); | ||
975 | if (!node->inode_dirty) { | ||
976 | mutex_unlock(&node->mutex); | ||
977 | return 0; | ||
978 | } | ||
979 | |||
980 | key.objectid = node->inode_id; | ||
981 | btrfs_set_key_type(&key, BTRFS_INODE_ITEM_KEY); | ||
982 | key.offset = 0; | ||
983 | ret = btrfs_lookup_inode(trans, root, path, &key, 1); | ||
984 | if (ret > 0) { | ||
985 | btrfs_release_path(path); | ||
986 | mutex_unlock(&node->mutex); | ||
987 | return -ENOENT; | ||
988 | } else if (ret < 0) { | ||
989 | mutex_unlock(&node->mutex); | ||
990 | return ret; | ||
991 | } | ||
992 | |||
993 | btrfs_unlock_up_safe(path, 1); | ||
994 | leaf = path->nodes[0]; | ||
995 | inode_item = btrfs_item_ptr(leaf, path->slots[0], | ||
996 | struct btrfs_inode_item); | ||
997 | write_extent_buffer(leaf, &node->inode_item, (unsigned long)inode_item, | ||
998 | sizeof(struct btrfs_inode_item)); | ||
999 | btrfs_mark_buffer_dirty(leaf); | ||
1000 | btrfs_release_path(path); | ||
1001 | |||
1002 | btrfs_delayed_inode_release_metadata(root, node); | ||
1003 | btrfs_release_delayed_inode(node); | ||
1004 | mutex_unlock(&node->mutex); | ||
1005 | |||
1006 | return 0; | ||
1007 | } | ||
1008 | |||
1009 | /* Called when committing the transaction. */ | ||
1010 | int btrfs_run_delayed_items(struct btrfs_trans_handle *trans, | ||
1011 | struct btrfs_root *root) | ||
1012 | { | ||
1013 | struct btrfs_delayed_root *delayed_root; | ||
1014 | struct btrfs_delayed_node *curr_node, *prev_node; | ||
1015 | struct btrfs_path *path; | ||
1016 | int ret = 0; | ||
1017 | |||
1018 | path = btrfs_alloc_path(); | ||
1019 | if (!path) | ||
1020 | return -ENOMEM; | ||
1021 | path->leave_spinning = 1; | ||
1022 | |||
1023 | delayed_root = btrfs_get_delayed_root(root); | ||
1024 | |||
1025 | curr_node = btrfs_first_delayed_node(delayed_root); | ||
1026 | while (curr_node) { | ||
1027 | root = curr_node->root; | ||
1028 | ret = btrfs_insert_delayed_items(trans, path, root, | ||
1029 | curr_node); | ||
1030 | if (!ret) | ||
1031 | ret = btrfs_delete_delayed_items(trans, path, root, | ||
1032 | curr_node); | ||
1033 | if (!ret) | ||
1034 | ret = btrfs_update_delayed_inode(trans, root, path, | ||
1035 | curr_node); | ||
1036 | if (ret) { | ||
1037 | btrfs_release_delayed_node(curr_node); | ||
1038 | break; | ||
1039 | } | ||
1040 | |||
1041 | prev_node = curr_node; | ||
1042 | curr_node = btrfs_next_delayed_node(curr_node); | ||
1043 | btrfs_release_delayed_node(prev_node); | ||
1044 | } | ||
1045 | |||
1046 | btrfs_free_path(path); | ||
1047 | return ret; | ||
1048 | } | ||
1049 | |||
1050 | static int __btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans, | ||
1051 | struct btrfs_delayed_node *node) | ||
1052 | { | ||
1053 | struct btrfs_path *path; | ||
1054 | int ret; | ||
1055 | |||
1056 | path = btrfs_alloc_path(); | ||
1057 | if (!path) | ||
1058 | return -ENOMEM; | ||
1059 | path->leave_spinning = 1; | ||
1060 | |||
1061 | ret = btrfs_insert_delayed_items(trans, path, node->root, node); | ||
1062 | if (!ret) | ||
1063 | ret = btrfs_delete_delayed_items(trans, path, node->root, node); | ||
1064 | if (!ret) | ||
1065 | ret = btrfs_update_delayed_inode(trans, node->root, path, node); | ||
1066 | btrfs_free_path(path); | ||
1067 | |||
1068 | return ret; | ||
1069 | } | ||
1070 | |||
1071 | int btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans, | ||
1072 | struct inode *inode) | ||
1073 | { | ||
1074 | struct btrfs_delayed_node *delayed_node = btrfs_get_delayed_node(inode); | ||
1075 | int ret; | ||
1076 | |||
1077 | if (!delayed_node) | ||
1078 | return 0; | ||
1079 | |||
1080 | mutex_lock(&delayed_node->mutex); | ||
1081 | if (!delayed_node->count) { | ||
1082 | mutex_unlock(&delayed_node->mutex); | ||
1083 | btrfs_release_delayed_node(delayed_node); | ||
1084 | return 0; | ||
1085 | } | ||
1086 | mutex_unlock(&delayed_node->mutex); | ||
1087 | |||
1088 | ret = __btrfs_commit_inode_delayed_items(trans, delayed_node); | ||
1089 | btrfs_release_delayed_node(delayed_node); | ||
1090 | return ret; | ||
1091 | } | ||
1092 | |||
1093 | void btrfs_remove_delayed_node(struct inode *inode) | ||
1094 | { | ||
1095 | struct btrfs_delayed_node *delayed_node; | ||
1096 | |||
1097 | delayed_node = ACCESS_ONCE(BTRFS_I(inode)->delayed_node); | ||
1098 | if (!delayed_node) | ||
1099 | return; | ||
1100 | |||
1101 | BTRFS_I(inode)->delayed_node = NULL; | ||
1102 | btrfs_release_delayed_node(delayed_node); | ||
1103 | } | ||
1104 | |||
1105 | struct btrfs_async_delayed_node { | ||
1106 | struct btrfs_root *root; | ||
1107 | struct btrfs_delayed_node *delayed_node; | ||
1108 | struct btrfs_work work; | ||
1109 | }; | ||
1110 | |||
1111 | static void btrfs_async_run_delayed_node_done(struct btrfs_work *work) | ||
1112 | { | ||
1113 | struct btrfs_async_delayed_node *async_node; | ||
1114 | struct btrfs_trans_handle *trans; | ||
1115 | struct btrfs_path *path; | ||
1116 | struct btrfs_delayed_node *delayed_node = NULL; | ||
1117 | struct btrfs_root *root; | ||
1118 | unsigned long nr = 0; | ||
1119 | int need_requeue = 0; | ||
1120 | int ret; | ||
1121 | |||
1122 | async_node = container_of(work, struct btrfs_async_delayed_node, work); | ||
1123 | |||
1124 | path = btrfs_alloc_path(); | ||
1125 | if (!path) | ||
1126 | goto out; | ||
1127 | path->leave_spinning = 1; | ||
1128 | |||
1129 | delayed_node = async_node->delayed_node; | ||
1130 | root = delayed_node->root; | ||
1131 | |||
1132 | trans = btrfs_join_transaction(root, 0); | ||
1133 | if (IS_ERR(trans)) | ||
1134 | goto free_path; | ||
1135 | |||
1136 | ret = btrfs_insert_delayed_items(trans, path, root, delayed_node); | ||
1137 | if (!ret) | ||
1138 | ret = btrfs_delete_delayed_items(trans, path, root, | ||
1139 | delayed_node); | ||
1140 | |||
1141 | if (!ret) | ||
1142 | btrfs_update_delayed_inode(trans, root, path, delayed_node); | ||
1143 | |||
1144 | /* | ||
1145 | * Maybe new delayed items have been inserted, so we need requeue | ||
1146 | * the work. Besides that, we must dequeue the empty delayed nodes | ||
1147 | * to avoid the race between delayed items balance and the worker. | ||
1148 | * The race like this: | ||
1149 | * Task1 Worker thread | ||
1150 | * count == 0, needn't requeue | ||
1151 | * also needn't insert the | ||
1152 | * delayed node into prepare | ||
1153 | * list again. | ||
1154 | * add lots of delayed items | ||
1155 | * queue the delayed node | ||
1156 | * already in the list, | ||
1157 | * and not in the prepare | ||
1158 | * list, it means the delayed | ||
1159 | * node is being dealt with | ||
1160 | * by the worker. | ||
1161 | * do delayed items balance | ||
1162 | * the delayed node is being | ||
1163 | * dealt with by the worker | ||
1164 | * now, just wait. | ||
1165 | * the worker goto idle. | ||
1166 | * Task1 will sleep until the transaction is commited. | ||
1167 | */ | ||
1168 | mutex_lock(&delayed_node->mutex); | ||
1169 | if (delayed_node->count) | ||
1170 | need_requeue = 1; | ||
1171 | else | ||
1172 | btrfs_dequeue_delayed_node(root->fs_info->delayed_root, | ||
1173 | delayed_node); | ||
1174 | mutex_unlock(&delayed_node->mutex); | ||
1175 | |||
1176 | nr = trans->blocks_used; | ||
1177 | |||
1178 | btrfs_end_transaction_dmeta(trans, root); | ||
1179 | __btrfs_btree_balance_dirty(root, nr); | ||
1180 | free_path: | ||
1181 | btrfs_free_path(path); | ||
1182 | out: | ||
1183 | if (need_requeue) | ||
1184 | btrfs_requeue_work(&async_node->work); | ||
1185 | else { | ||
1186 | btrfs_release_prepared_delayed_node(delayed_node); | ||
1187 | kfree(async_node); | ||
1188 | } | ||
1189 | } | ||
1190 | |||
1191 | static int btrfs_wq_run_delayed_node(struct btrfs_delayed_root *delayed_root, | ||
1192 | struct btrfs_root *root, int all) | ||
1193 | { | ||
1194 | struct btrfs_async_delayed_node *async_node; | ||
1195 | struct btrfs_delayed_node *curr; | ||
1196 | int count = 0; | ||
1197 | |||
1198 | again: | ||
1199 | curr = btrfs_first_prepared_delayed_node(delayed_root); | ||
1200 | if (!curr) | ||
1201 | return 0; | ||
1202 | |||
1203 | async_node = kmalloc(sizeof(*async_node), GFP_NOFS); | ||
1204 | if (!async_node) { | ||
1205 | btrfs_release_prepared_delayed_node(curr); | ||
1206 | return -ENOMEM; | ||
1207 | } | ||
1208 | |||
1209 | async_node->root = root; | ||
1210 | async_node->delayed_node = curr; | ||
1211 | |||
1212 | async_node->work.func = btrfs_async_run_delayed_node_done; | ||
1213 | async_node->work.flags = 0; | ||
1214 | |||
1215 | btrfs_queue_worker(&root->fs_info->delayed_workers, &async_node->work); | ||
1216 | count++; | ||
1217 | |||
1218 | if (all || count < 4) | ||
1219 | goto again; | ||
1220 | |||
1221 | return 0; | ||
1222 | } | ||
1223 | |||
1224 | void btrfs_balance_delayed_items(struct btrfs_root *root) | ||
1225 | { | ||
1226 | struct btrfs_delayed_root *delayed_root; | ||
1227 | |||
1228 | delayed_root = btrfs_get_delayed_root(root); | ||
1229 | |||
1230 | if (atomic_read(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND) | ||
1231 | return; | ||
1232 | |||
1233 | if (atomic_read(&delayed_root->items) >= BTRFS_DELAYED_WRITEBACK) { | ||
1234 | int ret; | ||
1235 | ret = btrfs_wq_run_delayed_node(delayed_root, root, 1); | ||
1236 | if (ret) | ||
1237 | return; | ||
1238 | |||
1239 | wait_event_interruptible_timeout( | ||
1240 | delayed_root->wait, | ||
1241 | (atomic_read(&delayed_root->items) < | ||
1242 | BTRFS_DELAYED_BACKGROUND), | ||
1243 | HZ); | ||
1244 | return; | ||
1245 | } | ||
1246 | |||
1247 | btrfs_wq_run_delayed_node(delayed_root, root, 0); | ||
1248 | } | ||
1249 | |||
1250 | int btrfs_insert_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
1251 | struct btrfs_root *root, const char *name, | ||
1252 | int name_len, struct inode *dir, | ||
1253 | struct btrfs_disk_key *disk_key, u8 type, | ||
1254 | u64 index) | ||
1255 | { | ||
1256 | struct btrfs_delayed_node *delayed_node; | ||
1257 | struct btrfs_delayed_item *delayed_item; | ||
1258 | struct btrfs_dir_item *dir_item; | ||
1259 | int ret; | ||
1260 | |||
1261 | delayed_node = btrfs_get_or_create_delayed_node(dir); | ||
1262 | if (IS_ERR(delayed_node)) | ||
1263 | return PTR_ERR(delayed_node); | ||
1264 | |||
1265 | delayed_item = btrfs_alloc_delayed_item(sizeof(*dir_item) + name_len); | ||
1266 | if (!delayed_item) { | ||
1267 | ret = -ENOMEM; | ||
1268 | goto release_node; | ||
1269 | } | ||
1270 | |||
1271 | ret = btrfs_delayed_item_reserve_metadata(trans, root, delayed_item); | ||
1272 | /* | ||
1273 | * we have reserved enough space when we start a new transaction, | ||
1274 | * so reserving metadata failure is impossible | ||
1275 | */ | ||
1276 | BUG_ON(ret); | ||
1277 | |||
1278 | delayed_item->key.objectid = btrfs_ino(dir); | ||
1279 | btrfs_set_key_type(&delayed_item->key, BTRFS_DIR_INDEX_KEY); | ||
1280 | delayed_item->key.offset = index; | ||
1281 | |||
1282 | dir_item = (struct btrfs_dir_item *)delayed_item->data; | ||
1283 | dir_item->location = *disk_key; | ||
1284 | dir_item->transid = cpu_to_le64(trans->transid); | ||
1285 | dir_item->data_len = 0; | ||
1286 | dir_item->name_len = cpu_to_le16(name_len); | ||
1287 | dir_item->type = type; | ||
1288 | memcpy((char *)(dir_item + 1), name, name_len); | ||
1289 | |||
1290 | mutex_lock(&delayed_node->mutex); | ||
1291 | ret = __btrfs_add_delayed_insertion_item(delayed_node, delayed_item); | ||
1292 | if (unlikely(ret)) { | ||
1293 | printk(KERN_ERR "err add delayed dir index item(name: %s) into " | ||
1294 | "the insertion tree of the delayed node" | ||
1295 | "(root id: %llu, inode id: %llu, errno: %d)\n", | ||
1296 | name, | ||
1297 | (unsigned long long)delayed_node->root->objectid, | ||
1298 | (unsigned long long)delayed_node->inode_id, | ||
1299 | ret); | ||
1300 | BUG(); | ||
1301 | } | ||
1302 | mutex_unlock(&delayed_node->mutex); | ||
1303 | |||
1304 | release_node: | ||
1305 | btrfs_release_delayed_node(delayed_node); | ||
1306 | return ret; | ||
1307 | } | ||
1308 | |||
1309 | static int btrfs_delete_delayed_insertion_item(struct btrfs_root *root, | ||
1310 | struct btrfs_delayed_node *node, | ||
1311 | struct btrfs_key *key) | ||
1312 | { | ||
1313 | struct btrfs_delayed_item *item; | ||
1314 | |||
1315 | mutex_lock(&node->mutex); | ||
1316 | item = __btrfs_lookup_delayed_insertion_item(node, key); | ||
1317 | if (!item) { | ||
1318 | mutex_unlock(&node->mutex); | ||
1319 | return 1; | ||
1320 | } | ||
1321 | |||
1322 | btrfs_delayed_item_release_metadata(root, item); | ||
1323 | btrfs_release_delayed_item(item); | ||
1324 | mutex_unlock(&node->mutex); | ||
1325 | return 0; | ||
1326 | } | ||
1327 | |||
1328 | int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
1329 | struct btrfs_root *root, struct inode *dir, | ||
1330 | u64 index) | ||
1331 | { | ||
1332 | struct btrfs_delayed_node *node; | ||
1333 | struct btrfs_delayed_item *item; | ||
1334 | struct btrfs_key item_key; | ||
1335 | int ret; | ||
1336 | |||
1337 | node = btrfs_get_or_create_delayed_node(dir); | ||
1338 | if (IS_ERR(node)) | ||
1339 | return PTR_ERR(node); | ||
1340 | |||
1341 | item_key.objectid = btrfs_ino(dir); | ||
1342 | btrfs_set_key_type(&item_key, BTRFS_DIR_INDEX_KEY); | ||
1343 | item_key.offset = index; | ||
1344 | |||
1345 | ret = btrfs_delete_delayed_insertion_item(root, node, &item_key); | ||
1346 | if (!ret) | ||
1347 | goto end; | ||
1348 | |||
1349 | item = btrfs_alloc_delayed_item(0); | ||
1350 | if (!item) { | ||
1351 | ret = -ENOMEM; | ||
1352 | goto end; | ||
1353 | } | ||
1354 | |||
1355 | item->key = item_key; | ||
1356 | |||
1357 | ret = btrfs_delayed_item_reserve_metadata(trans, root, item); | ||
1358 | /* | ||
1359 | * we have reserved enough space when we start a new transaction, | ||
1360 | * so reserving metadata failure is impossible. | ||
1361 | */ | ||
1362 | BUG_ON(ret); | ||
1363 | |||
1364 | mutex_lock(&node->mutex); | ||
1365 | ret = __btrfs_add_delayed_deletion_item(node, item); | ||
1366 | if (unlikely(ret)) { | ||
1367 | printk(KERN_ERR "err add delayed dir index item(index: %llu) " | ||
1368 | "into the deletion tree of the delayed node" | ||
1369 | "(root id: %llu, inode id: %llu, errno: %d)\n", | ||
1370 | (unsigned long long)index, | ||
1371 | (unsigned long long)node->root->objectid, | ||
1372 | (unsigned long long)node->inode_id, | ||
1373 | ret); | ||
1374 | BUG(); | ||
1375 | } | ||
1376 | mutex_unlock(&node->mutex); | ||
1377 | end: | ||
1378 | btrfs_release_delayed_node(node); | ||
1379 | return ret; | ||
1380 | } | ||
1381 | |||
1382 | int btrfs_inode_delayed_dir_index_count(struct inode *inode) | ||
1383 | { | ||
1384 | struct btrfs_delayed_node *delayed_node = BTRFS_I(inode)->delayed_node; | ||
1385 | int ret = 0; | ||
1386 | |||
1387 | if (!delayed_node) | ||
1388 | return -ENOENT; | ||
1389 | |||
1390 | /* | ||
1391 | * Since we have held i_mutex of this directory, it is impossible that | ||
1392 | * a new directory index is added into the delayed node and index_cnt | ||
1393 | * is updated now. So we needn't lock the delayed node. | ||
1394 | */ | ||
1395 | if (!delayed_node->index_cnt) | ||
1396 | return -EINVAL; | ||
1397 | |||
1398 | BTRFS_I(inode)->index_cnt = delayed_node->index_cnt; | ||
1399 | return ret; | ||
1400 | } | ||
1401 | |||
1402 | void btrfs_get_delayed_items(struct inode *inode, struct list_head *ins_list, | ||
1403 | struct list_head *del_list) | ||
1404 | { | ||
1405 | struct btrfs_delayed_node *delayed_node; | ||
1406 | struct btrfs_delayed_item *item; | ||
1407 | |||
1408 | delayed_node = btrfs_get_delayed_node(inode); | ||
1409 | if (!delayed_node) | ||
1410 | return; | ||
1411 | |||
1412 | mutex_lock(&delayed_node->mutex); | ||
1413 | item = __btrfs_first_delayed_insertion_item(delayed_node); | ||
1414 | while (item) { | ||
1415 | atomic_inc(&item->refs); | ||
1416 | list_add_tail(&item->readdir_list, ins_list); | ||
1417 | item = __btrfs_next_delayed_item(item); | ||
1418 | } | ||
1419 | |||
1420 | item = __btrfs_first_delayed_deletion_item(delayed_node); | ||
1421 | while (item) { | ||
1422 | atomic_inc(&item->refs); | ||
1423 | list_add_tail(&item->readdir_list, del_list); | ||
1424 | item = __btrfs_next_delayed_item(item); | ||
1425 | } | ||
1426 | mutex_unlock(&delayed_node->mutex); | ||
1427 | /* | ||
1428 | * This delayed node is still cached in the btrfs inode, so refs | ||
1429 | * must be > 1 now, and we needn't check it is going to be freed | ||
1430 | * or not. | ||
1431 | * | ||
1432 | * Besides that, this function is used to read dir, we do not | ||
1433 | * insert/delete delayed items in this period. So we also needn't | ||
1434 | * requeue or dequeue this delayed node. | ||
1435 | */ | ||
1436 | atomic_dec(&delayed_node->refs); | ||
1437 | } | ||
1438 | |||
1439 | void btrfs_put_delayed_items(struct list_head *ins_list, | ||
1440 | struct list_head *del_list) | ||
1441 | { | ||
1442 | struct btrfs_delayed_item *curr, *next; | ||
1443 | |||
1444 | list_for_each_entry_safe(curr, next, ins_list, readdir_list) { | ||
1445 | list_del(&curr->readdir_list); | ||
1446 | if (atomic_dec_and_test(&curr->refs)) | ||
1447 | kfree(curr); | ||
1448 | } | ||
1449 | |||
1450 | list_for_each_entry_safe(curr, next, del_list, readdir_list) { | ||
1451 | list_del(&curr->readdir_list); | ||
1452 | if (atomic_dec_and_test(&curr->refs)) | ||
1453 | kfree(curr); | ||
1454 | } | ||
1455 | } | ||
1456 | |||
1457 | int btrfs_should_delete_dir_index(struct list_head *del_list, | ||
1458 | u64 index) | ||
1459 | { | ||
1460 | struct btrfs_delayed_item *curr, *next; | ||
1461 | int ret; | ||
1462 | |||
1463 | if (list_empty(del_list)) | ||
1464 | return 0; | ||
1465 | |||
1466 | list_for_each_entry_safe(curr, next, del_list, readdir_list) { | ||
1467 | if (curr->key.offset > index) | ||
1468 | break; | ||
1469 | |||
1470 | list_del(&curr->readdir_list); | ||
1471 | ret = (curr->key.offset == index); | ||
1472 | |||
1473 | if (atomic_dec_and_test(&curr->refs)) | ||
1474 | kfree(curr); | ||
1475 | |||
1476 | if (ret) | ||
1477 | return 1; | ||
1478 | else | ||
1479 | continue; | ||
1480 | } | ||
1481 | return 0; | ||
1482 | } | ||
1483 | |||
1484 | /* | ||
1485 | * btrfs_readdir_delayed_dir_index - read dir info stored in the delayed tree | ||
1486 | * | ||
1487 | */ | ||
1488 | int btrfs_readdir_delayed_dir_index(struct file *filp, void *dirent, | ||
1489 | filldir_t filldir, | ||
1490 | struct list_head *ins_list) | ||
1491 | { | ||
1492 | struct btrfs_dir_item *di; | ||
1493 | struct btrfs_delayed_item *curr, *next; | ||
1494 | struct btrfs_key location; | ||
1495 | char *name; | ||
1496 | int name_len; | ||
1497 | int over = 0; | ||
1498 | unsigned char d_type; | ||
1499 | |||
1500 | if (list_empty(ins_list)) | ||
1501 | return 0; | ||
1502 | |||
1503 | /* | ||
1504 | * Changing the data of the delayed item is impossible. So | ||
1505 | * we needn't lock them. And we have held i_mutex of the | ||
1506 | * directory, nobody can delete any directory indexes now. | ||
1507 | */ | ||
1508 | list_for_each_entry_safe(curr, next, ins_list, readdir_list) { | ||
1509 | list_del(&curr->readdir_list); | ||
1510 | |||
1511 | if (curr->key.offset < filp->f_pos) { | ||
1512 | if (atomic_dec_and_test(&curr->refs)) | ||
1513 | kfree(curr); | ||
1514 | continue; | ||
1515 | } | ||
1516 | |||
1517 | filp->f_pos = curr->key.offset; | ||
1518 | |||
1519 | di = (struct btrfs_dir_item *)curr->data; | ||
1520 | name = (char *)(di + 1); | ||
1521 | name_len = le16_to_cpu(di->name_len); | ||
1522 | |||
1523 | d_type = btrfs_filetype_table[di->type]; | ||
1524 | btrfs_disk_key_to_cpu(&location, &di->location); | ||
1525 | |||
1526 | over = filldir(dirent, name, name_len, curr->key.offset, | ||
1527 | location.objectid, d_type); | ||
1528 | |||
1529 | if (atomic_dec_and_test(&curr->refs)) | ||
1530 | kfree(curr); | ||
1531 | |||
1532 | if (over) | ||
1533 | return 1; | ||
1534 | } | ||
1535 | return 0; | ||
1536 | } | ||
1537 | |||
1538 | BTRFS_SETGET_STACK_FUNCS(stack_inode_generation, struct btrfs_inode_item, | ||
1539 | generation, 64); | ||
1540 | BTRFS_SETGET_STACK_FUNCS(stack_inode_sequence, struct btrfs_inode_item, | ||
1541 | sequence, 64); | ||
1542 | BTRFS_SETGET_STACK_FUNCS(stack_inode_transid, struct btrfs_inode_item, | ||
1543 | transid, 64); | ||
1544 | BTRFS_SETGET_STACK_FUNCS(stack_inode_size, struct btrfs_inode_item, size, 64); | ||
1545 | BTRFS_SETGET_STACK_FUNCS(stack_inode_nbytes, struct btrfs_inode_item, | ||
1546 | nbytes, 64); | ||
1547 | BTRFS_SETGET_STACK_FUNCS(stack_inode_block_group, struct btrfs_inode_item, | ||
1548 | block_group, 64); | ||
1549 | BTRFS_SETGET_STACK_FUNCS(stack_inode_nlink, struct btrfs_inode_item, nlink, 32); | ||
1550 | BTRFS_SETGET_STACK_FUNCS(stack_inode_uid, struct btrfs_inode_item, uid, 32); | ||
1551 | BTRFS_SETGET_STACK_FUNCS(stack_inode_gid, struct btrfs_inode_item, gid, 32); | ||
1552 | BTRFS_SETGET_STACK_FUNCS(stack_inode_mode, struct btrfs_inode_item, mode, 32); | ||
1553 | BTRFS_SETGET_STACK_FUNCS(stack_inode_rdev, struct btrfs_inode_item, rdev, 64); | ||
1554 | BTRFS_SETGET_STACK_FUNCS(stack_inode_flags, struct btrfs_inode_item, flags, 64); | ||
1555 | |||
1556 | BTRFS_SETGET_STACK_FUNCS(stack_timespec_sec, struct btrfs_timespec, sec, 64); | ||
1557 | BTRFS_SETGET_STACK_FUNCS(stack_timespec_nsec, struct btrfs_timespec, nsec, 32); | ||
1558 | |||
1559 | static void fill_stack_inode_item(struct btrfs_trans_handle *trans, | ||
1560 | struct btrfs_inode_item *inode_item, | ||
1561 | struct inode *inode) | ||
1562 | { | ||
1563 | btrfs_set_stack_inode_uid(inode_item, inode->i_uid); | ||
1564 | btrfs_set_stack_inode_gid(inode_item, inode->i_gid); | ||
1565 | btrfs_set_stack_inode_size(inode_item, BTRFS_I(inode)->disk_i_size); | ||
1566 | btrfs_set_stack_inode_mode(inode_item, inode->i_mode); | ||
1567 | btrfs_set_stack_inode_nlink(inode_item, inode->i_nlink); | ||
1568 | btrfs_set_stack_inode_nbytes(inode_item, inode_get_bytes(inode)); | ||
1569 | btrfs_set_stack_inode_generation(inode_item, | ||
1570 | BTRFS_I(inode)->generation); | ||
1571 | btrfs_set_stack_inode_sequence(inode_item, BTRFS_I(inode)->sequence); | ||
1572 | btrfs_set_stack_inode_transid(inode_item, trans->transid); | ||
1573 | btrfs_set_stack_inode_rdev(inode_item, inode->i_rdev); | ||
1574 | btrfs_set_stack_inode_flags(inode_item, BTRFS_I(inode)->flags); | ||
1575 | btrfs_set_stack_inode_block_group(inode_item, | ||
1576 | BTRFS_I(inode)->block_group); | ||
1577 | |||
1578 | btrfs_set_stack_timespec_sec(btrfs_inode_atime(inode_item), | ||
1579 | inode->i_atime.tv_sec); | ||
1580 | btrfs_set_stack_timespec_nsec(btrfs_inode_atime(inode_item), | ||
1581 | inode->i_atime.tv_nsec); | ||
1582 | |||
1583 | btrfs_set_stack_timespec_sec(btrfs_inode_mtime(inode_item), | ||
1584 | inode->i_mtime.tv_sec); | ||
1585 | btrfs_set_stack_timespec_nsec(btrfs_inode_mtime(inode_item), | ||
1586 | inode->i_mtime.tv_nsec); | ||
1587 | |||
1588 | btrfs_set_stack_timespec_sec(btrfs_inode_ctime(inode_item), | ||
1589 | inode->i_ctime.tv_sec); | ||
1590 | btrfs_set_stack_timespec_nsec(btrfs_inode_ctime(inode_item), | ||
1591 | inode->i_ctime.tv_nsec); | ||
1592 | } | ||
1593 | |||
1594 | int btrfs_delayed_update_inode(struct btrfs_trans_handle *trans, | ||
1595 | struct btrfs_root *root, struct inode *inode) | ||
1596 | { | ||
1597 | struct btrfs_delayed_node *delayed_node; | ||
1598 | int ret; | ||
1599 | |||
1600 | delayed_node = btrfs_get_or_create_delayed_node(inode); | ||
1601 | if (IS_ERR(delayed_node)) | ||
1602 | return PTR_ERR(delayed_node); | ||
1603 | |||
1604 | mutex_lock(&delayed_node->mutex); | ||
1605 | if (delayed_node->inode_dirty) { | ||
1606 | fill_stack_inode_item(trans, &delayed_node->inode_item, inode); | ||
1607 | goto release_node; | ||
1608 | } | ||
1609 | |||
1610 | ret = btrfs_delayed_inode_reserve_metadata(trans, root, delayed_node); | ||
1611 | /* | ||
1612 | * we must reserve enough space when we start a new transaction, | ||
1613 | * so reserving metadata failure is impossible | ||
1614 | */ | ||
1615 | BUG_ON(ret); | ||
1616 | |||
1617 | fill_stack_inode_item(trans, &delayed_node->inode_item, inode); | ||
1618 | delayed_node->inode_dirty = 1; | ||
1619 | delayed_node->count++; | ||
1620 | atomic_inc(&root->fs_info->delayed_root->items); | ||
1621 | release_node: | ||
1622 | mutex_unlock(&delayed_node->mutex); | ||
1623 | btrfs_release_delayed_node(delayed_node); | ||
1624 | return ret; | ||
1625 | } | ||
1626 | |||
1627 | static void __btrfs_kill_delayed_node(struct btrfs_delayed_node *delayed_node) | ||
1628 | { | ||
1629 | struct btrfs_root *root = delayed_node->root; | ||
1630 | struct btrfs_delayed_item *curr_item, *prev_item; | ||
1631 | |||
1632 | mutex_lock(&delayed_node->mutex); | ||
1633 | curr_item = __btrfs_first_delayed_insertion_item(delayed_node); | ||
1634 | while (curr_item) { | ||
1635 | btrfs_delayed_item_release_metadata(root, curr_item); | ||
1636 | prev_item = curr_item; | ||
1637 | curr_item = __btrfs_next_delayed_item(prev_item); | ||
1638 | btrfs_release_delayed_item(prev_item); | ||
1639 | } | ||
1640 | |||
1641 | curr_item = __btrfs_first_delayed_deletion_item(delayed_node); | ||
1642 | while (curr_item) { | ||
1643 | btrfs_delayed_item_release_metadata(root, curr_item); | ||
1644 | prev_item = curr_item; | ||
1645 | curr_item = __btrfs_next_delayed_item(prev_item); | ||
1646 | btrfs_release_delayed_item(prev_item); | ||
1647 | } | ||
1648 | |||
1649 | if (delayed_node->inode_dirty) { | ||
1650 | btrfs_delayed_inode_release_metadata(root, delayed_node); | ||
1651 | btrfs_release_delayed_inode(delayed_node); | ||
1652 | } | ||
1653 | mutex_unlock(&delayed_node->mutex); | ||
1654 | } | ||
1655 | |||
1656 | void btrfs_kill_delayed_inode_items(struct inode *inode) | ||
1657 | { | ||
1658 | struct btrfs_delayed_node *delayed_node; | ||
1659 | |||
1660 | delayed_node = btrfs_get_delayed_node(inode); | ||
1661 | if (!delayed_node) | ||
1662 | return; | ||
1663 | |||
1664 | __btrfs_kill_delayed_node(delayed_node); | ||
1665 | btrfs_release_delayed_node(delayed_node); | ||
1666 | } | ||
1667 | |||
1668 | void btrfs_kill_all_delayed_nodes(struct btrfs_root *root) | ||
1669 | { | ||
1670 | u64 inode_id = 0; | ||
1671 | struct btrfs_delayed_node *delayed_nodes[8]; | ||
1672 | int i, n; | ||
1673 | |||
1674 | while (1) { | ||
1675 | spin_lock(&root->inode_lock); | ||
1676 | n = radix_tree_gang_lookup(&root->delayed_nodes_tree, | ||
1677 | (void **)delayed_nodes, inode_id, | ||
1678 | ARRAY_SIZE(delayed_nodes)); | ||
1679 | if (!n) { | ||
1680 | spin_unlock(&root->inode_lock); | ||
1681 | break; | ||
1682 | } | ||
1683 | |||
1684 | inode_id = delayed_nodes[n - 1]->inode_id + 1; | ||
1685 | |||
1686 | for (i = 0; i < n; i++) | ||
1687 | atomic_inc(&delayed_nodes[i]->refs); | ||
1688 | spin_unlock(&root->inode_lock); | ||
1689 | |||
1690 | for (i = 0; i < n; i++) { | ||
1691 | __btrfs_kill_delayed_node(delayed_nodes[i]); | ||
1692 | btrfs_release_delayed_node(delayed_nodes[i]); | ||
1693 | } | ||
1694 | } | ||
1695 | } | ||
diff --git a/fs/btrfs/delayed-inode.h b/fs/btrfs/delayed-inode.h new file mode 100644 index 000000000000..eb7d240aa648 --- /dev/null +++ b/fs/btrfs/delayed-inode.h | |||
@@ -0,0 +1,141 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2011 Fujitsu. All rights reserved. | ||
3 | * Written by Miao Xie <miaox@cn.fujitsu.com> | ||
4 | * | ||
5 | * This program is free software; you can redistribute it and/or | ||
6 | * modify it under the terms of the GNU General Public | ||
7 | * License v2 as published by the Free Software Foundation. | ||
8 | * | ||
9 | * This program is distributed in the hope that it will be useful, | ||
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | ||
12 | * General Public License for more details. | ||
13 | * | ||
14 | * You should have received a copy of the GNU General Public | ||
15 | * License along with this program; if not, write to the | ||
16 | * Free Software Foundation, Inc., 59 Temple Place - Suite 330, | ||
17 | * Boston, MA 021110-1307, USA. | ||
18 | */ | ||
19 | |||
20 | #ifndef __DELAYED_TREE_OPERATION_H | ||
21 | #define __DELAYED_TREE_OPERATION_H | ||
22 | |||
23 | #include <linux/rbtree.h> | ||
24 | #include <linux/spinlock.h> | ||
25 | #include <linux/mutex.h> | ||
26 | #include <linux/list.h> | ||
27 | #include <linux/wait.h> | ||
28 | #include <asm/atomic.h> | ||
29 | |||
30 | #include "ctree.h" | ||
31 | |||
32 | /* types of the delayed item */ | ||
33 | #define BTRFS_DELAYED_INSERTION_ITEM 1 | ||
34 | #define BTRFS_DELAYED_DELETION_ITEM 2 | ||
35 | |||
36 | struct btrfs_delayed_root { | ||
37 | spinlock_t lock; | ||
38 | struct list_head node_list; | ||
39 | /* | ||
40 | * Used for delayed nodes which is waiting to be dealt with by the | ||
41 | * worker. If the delayed node is inserted into the work queue, we | ||
42 | * drop it from this list. | ||
43 | */ | ||
44 | struct list_head prepare_list; | ||
45 | atomic_t items; /* for delayed items */ | ||
46 | int nodes; /* for delayed nodes */ | ||
47 | wait_queue_head_t wait; | ||
48 | }; | ||
49 | |||
50 | struct btrfs_delayed_node { | ||
51 | u64 inode_id; | ||
52 | u64 bytes_reserved; | ||
53 | struct btrfs_root *root; | ||
54 | /* Used to add the node into the delayed root's node list. */ | ||
55 | struct list_head n_list; | ||
56 | /* | ||
57 | * Used to add the node into the prepare list, the nodes in this list | ||
58 | * is waiting to be dealt with by the async worker. | ||
59 | */ | ||
60 | struct list_head p_list; | ||
61 | struct rb_root ins_root; | ||
62 | struct rb_root del_root; | ||
63 | struct mutex mutex; | ||
64 | struct btrfs_inode_item inode_item; | ||
65 | atomic_t refs; | ||
66 | u64 index_cnt; | ||
67 | bool in_list; | ||
68 | bool inode_dirty; | ||
69 | int count; | ||
70 | }; | ||
71 | |||
72 | struct btrfs_delayed_item { | ||
73 | struct rb_node rb_node; | ||
74 | struct btrfs_key key; | ||
75 | struct list_head tree_list; /* used for batch insert/delete items */ | ||
76 | struct list_head readdir_list; /* used for readdir items */ | ||
77 | u64 bytes_reserved; | ||
78 | struct btrfs_block_rsv *block_rsv; | ||
79 | struct btrfs_delayed_node *delayed_node; | ||
80 | atomic_t refs; | ||
81 | int ins_or_del; | ||
82 | u32 data_len; | ||
83 | char data[0]; | ||
84 | }; | ||
85 | |||
86 | static inline void btrfs_init_delayed_root( | ||
87 | struct btrfs_delayed_root *delayed_root) | ||
88 | { | ||
89 | atomic_set(&delayed_root->items, 0); | ||
90 | delayed_root->nodes = 0; | ||
91 | spin_lock_init(&delayed_root->lock); | ||
92 | init_waitqueue_head(&delayed_root->wait); | ||
93 | INIT_LIST_HEAD(&delayed_root->node_list); | ||
94 | INIT_LIST_HEAD(&delayed_root->prepare_list); | ||
95 | } | ||
96 | |||
97 | int btrfs_insert_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
98 | struct btrfs_root *root, const char *name, | ||
99 | int name_len, struct inode *dir, | ||
100 | struct btrfs_disk_key *disk_key, u8 type, | ||
101 | u64 index); | ||
102 | |||
103 | int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans, | ||
104 | struct btrfs_root *root, struct inode *dir, | ||
105 | u64 index); | ||
106 | |||
107 | int btrfs_inode_delayed_dir_index_count(struct inode *inode); | ||
108 | |||
109 | int btrfs_run_delayed_items(struct btrfs_trans_handle *trans, | ||
110 | struct btrfs_root *root); | ||
111 | |||
112 | void btrfs_balance_delayed_items(struct btrfs_root *root); | ||
113 | |||
114 | int btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans, | ||
115 | struct inode *inode); | ||
116 | /* Used for evicting the inode. */ | ||
117 | void btrfs_remove_delayed_node(struct inode *inode); | ||
118 | void btrfs_kill_delayed_inode_items(struct inode *inode); | ||
119 | |||
120 | |||
121 | int btrfs_delayed_update_inode(struct btrfs_trans_handle *trans, | ||
122 | struct btrfs_root *root, struct inode *inode); | ||
123 | |||
124 | /* Used for drop dead root */ | ||
125 | void btrfs_kill_all_delayed_nodes(struct btrfs_root *root); | ||
126 | |||
127 | /* Used for readdir() */ | ||
128 | void btrfs_get_delayed_items(struct inode *inode, struct list_head *ins_list, | ||
129 | struct list_head *del_list); | ||
130 | void btrfs_put_delayed_items(struct list_head *ins_list, | ||
131 | struct list_head *del_list); | ||
132 | int btrfs_should_delete_dir_index(struct list_head *del_list, | ||
133 | u64 index); | ||
134 | int btrfs_readdir_delayed_dir_index(struct file *filp, void *dirent, | ||
135 | filldir_t filldir, | ||
136 | struct list_head *ins_list); | ||
137 | |||
138 | /* for init */ | ||
139 | int __init btrfs_delayed_inode_init(void); | ||
140 | void btrfs_delayed_inode_exit(void); | ||
141 | #endif | ||
diff --git a/fs/btrfs/dir-item.c b/fs/btrfs/dir-item.c index ab8afed671a0..1ddfca78e910 100644 --- a/fs/btrfs/dir-item.c +++ b/fs/btrfs/dir-item.c | |||
@@ -124,8 +124,9 @@ int btrfs_insert_xattr_item(struct btrfs_trans_handle *trans, | |||
124 | * to use for the second index (if one is created). | 124 | * to use for the second index (if one is created). |
125 | */ | 125 | */ |
126 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root | 126 | int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root |
127 | *root, const char *name, int name_len, u64 dir, | 127 | *root, const char *name, int name_len, |
128 | struct btrfs_key *location, u8 type, u64 index) | 128 | struct inode *dir, struct btrfs_key *location, |
129 | u8 type, u64 index) | ||
129 | { | 130 | { |
130 | int ret = 0; | 131 | int ret = 0; |
131 | int ret2 = 0; | 132 | int ret2 = 0; |
@@ -137,13 +138,17 @@ int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root | |||
137 | struct btrfs_disk_key disk_key; | 138 | struct btrfs_disk_key disk_key; |
138 | u32 data_size; | 139 | u32 data_size; |
139 | 140 | ||
140 | key.objectid = dir; | 141 | key.objectid = btrfs_ino(dir); |
141 | btrfs_set_key_type(&key, BTRFS_DIR_ITEM_KEY); | 142 | btrfs_set_key_type(&key, BTRFS_DIR_ITEM_KEY); |
142 | key.offset = btrfs_name_hash(name, name_len); | 143 | key.offset = btrfs_name_hash(name, name_len); |
143 | 144 | ||
144 | path = btrfs_alloc_path(); | 145 | path = btrfs_alloc_path(); |
146 | if (!path) | ||
147 | return -ENOMEM; | ||
145 | path->leave_spinning = 1; | 148 | path->leave_spinning = 1; |
146 | 149 | ||
150 | btrfs_cpu_key_to_disk(&disk_key, location); | ||
151 | |||
147 | data_size = sizeof(*dir_item) + name_len; | 152 | data_size = sizeof(*dir_item) + name_len; |
148 | dir_item = insert_with_overflow(trans, root, path, &key, data_size, | 153 | dir_item = insert_with_overflow(trans, root, path, &key, data_size, |
149 | name, name_len); | 154 | name, name_len); |
@@ -155,7 +160,6 @@ int btrfs_insert_dir_item(struct btrfs_trans_handle *trans, struct btrfs_root | |||
155 | } | 160 | } |
156 | 161 | ||
157 | leaf = path->nodes[0]; | 162 | leaf = path->nodes[0]; |
158 | btrfs_cpu_key_to_disk(&disk_key, location); | ||
159 | btrfs_set_dir_item_key(leaf, dir_item, &disk_key); | 163 | btrfs_set_dir_item_key(leaf, dir_item, &disk_key); |
160 | btrfs_set_dir_type(leaf, dir_item, type); | 164 | btrfs_set_dir_type(leaf, dir_item, type); |
161 | btrfs_set_dir_data_len(leaf, dir_item, 0); | 165 | btrfs_set_dir_data_len(leaf, dir_item, 0); |
@@ -174,27 +178,9 @@ second_insert: | |||
174 | } | 178 | } |
175 | btrfs_release_path(path); | 179 | btrfs_release_path(path); |
176 | 180 | ||
177 | btrfs_set_key_type(&key, BTRFS_DIR_INDEX_KEY); | 181 | ret2 = btrfs_insert_delayed_dir_index(trans, root, name, name_len, dir, |
178 | key.offset = index; | 182 | &disk_key, type, index); |
179 | dir_item = insert_with_overflow(trans, root, path, &key, data_size, | ||
180 | name, name_len); | ||
181 | if (IS_ERR(dir_item)) { | ||
182 | ret2 = PTR_ERR(dir_item); | ||
183 | goto out_free; | ||
184 | } | ||
185 | leaf = path->nodes[0]; | ||
186 | btrfs_cpu_key_to_disk(&disk_key, location); | ||
187 | btrfs_set_dir_item_key(leaf, dir_item, &disk_key); | ||
188 | btrfs_set_dir_type(leaf, dir_item, type); | ||
189 | btrfs_set_dir_data_len(leaf, dir_item, 0); | ||
190 | btrfs_set_dir_name_len(leaf, dir_item, name_len); | ||
191 | btrfs_set_dir_transid(leaf, dir_item, trans->transid); | ||
192 | name_ptr = (unsigned long)(dir_item + 1); | ||
193 | write_extent_buffer(leaf, name, name_ptr, name_len); | ||
194 | btrfs_mark_buffer_dirty(leaf); | ||
195 | |||
196 | out_free: | 183 | out_free: |
197 | |||
198 | btrfs_free_path(path); | 184 | btrfs_free_path(path); |
199 | if (ret) | 185 | if (ret) |
200 | return ret; | 186 | return ret; |
diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c index cb9d1b8bfe74..a2eb3a3755db 100644 --- a/fs/btrfs/disk-io.c +++ b/fs/btrfs/disk-io.c | |||
@@ -42,6 +42,7 @@ | |||
42 | #include "locking.h" | 42 | #include "locking.h" |
43 | #include "tree-log.h" | 43 | #include "tree-log.h" |
44 | #include "free-space-cache.h" | 44 | #include "free-space-cache.h" |
45 | #include "inode-map.h" | ||
45 | 46 | ||
46 | static struct extent_io_ops btree_extent_io_ops; | 47 | static struct extent_io_ops btree_extent_io_ops; |
47 | static void end_workqueue_fn(struct btrfs_work *work); | 48 | static void end_workqueue_fn(struct btrfs_work *work); |
@@ -1045,6 +1046,7 @@ static int __setup_root(u32 nodesize, u32 leafsize, u32 sectorsize, | |||
1045 | root->name = NULL; | 1046 | root->name = NULL; |
1046 | root->in_sysfs = 0; | 1047 | root->in_sysfs = 0; |
1047 | root->inode_tree = RB_ROOT; | 1048 | root->inode_tree = RB_ROOT; |
1049 | INIT_RADIX_TREE(&root->delayed_nodes_tree, GFP_ATOMIC); | ||
1048 | root->block_rsv = NULL; | 1050 | root->block_rsv = NULL; |
1049 | root->orphan_block_rsv = NULL; | 1051 | root->orphan_block_rsv = NULL; |
1050 | 1052 | ||
@@ -1298,6 +1300,19 @@ again: | |||
1298 | if (IS_ERR(root)) | 1300 | if (IS_ERR(root)) |
1299 | return root; | 1301 | return root; |
1300 | 1302 | ||
1303 | root->free_ino_ctl = kzalloc(sizeof(*root->free_ino_ctl), GFP_NOFS); | ||
1304 | if (!root->free_ino_ctl) | ||
1305 | goto fail; | ||
1306 | root->free_ino_pinned = kzalloc(sizeof(*root->free_ino_pinned), | ||
1307 | GFP_NOFS); | ||
1308 | if (!root->free_ino_pinned) | ||
1309 | goto fail; | ||
1310 | |||
1311 | btrfs_init_free_ino_ctl(root); | ||
1312 | mutex_init(&root->fs_commit_mutex); | ||
1313 | spin_lock_init(&root->cache_lock); | ||
1314 | init_waitqueue_head(&root->cache_wait); | ||
1315 | |||
1301 | set_anon_super(&root->anon_super, NULL); | 1316 | set_anon_super(&root->anon_super, NULL); |
1302 | 1317 | ||
1303 | if (btrfs_root_refs(&root->root_item) == 0) { | 1318 | if (btrfs_root_refs(&root->root_item) == 0) { |
@@ -1631,6 +1646,13 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1631 | 1646 | ||
1632 | INIT_LIST_HEAD(&fs_info->ordered_extents); | 1647 | INIT_LIST_HEAD(&fs_info->ordered_extents); |
1633 | spin_lock_init(&fs_info->ordered_extent_lock); | 1648 | spin_lock_init(&fs_info->ordered_extent_lock); |
1649 | fs_info->delayed_root = kmalloc(sizeof(struct btrfs_delayed_root), | ||
1650 | GFP_NOFS); | ||
1651 | if (!fs_info->delayed_root) { | ||
1652 | err = -ENOMEM; | ||
1653 | goto fail_iput; | ||
1654 | } | ||
1655 | btrfs_init_delayed_root(fs_info->delayed_root); | ||
1634 | 1656 | ||
1635 | sb->s_blocksize = 4096; | 1657 | sb->s_blocksize = 4096; |
1636 | sb->s_blocksize_bits = blksize_bits(4096); | 1658 | sb->s_blocksize_bits = blksize_bits(4096); |
@@ -1696,7 +1718,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1696 | bh = btrfs_read_dev_super(fs_devices->latest_bdev); | 1718 | bh = btrfs_read_dev_super(fs_devices->latest_bdev); |
1697 | if (!bh) { | 1719 | if (!bh) { |
1698 | err = -EINVAL; | 1720 | err = -EINVAL; |
1699 | goto fail_iput; | 1721 | goto fail_alloc; |
1700 | } | 1722 | } |
1701 | 1723 | ||
1702 | memcpy(&fs_info->super_copy, bh->b_data, sizeof(fs_info->super_copy)); | 1724 | memcpy(&fs_info->super_copy, bh->b_data, sizeof(fs_info->super_copy)); |
@@ -1708,7 +1730,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1708 | 1730 | ||
1709 | disk_super = &fs_info->super_copy; | 1731 | disk_super = &fs_info->super_copy; |
1710 | if (!btrfs_super_root(disk_super)) | 1732 | if (!btrfs_super_root(disk_super)) |
1711 | goto fail_iput; | 1733 | goto fail_alloc; |
1712 | 1734 | ||
1713 | /* check FS state, whether FS is broken. */ | 1735 | /* check FS state, whether FS is broken. */ |
1714 | fs_info->fs_state |= btrfs_super_flags(disk_super); | 1736 | fs_info->fs_state |= btrfs_super_flags(disk_super); |
@@ -1724,7 +1746,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1724 | ret = btrfs_parse_options(tree_root, options); | 1746 | ret = btrfs_parse_options(tree_root, options); |
1725 | if (ret) { | 1747 | if (ret) { |
1726 | err = ret; | 1748 | err = ret; |
1727 | goto fail_iput; | 1749 | goto fail_alloc; |
1728 | } | 1750 | } |
1729 | 1751 | ||
1730 | features = btrfs_super_incompat_flags(disk_super) & | 1752 | features = btrfs_super_incompat_flags(disk_super) & |
@@ -1734,7 +1756,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1734 | "unsupported optional features (%Lx).\n", | 1756 | "unsupported optional features (%Lx).\n", |
1735 | (unsigned long long)features); | 1757 | (unsigned long long)features); |
1736 | err = -EINVAL; | 1758 | err = -EINVAL; |
1737 | goto fail_iput; | 1759 | goto fail_alloc; |
1738 | } | 1760 | } |
1739 | 1761 | ||
1740 | features = btrfs_super_incompat_flags(disk_super); | 1762 | features = btrfs_super_incompat_flags(disk_super); |
@@ -1750,7 +1772,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1750 | "unsupported option features (%Lx).\n", | 1772 | "unsupported option features (%Lx).\n", |
1751 | (unsigned long long)features); | 1773 | (unsigned long long)features); |
1752 | err = -EINVAL; | 1774 | err = -EINVAL; |
1753 | goto fail_iput; | 1775 | goto fail_alloc; |
1754 | } | 1776 | } |
1755 | 1777 | ||
1756 | btrfs_init_workers(&fs_info->generic_worker, | 1778 | btrfs_init_workers(&fs_info->generic_worker, |
@@ -1797,6 +1819,9 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1797 | &fs_info->generic_worker); | 1819 | &fs_info->generic_worker); |
1798 | btrfs_init_workers(&fs_info->endio_freespace_worker, "freespace-write", | 1820 | btrfs_init_workers(&fs_info->endio_freespace_worker, "freespace-write", |
1799 | 1, &fs_info->generic_worker); | 1821 | 1, &fs_info->generic_worker); |
1822 | btrfs_init_workers(&fs_info->delayed_workers, "delayed-meta", | ||
1823 | fs_info->thread_pool_size, | ||
1824 | &fs_info->generic_worker); | ||
1800 | 1825 | ||
1801 | /* | 1826 | /* |
1802 | * endios are largely parallel and should have a very | 1827 | * endios are largely parallel and should have a very |
@@ -1818,6 +1843,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, | |||
1818 | btrfs_start_workers(&fs_info->endio_meta_write_workers, 1); | 1843 | btrfs_start_workers(&fs_info->endio_meta_write_workers, 1); |
1819 | btrfs_start_workers(&fs_info->endio_write_workers, 1); | 1844 | btrfs_start_workers(&fs_info->endio_write_workers, 1); |
1820 | btrfs_start_workers(&fs_info->endio_freespace_worker, 1); | 1845 | btrfs_start_workers(&fs_info->endio_freespace_worker, 1); |
1846 | btrfs_start_workers(&fs_info->delayed_workers, 1); | ||
1821 | 1847 | ||
1822 | fs_info->bdi.ra_pages *= btrfs_super_num_devices(disk_super); | 1848 | fs_info->bdi.ra_pages *= btrfs_super_num_devices(disk_super); |
1823 | fs_info->bdi.ra_pages = max(fs_info->bdi.ra_pages, | 1849 | fs_info->bdi.ra_pages = max(fs_info->bdi.ra_pages, |
@@ -2074,6 +2100,9 @@ fail_sb_buffer: | |||
2074 | btrfs_stop_workers(&fs_info->endio_write_workers); | 2100 | btrfs_stop_workers(&fs_info->endio_write_workers); |
2075 | btrfs_stop_workers(&fs_info->endio_freespace_worker); | 2101 | btrfs_stop_workers(&fs_info->endio_freespace_worker); |
2076 | btrfs_stop_workers(&fs_info->submit_workers); | 2102 | btrfs_stop_workers(&fs_info->submit_workers); |
2103 | btrfs_stop_workers(&fs_info->delayed_workers); | ||
2104 | fail_alloc: | ||
2105 | kfree(fs_info->delayed_root); | ||
2077 | fail_iput: | 2106 | fail_iput: |
2078 | invalidate_inode_pages2(fs_info->btree_inode->i_mapping); | 2107 | invalidate_inode_pages2(fs_info->btree_inode->i_mapping); |
2079 | iput(fs_info->btree_inode); | 2108 | iput(fs_info->btree_inode); |
@@ -2338,12 +2367,15 @@ int btrfs_free_fs_root(struct btrfs_fs_info *fs_info, struct btrfs_root *root) | |||
2338 | if (btrfs_root_refs(&root->root_item) == 0) | 2367 | if (btrfs_root_refs(&root->root_item) == 0) |
2339 | synchronize_srcu(&fs_info->subvol_srcu); | 2368 | synchronize_srcu(&fs_info->subvol_srcu); |
2340 | 2369 | ||
2370 | __btrfs_remove_free_space_cache(root->free_ino_pinned); | ||
2371 | __btrfs_remove_free_space_cache(root->free_ino_ctl); | ||
2341 | free_fs_root(root); | 2372 | free_fs_root(root); |
2342 | return 0; | 2373 | return 0; |
2343 | } | 2374 | } |
2344 | 2375 | ||
2345 | static void free_fs_root(struct btrfs_root *root) | 2376 | static void free_fs_root(struct btrfs_root *root) |
2346 | { | 2377 | { |
2378 | iput(root->cache_inode); | ||
2347 | WARN_ON(!RB_EMPTY_ROOT(&root->inode_tree)); | 2379 | WARN_ON(!RB_EMPTY_ROOT(&root->inode_tree)); |
2348 | if (root->anon_super.s_dev) { | 2380 | if (root->anon_super.s_dev) { |
2349 | down_write(&root->anon_super.s_umount); | 2381 | down_write(&root->anon_super.s_umount); |
@@ -2351,6 +2383,8 @@ static void free_fs_root(struct btrfs_root *root) | |||
2351 | } | 2383 | } |
2352 | free_extent_buffer(root->node); | 2384 | free_extent_buffer(root->node); |
2353 | free_extent_buffer(root->commit_root); | 2385 | free_extent_buffer(root->commit_root); |
2386 | kfree(root->free_ino_ctl); | ||
2387 | kfree(root->free_ino_pinned); | ||
2354 | kfree(root->name); | 2388 | kfree(root->name); |
2355 | kfree(root); | 2389 | kfree(root); |
2356 | } | 2390 | } |
@@ -2512,6 +2546,7 @@ int close_ctree(struct btrfs_root *root) | |||
2512 | del_fs_roots(fs_info); | 2546 | del_fs_roots(fs_info); |
2513 | 2547 | ||
2514 | iput(fs_info->btree_inode); | 2548 | iput(fs_info->btree_inode); |
2549 | kfree(fs_info->delayed_root); | ||
2515 | 2550 | ||
2516 | btrfs_stop_workers(&fs_info->generic_worker); | 2551 | btrfs_stop_workers(&fs_info->generic_worker); |
2517 | btrfs_stop_workers(&fs_info->fixup_workers); | 2552 | btrfs_stop_workers(&fs_info->fixup_workers); |
@@ -2523,6 +2558,7 @@ int close_ctree(struct btrfs_root *root) | |||
2523 | btrfs_stop_workers(&fs_info->endio_write_workers); | 2558 | btrfs_stop_workers(&fs_info->endio_write_workers); |
2524 | btrfs_stop_workers(&fs_info->endio_freespace_worker); | 2559 | btrfs_stop_workers(&fs_info->endio_freespace_worker); |
2525 | btrfs_stop_workers(&fs_info->submit_workers); | 2560 | btrfs_stop_workers(&fs_info->submit_workers); |
2561 | btrfs_stop_workers(&fs_info->delayed_workers); | ||
2526 | 2562 | ||
2527 | btrfs_close_devices(fs_info->fs_devices); | 2563 | btrfs_close_devices(fs_info->fs_devices); |
2528 | btrfs_mapping_tree_free(&fs_info->mapping_tree); | 2564 | btrfs_mapping_tree_free(&fs_info->mapping_tree); |
@@ -2599,6 +2635,29 @@ void btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr) | |||
2599 | if (current->flags & PF_MEMALLOC) | 2635 | if (current->flags & PF_MEMALLOC) |
2600 | return; | 2636 | return; |
2601 | 2637 | ||
2638 | btrfs_balance_delayed_items(root); | ||
2639 | |||
2640 | num_dirty = root->fs_info->dirty_metadata_bytes; | ||
2641 | |||
2642 | if (num_dirty > thresh) { | ||
2643 | balance_dirty_pages_ratelimited_nr( | ||
2644 | root->fs_info->btree_inode->i_mapping, 1); | ||
2645 | } | ||
2646 | return; | ||
2647 | } | ||
2648 | |||
2649 | void __btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr) | ||
2650 | { | ||
2651 | /* | ||
2652 | * looks as though older kernels can get into trouble with | ||
2653 | * this code, they end up stuck in balance_dirty_pages forever | ||
2654 | */ | ||
2655 | u64 num_dirty; | ||
2656 | unsigned long thresh = 32 * 1024 * 1024; | ||
2657 | |||
2658 | if (current->flags & PF_MEMALLOC) | ||
2659 | return; | ||
2660 | |||
2602 | num_dirty = root->fs_info->dirty_metadata_bytes; | 2661 | num_dirty = root->fs_info->dirty_metadata_bytes; |
2603 | 2662 | ||
2604 | if (num_dirty > thresh) { | 2663 | if (num_dirty > thresh) { |
diff --git a/fs/btrfs/disk-io.h b/fs/btrfs/disk-io.h index 2d75f9e896f6..a0b610a67aae 100644 --- a/fs/btrfs/disk-io.h +++ b/fs/btrfs/disk-io.h | |||
@@ -61,6 +61,7 @@ struct btrfs_root *btrfs_read_fs_root_no_name(struct btrfs_fs_info *fs_info, | |||
61 | struct btrfs_key *location); | 61 | struct btrfs_key *location); |
62 | int btrfs_cleanup_fs_roots(struct btrfs_fs_info *fs_info); | 62 | int btrfs_cleanup_fs_roots(struct btrfs_fs_info *fs_info); |
63 | void btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr); | 63 | void btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr); |
64 | void __btrfs_btree_balance_dirty(struct btrfs_root *root, unsigned long nr); | ||
64 | int btrfs_free_fs_root(struct btrfs_fs_info *fs_info, struct btrfs_root *root); | 65 | int btrfs_free_fs_root(struct btrfs_fs_info *fs_info, struct btrfs_root *root); |
65 | void btrfs_mark_buffer_dirty(struct extent_buffer *buf); | 66 | void btrfs_mark_buffer_dirty(struct extent_buffer *buf); |
66 | int btrfs_buffer_uptodate(struct extent_buffer *buf, u64 parent_transid); | 67 | int btrfs_buffer_uptodate(struct extent_buffer *buf, u64 parent_transid); |
diff --git a/fs/btrfs/export.c b/fs/btrfs/export.c index b4ffad859adb..1b8dc33778f9 100644 --- a/fs/btrfs/export.c +++ b/fs/btrfs/export.c | |||
@@ -32,7 +32,7 @@ static int btrfs_encode_fh(struct dentry *dentry, u32 *fh, int *max_len, | |||
32 | len = BTRFS_FID_SIZE_NON_CONNECTABLE; | 32 | len = BTRFS_FID_SIZE_NON_CONNECTABLE; |
33 | type = FILEID_BTRFS_WITHOUT_PARENT; | 33 | type = FILEID_BTRFS_WITHOUT_PARENT; |
34 | 34 | ||
35 | fid->objectid = inode->i_ino; | 35 | fid->objectid = btrfs_ino(inode); |
36 | fid->root_objectid = BTRFS_I(inode)->root->objectid; | 36 | fid->root_objectid = BTRFS_I(inode)->root->objectid; |
37 | fid->gen = inode->i_generation; | 37 | fid->gen = inode->i_generation; |
38 | 38 | ||
@@ -178,13 +178,13 @@ static struct dentry *btrfs_get_parent(struct dentry *child) | |||
178 | if (!path) | 178 | if (!path) |
179 | return ERR_PTR(-ENOMEM); | 179 | return ERR_PTR(-ENOMEM); |
180 | 180 | ||
181 | if (dir->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 181 | if (btrfs_ino(dir) == BTRFS_FIRST_FREE_OBJECTID) { |
182 | key.objectid = root->root_key.objectid; | 182 | key.objectid = root->root_key.objectid; |
183 | key.type = BTRFS_ROOT_BACKREF_KEY; | 183 | key.type = BTRFS_ROOT_BACKREF_KEY; |
184 | key.offset = (u64)-1; | 184 | key.offset = (u64)-1; |
185 | root = root->fs_info->tree_root; | 185 | root = root->fs_info->tree_root; |
186 | } else { | 186 | } else { |
187 | key.objectid = dir->i_ino; | 187 | key.objectid = btrfs_ino(dir); |
188 | key.type = BTRFS_INODE_REF_KEY; | 188 | key.type = BTRFS_INODE_REF_KEY; |
189 | key.offset = (u64)-1; | 189 | key.offset = (u64)-1; |
190 | } | 190 | } |
@@ -244,6 +244,7 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
244 | struct btrfs_key key; | 244 | struct btrfs_key key; |
245 | int name_len; | 245 | int name_len; |
246 | int ret; | 246 | int ret; |
247 | u64 ino; | ||
247 | 248 | ||
248 | if (!dir || !inode) | 249 | if (!dir || !inode) |
249 | return -EINVAL; | 250 | return -EINVAL; |
@@ -251,19 +252,21 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
251 | if (!S_ISDIR(dir->i_mode)) | 252 | if (!S_ISDIR(dir->i_mode)) |
252 | return -EINVAL; | 253 | return -EINVAL; |
253 | 254 | ||
255 | ino = btrfs_ino(inode); | ||
256 | |||
254 | path = btrfs_alloc_path(); | 257 | path = btrfs_alloc_path(); |
255 | if (!path) | 258 | if (!path) |
256 | return -ENOMEM; | 259 | return -ENOMEM; |
257 | path->leave_spinning = 1; | 260 | path->leave_spinning = 1; |
258 | 261 | ||
259 | if (inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 262 | if (ino == BTRFS_FIRST_FREE_OBJECTID) { |
260 | key.objectid = BTRFS_I(inode)->root->root_key.objectid; | 263 | key.objectid = BTRFS_I(inode)->root->root_key.objectid; |
261 | key.type = BTRFS_ROOT_BACKREF_KEY; | 264 | key.type = BTRFS_ROOT_BACKREF_KEY; |
262 | key.offset = (u64)-1; | 265 | key.offset = (u64)-1; |
263 | root = root->fs_info->tree_root; | 266 | root = root->fs_info->tree_root; |
264 | } else { | 267 | } else { |
265 | key.objectid = inode->i_ino; | 268 | key.objectid = ino; |
266 | key.offset = dir->i_ino; | 269 | key.offset = btrfs_ino(dir); |
267 | key.type = BTRFS_INODE_REF_KEY; | 270 | key.type = BTRFS_INODE_REF_KEY; |
268 | } | 271 | } |
269 | 272 | ||
@@ -272,7 +275,7 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
272 | btrfs_free_path(path); | 275 | btrfs_free_path(path); |
273 | return ret; | 276 | return ret; |
274 | } else if (ret > 0) { | 277 | } else if (ret > 0) { |
275 | if (inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 278 | if (ino == BTRFS_FIRST_FREE_OBJECTID) { |
276 | path->slots[0]--; | 279 | path->slots[0]--; |
277 | } else { | 280 | } else { |
278 | btrfs_free_path(path); | 281 | btrfs_free_path(path); |
@@ -281,11 +284,11 @@ static int btrfs_get_name(struct dentry *parent, char *name, | |||
281 | } | 284 | } |
282 | leaf = path->nodes[0]; | 285 | leaf = path->nodes[0]; |
283 | 286 | ||
284 | if (inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) { | 287 | if (ino == BTRFS_FIRST_FREE_OBJECTID) { |
285 | rref = btrfs_item_ptr(leaf, path->slots[0], | 288 | rref = btrfs_item_ptr(leaf, path->slots[0], |
286 | struct btrfs_root_ref); | 289 | struct btrfs_root_ref); |
287 | name_ptr = (unsigned long)(rref + 1); | 290 | name_ptr = (unsigned long)(rref + 1); |
288 | name_len = btrfs_root_ref_name_len(leaf, rref); | 291 | name_len = btrfs_root_ref_name_len(leaf, rref); |
289 | } else { | 292 | } else { |
290 | iref = btrfs_item_ptr(leaf, path->slots[0], | 293 | iref = btrfs_item_ptr(leaf, path->slots[0], |
291 | struct btrfs_inode_ref); | 294 | struct btrfs_inode_ref); |
diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c index b457f195636e..98ca149bdbc8 100644 --- a/fs/btrfs/extent-tree.c +++ b/fs/btrfs/extent-tree.c | |||
@@ -105,6 +105,7 @@ void btrfs_put_block_group(struct btrfs_block_group_cache *cache) | |||
105 | WARN_ON(cache->pinned > 0); | 105 | WARN_ON(cache->pinned > 0); |
106 | WARN_ON(cache->reserved > 0); | 106 | WARN_ON(cache->reserved > 0); |
107 | WARN_ON(cache->reserved_pinned > 0); | 107 | WARN_ON(cache->reserved_pinned > 0); |
108 | kfree(cache->free_space_ctl); | ||
108 | kfree(cache); | 109 | kfree(cache); |
109 | } | 110 | } |
110 | } | 111 | } |
@@ -3036,7 +3037,8 @@ int btrfs_check_data_free_space(struct inode *inode, u64 bytes) | |||
3036 | /* make sure bytes are sectorsize aligned */ | 3037 | /* make sure bytes are sectorsize aligned */ |
3037 | bytes = (bytes + root->sectorsize - 1) & ~((u64)root->sectorsize - 1); | 3038 | bytes = (bytes + root->sectorsize - 1) & ~((u64)root->sectorsize - 1); |
3038 | 3039 | ||
3039 | if (root == root->fs_info->tree_root) { | 3040 | if (root == root->fs_info->tree_root || |
3041 | BTRFS_I(inode)->location.objectid == BTRFS_FREE_INO_OBJECTID) { | ||
3040 | alloc_chunk = 0; | 3042 | alloc_chunk = 0; |
3041 | committed = 1; | 3043 | committed = 1; |
3042 | } | 3044 | } |
@@ -3834,12 +3836,6 @@ static void release_global_block_rsv(struct btrfs_fs_info *fs_info) | |||
3834 | WARN_ON(fs_info->chunk_block_rsv.reserved > 0); | 3836 | WARN_ON(fs_info->chunk_block_rsv.reserved > 0); |
3835 | } | 3837 | } |
3836 | 3838 | ||
3837 | static u64 calc_trans_metadata_size(struct btrfs_root *root, int num_items) | ||
3838 | { | ||
3839 | return (root->leafsize + root->nodesize * (BTRFS_MAX_LEVEL - 1)) * | ||
3840 | 3 * num_items; | ||
3841 | } | ||
3842 | |||
3843 | int btrfs_trans_reserve_metadata(struct btrfs_trans_handle *trans, | 3839 | int btrfs_trans_reserve_metadata(struct btrfs_trans_handle *trans, |
3844 | struct btrfs_root *root, | 3840 | struct btrfs_root *root, |
3845 | int num_items) | 3841 | int num_items) |
@@ -3850,7 +3846,7 @@ int btrfs_trans_reserve_metadata(struct btrfs_trans_handle *trans, | |||
3850 | if (num_items == 0 || root->fs_info->chunk_root == root) | 3846 | if (num_items == 0 || root->fs_info->chunk_root == root) |
3851 | return 0; | 3847 | return 0; |
3852 | 3848 | ||
3853 | num_bytes = calc_trans_metadata_size(root, num_items); | 3849 | num_bytes = btrfs_calc_trans_metadata_size(root, num_items); |
3854 | ret = btrfs_block_rsv_add(trans, root, &root->fs_info->trans_block_rsv, | 3850 | ret = btrfs_block_rsv_add(trans, root, &root->fs_info->trans_block_rsv, |
3855 | num_bytes); | 3851 | num_bytes); |
3856 | if (!ret) { | 3852 | if (!ret) { |
@@ -3889,14 +3885,14 @@ int btrfs_orphan_reserve_metadata(struct btrfs_trans_handle *trans, | |||
3889 | * If all of the metadata space is used, we can commit | 3885 | * If all of the metadata space is used, we can commit |
3890 | * transaction and use space it freed. | 3886 | * transaction and use space it freed. |
3891 | */ | 3887 | */ |
3892 | u64 num_bytes = calc_trans_metadata_size(root, 4); | 3888 | u64 num_bytes = btrfs_calc_trans_metadata_size(root, 4); |
3893 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); | 3889 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); |
3894 | } | 3890 | } |
3895 | 3891 | ||
3896 | void btrfs_orphan_release_metadata(struct inode *inode) | 3892 | void btrfs_orphan_release_metadata(struct inode *inode) |
3897 | { | 3893 | { |
3898 | struct btrfs_root *root = BTRFS_I(inode)->root; | 3894 | struct btrfs_root *root = BTRFS_I(inode)->root; |
3899 | u64 num_bytes = calc_trans_metadata_size(root, 4); | 3895 | u64 num_bytes = btrfs_calc_trans_metadata_size(root, 4); |
3900 | btrfs_block_rsv_release(root, root->orphan_block_rsv, num_bytes); | 3896 | btrfs_block_rsv_release(root, root->orphan_block_rsv, num_bytes); |
3901 | } | 3897 | } |
3902 | 3898 | ||
@@ -3910,7 +3906,7 @@ int btrfs_snap_reserve_metadata(struct btrfs_trans_handle *trans, | |||
3910 | * two for root back/forward refs, two for directory entries | 3906 | * two for root back/forward refs, two for directory entries |
3911 | * and one for root of the snapshot. | 3907 | * and one for root of the snapshot. |
3912 | */ | 3908 | */ |
3913 | u64 num_bytes = calc_trans_metadata_size(root, 5); | 3909 | u64 num_bytes = btrfs_calc_trans_metadata_size(root, 5); |
3914 | dst_rsv->space_info = src_rsv->space_info; | 3910 | dst_rsv->space_info = src_rsv->space_info; |
3915 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); | 3911 | return block_rsv_migrate_bytes(src_rsv, dst_rsv, num_bytes); |
3916 | } | 3912 | } |
@@ -3939,7 +3935,7 @@ int btrfs_delalloc_reserve_metadata(struct inode *inode, u64 num_bytes) | |||
3939 | 3935 | ||
3940 | if (nr_extents > reserved_extents) { | 3936 | if (nr_extents > reserved_extents) { |
3941 | nr_extents -= reserved_extents; | 3937 | nr_extents -= reserved_extents; |
3942 | to_reserve = calc_trans_metadata_size(root, nr_extents); | 3938 | to_reserve = btrfs_calc_trans_metadata_size(root, nr_extents); |
3943 | } else { | 3939 | } else { |
3944 | nr_extents = 0; | 3940 | nr_extents = 0; |
3945 | to_reserve = 0; | 3941 | to_reserve = 0; |
@@ -3993,7 +3989,7 @@ void btrfs_delalloc_release_metadata(struct inode *inode, u64 num_bytes) | |||
3993 | 3989 | ||
3994 | to_free = calc_csum_metadata_size(inode, num_bytes); | 3990 | to_free = calc_csum_metadata_size(inode, num_bytes); |
3995 | if (nr_extents > 0) | 3991 | if (nr_extents > 0) |
3996 | to_free += calc_trans_metadata_size(root, nr_extents); | 3992 | to_free += btrfs_calc_trans_metadata_size(root, nr_extents); |
3997 | 3993 | ||
3998 | btrfs_block_rsv_release(root, &root->fs_info->delalloc_block_rsv, | 3994 | btrfs_block_rsv_release(root, &root->fs_info->delalloc_block_rsv, |
3999 | to_free); | 3995 | to_free); |
@@ -4754,7 +4750,7 @@ wait_block_group_cache_progress(struct btrfs_block_group_cache *cache, | |||
4754 | return 0; | 4750 | return 0; |
4755 | 4751 | ||
4756 | wait_event(caching_ctl->wait, block_group_cache_done(cache) || | 4752 | wait_event(caching_ctl->wait, block_group_cache_done(cache) || |
4757 | (cache->free_space >= num_bytes)); | 4753 | (cache->free_space_ctl->free_space >= num_bytes)); |
4758 | 4754 | ||
4759 | put_caching_control(caching_ctl); | 4755 | put_caching_control(caching_ctl); |
4760 | return 0; | 4756 | return 0; |
@@ -6908,10 +6904,16 @@ int btrfs_read_block_groups(struct btrfs_root *root) | |||
6908 | ret = -ENOMEM; | 6904 | ret = -ENOMEM; |
6909 | goto error; | 6905 | goto error; |
6910 | } | 6906 | } |
6907 | cache->free_space_ctl = kzalloc(sizeof(*cache->free_space_ctl), | ||
6908 | GFP_NOFS); | ||
6909 | if (!cache->free_space_ctl) { | ||
6910 | kfree(cache); | ||
6911 | ret = -ENOMEM; | ||
6912 | goto error; | ||
6913 | } | ||
6911 | 6914 | ||
6912 | atomic_set(&cache->count, 1); | 6915 | atomic_set(&cache->count, 1); |
6913 | spin_lock_init(&cache->lock); | 6916 | spin_lock_init(&cache->lock); |
6914 | spin_lock_init(&cache->tree_lock); | ||
6915 | cache->fs_info = info; | 6917 | cache->fs_info = info; |
6916 | INIT_LIST_HEAD(&cache->list); | 6918 | INIT_LIST_HEAD(&cache->list); |
6917 | INIT_LIST_HEAD(&cache->cluster_list); | 6919 | INIT_LIST_HEAD(&cache->cluster_list); |
@@ -6919,14 +6921,6 @@ int btrfs_read_block_groups(struct btrfs_root *root) | |||
6919 | if (need_clear) | 6921 | if (need_clear) |
6920 | cache->disk_cache_state = BTRFS_DC_CLEAR; | 6922 | cache->disk_cache_state = BTRFS_DC_CLEAR; |
6921 | 6923 | ||
6922 | /* | ||
6923 | * we only want to have 32k of ram per block group for keeping | ||
6924 | * track of free space, and if we pass 1/2 of that we want to | ||
6925 | * start converting things over to using bitmaps | ||
6926 | */ | ||
6927 | cache->extents_thresh = ((1024 * 32) / 2) / | ||
6928 | sizeof(struct btrfs_free_space); | ||
6929 | |||
6930 | read_extent_buffer(leaf, &cache->item, | 6924 | read_extent_buffer(leaf, &cache->item, |
6931 | btrfs_item_ptr_offset(leaf, path->slots[0]), | 6925 | btrfs_item_ptr_offset(leaf, path->slots[0]), |
6932 | sizeof(cache->item)); | 6926 | sizeof(cache->item)); |
@@ -6937,6 +6931,8 @@ int btrfs_read_block_groups(struct btrfs_root *root) | |||
6937 | cache->flags = btrfs_block_group_flags(&cache->item); | 6931 | cache->flags = btrfs_block_group_flags(&cache->item); |
6938 | cache->sectorsize = root->sectorsize; | 6932 | cache->sectorsize = root->sectorsize; |
6939 | 6933 | ||
6934 | btrfs_init_free_space_ctl(cache); | ||
6935 | |||
6940 | /* | 6936 | /* |
6941 | * We need to exclude the super stripes now so that the space | 6937 | * We need to exclude the super stripes now so that the space |
6942 | * info has super bytes accounted for, otherwise we'll think | 6938 | * info has super bytes accounted for, otherwise we'll think |
@@ -7023,6 +7019,12 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans, | |||
7023 | cache = kzalloc(sizeof(*cache), GFP_NOFS); | 7019 | cache = kzalloc(sizeof(*cache), GFP_NOFS); |
7024 | if (!cache) | 7020 | if (!cache) |
7025 | return -ENOMEM; | 7021 | return -ENOMEM; |
7022 | cache->free_space_ctl = kzalloc(sizeof(*cache->free_space_ctl), | ||
7023 | GFP_NOFS); | ||
7024 | if (!cache->free_space_ctl) { | ||
7025 | kfree(cache); | ||
7026 | return -ENOMEM; | ||
7027 | } | ||
7026 | 7028 | ||
7027 | cache->key.objectid = chunk_offset; | 7029 | cache->key.objectid = chunk_offset; |
7028 | cache->key.offset = size; | 7030 | cache->key.offset = size; |
@@ -7030,19 +7032,13 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans, | |||
7030 | cache->sectorsize = root->sectorsize; | 7032 | cache->sectorsize = root->sectorsize; |
7031 | cache->fs_info = root->fs_info; | 7033 | cache->fs_info = root->fs_info; |
7032 | 7034 | ||
7033 | /* | ||
7034 | * we only want to have 32k of ram per block group for keeping track | ||
7035 | * of free space, and if we pass 1/2 of that we want to start | ||
7036 | * converting things over to using bitmaps | ||
7037 | */ | ||
7038 | cache->extents_thresh = ((1024 * 32) / 2) / | ||
7039 | sizeof(struct btrfs_free_space); | ||
7040 | atomic_set(&cache->count, 1); | 7035 | atomic_set(&cache->count, 1); |
7041 | spin_lock_init(&cache->lock); | 7036 | spin_lock_init(&cache->lock); |
7042 | spin_lock_init(&cache->tree_lock); | ||
7043 | INIT_LIST_HEAD(&cache->list); | 7037 | INIT_LIST_HEAD(&cache->list); |
7044 | INIT_LIST_HEAD(&cache->cluster_list); | 7038 | INIT_LIST_HEAD(&cache->cluster_list); |
7045 | 7039 | ||
7040 | btrfs_init_free_space_ctl(cache); | ||
7041 | |||
7046 | btrfs_set_block_group_used(&cache->item, bytes_used); | 7042 | btrfs_set_block_group_used(&cache->item, bytes_used); |
7047 | btrfs_set_block_group_chunk_objectid(&cache->item, chunk_objectid); | 7043 | btrfs_set_block_group_chunk_objectid(&cache->item, chunk_objectid); |
7048 | cache->flags = type; | 7044 | cache->flags = type; |
@@ -7209,23 +7205,38 @@ out: | |||
7209 | int btrfs_init_space_info(struct btrfs_fs_info *fs_info) | 7205 | int btrfs_init_space_info(struct btrfs_fs_info *fs_info) |
7210 | { | 7206 | { |
7211 | struct btrfs_space_info *space_info; | 7207 | struct btrfs_space_info *space_info; |
7208 | struct btrfs_super_block *disk_super; | ||
7209 | u64 features; | ||
7210 | u64 flags; | ||
7211 | int mixed = 0; | ||
7212 | int ret; | 7212 | int ret; |
7213 | 7213 | ||
7214 | ret = update_space_info(fs_info, BTRFS_BLOCK_GROUP_SYSTEM, 0, 0, | 7214 | disk_super = &fs_info->super_copy; |
7215 | &space_info); | 7215 | if (!btrfs_super_root(disk_super)) |
7216 | if (ret) | 7216 | return 1; |
7217 | return ret; | ||
7218 | 7217 | ||
7219 | ret = update_space_info(fs_info, BTRFS_BLOCK_GROUP_METADATA, 0, 0, | 7218 | features = btrfs_super_incompat_flags(disk_super); |
7220 | &space_info); | 7219 | if (features & BTRFS_FEATURE_INCOMPAT_MIXED_GROUPS) |
7221 | if (ret) | 7220 | mixed = 1; |
7222 | return ret; | ||
7223 | 7221 | ||
7224 | ret = update_space_info(fs_info, BTRFS_BLOCK_GROUP_DATA, 0, 0, | 7222 | flags = BTRFS_BLOCK_GROUP_SYSTEM; |
7225 | &space_info); | 7223 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); |
7226 | if (ret) | 7224 | if (ret) |
7227 | return ret; | 7225 | goto out; |
7228 | 7226 | ||
7227 | if (mixed) { | ||
7228 | flags = BTRFS_BLOCK_GROUP_METADATA | BTRFS_BLOCK_GROUP_DATA; | ||
7229 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); | ||
7230 | } else { | ||
7231 | flags = BTRFS_BLOCK_GROUP_METADATA; | ||
7232 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); | ||
7233 | if (ret) | ||
7234 | goto out; | ||
7235 | |||
7236 | flags = BTRFS_BLOCK_GROUP_DATA; | ||
7237 | ret = update_space_info(fs_info, flags, 0, 0, &space_info); | ||
7238 | } | ||
7239 | out: | ||
7229 | return ret; | 7240 | return ret; |
7230 | } | 7241 | } |
7231 | 7242 | ||
diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c index 91208296ff2b..64c8b361b539 100644 --- a/fs/btrfs/extent_io.c +++ b/fs/btrfs/extent_io.c | |||
@@ -2810,7 +2810,7 @@ int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
2810 | * because there might be preallocation past i_size | 2810 | * because there might be preallocation past i_size |
2811 | */ | 2811 | */ |
2812 | ret = btrfs_lookup_file_extent(NULL, BTRFS_I(inode)->root, | 2812 | ret = btrfs_lookup_file_extent(NULL, BTRFS_I(inode)->root, |
2813 | path, inode->i_ino, -1, 0); | 2813 | path, btrfs_ino(inode), -1, 0); |
2814 | if (ret < 0) { | 2814 | if (ret < 0) { |
2815 | btrfs_free_path(path); | 2815 | btrfs_free_path(path); |
2816 | return ret; | 2816 | return ret; |
@@ -2823,7 +2823,7 @@ int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, | |||
2823 | found_type = btrfs_key_type(&found_key); | 2823 | found_type = btrfs_key_type(&found_key); |
2824 | 2824 | ||
2825 | /* No extents, but there might be delalloc bits */ | 2825 | /* No extents, but there might be delalloc bits */ |
2826 | if (found_key.objectid != inode->i_ino || | 2826 | if (found_key.objectid != btrfs_ino(inode) || |
2827 | found_type != BTRFS_EXTENT_DATA_KEY) { | 2827 | found_type != BTRFS_EXTENT_DATA_KEY) { |
2828 | /* have to trust i_size as the end */ | 2828 | /* have to trust i_size as the end */ |
2829 | last = (u64)-1; | 2829 | last = (u64)-1; |
diff --git a/fs/btrfs/file-item.c b/fs/btrfs/file-item.c index f47e43d855aa..29e014984f60 100644 --- a/fs/btrfs/file-item.c +++ b/fs/btrfs/file-item.c | |||
@@ -208,8 +208,9 @@ static int __btrfs_lookup_bio_sums(struct btrfs_root *root, | |||
208 | EXTENT_NODATASUM, GFP_NOFS); | 208 | EXTENT_NODATASUM, GFP_NOFS); |
209 | } else { | 209 | } else { |
210 | printk(KERN_INFO "btrfs no csum found " | 210 | printk(KERN_INFO "btrfs no csum found " |
211 | "for inode %lu start %llu\n", | 211 | "for inode %llu start %llu\n", |
212 | inode->i_ino, | 212 | (unsigned long long) |
213 | btrfs_ino(inode), | ||
213 | (unsigned long long)offset); | 214 | (unsigned long long)offset); |
214 | } | 215 | } |
215 | item = NULL; | 216 | item = NULL; |
diff --git a/fs/btrfs/file.c b/fs/btrfs/file.c index 566bdf298ea8..58ddc4442159 100644 --- a/fs/btrfs/file.c +++ b/fs/btrfs/file.c | |||
@@ -298,6 +298,7 @@ int btrfs_drop_extents(struct btrfs_trans_handle *trans, struct inode *inode, | |||
298 | struct btrfs_path *path; | 298 | struct btrfs_path *path; |
299 | struct btrfs_key key; | 299 | struct btrfs_key key; |
300 | struct btrfs_key new_key; | 300 | struct btrfs_key new_key; |
301 | u64 ino = btrfs_ino(inode); | ||
301 | u64 search_start = start; | 302 | u64 search_start = start; |
302 | u64 disk_bytenr = 0; | 303 | u64 disk_bytenr = 0; |
303 | u64 num_bytes = 0; | 304 | u64 num_bytes = 0; |
@@ -318,14 +319,14 @@ int btrfs_drop_extents(struct btrfs_trans_handle *trans, struct inode *inode, | |||
318 | 319 | ||
319 | while (1) { | 320 | while (1) { |
320 | recow = 0; | 321 | recow = 0; |
321 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 322 | ret = btrfs_lookup_file_extent(trans, root, path, ino, |
322 | search_start, -1); | 323 | search_start, -1); |
323 | if (ret < 0) | 324 | if (ret < 0) |
324 | break; | 325 | break; |
325 | if (ret > 0 && path->slots[0] > 0 && search_start == start) { | 326 | if (ret > 0 && path->slots[0] > 0 && search_start == start) { |
326 | leaf = path->nodes[0]; | 327 | leaf = path->nodes[0]; |
327 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0] - 1); | 328 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0] - 1); |
328 | if (key.objectid == inode->i_ino && | 329 | if (key.objectid == ino && |
329 | key.type == BTRFS_EXTENT_DATA_KEY) | 330 | key.type == BTRFS_EXTENT_DATA_KEY) |
330 | path->slots[0]--; | 331 | path->slots[0]--; |
331 | } | 332 | } |
@@ -346,7 +347,7 @@ next_slot: | |||
346 | } | 347 | } |
347 | 348 | ||
348 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); | 349 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); |
349 | if (key.objectid > inode->i_ino || | 350 | if (key.objectid > ino || |
350 | key.type > BTRFS_EXTENT_DATA_KEY || key.offset >= end) | 351 | key.type > BTRFS_EXTENT_DATA_KEY || key.offset >= end) |
351 | break; | 352 | break; |
352 | 353 | ||
@@ -592,6 +593,7 @@ int btrfs_mark_extent_written(struct btrfs_trans_handle *trans, | |||
592 | int del_slot = 0; | 593 | int del_slot = 0; |
593 | int recow; | 594 | int recow; |
594 | int ret; | 595 | int ret; |
596 | u64 ino = btrfs_ino(inode); | ||
595 | 597 | ||
596 | btrfs_drop_extent_cache(inode, start, end - 1, 0); | 598 | btrfs_drop_extent_cache(inode, start, end - 1, 0); |
597 | 599 | ||
@@ -600,7 +602,7 @@ int btrfs_mark_extent_written(struct btrfs_trans_handle *trans, | |||
600 | again: | 602 | again: |
601 | recow = 0; | 603 | recow = 0; |
602 | split = start; | 604 | split = start; |
603 | key.objectid = inode->i_ino; | 605 | key.objectid = ino; |
604 | key.type = BTRFS_EXTENT_DATA_KEY; | 606 | key.type = BTRFS_EXTENT_DATA_KEY; |
605 | key.offset = split; | 607 | key.offset = split; |
606 | 608 | ||
@@ -612,8 +614,7 @@ again: | |||
612 | 614 | ||
613 | leaf = path->nodes[0]; | 615 | leaf = path->nodes[0]; |
614 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); | 616 | btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); |
615 | BUG_ON(key.objectid != inode->i_ino || | 617 | BUG_ON(key.objectid != ino || key.type != BTRFS_EXTENT_DATA_KEY); |
616 | key.type != BTRFS_EXTENT_DATA_KEY); | ||
617 | fi = btrfs_item_ptr(leaf, path->slots[0], | 618 | fi = btrfs_item_ptr(leaf, path->slots[0], |
618 | struct btrfs_file_extent_item); | 619 | struct btrfs_file_extent_item); |
619 | BUG_ON(btrfs_file_extent_type(leaf, fi) != | 620 | BUG_ON(btrfs_file_extent_type(leaf, fi) != |
@@ -630,7 +631,7 @@ again: | |||
630 | other_start = 0; | 631 | other_start = 0; |
631 | other_end = start; | 632 | other_end = start; |
632 | if (extent_mergeable(leaf, path->slots[0] - 1, | 633 | if (extent_mergeable(leaf, path->slots[0] - 1, |
633 | inode->i_ino, bytenr, orig_offset, | 634 | ino, bytenr, orig_offset, |
634 | &other_start, &other_end)) { | 635 | &other_start, &other_end)) { |
635 | new_key.offset = end; | 636 | new_key.offset = end; |
636 | btrfs_set_item_key_safe(trans, root, path, &new_key); | 637 | btrfs_set_item_key_safe(trans, root, path, &new_key); |
@@ -653,7 +654,7 @@ again: | |||
653 | other_start = end; | 654 | other_start = end; |
654 | other_end = 0; | 655 | other_end = 0; |
655 | if (extent_mergeable(leaf, path->slots[0] + 1, | 656 | if (extent_mergeable(leaf, path->slots[0] + 1, |
656 | inode->i_ino, bytenr, orig_offset, | 657 | ino, bytenr, orig_offset, |
657 | &other_start, &other_end)) { | 658 | &other_start, &other_end)) { |
658 | fi = btrfs_item_ptr(leaf, path->slots[0], | 659 | fi = btrfs_item_ptr(leaf, path->slots[0], |
659 | struct btrfs_file_extent_item); | 660 | struct btrfs_file_extent_item); |
@@ -702,7 +703,7 @@ again: | |||
702 | 703 | ||
703 | ret = btrfs_inc_extent_ref(trans, root, bytenr, num_bytes, 0, | 704 | ret = btrfs_inc_extent_ref(trans, root, bytenr, num_bytes, 0, |
704 | root->root_key.objectid, | 705 | root->root_key.objectid, |
705 | inode->i_ino, orig_offset); | 706 | ino, orig_offset); |
706 | BUG_ON(ret); | 707 | BUG_ON(ret); |
707 | 708 | ||
708 | if (split == start) { | 709 | if (split == start) { |
@@ -718,7 +719,7 @@ again: | |||
718 | other_start = end; | 719 | other_start = end; |
719 | other_end = 0; | 720 | other_end = 0; |
720 | if (extent_mergeable(leaf, path->slots[0] + 1, | 721 | if (extent_mergeable(leaf, path->slots[0] + 1, |
721 | inode->i_ino, bytenr, orig_offset, | 722 | ino, bytenr, orig_offset, |
722 | &other_start, &other_end)) { | 723 | &other_start, &other_end)) { |
723 | if (recow) { | 724 | if (recow) { |
724 | btrfs_release_path(path); | 725 | btrfs_release_path(path); |
@@ -729,13 +730,13 @@ again: | |||
729 | del_nr++; | 730 | del_nr++; |
730 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, | 731 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, |
731 | 0, root->root_key.objectid, | 732 | 0, root->root_key.objectid, |
732 | inode->i_ino, orig_offset); | 733 | ino, orig_offset); |
733 | BUG_ON(ret); | 734 | BUG_ON(ret); |
734 | } | 735 | } |
735 | other_start = 0; | 736 | other_start = 0; |
736 | other_end = start; | 737 | other_end = start; |
737 | if (extent_mergeable(leaf, path->slots[0] - 1, | 738 | if (extent_mergeable(leaf, path->slots[0] - 1, |
738 | inode->i_ino, bytenr, orig_offset, | 739 | ino, bytenr, orig_offset, |
739 | &other_start, &other_end)) { | 740 | &other_start, &other_end)) { |
740 | if (recow) { | 741 | if (recow) { |
741 | btrfs_release_path(path); | 742 | btrfs_release_path(path); |
@@ -746,7 +747,7 @@ again: | |||
746 | del_nr++; | 747 | del_nr++; |
747 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, | 748 | ret = btrfs_free_extent(trans, root, bytenr, num_bytes, |
748 | 0, root->root_key.objectid, | 749 | 0, root->root_key.objectid, |
749 | inode->i_ino, orig_offset); | 750 | ino, orig_offset); |
750 | BUG_ON(ret); | 751 | BUG_ON(ret); |
751 | } | 752 | } |
752 | if (del_nr == 0) { | 753 | if (del_nr == 0) { |
diff --git a/fs/btrfs/free-space-cache.c b/fs/btrfs/free-space-cache.c index 0290b0c7b003..70d45795d758 100644 --- a/fs/btrfs/free-space-cache.c +++ b/fs/btrfs/free-space-cache.c | |||
@@ -25,18 +25,17 @@ | |||
25 | #include "transaction.h" | 25 | #include "transaction.h" |
26 | #include "disk-io.h" | 26 | #include "disk-io.h" |
27 | #include "extent_io.h" | 27 | #include "extent_io.h" |
28 | #include "inode-map.h" | ||
28 | 29 | ||
29 | #define BITS_PER_BITMAP (PAGE_CACHE_SIZE * 8) | 30 | #define BITS_PER_BITMAP (PAGE_CACHE_SIZE * 8) |
30 | #define MAX_CACHE_BYTES_PER_GIG (32 * 1024) | 31 | #define MAX_CACHE_BYTES_PER_GIG (32 * 1024) |
31 | 32 | ||
32 | static void recalculate_thresholds(struct btrfs_block_group_cache | 33 | static int link_free_space(struct btrfs_free_space_ctl *ctl, |
33 | *block_group); | ||
34 | static int link_free_space(struct btrfs_block_group_cache *block_group, | ||
35 | struct btrfs_free_space *info); | 34 | struct btrfs_free_space *info); |
36 | 35 | ||
37 | struct inode *lookup_free_space_inode(struct btrfs_root *root, | 36 | static struct inode *__lookup_free_space_inode(struct btrfs_root *root, |
38 | struct btrfs_block_group_cache | 37 | struct btrfs_path *path, |
39 | *block_group, struct btrfs_path *path) | 38 | u64 offset) |
40 | { | 39 | { |
41 | struct btrfs_key key; | 40 | struct btrfs_key key; |
42 | struct btrfs_key location; | 41 | struct btrfs_key location; |
@@ -46,15 +45,8 @@ struct inode *lookup_free_space_inode(struct btrfs_root *root, | |||
46 | struct inode *inode = NULL; | 45 | struct inode *inode = NULL; |
47 | int ret; | 46 | int ret; |
48 | 47 | ||
49 | spin_lock(&block_group->lock); | ||
50 | if (block_group->inode) | ||
51 | inode = igrab(block_group->inode); | ||
52 | spin_unlock(&block_group->lock); | ||
53 | if (inode) | ||
54 | return inode; | ||
55 | |||
56 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 48 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
57 | key.offset = block_group->key.objectid; | 49 | key.offset = offset; |
58 | key.type = 0; | 50 | key.type = 0; |
59 | 51 | ||
60 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | 52 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); |
@@ -84,6 +76,27 @@ struct inode *lookup_free_space_inode(struct btrfs_root *root, | |||
84 | 76 | ||
85 | inode->i_mapping->flags &= ~__GFP_FS; | 77 | inode->i_mapping->flags &= ~__GFP_FS; |
86 | 78 | ||
79 | return inode; | ||
80 | } | ||
81 | |||
82 | struct inode *lookup_free_space_inode(struct btrfs_root *root, | ||
83 | struct btrfs_block_group_cache | ||
84 | *block_group, struct btrfs_path *path) | ||
85 | { | ||
86 | struct inode *inode = NULL; | ||
87 | |||
88 | spin_lock(&block_group->lock); | ||
89 | if (block_group->inode) | ||
90 | inode = igrab(block_group->inode); | ||
91 | spin_unlock(&block_group->lock); | ||
92 | if (inode) | ||
93 | return inode; | ||
94 | |||
95 | inode = __lookup_free_space_inode(root, path, | ||
96 | block_group->key.objectid); | ||
97 | if (IS_ERR(inode)) | ||
98 | return inode; | ||
99 | |||
87 | spin_lock(&block_group->lock); | 100 | spin_lock(&block_group->lock); |
88 | if (!root->fs_info->closing) { | 101 | if (!root->fs_info->closing) { |
89 | block_group->inode = igrab(inode); | 102 | block_group->inode = igrab(inode); |
@@ -94,24 +107,18 @@ struct inode *lookup_free_space_inode(struct btrfs_root *root, | |||
94 | return inode; | 107 | return inode; |
95 | } | 108 | } |
96 | 109 | ||
97 | int create_free_space_inode(struct btrfs_root *root, | 110 | int __create_free_space_inode(struct btrfs_root *root, |
98 | struct btrfs_trans_handle *trans, | 111 | struct btrfs_trans_handle *trans, |
99 | struct btrfs_block_group_cache *block_group, | 112 | struct btrfs_path *path, u64 ino, u64 offset) |
100 | struct btrfs_path *path) | ||
101 | { | 113 | { |
102 | struct btrfs_key key; | 114 | struct btrfs_key key; |
103 | struct btrfs_disk_key disk_key; | 115 | struct btrfs_disk_key disk_key; |
104 | struct btrfs_free_space_header *header; | 116 | struct btrfs_free_space_header *header; |
105 | struct btrfs_inode_item *inode_item; | 117 | struct btrfs_inode_item *inode_item; |
106 | struct extent_buffer *leaf; | 118 | struct extent_buffer *leaf; |
107 | u64 objectid; | ||
108 | int ret; | 119 | int ret; |
109 | 120 | ||
110 | ret = btrfs_find_free_objectid(trans, root, 0, &objectid); | 121 | ret = btrfs_insert_empty_inode(trans, root, path, ino); |
111 | if (ret < 0) | ||
112 | return ret; | ||
113 | |||
114 | ret = btrfs_insert_empty_inode(trans, root, path, objectid); | ||
115 | if (ret) | 122 | if (ret) |
116 | return ret; | 123 | return ret; |
117 | 124 | ||
@@ -131,13 +138,12 @@ int create_free_space_inode(struct btrfs_root *root, | |||
131 | BTRFS_INODE_PREALLOC | BTRFS_INODE_NODATASUM); | 138 | BTRFS_INODE_PREALLOC | BTRFS_INODE_NODATASUM); |
132 | btrfs_set_inode_nlink(leaf, inode_item, 1); | 139 | btrfs_set_inode_nlink(leaf, inode_item, 1); |
133 | btrfs_set_inode_transid(leaf, inode_item, trans->transid); | 140 | btrfs_set_inode_transid(leaf, inode_item, trans->transid); |
134 | btrfs_set_inode_block_group(leaf, inode_item, | 141 | btrfs_set_inode_block_group(leaf, inode_item, offset); |
135 | block_group->key.objectid); | ||
136 | btrfs_mark_buffer_dirty(leaf); | 142 | btrfs_mark_buffer_dirty(leaf); |
137 | btrfs_release_path(path); | 143 | btrfs_release_path(path); |
138 | 144 | ||
139 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 145 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
140 | key.offset = block_group->key.objectid; | 146 | key.offset = offset; |
141 | key.type = 0; | 147 | key.type = 0; |
142 | 148 | ||
143 | ret = btrfs_insert_empty_item(trans, root, path, &key, | 149 | ret = btrfs_insert_empty_item(trans, root, path, &key, |
@@ -157,6 +163,22 @@ int create_free_space_inode(struct btrfs_root *root, | |||
157 | return 0; | 163 | return 0; |
158 | } | 164 | } |
159 | 165 | ||
166 | int create_free_space_inode(struct btrfs_root *root, | ||
167 | struct btrfs_trans_handle *trans, | ||
168 | struct btrfs_block_group_cache *block_group, | ||
169 | struct btrfs_path *path) | ||
170 | { | ||
171 | int ret; | ||
172 | u64 ino; | ||
173 | |||
174 | ret = btrfs_find_free_objectid(root, &ino); | ||
175 | if (ret < 0) | ||
176 | return ret; | ||
177 | |||
178 | return __create_free_space_inode(root, trans, path, ino, | ||
179 | block_group->key.objectid); | ||
180 | } | ||
181 | |||
160 | int btrfs_truncate_free_space_cache(struct btrfs_root *root, | 182 | int btrfs_truncate_free_space_cache(struct btrfs_root *root, |
161 | struct btrfs_trans_handle *trans, | 183 | struct btrfs_trans_handle *trans, |
162 | struct btrfs_path *path, | 184 | struct btrfs_path *path, |
@@ -187,7 +209,8 @@ int btrfs_truncate_free_space_cache(struct btrfs_root *root, | |||
187 | return ret; | 209 | return ret; |
188 | } | 210 | } |
189 | 211 | ||
190 | return btrfs_update_inode(trans, root, inode); | 212 | ret = btrfs_update_inode(trans, root, inode); |
213 | return ret; | ||
191 | } | 214 | } |
192 | 215 | ||
193 | static int readahead_cache(struct inode *inode) | 216 | static int readahead_cache(struct inode *inode) |
@@ -209,15 +232,13 @@ static int readahead_cache(struct inode *inode) | |||
209 | return 0; | 232 | return 0; |
210 | } | 233 | } |
211 | 234 | ||
212 | int load_free_space_cache(struct btrfs_fs_info *fs_info, | 235 | int __load_free_space_cache(struct btrfs_root *root, struct inode *inode, |
213 | struct btrfs_block_group_cache *block_group) | 236 | struct btrfs_free_space_ctl *ctl, |
237 | struct btrfs_path *path, u64 offset) | ||
214 | { | 238 | { |
215 | struct btrfs_root *root = fs_info->tree_root; | ||
216 | struct inode *inode; | ||
217 | struct btrfs_free_space_header *header; | 239 | struct btrfs_free_space_header *header; |
218 | struct extent_buffer *leaf; | 240 | struct extent_buffer *leaf; |
219 | struct page *page; | 241 | struct page *page; |
220 | struct btrfs_path *path; | ||
221 | u32 *checksums = NULL, *crc; | 242 | u32 *checksums = NULL, *crc; |
222 | char *disk_crcs = NULL; | 243 | char *disk_crcs = NULL; |
223 | struct btrfs_key key; | 244 | struct btrfs_key key; |
@@ -225,76 +246,47 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
225 | u64 num_entries; | 246 | u64 num_entries; |
226 | u64 num_bitmaps; | 247 | u64 num_bitmaps; |
227 | u64 generation; | 248 | u64 generation; |
228 | u64 used = btrfs_block_group_used(&block_group->item); | ||
229 | u32 cur_crc = ~(u32)0; | 249 | u32 cur_crc = ~(u32)0; |
230 | pgoff_t index = 0; | 250 | pgoff_t index = 0; |
231 | unsigned long first_page_offset; | 251 | unsigned long first_page_offset; |
232 | int num_checksums; | 252 | int num_checksums; |
233 | int ret = 0; | 253 | int ret = 0, ret2; |
234 | |||
235 | /* | ||
236 | * If we're unmounting then just return, since this does a search on the | ||
237 | * normal root and not the commit root and we could deadlock. | ||
238 | */ | ||
239 | smp_mb(); | ||
240 | if (fs_info->closing) | ||
241 | return 0; | ||
242 | |||
243 | /* | ||
244 | * If this block group has been marked to be cleared for one reason or | ||
245 | * another then we can't trust the on disk cache, so just return. | ||
246 | */ | ||
247 | spin_lock(&block_group->lock); | ||
248 | if (block_group->disk_cache_state != BTRFS_DC_WRITTEN) { | ||
249 | spin_unlock(&block_group->lock); | ||
250 | return 0; | ||
251 | } | ||
252 | spin_unlock(&block_group->lock); | ||
253 | 254 | ||
254 | INIT_LIST_HEAD(&bitmaps); | 255 | INIT_LIST_HEAD(&bitmaps); |
255 | 256 | ||
256 | path = btrfs_alloc_path(); | ||
257 | if (!path) | ||
258 | return 0; | ||
259 | |||
260 | inode = lookup_free_space_inode(root, block_group, path); | ||
261 | if (IS_ERR(inode)) { | ||
262 | btrfs_free_path(path); | ||
263 | return 0; | ||
264 | } | ||
265 | |||
266 | /* Nothing in the space cache, goodbye */ | 257 | /* Nothing in the space cache, goodbye */ |
267 | if (!i_size_read(inode)) { | 258 | if (!i_size_read(inode)) |
268 | btrfs_free_path(path); | ||
269 | goto out; | 259 | goto out; |
270 | } | ||
271 | 260 | ||
272 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 261 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
273 | key.offset = block_group->key.objectid; | 262 | key.offset = offset; |
274 | key.type = 0; | 263 | key.type = 0; |
275 | 264 | ||
276 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | 265 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); |
277 | if (ret) { | 266 | if (ret < 0) |
278 | btrfs_free_path(path); | 267 | goto out; |
268 | else if (ret > 0) { | ||
269 | btrfs_release_path(path); | ||
270 | ret = 0; | ||
279 | goto out; | 271 | goto out; |
280 | } | 272 | } |
281 | 273 | ||
274 | ret = -1; | ||
275 | |||
282 | leaf = path->nodes[0]; | 276 | leaf = path->nodes[0]; |
283 | header = btrfs_item_ptr(leaf, path->slots[0], | 277 | header = btrfs_item_ptr(leaf, path->slots[0], |
284 | struct btrfs_free_space_header); | 278 | struct btrfs_free_space_header); |
285 | num_entries = btrfs_free_space_entries(leaf, header); | 279 | num_entries = btrfs_free_space_entries(leaf, header); |
286 | num_bitmaps = btrfs_free_space_bitmaps(leaf, header); | 280 | num_bitmaps = btrfs_free_space_bitmaps(leaf, header); |
287 | generation = btrfs_free_space_generation(leaf, header); | 281 | generation = btrfs_free_space_generation(leaf, header); |
288 | btrfs_free_path(path); | 282 | btrfs_release_path(path); |
289 | 283 | ||
290 | if (BTRFS_I(inode)->generation != generation) { | 284 | if (BTRFS_I(inode)->generation != generation) { |
291 | printk(KERN_ERR "btrfs: free space inode generation (%llu) did" | 285 | printk(KERN_ERR "btrfs: free space inode generation (%llu) did" |
292 | " not match free space cache generation (%llu) for " | 286 | " not match free space cache generation (%llu)\n", |
293 | "block group %llu\n", | ||
294 | (unsigned long long)BTRFS_I(inode)->generation, | 287 | (unsigned long long)BTRFS_I(inode)->generation, |
295 | (unsigned long long)generation, | 288 | (unsigned long long)generation); |
296 | (unsigned long long)block_group->key.objectid); | 289 | goto out; |
297 | goto free_cache; | ||
298 | } | 290 | } |
299 | 291 | ||
300 | if (!num_entries) | 292 | if (!num_entries) |
@@ -311,10 +303,8 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
311 | goto out; | 303 | goto out; |
312 | 304 | ||
313 | ret = readahead_cache(inode); | 305 | ret = readahead_cache(inode); |
314 | if (ret) { | 306 | if (ret) |
315 | ret = 0; | ||
316 | goto out; | 307 | goto out; |
317 | } | ||
318 | 308 | ||
319 | while (1) { | 309 | while (1) { |
320 | struct btrfs_free_space_entry *entry; | 310 | struct btrfs_free_space_entry *entry; |
@@ -333,10 +323,8 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
333 | } | 323 | } |
334 | 324 | ||
335 | page = grab_cache_page(inode->i_mapping, index); | 325 | page = grab_cache_page(inode->i_mapping, index); |
336 | if (!page) { | 326 | if (!page) |
337 | ret = 0; | ||
338 | goto free_cache; | 327 | goto free_cache; |
339 | } | ||
340 | 328 | ||
341 | if (!PageUptodate(page)) { | 329 | if (!PageUptodate(page)) { |
342 | btrfs_readpage(NULL, page); | 330 | btrfs_readpage(NULL, page); |
@@ -345,9 +333,7 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
345 | unlock_page(page); | 333 | unlock_page(page); |
346 | page_cache_release(page); | 334 | page_cache_release(page); |
347 | printk(KERN_ERR "btrfs: error reading free " | 335 | printk(KERN_ERR "btrfs: error reading free " |
348 | "space cache: %llu\n", | 336 | "space cache\n"); |
349 | (unsigned long long) | ||
350 | block_group->key.objectid); | ||
351 | goto free_cache; | 337 | goto free_cache; |
352 | } | 338 | } |
353 | } | 339 | } |
@@ -360,13 +346,10 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
360 | gen = addr + (sizeof(u32) * num_checksums); | 346 | gen = addr + (sizeof(u32) * num_checksums); |
361 | if (*gen != BTRFS_I(inode)->generation) { | 347 | if (*gen != BTRFS_I(inode)->generation) { |
362 | printk(KERN_ERR "btrfs: space cache generation" | 348 | printk(KERN_ERR "btrfs: space cache generation" |
363 | " (%llu) does not match inode (%llu) " | 349 | " (%llu) does not match inode (%llu)\n", |
364 | "for block group %llu\n", | ||
365 | (unsigned long long)*gen, | 350 | (unsigned long long)*gen, |
366 | (unsigned long long) | 351 | (unsigned long long) |
367 | BTRFS_I(inode)->generation, | 352 | BTRFS_I(inode)->generation); |
368 | (unsigned long long) | ||
369 | block_group->key.objectid); | ||
370 | kunmap(page); | 353 | kunmap(page); |
371 | unlock_page(page); | 354 | unlock_page(page); |
372 | page_cache_release(page); | 355 | page_cache_release(page); |
@@ -382,9 +365,8 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
382 | PAGE_CACHE_SIZE - start_offset); | 365 | PAGE_CACHE_SIZE - start_offset); |
383 | btrfs_csum_final(cur_crc, (char *)&cur_crc); | 366 | btrfs_csum_final(cur_crc, (char *)&cur_crc); |
384 | if (cur_crc != *crc) { | 367 | if (cur_crc != *crc) { |
385 | printk(KERN_ERR "btrfs: crc mismatch for page %lu in " | 368 | printk(KERN_ERR "btrfs: crc mismatch for page %lu\n", |
386 | "block group %llu\n", index, | 369 | index); |
387 | (unsigned long long)block_group->key.objectid); | ||
388 | kunmap(page); | 370 | kunmap(page); |
389 | unlock_page(page); | 371 | unlock_page(page); |
390 | page_cache_release(page); | 372 | page_cache_release(page); |
@@ -417,9 +399,9 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
417 | } | 399 | } |
418 | 400 | ||
419 | if (entry->type == BTRFS_FREE_SPACE_EXTENT) { | 401 | if (entry->type == BTRFS_FREE_SPACE_EXTENT) { |
420 | spin_lock(&block_group->tree_lock); | 402 | spin_lock(&ctl->tree_lock); |
421 | ret = link_free_space(block_group, e); | 403 | ret = link_free_space(ctl, e); |
422 | spin_unlock(&block_group->tree_lock); | 404 | spin_unlock(&ctl->tree_lock); |
423 | BUG_ON(ret); | 405 | BUG_ON(ret); |
424 | } else { | 406 | } else { |
425 | e->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); | 407 | e->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); |
@@ -431,11 +413,11 @@ int load_free_space_cache(struct btrfs_fs_info *fs_info, | |||
431 | page_cache_release(page); | 413 | page_cache_release(page); |
432 | goto free_cache; | 414 | goto free_cache; |
433 | } | 415 | } |
434 | spin_lock(&block_group->tree_lock); | 416 | spin_lock(&ctl->tree_lock); |
435 | ret = link_free_space(block_group, e); | 417 | ret2 = link_free_space(ctl, e); |
436 | block_group->total_bitmaps++; | 418 | ctl->total_bitmaps++; |
437 | recalculate_thresholds(block_group); | 419 | ctl->op->recalc_thresholds(ctl); |
438 | spin_unlock(&block_group->tree_lock); | 420 | spin_unlock(&ctl->tree_lock); |
439 | list_add_tail(&e->list, &bitmaps); | 421 | list_add_tail(&e->list, &bitmaps); |
440 | } | 422 | } |
441 | 423 | ||
@@ -471,41 +453,97 @@ next: | |||
471 | index++; | 453 | index++; |
472 | } | 454 | } |
473 | 455 | ||
474 | spin_lock(&block_group->tree_lock); | ||
475 | if (block_group->free_space != (block_group->key.offset - used - | ||
476 | block_group->bytes_super)) { | ||
477 | spin_unlock(&block_group->tree_lock); | ||
478 | printk(KERN_ERR "block group %llu has an wrong amount of free " | ||
479 | "space\n", block_group->key.objectid); | ||
480 | ret = 0; | ||
481 | goto free_cache; | ||
482 | } | ||
483 | spin_unlock(&block_group->tree_lock); | ||
484 | |||
485 | ret = 1; | 456 | ret = 1; |
486 | out: | 457 | out: |
487 | kfree(checksums); | 458 | kfree(checksums); |
488 | kfree(disk_crcs); | 459 | kfree(disk_crcs); |
489 | iput(inode); | ||
490 | return ret; | 460 | return ret; |
491 | |||
492 | free_cache: | 461 | free_cache: |
493 | /* This cache is bogus, make sure it gets cleared */ | 462 | __btrfs_remove_free_space_cache(ctl); |
463 | goto out; | ||
464 | } | ||
465 | |||
466 | int load_free_space_cache(struct btrfs_fs_info *fs_info, | ||
467 | struct btrfs_block_group_cache *block_group) | ||
468 | { | ||
469 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
470 | struct btrfs_root *root = fs_info->tree_root; | ||
471 | struct inode *inode; | ||
472 | struct btrfs_path *path; | ||
473 | int ret; | ||
474 | bool matched; | ||
475 | u64 used = btrfs_block_group_used(&block_group->item); | ||
476 | |||
477 | /* | ||
478 | * If we're unmounting then just return, since this does a search on the | ||
479 | * normal root and not the commit root and we could deadlock. | ||
480 | */ | ||
481 | smp_mb(); | ||
482 | if (fs_info->closing) | ||
483 | return 0; | ||
484 | |||
485 | /* | ||
486 | * If this block group has been marked to be cleared for one reason or | ||
487 | * another then we can't trust the on disk cache, so just return. | ||
488 | */ | ||
494 | spin_lock(&block_group->lock); | 489 | spin_lock(&block_group->lock); |
495 | block_group->disk_cache_state = BTRFS_DC_CLEAR; | 490 | if (block_group->disk_cache_state != BTRFS_DC_WRITTEN) { |
491 | spin_unlock(&block_group->lock); | ||
492 | return 0; | ||
493 | } | ||
496 | spin_unlock(&block_group->lock); | 494 | spin_unlock(&block_group->lock); |
497 | btrfs_remove_free_space_cache(block_group); | 495 | |
498 | goto out; | 496 | path = btrfs_alloc_path(); |
497 | if (!path) | ||
498 | return 0; | ||
499 | |||
500 | inode = lookup_free_space_inode(root, block_group, path); | ||
501 | if (IS_ERR(inode)) { | ||
502 | btrfs_free_path(path); | ||
503 | return 0; | ||
504 | } | ||
505 | |||
506 | ret = __load_free_space_cache(fs_info->tree_root, inode, ctl, | ||
507 | path, block_group->key.objectid); | ||
508 | btrfs_free_path(path); | ||
509 | if (ret <= 0) | ||
510 | goto out; | ||
511 | |||
512 | spin_lock(&ctl->tree_lock); | ||
513 | matched = (ctl->free_space == (block_group->key.offset - used - | ||
514 | block_group->bytes_super)); | ||
515 | spin_unlock(&ctl->tree_lock); | ||
516 | |||
517 | if (!matched) { | ||
518 | __btrfs_remove_free_space_cache(ctl); | ||
519 | printk(KERN_ERR "block group %llu has an wrong amount of free " | ||
520 | "space\n", block_group->key.objectid); | ||
521 | ret = -1; | ||
522 | } | ||
523 | out: | ||
524 | if (ret < 0) { | ||
525 | /* This cache is bogus, make sure it gets cleared */ | ||
526 | spin_lock(&block_group->lock); | ||
527 | block_group->disk_cache_state = BTRFS_DC_CLEAR; | ||
528 | spin_unlock(&block_group->lock); | ||
529 | ret = 0; | ||
530 | |||
531 | printk(KERN_ERR "btrfs: failed to load free space cache " | ||
532 | "for block group %llu\n", block_group->key.objectid); | ||
533 | } | ||
534 | |||
535 | iput(inode); | ||
536 | return ret; | ||
499 | } | 537 | } |
500 | 538 | ||
501 | int btrfs_write_out_cache(struct btrfs_root *root, | 539 | int __btrfs_write_out_cache(struct btrfs_root *root, struct inode *inode, |
502 | struct btrfs_trans_handle *trans, | 540 | struct btrfs_free_space_ctl *ctl, |
503 | struct btrfs_block_group_cache *block_group, | 541 | struct btrfs_block_group_cache *block_group, |
504 | struct btrfs_path *path) | 542 | struct btrfs_trans_handle *trans, |
543 | struct btrfs_path *path, u64 offset) | ||
505 | { | 544 | { |
506 | struct btrfs_free_space_header *header; | 545 | struct btrfs_free_space_header *header; |
507 | struct extent_buffer *leaf; | 546 | struct extent_buffer *leaf; |
508 | struct inode *inode; | ||
509 | struct rb_node *node; | 547 | struct rb_node *node; |
510 | struct list_head *pos, *n; | 548 | struct list_head *pos, *n; |
511 | struct page **pages; | 549 | struct page **pages; |
@@ -522,35 +560,18 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
522 | int index = 0, num_pages = 0; | 560 | int index = 0, num_pages = 0; |
523 | int entries = 0; | 561 | int entries = 0; |
524 | int bitmaps = 0; | 562 | int bitmaps = 0; |
525 | int ret = 0; | 563 | int ret = -1; |
526 | bool next_page = false; | 564 | bool next_page = false; |
527 | bool out_of_space = false; | 565 | bool out_of_space = false; |
528 | 566 | ||
529 | root = root->fs_info->tree_root; | ||
530 | |||
531 | INIT_LIST_HEAD(&bitmap_list); | 567 | INIT_LIST_HEAD(&bitmap_list); |
532 | 568 | ||
533 | spin_lock(&block_group->lock); | 569 | node = rb_first(&ctl->free_space_offset); |
534 | if (block_group->disk_cache_state < BTRFS_DC_SETUP) { | 570 | if (!node) |
535 | spin_unlock(&block_group->lock); | ||
536 | return 0; | ||
537 | } | ||
538 | spin_unlock(&block_group->lock); | ||
539 | |||
540 | inode = lookup_free_space_inode(root, block_group, path); | ||
541 | if (IS_ERR(inode)) | ||
542 | return 0; | ||
543 | |||
544 | if (!i_size_read(inode)) { | ||
545 | iput(inode); | ||
546 | return 0; | 571 | return 0; |
547 | } | ||
548 | 572 | ||
549 | node = rb_first(&block_group->free_space_offset); | 573 | if (!i_size_read(inode)) |
550 | if (!node) { | 574 | return -1; |
551 | iput(inode); | ||
552 | return 0; | ||
553 | } | ||
554 | 575 | ||
555 | num_pages = (i_size_read(inode) + PAGE_CACHE_SIZE - 1) >> | 576 | num_pages = (i_size_read(inode) + PAGE_CACHE_SIZE - 1) >> |
556 | PAGE_CACHE_SHIFT; | 577 | PAGE_CACHE_SHIFT; |
@@ -560,16 +581,13 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
560 | 581 | ||
561 | /* We need a checksum per page. */ | 582 | /* We need a checksum per page. */ |
562 | crc = checksums = kzalloc(sizeof(u32) * num_pages, GFP_NOFS); | 583 | crc = checksums = kzalloc(sizeof(u32) * num_pages, GFP_NOFS); |
563 | if (!crc) { | 584 | if (!crc) |
564 | iput(inode); | 585 | return -1; |
565 | return 0; | ||
566 | } | ||
567 | 586 | ||
568 | pages = kzalloc(sizeof(struct page *) * num_pages, GFP_NOFS); | 587 | pages = kzalloc(sizeof(struct page *) * num_pages, GFP_NOFS); |
569 | if (!pages) { | 588 | if (!pages) { |
570 | kfree(crc); | 589 | kfree(crc); |
571 | iput(inode); | 590 | return -1; |
572 | return 0; | ||
573 | } | 591 | } |
574 | 592 | ||
575 | /* Since the first page has all of our checksums and our generation we | 593 | /* Since the first page has all of our checksums and our generation we |
@@ -579,7 +597,7 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
579 | first_page_offset = (sizeof(u32) * num_pages) + sizeof(u64); | 597 | first_page_offset = (sizeof(u32) * num_pages) + sizeof(u64); |
580 | 598 | ||
581 | /* Get the cluster for this block_group if it exists */ | 599 | /* Get the cluster for this block_group if it exists */ |
582 | if (!list_empty(&block_group->cluster_list)) | 600 | if (block_group && !list_empty(&block_group->cluster_list)) |
583 | cluster = list_entry(block_group->cluster_list.next, | 601 | cluster = list_entry(block_group->cluster_list.next, |
584 | struct btrfs_free_cluster, | 602 | struct btrfs_free_cluster, |
585 | block_group_list); | 603 | block_group_list); |
@@ -621,7 +639,8 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
621 | * When searching for pinned extents, we need to start at our start | 639 | * When searching for pinned extents, we need to start at our start |
622 | * offset. | 640 | * offset. |
623 | */ | 641 | */ |
624 | start = block_group->key.objectid; | 642 | if (block_group) |
643 | start = block_group->key.objectid; | ||
625 | 644 | ||
626 | /* Write out the extent entries */ | 645 | /* Write out the extent entries */ |
627 | do { | 646 | do { |
@@ -679,8 +698,9 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
679 | * We want to add any pinned extents to our free space cache | 698 | * We want to add any pinned extents to our free space cache |
680 | * so we don't leak the space | 699 | * so we don't leak the space |
681 | */ | 700 | */ |
682 | while (!next_page && (start < block_group->key.objectid + | 701 | while (block_group && !next_page && |
683 | block_group->key.offset)) { | 702 | (start < block_group->key.objectid + |
703 | block_group->key.offset)) { | ||
684 | ret = find_first_extent_bit(unpin, start, &start, &end, | 704 | ret = find_first_extent_bit(unpin, start, &start, &end, |
685 | EXTENT_DIRTY); | 705 | EXTENT_DIRTY); |
686 | if (ret) { | 706 | if (ret) { |
@@ -798,12 +818,12 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
798 | filemap_write_and_wait(inode->i_mapping); | 818 | filemap_write_and_wait(inode->i_mapping); |
799 | 819 | ||
800 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; | 820 | key.objectid = BTRFS_FREE_SPACE_OBJECTID; |
801 | key.offset = block_group->key.objectid; | 821 | key.offset = offset; |
802 | key.type = 0; | 822 | key.type = 0; |
803 | 823 | ||
804 | ret = btrfs_search_slot(trans, root, &key, path, 1, 1); | 824 | ret = btrfs_search_slot(trans, root, &key, path, 1, 1); |
805 | if (ret < 0) { | 825 | if (ret < 0) { |
806 | ret = 0; | 826 | ret = -1; |
807 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, | 827 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, |
808 | EXTENT_DIRTY | EXTENT_DELALLOC | | 828 | EXTENT_DIRTY | EXTENT_DELALLOC | |
809 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, GFP_NOFS); | 829 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, GFP_NOFS); |
@@ -816,8 +836,8 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
816 | path->slots[0]--; | 836 | path->slots[0]--; |
817 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); | 837 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); |
818 | if (found_key.objectid != BTRFS_FREE_SPACE_OBJECTID || | 838 | if (found_key.objectid != BTRFS_FREE_SPACE_OBJECTID || |
819 | found_key.offset != block_group->key.objectid) { | 839 | found_key.offset != offset) { |
820 | ret = 0; | 840 | ret = -1; |
821 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, | 841 | clear_extent_bit(&BTRFS_I(inode)->io_tree, 0, bytes - 1, |
822 | EXTENT_DIRTY | EXTENT_DELALLOC | | 842 | EXTENT_DIRTY | EXTENT_DELALLOC | |
823 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, | 843 | EXTENT_DO_ACCOUNTING, 0, 0, NULL, |
@@ -837,44 +857,78 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
837 | ret = 1; | 857 | ret = 1; |
838 | 858 | ||
839 | out_free: | 859 | out_free: |
840 | if (ret == 0) { | 860 | if (ret != 1) { |
841 | invalidate_inode_pages2_range(inode->i_mapping, 0, index); | 861 | invalidate_inode_pages2_range(inode->i_mapping, 0, index); |
842 | spin_lock(&block_group->lock); | ||
843 | block_group->disk_cache_state = BTRFS_DC_ERROR; | ||
844 | spin_unlock(&block_group->lock); | ||
845 | BTRFS_I(inode)->generation = 0; | 862 | BTRFS_I(inode)->generation = 0; |
846 | } | 863 | } |
847 | kfree(checksums); | 864 | kfree(checksums); |
848 | kfree(pages); | 865 | kfree(pages); |
849 | btrfs_update_inode(trans, root, inode); | 866 | btrfs_update_inode(trans, root, inode); |
867 | return ret; | ||
868 | } | ||
869 | |||
870 | int btrfs_write_out_cache(struct btrfs_root *root, | ||
871 | struct btrfs_trans_handle *trans, | ||
872 | struct btrfs_block_group_cache *block_group, | ||
873 | struct btrfs_path *path) | ||
874 | { | ||
875 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
876 | struct inode *inode; | ||
877 | int ret = 0; | ||
878 | |||
879 | root = root->fs_info->tree_root; | ||
880 | |||
881 | spin_lock(&block_group->lock); | ||
882 | if (block_group->disk_cache_state < BTRFS_DC_SETUP) { | ||
883 | spin_unlock(&block_group->lock); | ||
884 | return 0; | ||
885 | } | ||
886 | spin_unlock(&block_group->lock); | ||
887 | |||
888 | inode = lookup_free_space_inode(root, block_group, path); | ||
889 | if (IS_ERR(inode)) | ||
890 | return 0; | ||
891 | |||
892 | ret = __btrfs_write_out_cache(root, inode, ctl, block_group, trans, | ||
893 | path, block_group->key.objectid); | ||
894 | if (ret < 0) { | ||
895 | spin_lock(&block_group->lock); | ||
896 | block_group->disk_cache_state = BTRFS_DC_ERROR; | ||
897 | spin_unlock(&block_group->lock); | ||
898 | ret = 0; | ||
899 | |||
900 | printk(KERN_ERR "btrfs: failed to write free space cace " | ||
901 | "for block group %llu\n", block_group->key.objectid); | ||
902 | } | ||
903 | |||
850 | iput(inode); | 904 | iput(inode); |
851 | return ret; | 905 | return ret; |
852 | } | 906 | } |
853 | 907 | ||
854 | static inline unsigned long offset_to_bit(u64 bitmap_start, u64 sectorsize, | 908 | static inline unsigned long offset_to_bit(u64 bitmap_start, u32 unit, |
855 | u64 offset) | 909 | u64 offset) |
856 | { | 910 | { |
857 | BUG_ON(offset < bitmap_start); | 911 | BUG_ON(offset < bitmap_start); |
858 | offset -= bitmap_start; | 912 | offset -= bitmap_start; |
859 | return (unsigned long)(div64_u64(offset, sectorsize)); | 913 | return (unsigned long)(div_u64(offset, unit)); |
860 | } | 914 | } |
861 | 915 | ||
862 | static inline unsigned long bytes_to_bits(u64 bytes, u64 sectorsize) | 916 | static inline unsigned long bytes_to_bits(u64 bytes, u32 unit) |
863 | { | 917 | { |
864 | return (unsigned long)(div64_u64(bytes, sectorsize)); | 918 | return (unsigned long)(div_u64(bytes, unit)); |
865 | } | 919 | } |
866 | 920 | ||
867 | static inline u64 offset_to_bitmap(struct btrfs_block_group_cache *block_group, | 921 | static inline u64 offset_to_bitmap(struct btrfs_free_space_ctl *ctl, |
868 | u64 offset) | 922 | u64 offset) |
869 | { | 923 | { |
870 | u64 bitmap_start; | 924 | u64 bitmap_start; |
871 | u64 bytes_per_bitmap; | 925 | u64 bytes_per_bitmap; |
872 | 926 | ||
873 | bytes_per_bitmap = BITS_PER_BITMAP * block_group->sectorsize; | 927 | bytes_per_bitmap = BITS_PER_BITMAP * ctl->unit; |
874 | bitmap_start = offset - block_group->key.objectid; | 928 | bitmap_start = offset - ctl->start; |
875 | bitmap_start = div64_u64(bitmap_start, bytes_per_bitmap); | 929 | bitmap_start = div64_u64(bitmap_start, bytes_per_bitmap); |
876 | bitmap_start *= bytes_per_bitmap; | 930 | bitmap_start *= bytes_per_bitmap; |
877 | bitmap_start += block_group->key.objectid; | 931 | bitmap_start += ctl->start; |
878 | 932 | ||
879 | return bitmap_start; | 933 | return bitmap_start; |
880 | } | 934 | } |
@@ -932,10 +986,10 @@ static int tree_insert_offset(struct rb_root *root, u64 offset, | |||
932 | * offset. | 986 | * offset. |
933 | */ | 987 | */ |
934 | static struct btrfs_free_space * | 988 | static struct btrfs_free_space * |
935 | tree_search_offset(struct btrfs_block_group_cache *block_group, | 989 | tree_search_offset(struct btrfs_free_space_ctl *ctl, |
936 | u64 offset, int bitmap_only, int fuzzy) | 990 | u64 offset, int bitmap_only, int fuzzy) |
937 | { | 991 | { |
938 | struct rb_node *n = block_group->free_space_offset.rb_node; | 992 | struct rb_node *n = ctl->free_space_offset.rb_node; |
939 | struct btrfs_free_space *entry, *prev = NULL; | 993 | struct btrfs_free_space *entry, *prev = NULL; |
940 | 994 | ||
941 | /* find entry that is closest to the 'offset' */ | 995 | /* find entry that is closest to the 'offset' */ |
@@ -1031,8 +1085,7 @@ tree_search_offset(struct btrfs_block_group_cache *block_group, | |||
1031 | break; | 1085 | break; |
1032 | } | 1086 | } |
1033 | } | 1087 | } |
1034 | if (entry->offset + BITS_PER_BITMAP * | 1088 | if (entry->offset + BITS_PER_BITMAP * ctl->unit > offset) |
1035 | block_group->sectorsize > offset) | ||
1036 | return entry; | 1089 | return entry; |
1037 | } else if (entry->offset + entry->bytes > offset) | 1090 | } else if (entry->offset + entry->bytes > offset) |
1038 | return entry; | 1091 | return entry; |
@@ -1043,7 +1096,7 @@ tree_search_offset(struct btrfs_block_group_cache *block_group, | |||
1043 | while (1) { | 1096 | while (1) { |
1044 | if (entry->bitmap) { | 1097 | if (entry->bitmap) { |
1045 | if (entry->offset + BITS_PER_BITMAP * | 1098 | if (entry->offset + BITS_PER_BITMAP * |
1046 | block_group->sectorsize > offset) | 1099 | ctl->unit > offset) |
1047 | break; | 1100 | break; |
1048 | } else { | 1101 | } else { |
1049 | if (entry->offset + entry->bytes > offset) | 1102 | if (entry->offset + entry->bytes > offset) |
@@ -1059,42 +1112,47 @@ tree_search_offset(struct btrfs_block_group_cache *block_group, | |||
1059 | } | 1112 | } |
1060 | 1113 | ||
1061 | static inline void | 1114 | static inline void |
1062 | __unlink_free_space(struct btrfs_block_group_cache *block_group, | 1115 | __unlink_free_space(struct btrfs_free_space_ctl *ctl, |
1063 | struct btrfs_free_space *info) | 1116 | struct btrfs_free_space *info) |
1064 | { | 1117 | { |
1065 | rb_erase(&info->offset_index, &block_group->free_space_offset); | 1118 | rb_erase(&info->offset_index, &ctl->free_space_offset); |
1066 | block_group->free_extents--; | 1119 | ctl->free_extents--; |
1067 | } | 1120 | } |
1068 | 1121 | ||
1069 | static void unlink_free_space(struct btrfs_block_group_cache *block_group, | 1122 | static void unlink_free_space(struct btrfs_free_space_ctl *ctl, |
1070 | struct btrfs_free_space *info) | 1123 | struct btrfs_free_space *info) |
1071 | { | 1124 | { |
1072 | __unlink_free_space(block_group, info); | 1125 | __unlink_free_space(ctl, info); |
1073 | block_group->free_space -= info->bytes; | 1126 | ctl->free_space -= info->bytes; |
1074 | } | 1127 | } |
1075 | 1128 | ||
1076 | static int link_free_space(struct btrfs_block_group_cache *block_group, | 1129 | static int link_free_space(struct btrfs_free_space_ctl *ctl, |
1077 | struct btrfs_free_space *info) | 1130 | struct btrfs_free_space *info) |
1078 | { | 1131 | { |
1079 | int ret = 0; | 1132 | int ret = 0; |
1080 | 1133 | ||
1081 | BUG_ON(!info->bitmap && !info->bytes); | 1134 | BUG_ON(!info->bitmap && !info->bytes); |
1082 | ret = tree_insert_offset(&block_group->free_space_offset, info->offset, | 1135 | ret = tree_insert_offset(&ctl->free_space_offset, info->offset, |
1083 | &info->offset_index, (info->bitmap != NULL)); | 1136 | &info->offset_index, (info->bitmap != NULL)); |
1084 | if (ret) | 1137 | if (ret) |
1085 | return ret; | 1138 | return ret; |
1086 | 1139 | ||
1087 | block_group->free_space += info->bytes; | 1140 | ctl->free_space += info->bytes; |
1088 | block_group->free_extents++; | 1141 | ctl->free_extents++; |
1089 | return ret; | 1142 | return ret; |
1090 | } | 1143 | } |
1091 | 1144 | ||
1092 | static void recalculate_thresholds(struct btrfs_block_group_cache *block_group) | 1145 | static void recalculate_thresholds(struct btrfs_free_space_ctl *ctl) |
1093 | { | 1146 | { |
1147 | struct btrfs_block_group_cache *block_group = ctl->private; | ||
1094 | u64 max_bytes; | 1148 | u64 max_bytes; |
1095 | u64 bitmap_bytes; | 1149 | u64 bitmap_bytes; |
1096 | u64 extent_bytes; | 1150 | u64 extent_bytes; |
1097 | u64 size = block_group->key.offset; | 1151 | u64 size = block_group->key.offset; |
1152 | u64 bytes_per_bg = BITS_PER_BITMAP * block_group->sectorsize; | ||
1153 | int max_bitmaps = div64_u64(size + bytes_per_bg - 1, bytes_per_bg); | ||
1154 | |||
1155 | BUG_ON(ctl->total_bitmaps > max_bitmaps); | ||
1098 | 1156 | ||
1099 | /* | 1157 | /* |
1100 | * The goal is to keep the total amount of memory used per 1gb of space | 1158 | * The goal is to keep the total amount of memory used per 1gb of space |
@@ -1112,10 +1170,10 @@ static void recalculate_thresholds(struct btrfs_block_group_cache *block_group) | |||
1112 | * sure we don't go over our overall goal of MAX_CACHE_BYTES_PER_GIG as | 1170 | * sure we don't go over our overall goal of MAX_CACHE_BYTES_PER_GIG as |
1113 | * we add more bitmaps. | 1171 | * we add more bitmaps. |
1114 | */ | 1172 | */ |
1115 | bitmap_bytes = (block_group->total_bitmaps + 1) * PAGE_CACHE_SIZE; | 1173 | bitmap_bytes = (ctl->total_bitmaps + 1) * PAGE_CACHE_SIZE; |
1116 | 1174 | ||
1117 | if (bitmap_bytes >= max_bytes) { | 1175 | if (bitmap_bytes >= max_bytes) { |
1118 | block_group->extents_thresh = 0; | 1176 | ctl->extents_thresh = 0; |
1119 | return; | 1177 | return; |
1120 | } | 1178 | } |
1121 | 1179 | ||
@@ -1126,47 +1184,43 @@ static void recalculate_thresholds(struct btrfs_block_group_cache *block_group) | |||
1126 | extent_bytes = max_bytes - bitmap_bytes; | 1184 | extent_bytes = max_bytes - bitmap_bytes; |
1127 | extent_bytes = min_t(u64, extent_bytes, div64_u64(max_bytes, 2)); | 1185 | extent_bytes = min_t(u64, extent_bytes, div64_u64(max_bytes, 2)); |
1128 | 1186 | ||
1129 | block_group->extents_thresh = | 1187 | ctl->extents_thresh = |
1130 | div64_u64(extent_bytes, (sizeof(struct btrfs_free_space))); | 1188 | div64_u64(extent_bytes, (sizeof(struct btrfs_free_space))); |
1131 | } | 1189 | } |
1132 | 1190 | ||
1133 | static void bitmap_clear_bits(struct btrfs_block_group_cache *block_group, | 1191 | static void bitmap_clear_bits(struct btrfs_free_space_ctl *ctl, |
1134 | struct btrfs_free_space *info, u64 offset, | 1192 | struct btrfs_free_space *info, u64 offset, |
1135 | u64 bytes) | 1193 | u64 bytes) |
1136 | { | 1194 | { |
1137 | unsigned long start, end; | 1195 | unsigned long start, count; |
1138 | unsigned long i; | ||
1139 | 1196 | ||
1140 | start = offset_to_bit(info->offset, block_group->sectorsize, offset); | 1197 | start = offset_to_bit(info->offset, ctl->unit, offset); |
1141 | end = start + bytes_to_bits(bytes, block_group->sectorsize); | 1198 | count = bytes_to_bits(bytes, ctl->unit); |
1142 | BUG_ON(end > BITS_PER_BITMAP); | 1199 | BUG_ON(start + count > BITS_PER_BITMAP); |
1143 | 1200 | ||
1144 | for (i = start; i < end; i++) | 1201 | bitmap_clear(info->bitmap, start, count); |
1145 | clear_bit(i, info->bitmap); | ||
1146 | 1202 | ||
1147 | info->bytes -= bytes; | 1203 | info->bytes -= bytes; |
1148 | block_group->free_space -= bytes; | 1204 | ctl->free_space -= bytes; |
1149 | } | 1205 | } |
1150 | 1206 | ||
1151 | static void bitmap_set_bits(struct btrfs_block_group_cache *block_group, | 1207 | static void bitmap_set_bits(struct btrfs_free_space_ctl *ctl, |
1152 | struct btrfs_free_space *info, u64 offset, | 1208 | struct btrfs_free_space *info, u64 offset, |
1153 | u64 bytes) | 1209 | u64 bytes) |
1154 | { | 1210 | { |
1155 | unsigned long start, end; | 1211 | unsigned long start, count; |
1156 | unsigned long i; | ||
1157 | 1212 | ||
1158 | start = offset_to_bit(info->offset, block_group->sectorsize, offset); | 1213 | start = offset_to_bit(info->offset, ctl->unit, offset); |
1159 | end = start + bytes_to_bits(bytes, block_group->sectorsize); | 1214 | count = bytes_to_bits(bytes, ctl->unit); |
1160 | BUG_ON(end > BITS_PER_BITMAP); | 1215 | BUG_ON(start + count > BITS_PER_BITMAP); |
1161 | 1216 | ||
1162 | for (i = start; i < end; i++) | 1217 | bitmap_set(info->bitmap, start, count); |
1163 | set_bit(i, info->bitmap); | ||
1164 | 1218 | ||
1165 | info->bytes += bytes; | 1219 | info->bytes += bytes; |
1166 | block_group->free_space += bytes; | 1220 | ctl->free_space += bytes; |
1167 | } | 1221 | } |
1168 | 1222 | ||
1169 | static int search_bitmap(struct btrfs_block_group_cache *block_group, | 1223 | static int search_bitmap(struct btrfs_free_space_ctl *ctl, |
1170 | struct btrfs_free_space *bitmap_info, u64 *offset, | 1224 | struct btrfs_free_space *bitmap_info, u64 *offset, |
1171 | u64 *bytes) | 1225 | u64 *bytes) |
1172 | { | 1226 | { |
@@ -1174,9 +1228,9 @@ static int search_bitmap(struct btrfs_block_group_cache *block_group, | |||
1174 | unsigned long bits, i; | 1228 | unsigned long bits, i; |
1175 | unsigned long next_zero; | 1229 | unsigned long next_zero; |
1176 | 1230 | ||
1177 | i = offset_to_bit(bitmap_info->offset, block_group->sectorsize, | 1231 | i = offset_to_bit(bitmap_info->offset, ctl->unit, |
1178 | max_t(u64, *offset, bitmap_info->offset)); | 1232 | max_t(u64, *offset, bitmap_info->offset)); |
1179 | bits = bytes_to_bits(*bytes, block_group->sectorsize); | 1233 | bits = bytes_to_bits(*bytes, ctl->unit); |
1180 | 1234 | ||
1181 | for (i = find_next_bit(bitmap_info->bitmap, BITS_PER_BITMAP, i); | 1235 | for (i = find_next_bit(bitmap_info->bitmap, BITS_PER_BITMAP, i); |
1182 | i < BITS_PER_BITMAP; | 1236 | i < BITS_PER_BITMAP; |
@@ -1191,29 +1245,25 @@ static int search_bitmap(struct btrfs_block_group_cache *block_group, | |||
1191 | } | 1245 | } |
1192 | 1246 | ||
1193 | if (found_bits) { | 1247 | if (found_bits) { |
1194 | *offset = (u64)(i * block_group->sectorsize) + | 1248 | *offset = (u64)(i * ctl->unit) + bitmap_info->offset; |
1195 | bitmap_info->offset; | 1249 | *bytes = (u64)(found_bits) * ctl->unit; |
1196 | *bytes = (u64)(found_bits) * block_group->sectorsize; | ||
1197 | return 0; | 1250 | return 0; |
1198 | } | 1251 | } |
1199 | 1252 | ||
1200 | return -1; | 1253 | return -1; |
1201 | } | 1254 | } |
1202 | 1255 | ||
1203 | static struct btrfs_free_space *find_free_space(struct btrfs_block_group_cache | 1256 | static struct btrfs_free_space * |
1204 | *block_group, u64 *offset, | 1257 | find_free_space(struct btrfs_free_space_ctl *ctl, u64 *offset, u64 *bytes) |
1205 | u64 *bytes, int debug) | ||
1206 | { | 1258 | { |
1207 | struct btrfs_free_space *entry; | 1259 | struct btrfs_free_space *entry; |
1208 | struct rb_node *node; | 1260 | struct rb_node *node; |
1209 | int ret; | 1261 | int ret; |
1210 | 1262 | ||
1211 | if (!block_group->free_space_offset.rb_node) | 1263 | if (!ctl->free_space_offset.rb_node) |
1212 | return NULL; | 1264 | return NULL; |
1213 | 1265 | ||
1214 | entry = tree_search_offset(block_group, | 1266 | entry = tree_search_offset(ctl, offset_to_bitmap(ctl, *offset), 0, 1); |
1215 | offset_to_bitmap(block_group, *offset), | ||
1216 | 0, 1); | ||
1217 | if (!entry) | 1267 | if (!entry) |
1218 | return NULL; | 1268 | return NULL; |
1219 | 1269 | ||
@@ -1223,7 +1273,7 @@ static struct btrfs_free_space *find_free_space(struct btrfs_block_group_cache | |||
1223 | continue; | 1273 | continue; |
1224 | 1274 | ||
1225 | if (entry->bitmap) { | 1275 | if (entry->bitmap) { |
1226 | ret = search_bitmap(block_group, entry, offset, bytes); | 1276 | ret = search_bitmap(ctl, entry, offset, bytes); |
1227 | if (!ret) | 1277 | if (!ret) |
1228 | return entry; | 1278 | return entry; |
1229 | continue; | 1279 | continue; |
@@ -1237,33 +1287,28 @@ static struct btrfs_free_space *find_free_space(struct btrfs_block_group_cache | |||
1237 | return NULL; | 1287 | return NULL; |
1238 | } | 1288 | } |
1239 | 1289 | ||
1240 | static void add_new_bitmap(struct btrfs_block_group_cache *block_group, | 1290 | static void add_new_bitmap(struct btrfs_free_space_ctl *ctl, |
1241 | struct btrfs_free_space *info, u64 offset) | 1291 | struct btrfs_free_space *info, u64 offset) |
1242 | { | 1292 | { |
1243 | u64 bytes_per_bg = BITS_PER_BITMAP * block_group->sectorsize; | 1293 | info->offset = offset_to_bitmap(ctl, offset); |
1244 | int max_bitmaps = (int)div64_u64(block_group->key.offset + | ||
1245 | bytes_per_bg - 1, bytes_per_bg); | ||
1246 | BUG_ON(block_group->total_bitmaps >= max_bitmaps); | ||
1247 | |||
1248 | info->offset = offset_to_bitmap(block_group, offset); | ||
1249 | info->bytes = 0; | 1294 | info->bytes = 0; |
1250 | link_free_space(block_group, info); | 1295 | link_free_space(ctl, info); |
1251 | block_group->total_bitmaps++; | 1296 | ctl->total_bitmaps++; |
1252 | 1297 | ||
1253 | recalculate_thresholds(block_group); | 1298 | ctl->op->recalc_thresholds(ctl); |
1254 | } | 1299 | } |
1255 | 1300 | ||
1256 | static void free_bitmap(struct btrfs_block_group_cache *block_group, | 1301 | static void free_bitmap(struct btrfs_free_space_ctl *ctl, |
1257 | struct btrfs_free_space *bitmap_info) | 1302 | struct btrfs_free_space *bitmap_info) |
1258 | { | 1303 | { |
1259 | unlink_free_space(block_group, bitmap_info); | 1304 | unlink_free_space(ctl, bitmap_info); |
1260 | kfree(bitmap_info->bitmap); | 1305 | kfree(bitmap_info->bitmap); |
1261 | kmem_cache_free(btrfs_free_space_cachep, bitmap_info); | 1306 | kmem_cache_free(btrfs_free_space_cachep, bitmap_info); |
1262 | block_group->total_bitmaps--; | 1307 | ctl->total_bitmaps--; |
1263 | recalculate_thresholds(block_group); | 1308 | ctl->op->recalc_thresholds(ctl); |
1264 | } | 1309 | } |
1265 | 1310 | ||
1266 | static noinline int remove_from_bitmap(struct btrfs_block_group_cache *block_group, | 1311 | static noinline int remove_from_bitmap(struct btrfs_free_space_ctl *ctl, |
1267 | struct btrfs_free_space *bitmap_info, | 1312 | struct btrfs_free_space *bitmap_info, |
1268 | u64 *offset, u64 *bytes) | 1313 | u64 *offset, u64 *bytes) |
1269 | { | 1314 | { |
@@ -1272,8 +1317,7 @@ static noinline int remove_from_bitmap(struct btrfs_block_group_cache *block_gro | |||
1272 | int ret; | 1317 | int ret; |
1273 | 1318 | ||
1274 | again: | 1319 | again: |
1275 | end = bitmap_info->offset + | 1320 | end = bitmap_info->offset + (u64)(BITS_PER_BITMAP * ctl->unit) - 1; |
1276 | (u64)(BITS_PER_BITMAP * block_group->sectorsize) - 1; | ||
1277 | 1321 | ||
1278 | /* | 1322 | /* |
1279 | * XXX - this can go away after a few releases. | 1323 | * XXX - this can go away after a few releases. |
@@ -1288,24 +1332,22 @@ again: | |||
1288 | search_start = *offset; | 1332 | search_start = *offset; |
1289 | search_bytes = *bytes; | 1333 | search_bytes = *bytes; |
1290 | search_bytes = min(search_bytes, end - search_start + 1); | 1334 | search_bytes = min(search_bytes, end - search_start + 1); |
1291 | ret = search_bitmap(block_group, bitmap_info, &search_start, | 1335 | ret = search_bitmap(ctl, bitmap_info, &search_start, &search_bytes); |
1292 | &search_bytes); | ||
1293 | BUG_ON(ret < 0 || search_start != *offset); | 1336 | BUG_ON(ret < 0 || search_start != *offset); |
1294 | 1337 | ||
1295 | if (*offset > bitmap_info->offset && *offset + *bytes > end) { | 1338 | if (*offset > bitmap_info->offset && *offset + *bytes > end) { |
1296 | bitmap_clear_bits(block_group, bitmap_info, *offset, | 1339 | bitmap_clear_bits(ctl, bitmap_info, *offset, end - *offset + 1); |
1297 | end - *offset + 1); | ||
1298 | *bytes -= end - *offset + 1; | 1340 | *bytes -= end - *offset + 1; |
1299 | *offset = end + 1; | 1341 | *offset = end + 1; |
1300 | } else if (*offset >= bitmap_info->offset && *offset + *bytes <= end) { | 1342 | } else if (*offset >= bitmap_info->offset && *offset + *bytes <= end) { |
1301 | bitmap_clear_bits(block_group, bitmap_info, *offset, *bytes); | 1343 | bitmap_clear_bits(ctl, bitmap_info, *offset, *bytes); |
1302 | *bytes = 0; | 1344 | *bytes = 0; |
1303 | } | 1345 | } |
1304 | 1346 | ||
1305 | if (*bytes) { | 1347 | if (*bytes) { |
1306 | struct rb_node *next = rb_next(&bitmap_info->offset_index); | 1348 | struct rb_node *next = rb_next(&bitmap_info->offset_index); |
1307 | if (!bitmap_info->bytes) | 1349 | if (!bitmap_info->bytes) |
1308 | free_bitmap(block_group, bitmap_info); | 1350 | free_bitmap(ctl, bitmap_info); |
1309 | 1351 | ||
1310 | /* | 1352 | /* |
1311 | * no entry after this bitmap, but we still have bytes to | 1353 | * no entry after this bitmap, but we still have bytes to |
@@ -1332,31 +1374,28 @@ again: | |||
1332 | */ | 1374 | */ |
1333 | search_start = *offset; | 1375 | search_start = *offset; |
1334 | search_bytes = *bytes; | 1376 | search_bytes = *bytes; |
1335 | ret = search_bitmap(block_group, bitmap_info, &search_start, | 1377 | ret = search_bitmap(ctl, bitmap_info, &search_start, |
1336 | &search_bytes); | 1378 | &search_bytes); |
1337 | if (ret < 0 || search_start != *offset) | 1379 | if (ret < 0 || search_start != *offset) |
1338 | return -EAGAIN; | 1380 | return -EAGAIN; |
1339 | 1381 | ||
1340 | goto again; | 1382 | goto again; |
1341 | } else if (!bitmap_info->bytes) | 1383 | } else if (!bitmap_info->bytes) |
1342 | free_bitmap(block_group, bitmap_info); | 1384 | free_bitmap(ctl, bitmap_info); |
1343 | 1385 | ||
1344 | return 0; | 1386 | return 0; |
1345 | } | 1387 | } |
1346 | 1388 | ||
1347 | static int insert_into_bitmap(struct btrfs_block_group_cache *block_group, | 1389 | static bool use_bitmap(struct btrfs_free_space_ctl *ctl, |
1348 | struct btrfs_free_space *info) | 1390 | struct btrfs_free_space *info) |
1349 | { | 1391 | { |
1350 | struct btrfs_free_space *bitmap_info; | 1392 | struct btrfs_block_group_cache *block_group = ctl->private; |
1351 | int added = 0; | ||
1352 | u64 bytes, offset, end; | ||
1353 | int ret; | ||
1354 | 1393 | ||
1355 | /* | 1394 | /* |
1356 | * If we are below the extents threshold then we can add this as an | 1395 | * If we are below the extents threshold then we can add this as an |
1357 | * extent, and don't have to deal with the bitmap | 1396 | * extent, and don't have to deal with the bitmap |
1358 | */ | 1397 | */ |
1359 | if (block_group->free_extents < block_group->extents_thresh) { | 1398 | if (ctl->free_extents < ctl->extents_thresh) { |
1360 | /* | 1399 | /* |
1361 | * If this block group has some small extents we don't want to | 1400 | * If this block group has some small extents we don't want to |
1362 | * use up all of our free slots in the cache with them, we want | 1401 | * use up all of our free slots in the cache with them, we want |
@@ -1365,11 +1404,10 @@ static int insert_into_bitmap(struct btrfs_block_group_cache *block_group, | |||
1365 | * the overhead of a bitmap if we don't have to. | 1404 | * the overhead of a bitmap if we don't have to. |
1366 | */ | 1405 | */ |
1367 | if (info->bytes <= block_group->sectorsize * 4) { | 1406 | if (info->bytes <= block_group->sectorsize * 4) { |
1368 | if (block_group->free_extents * 2 <= | 1407 | if (ctl->free_extents * 2 <= ctl->extents_thresh) |
1369 | block_group->extents_thresh) | 1408 | return false; |
1370 | return 0; | ||
1371 | } else { | 1409 | } else { |
1372 | return 0; | 1410 | return false; |
1373 | } | 1411 | } |
1374 | } | 1412 | } |
1375 | 1413 | ||
@@ -1379,31 +1417,42 @@ static int insert_into_bitmap(struct btrfs_block_group_cache *block_group, | |||
1379 | */ | 1417 | */ |
1380 | if (BITS_PER_BITMAP * block_group->sectorsize > | 1418 | if (BITS_PER_BITMAP * block_group->sectorsize > |
1381 | block_group->key.offset) | 1419 | block_group->key.offset) |
1382 | return 0; | 1420 | return false; |
1421 | |||
1422 | return true; | ||
1423 | } | ||
1424 | |||
1425 | static int insert_into_bitmap(struct btrfs_free_space_ctl *ctl, | ||
1426 | struct btrfs_free_space *info) | ||
1427 | { | ||
1428 | struct btrfs_free_space *bitmap_info; | ||
1429 | int added = 0; | ||
1430 | u64 bytes, offset, end; | ||
1431 | int ret; | ||
1383 | 1432 | ||
1384 | bytes = info->bytes; | 1433 | bytes = info->bytes; |
1385 | offset = info->offset; | 1434 | offset = info->offset; |
1386 | 1435 | ||
1436 | if (!ctl->op->use_bitmap(ctl, info)) | ||
1437 | return 0; | ||
1438 | |||
1387 | again: | 1439 | again: |
1388 | bitmap_info = tree_search_offset(block_group, | 1440 | bitmap_info = tree_search_offset(ctl, offset_to_bitmap(ctl, offset), |
1389 | offset_to_bitmap(block_group, offset), | ||
1390 | 1, 0); | 1441 | 1, 0); |
1391 | if (!bitmap_info) { | 1442 | if (!bitmap_info) { |
1392 | BUG_ON(added); | 1443 | BUG_ON(added); |
1393 | goto new_bitmap; | 1444 | goto new_bitmap; |
1394 | } | 1445 | } |
1395 | 1446 | ||
1396 | end = bitmap_info->offset + | 1447 | end = bitmap_info->offset + (u64)(BITS_PER_BITMAP * ctl->unit); |
1397 | (u64)(BITS_PER_BITMAP * block_group->sectorsize); | ||
1398 | 1448 | ||
1399 | if (offset >= bitmap_info->offset && offset + bytes > end) { | 1449 | if (offset >= bitmap_info->offset && offset + bytes > end) { |
1400 | bitmap_set_bits(block_group, bitmap_info, offset, | 1450 | bitmap_set_bits(ctl, bitmap_info, offset, end - offset); |
1401 | end - offset); | ||
1402 | bytes -= end - offset; | 1451 | bytes -= end - offset; |
1403 | offset = end; | 1452 | offset = end; |
1404 | added = 0; | 1453 | added = 0; |
1405 | } else if (offset >= bitmap_info->offset && offset + bytes <= end) { | 1454 | } else if (offset >= bitmap_info->offset && offset + bytes <= end) { |
1406 | bitmap_set_bits(block_group, bitmap_info, offset, bytes); | 1455 | bitmap_set_bits(ctl, bitmap_info, offset, bytes); |
1407 | bytes = 0; | 1456 | bytes = 0; |
1408 | } else { | 1457 | } else { |
1409 | BUG(); | 1458 | BUG(); |
@@ -1417,19 +1466,19 @@ again: | |||
1417 | 1466 | ||
1418 | new_bitmap: | 1467 | new_bitmap: |
1419 | if (info && info->bitmap) { | 1468 | if (info && info->bitmap) { |
1420 | add_new_bitmap(block_group, info, offset); | 1469 | add_new_bitmap(ctl, info, offset); |
1421 | added = 1; | 1470 | added = 1; |
1422 | info = NULL; | 1471 | info = NULL; |
1423 | goto again; | 1472 | goto again; |
1424 | } else { | 1473 | } else { |
1425 | spin_unlock(&block_group->tree_lock); | 1474 | spin_unlock(&ctl->tree_lock); |
1426 | 1475 | ||
1427 | /* no pre-allocated info, allocate a new one */ | 1476 | /* no pre-allocated info, allocate a new one */ |
1428 | if (!info) { | 1477 | if (!info) { |
1429 | info = kmem_cache_zalloc(btrfs_free_space_cachep, | 1478 | info = kmem_cache_zalloc(btrfs_free_space_cachep, |
1430 | GFP_NOFS); | 1479 | GFP_NOFS); |
1431 | if (!info) { | 1480 | if (!info) { |
1432 | spin_lock(&block_group->tree_lock); | 1481 | spin_lock(&ctl->tree_lock); |
1433 | ret = -ENOMEM; | 1482 | ret = -ENOMEM; |
1434 | goto out; | 1483 | goto out; |
1435 | } | 1484 | } |
@@ -1437,7 +1486,7 @@ new_bitmap: | |||
1437 | 1486 | ||
1438 | /* allocate the bitmap */ | 1487 | /* allocate the bitmap */ |
1439 | info->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); | 1488 | info->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); |
1440 | spin_lock(&block_group->tree_lock); | 1489 | spin_lock(&ctl->tree_lock); |
1441 | if (!info->bitmap) { | 1490 | if (!info->bitmap) { |
1442 | ret = -ENOMEM; | 1491 | ret = -ENOMEM; |
1443 | goto out; | 1492 | goto out; |
@@ -1455,8 +1504,8 @@ out: | |||
1455 | return ret; | 1504 | return ret; |
1456 | } | 1505 | } |
1457 | 1506 | ||
1458 | static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | 1507 | static bool try_merge_free_space(struct btrfs_free_space_ctl *ctl, |
1459 | struct btrfs_free_space *info, bool update_stat) | 1508 | struct btrfs_free_space *info, bool update_stat) |
1460 | { | 1509 | { |
1461 | struct btrfs_free_space *left_info; | 1510 | struct btrfs_free_space *left_info; |
1462 | struct btrfs_free_space *right_info; | 1511 | struct btrfs_free_space *right_info; |
@@ -1469,18 +1518,18 @@ static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | |||
1469 | * are adding, if there is remove that struct and add a new one to | 1518 | * are adding, if there is remove that struct and add a new one to |
1470 | * cover the entire range | 1519 | * cover the entire range |
1471 | */ | 1520 | */ |
1472 | right_info = tree_search_offset(block_group, offset + bytes, 0, 0); | 1521 | right_info = tree_search_offset(ctl, offset + bytes, 0, 0); |
1473 | if (right_info && rb_prev(&right_info->offset_index)) | 1522 | if (right_info && rb_prev(&right_info->offset_index)) |
1474 | left_info = rb_entry(rb_prev(&right_info->offset_index), | 1523 | left_info = rb_entry(rb_prev(&right_info->offset_index), |
1475 | struct btrfs_free_space, offset_index); | 1524 | struct btrfs_free_space, offset_index); |
1476 | else | 1525 | else |
1477 | left_info = tree_search_offset(block_group, offset - 1, 0, 0); | 1526 | left_info = tree_search_offset(ctl, offset - 1, 0, 0); |
1478 | 1527 | ||
1479 | if (right_info && !right_info->bitmap) { | 1528 | if (right_info && !right_info->bitmap) { |
1480 | if (update_stat) | 1529 | if (update_stat) |
1481 | unlink_free_space(block_group, right_info); | 1530 | unlink_free_space(ctl, right_info); |
1482 | else | 1531 | else |
1483 | __unlink_free_space(block_group, right_info); | 1532 | __unlink_free_space(ctl, right_info); |
1484 | info->bytes += right_info->bytes; | 1533 | info->bytes += right_info->bytes; |
1485 | kmem_cache_free(btrfs_free_space_cachep, right_info); | 1534 | kmem_cache_free(btrfs_free_space_cachep, right_info); |
1486 | merged = true; | 1535 | merged = true; |
@@ -1489,9 +1538,9 @@ static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | |||
1489 | if (left_info && !left_info->bitmap && | 1538 | if (left_info && !left_info->bitmap && |
1490 | left_info->offset + left_info->bytes == offset) { | 1539 | left_info->offset + left_info->bytes == offset) { |
1491 | if (update_stat) | 1540 | if (update_stat) |
1492 | unlink_free_space(block_group, left_info); | 1541 | unlink_free_space(ctl, left_info); |
1493 | else | 1542 | else |
1494 | __unlink_free_space(block_group, left_info); | 1543 | __unlink_free_space(ctl, left_info); |
1495 | info->offset = left_info->offset; | 1544 | info->offset = left_info->offset; |
1496 | info->bytes += left_info->bytes; | 1545 | info->bytes += left_info->bytes; |
1497 | kmem_cache_free(btrfs_free_space_cachep, left_info); | 1546 | kmem_cache_free(btrfs_free_space_cachep, left_info); |
@@ -1501,8 +1550,8 @@ static bool try_merge_free_space(struct btrfs_block_group_cache *block_group, | |||
1501 | return merged; | 1550 | return merged; |
1502 | } | 1551 | } |
1503 | 1552 | ||
1504 | int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | 1553 | int __btrfs_add_free_space(struct btrfs_free_space_ctl *ctl, |
1505 | u64 offset, u64 bytes) | 1554 | u64 offset, u64 bytes) |
1506 | { | 1555 | { |
1507 | struct btrfs_free_space *info; | 1556 | struct btrfs_free_space *info; |
1508 | int ret = 0; | 1557 | int ret = 0; |
@@ -1514,9 +1563,9 @@ int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | |||
1514 | info->offset = offset; | 1563 | info->offset = offset; |
1515 | info->bytes = bytes; | 1564 | info->bytes = bytes; |
1516 | 1565 | ||
1517 | spin_lock(&block_group->tree_lock); | 1566 | spin_lock(&ctl->tree_lock); |
1518 | 1567 | ||
1519 | if (try_merge_free_space(block_group, info, true)) | 1568 | if (try_merge_free_space(ctl, info, true)) |
1520 | goto link; | 1569 | goto link; |
1521 | 1570 | ||
1522 | /* | 1571 | /* |
@@ -1524,7 +1573,7 @@ int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | |||
1524 | * extent then we know we're going to have to allocate a new extent, so | 1573 | * extent then we know we're going to have to allocate a new extent, so |
1525 | * before we do that see if we need to drop this into a bitmap | 1574 | * before we do that see if we need to drop this into a bitmap |
1526 | */ | 1575 | */ |
1527 | ret = insert_into_bitmap(block_group, info); | 1576 | ret = insert_into_bitmap(ctl, info); |
1528 | if (ret < 0) { | 1577 | if (ret < 0) { |
1529 | goto out; | 1578 | goto out; |
1530 | } else if (ret) { | 1579 | } else if (ret) { |
@@ -1532,11 +1581,11 @@ int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | |||
1532 | goto out; | 1581 | goto out; |
1533 | } | 1582 | } |
1534 | link: | 1583 | link: |
1535 | ret = link_free_space(block_group, info); | 1584 | ret = link_free_space(ctl, info); |
1536 | if (ret) | 1585 | if (ret) |
1537 | kmem_cache_free(btrfs_free_space_cachep, info); | 1586 | kmem_cache_free(btrfs_free_space_cachep, info); |
1538 | out: | 1587 | out: |
1539 | spin_unlock(&block_group->tree_lock); | 1588 | spin_unlock(&ctl->tree_lock); |
1540 | 1589 | ||
1541 | if (ret) { | 1590 | if (ret) { |
1542 | printk(KERN_CRIT "btrfs: unable to add free space :%d\n", ret); | 1591 | printk(KERN_CRIT "btrfs: unable to add free space :%d\n", ret); |
@@ -1549,21 +1598,21 @@ out: | |||
1549 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, | 1598 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, |
1550 | u64 offset, u64 bytes) | 1599 | u64 offset, u64 bytes) |
1551 | { | 1600 | { |
1601 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1552 | struct btrfs_free_space *info; | 1602 | struct btrfs_free_space *info; |
1553 | struct btrfs_free_space *next_info = NULL; | 1603 | struct btrfs_free_space *next_info = NULL; |
1554 | int ret = 0; | 1604 | int ret = 0; |
1555 | 1605 | ||
1556 | spin_lock(&block_group->tree_lock); | 1606 | spin_lock(&ctl->tree_lock); |
1557 | 1607 | ||
1558 | again: | 1608 | again: |
1559 | info = tree_search_offset(block_group, offset, 0, 0); | 1609 | info = tree_search_offset(ctl, offset, 0, 0); |
1560 | if (!info) { | 1610 | if (!info) { |
1561 | /* | 1611 | /* |
1562 | * oops didn't find an extent that matched the space we wanted | 1612 | * oops didn't find an extent that matched the space we wanted |
1563 | * to remove, look for a bitmap instead | 1613 | * to remove, look for a bitmap instead |
1564 | */ | 1614 | */ |
1565 | info = tree_search_offset(block_group, | 1615 | info = tree_search_offset(ctl, offset_to_bitmap(ctl, offset), |
1566 | offset_to_bitmap(block_group, offset), | ||
1567 | 1, 0); | 1616 | 1, 0); |
1568 | if (!info) { | 1617 | if (!info) { |
1569 | WARN_ON(1); | 1618 | WARN_ON(1); |
@@ -1578,8 +1627,8 @@ again: | |||
1578 | offset_index); | 1627 | offset_index); |
1579 | 1628 | ||
1580 | if (next_info->bitmap) | 1629 | if (next_info->bitmap) |
1581 | end = next_info->offset + BITS_PER_BITMAP * | 1630 | end = next_info->offset + |
1582 | block_group->sectorsize - 1; | 1631 | BITS_PER_BITMAP * ctl->unit - 1; |
1583 | else | 1632 | else |
1584 | end = next_info->offset + next_info->bytes; | 1633 | end = next_info->offset + next_info->bytes; |
1585 | 1634 | ||
@@ -1599,20 +1648,20 @@ again: | |||
1599 | } | 1648 | } |
1600 | 1649 | ||
1601 | if (info->bytes == bytes) { | 1650 | if (info->bytes == bytes) { |
1602 | unlink_free_space(block_group, info); | 1651 | unlink_free_space(ctl, info); |
1603 | if (info->bitmap) { | 1652 | if (info->bitmap) { |
1604 | kfree(info->bitmap); | 1653 | kfree(info->bitmap); |
1605 | block_group->total_bitmaps--; | 1654 | ctl->total_bitmaps--; |
1606 | } | 1655 | } |
1607 | kmem_cache_free(btrfs_free_space_cachep, info); | 1656 | kmem_cache_free(btrfs_free_space_cachep, info); |
1608 | goto out_lock; | 1657 | goto out_lock; |
1609 | } | 1658 | } |
1610 | 1659 | ||
1611 | if (!info->bitmap && info->offset == offset) { | 1660 | if (!info->bitmap && info->offset == offset) { |
1612 | unlink_free_space(block_group, info); | 1661 | unlink_free_space(ctl, info); |
1613 | info->offset += bytes; | 1662 | info->offset += bytes; |
1614 | info->bytes -= bytes; | 1663 | info->bytes -= bytes; |
1615 | link_free_space(block_group, info); | 1664 | link_free_space(ctl, info); |
1616 | goto out_lock; | 1665 | goto out_lock; |
1617 | } | 1666 | } |
1618 | 1667 | ||
@@ -1626,13 +1675,13 @@ again: | |||
1626 | * first unlink the old info and then | 1675 | * first unlink the old info and then |
1627 | * insert it again after the hole we're creating | 1676 | * insert it again after the hole we're creating |
1628 | */ | 1677 | */ |
1629 | unlink_free_space(block_group, info); | 1678 | unlink_free_space(ctl, info); |
1630 | if (offset + bytes < info->offset + info->bytes) { | 1679 | if (offset + bytes < info->offset + info->bytes) { |
1631 | u64 old_end = info->offset + info->bytes; | 1680 | u64 old_end = info->offset + info->bytes; |
1632 | 1681 | ||
1633 | info->offset = offset + bytes; | 1682 | info->offset = offset + bytes; |
1634 | info->bytes = old_end - info->offset; | 1683 | info->bytes = old_end - info->offset; |
1635 | ret = link_free_space(block_group, info); | 1684 | ret = link_free_space(ctl, info); |
1636 | WARN_ON(ret); | 1685 | WARN_ON(ret); |
1637 | if (ret) | 1686 | if (ret) |
1638 | goto out_lock; | 1687 | goto out_lock; |
@@ -1642,7 +1691,7 @@ again: | |||
1642 | */ | 1691 | */ |
1643 | kmem_cache_free(btrfs_free_space_cachep, info); | 1692 | kmem_cache_free(btrfs_free_space_cachep, info); |
1644 | } | 1693 | } |
1645 | spin_unlock(&block_group->tree_lock); | 1694 | spin_unlock(&ctl->tree_lock); |
1646 | 1695 | ||
1647 | /* step two, insert a new info struct to cover | 1696 | /* step two, insert a new info struct to cover |
1648 | * anything before the hole | 1697 | * anything before the hole |
@@ -1653,12 +1702,12 @@ again: | |||
1653 | goto out; | 1702 | goto out; |
1654 | } | 1703 | } |
1655 | 1704 | ||
1656 | ret = remove_from_bitmap(block_group, info, &offset, &bytes); | 1705 | ret = remove_from_bitmap(ctl, info, &offset, &bytes); |
1657 | if (ret == -EAGAIN) | 1706 | if (ret == -EAGAIN) |
1658 | goto again; | 1707 | goto again; |
1659 | BUG_ON(ret); | 1708 | BUG_ON(ret); |
1660 | out_lock: | 1709 | out_lock: |
1661 | spin_unlock(&block_group->tree_lock); | 1710 | spin_unlock(&ctl->tree_lock); |
1662 | out: | 1711 | out: |
1663 | return ret; | 1712 | return ret; |
1664 | } | 1713 | } |
@@ -1666,11 +1715,12 @@ out: | |||
1666 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, | 1715 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, |
1667 | u64 bytes) | 1716 | u64 bytes) |
1668 | { | 1717 | { |
1718 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1669 | struct btrfs_free_space *info; | 1719 | struct btrfs_free_space *info; |
1670 | struct rb_node *n; | 1720 | struct rb_node *n; |
1671 | int count = 0; | 1721 | int count = 0; |
1672 | 1722 | ||
1673 | for (n = rb_first(&block_group->free_space_offset); n; n = rb_next(n)) { | 1723 | for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) { |
1674 | info = rb_entry(n, struct btrfs_free_space, offset_index); | 1724 | info = rb_entry(n, struct btrfs_free_space, offset_index); |
1675 | if (info->bytes >= bytes) | 1725 | if (info->bytes >= bytes) |
1676 | count++; | 1726 | count++; |
@@ -1685,6 +1735,30 @@ void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, | |||
1685 | "\n", count); | 1735 | "\n", count); |
1686 | } | 1736 | } |
1687 | 1737 | ||
1738 | static struct btrfs_free_space_op free_space_op = { | ||
1739 | .recalc_thresholds = recalculate_thresholds, | ||
1740 | .use_bitmap = use_bitmap, | ||
1741 | }; | ||
1742 | |||
1743 | void btrfs_init_free_space_ctl(struct btrfs_block_group_cache *block_group) | ||
1744 | { | ||
1745 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1746 | |||
1747 | spin_lock_init(&ctl->tree_lock); | ||
1748 | ctl->unit = block_group->sectorsize; | ||
1749 | ctl->start = block_group->key.objectid; | ||
1750 | ctl->private = block_group; | ||
1751 | ctl->op = &free_space_op; | ||
1752 | |||
1753 | /* | ||
1754 | * we only want to have 32k of ram per block group for keeping | ||
1755 | * track of free space, and if we pass 1/2 of that we want to | ||
1756 | * start converting things over to using bitmaps | ||
1757 | */ | ||
1758 | ctl->extents_thresh = ((1024 * 32) / 2) / | ||
1759 | sizeof(struct btrfs_free_space); | ||
1760 | } | ||
1761 | |||
1688 | /* | 1762 | /* |
1689 | * for a given cluster, put all of its extents back into the free | 1763 | * for a given cluster, put all of its extents back into the free |
1690 | * space cache. If the block group passed doesn't match the block group | 1764 | * space cache. If the block group passed doesn't match the block group |
@@ -1696,6 +1770,7 @@ __btrfs_return_cluster_to_free_space( | |||
1696 | struct btrfs_block_group_cache *block_group, | 1770 | struct btrfs_block_group_cache *block_group, |
1697 | struct btrfs_free_cluster *cluster) | 1771 | struct btrfs_free_cluster *cluster) |
1698 | { | 1772 | { |
1773 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1699 | struct btrfs_free_space *entry; | 1774 | struct btrfs_free_space *entry; |
1700 | struct rb_node *node; | 1775 | struct rb_node *node; |
1701 | 1776 | ||
@@ -1717,8 +1792,8 @@ __btrfs_return_cluster_to_free_space( | |||
1717 | 1792 | ||
1718 | bitmap = (entry->bitmap != NULL); | 1793 | bitmap = (entry->bitmap != NULL); |
1719 | if (!bitmap) | 1794 | if (!bitmap) |
1720 | try_merge_free_space(block_group, entry, false); | 1795 | try_merge_free_space(ctl, entry, false); |
1721 | tree_insert_offset(&block_group->free_space_offset, | 1796 | tree_insert_offset(&ctl->free_space_offset, |
1722 | entry->offset, &entry->offset_index, bitmap); | 1797 | entry->offset, &entry->offset_index, bitmap); |
1723 | } | 1798 | } |
1724 | cluster->root = RB_ROOT; | 1799 | cluster->root = RB_ROOT; |
@@ -1729,14 +1804,38 @@ out: | |||
1729 | return 0; | 1804 | return 0; |
1730 | } | 1805 | } |
1731 | 1806 | ||
1732 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache *block_group) | 1807 | void __btrfs_remove_free_space_cache_locked(struct btrfs_free_space_ctl *ctl) |
1733 | { | 1808 | { |
1734 | struct btrfs_free_space *info; | 1809 | struct btrfs_free_space *info; |
1735 | struct rb_node *node; | 1810 | struct rb_node *node; |
1811 | |||
1812 | while ((node = rb_last(&ctl->free_space_offset)) != NULL) { | ||
1813 | info = rb_entry(node, struct btrfs_free_space, offset_index); | ||
1814 | unlink_free_space(ctl, info); | ||
1815 | kfree(info->bitmap); | ||
1816 | kmem_cache_free(btrfs_free_space_cachep, info); | ||
1817 | if (need_resched()) { | ||
1818 | spin_unlock(&ctl->tree_lock); | ||
1819 | cond_resched(); | ||
1820 | spin_lock(&ctl->tree_lock); | ||
1821 | } | ||
1822 | } | ||
1823 | } | ||
1824 | |||
1825 | void __btrfs_remove_free_space_cache(struct btrfs_free_space_ctl *ctl) | ||
1826 | { | ||
1827 | spin_lock(&ctl->tree_lock); | ||
1828 | __btrfs_remove_free_space_cache_locked(ctl); | ||
1829 | spin_unlock(&ctl->tree_lock); | ||
1830 | } | ||
1831 | |||
1832 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache *block_group) | ||
1833 | { | ||
1834 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1736 | struct btrfs_free_cluster *cluster; | 1835 | struct btrfs_free_cluster *cluster; |
1737 | struct list_head *head; | 1836 | struct list_head *head; |
1738 | 1837 | ||
1739 | spin_lock(&block_group->tree_lock); | 1838 | spin_lock(&ctl->tree_lock); |
1740 | while ((head = block_group->cluster_list.next) != | 1839 | while ((head = block_group->cluster_list.next) != |
1741 | &block_group->cluster_list) { | 1840 | &block_group->cluster_list) { |
1742 | cluster = list_entry(head, struct btrfs_free_cluster, | 1841 | cluster = list_entry(head, struct btrfs_free_cluster, |
@@ -1745,60 +1844,46 @@ void btrfs_remove_free_space_cache(struct btrfs_block_group_cache *block_group) | |||
1745 | WARN_ON(cluster->block_group != block_group); | 1844 | WARN_ON(cluster->block_group != block_group); |
1746 | __btrfs_return_cluster_to_free_space(block_group, cluster); | 1845 | __btrfs_return_cluster_to_free_space(block_group, cluster); |
1747 | if (need_resched()) { | 1846 | if (need_resched()) { |
1748 | spin_unlock(&block_group->tree_lock); | 1847 | spin_unlock(&ctl->tree_lock); |
1749 | cond_resched(); | 1848 | cond_resched(); |
1750 | spin_lock(&block_group->tree_lock); | 1849 | spin_lock(&ctl->tree_lock); |
1751 | } | 1850 | } |
1752 | } | 1851 | } |
1852 | __btrfs_remove_free_space_cache_locked(ctl); | ||
1853 | spin_unlock(&ctl->tree_lock); | ||
1753 | 1854 | ||
1754 | while ((node = rb_last(&block_group->free_space_offset)) != NULL) { | ||
1755 | info = rb_entry(node, struct btrfs_free_space, offset_index); | ||
1756 | if (!info->bitmap) { | ||
1757 | unlink_free_space(block_group, info); | ||
1758 | kmem_cache_free(btrfs_free_space_cachep, info); | ||
1759 | } else { | ||
1760 | free_bitmap(block_group, info); | ||
1761 | } | ||
1762 | |||
1763 | if (need_resched()) { | ||
1764 | spin_unlock(&block_group->tree_lock); | ||
1765 | cond_resched(); | ||
1766 | spin_lock(&block_group->tree_lock); | ||
1767 | } | ||
1768 | } | ||
1769 | |||
1770 | spin_unlock(&block_group->tree_lock); | ||
1771 | } | 1855 | } |
1772 | 1856 | ||
1773 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, | 1857 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, |
1774 | u64 offset, u64 bytes, u64 empty_size) | 1858 | u64 offset, u64 bytes, u64 empty_size) |
1775 | { | 1859 | { |
1860 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1776 | struct btrfs_free_space *entry = NULL; | 1861 | struct btrfs_free_space *entry = NULL; |
1777 | u64 bytes_search = bytes + empty_size; | 1862 | u64 bytes_search = bytes + empty_size; |
1778 | u64 ret = 0; | 1863 | u64 ret = 0; |
1779 | 1864 | ||
1780 | spin_lock(&block_group->tree_lock); | 1865 | spin_lock(&ctl->tree_lock); |
1781 | entry = find_free_space(block_group, &offset, &bytes_search, 0); | 1866 | entry = find_free_space(ctl, &offset, &bytes_search); |
1782 | if (!entry) | 1867 | if (!entry) |
1783 | goto out; | 1868 | goto out; |
1784 | 1869 | ||
1785 | ret = offset; | 1870 | ret = offset; |
1786 | if (entry->bitmap) { | 1871 | if (entry->bitmap) { |
1787 | bitmap_clear_bits(block_group, entry, offset, bytes); | 1872 | bitmap_clear_bits(ctl, entry, offset, bytes); |
1788 | if (!entry->bytes) | 1873 | if (!entry->bytes) |
1789 | free_bitmap(block_group, entry); | 1874 | free_bitmap(ctl, entry); |
1790 | } else { | 1875 | } else { |
1791 | unlink_free_space(block_group, entry); | 1876 | unlink_free_space(ctl, entry); |
1792 | entry->offset += bytes; | 1877 | entry->offset += bytes; |
1793 | entry->bytes -= bytes; | 1878 | entry->bytes -= bytes; |
1794 | if (!entry->bytes) | 1879 | if (!entry->bytes) |
1795 | kmem_cache_free(btrfs_free_space_cachep, entry); | 1880 | kmem_cache_free(btrfs_free_space_cachep, entry); |
1796 | else | 1881 | else |
1797 | link_free_space(block_group, entry); | 1882 | link_free_space(ctl, entry); |
1798 | } | 1883 | } |
1799 | 1884 | ||
1800 | out: | 1885 | out: |
1801 | spin_unlock(&block_group->tree_lock); | 1886 | spin_unlock(&ctl->tree_lock); |
1802 | 1887 | ||
1803 | return ret; | 1888 | return ret; |
1804 | } | 1889 | } |
@@ -1815,6 +1900,7 @@ int btrfs_return_cluster_to_free_space( | |||
1815 | struct btrfs_block_group_cache *block_group, | 1900 | struct btrfs_block_group_cache *block_group, |
1816 | struct btrfs_free_cluster *cluster) | 1901 | struct btrfs_free_cluster *cluster) |
1817 | { | 1902 | { |
1903 | struct btrfs_free_space_ctl *ctl; | ||
1818 | int ret; | 1904 | int ret; |
1819 | 1905 | ||
1820 | /* first, get a safe pointer to the block group */ | 1906 | /* first, get a safe pointer to the block group */ |
@@ -1833,10 +1919,12 @@ int btrfs_return_cluster_to_free_space( | |||
1833 | atomic_inc(&block_group->count); | 1919 | atomic_inc(&block_group->count); |
1834 | spin_unlock(&cluster->lock); | 1920 | spin_unlock(&cluster->lock); |
1835 | 1921 | ||
1922 | ctl = block_group->free_space_ctl; | ||
1923 | |||
1836 | /* now return any extents the cluster had on it */ | 1924 | /* now return any extents the cluster had on it */ |
1837 | spin_lock(&block_group->tree_lock); | 1925 | spin_lock(&ctl->tree_lock); |
1838 | ret = __btrfs_return_cluster_to_free_space(block_group, cluster); | 1926 | ret = __btrfs_return_cluster_to_free_space(block_group, cluster); |
1839 | spin_unlock(&block_group->tree_lock); | 1927 | spin_unlock(&ctl->tree_lock); |
1840 | 1928 | ||
1841 | /* finally drop our ref */ | 1929 | /* finally drop our ref */ |
1842 | btrfs_put_block_group(block_group); | 1930 | btrfs_put_block_group(block_group); |
@@ -1848,6 +1936,7 @@ static u64 btrfs_alloc_from_bitmap(struct btrfs_block_group_cache *block_group, | |||
1848 | struct btrfs_free_space *entry, | 1936 | struct btrfs_free_space *entry, |
1849 | u64 bytes, u64 min_start) | 1937 | u64 bytes, u64 min_start) |
1850 | { | 1938 | { |
1939 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1851 | int err; | 1940 | int err; |
1852 | u64 search_start = cluster->window_start; | 1941 | u64 search_start = cluster->window_start; |
1853 | u64 search_bytes = bytes; | 1942 | u64 search_bytes = bytes; |
@@ -1856,13 +1945,12 @@ static u64 btrfs_alloc_from_bitmap(struct btrfs_block_group_cache *block_group, | |||
1856 | search_start = min_start; | 1945 | search_start = min_start; |
1857 | search_bytes = bytes; | 1946 | search_bytes = bytes; |
1858 | 1947 | ||
1859 | err = search_bitmap(block_group, entry, &search_start, | 1948 | err = search_bitmap(ctl, entry, &search_start, &search_bytes); |
1860 | &search_bytes); | ||
1861 | if (err) | 1949 | if (err) |
1862 | return 0; | 1950 | return 0; |
1863 | 1951 | ||
1864 | ret = search_start; | 1952 | ret = search_start; |
1865 | bitmap_clear_bits(block_group, entry, ret, bytes); | 1953 | bitmap_clear_bits(ctl, entry, ret, bytes); |
1866 | 1954 | ||
1867 | return ret; | 1955 | return ret; |
1868 | } | 1956 | } |
@@ -1876,6 +1964,7 @@ u64 btrfs_alloc_from_cluster(struct btrfs_block_group_cache *block_group, | |||
1876 | struct btrfs_free_cluster *cluster, u64 bytes, | 1964 | struct btrfs_free_cluster *cluster, u64 bytes, |
1877 | u64 min_start) | 1965 | u64 min_start) |
1878 | { | 1966 | { |
1967 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1879 | struct btrfs_free_space *entry = NULL; | 1968 | struct btrfs_free_space *entry = NULL; |
1880 | struct rb_node *node; | 1969 | struct rb_node *node; |
1881 | u64 ret = 0; | 1970 | u64 ret = 0; |
@@ -1933,20 +2022,20 @@ out: | |||
1933 | if (!ret) | 2022 | if (!ret) |
1934 | return 0; | 2023 | return 0; |
1935 | 2024 | ||
1936 | spin_lock(&block_group->tree_lock); | 2025 | spin_lock(&ctl->tree_lock); |
1937 | 2026 | ||
1938 | block_group->free_space -= bytes; | 2027 | ctl->free_space -= bytes; |
1939 | if (entry->bytes == 0) { | 2028 | if (entry->bytes == 0) { |
1940 | block_group->free_extents--; | 2029 | ctl->free_extents--; |
1941 | if (entry->bitmap) { | 2030 | if (entry->bitmap) { |
1942 | kfree(entry->bitmap); | 2031 | kfree(entry->bitmap); |
1943 | block_group->total_bitmaps--; | 2032 | ctl->total_bitmaps--; |
1944 | recalculate_thresholds(block_group); | 2033 | ctl->op->recalc_thresholds(ctl); |
1945 | } | 2034 | } |
1946 | kmem_cache_free(btrfs_free_space_cachep, entry); | 2035 | kmem_cache_free(btrfs_free_space_cachep, entry); |
1947 | } | 2036 | } |
1948 | 2037 | ||
1949 | spin_unlock(&block_group->tree_lock); | 2038 | spin_unlock(&ctl->tree_lock); |
1950 | 2039 | ||
1951 | return ret; | 2040 | return ret; |
1952 | } | 2041 | } |
@@ -1956,6 +2045,7 @@ static int btrfs_bitmap_cluster(struct btrfs_block_group_cache *block_group, | |||
1956 | struct btrfs_free_cluster *cluster, | 2045 | struct btrfs_free_cluster *cluster, |
1957 | u64 offset, u64 bytes, u64 min_bytes) | 2046 | u64 offset, u64 bytes, u64 min_bytes) |
1958 | { | 2047 | { |
2048 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
1959 | unsigned long next_zero; | 2049 | unsigned long next_zero; |
1960 | unsigned long i; | 2050 | unsigned long i; |
1961 | unsigned long search_bits; | 2051 | unsigned long search_bits; |
@@ -2010,7 +2100,7 @@ again: | |||
2010 | 2100 | ||
2011 | cluster->window_start = start * block_group->sectorsize + | 2101 | cluster->window_start = start * block_group->sectorsize + |
2012 | entry->offset; | 2102 | entry->offset; |
2013 | rb_erase(&entry->offset_index, &block_group->free_space_offset); | 2103 | rb_erase(&entry->offset_index, &ctl->free_space_offset); |
2014 | ret = tree_insert_offset(&cluster->root, entry->offset, | 2104 | ret = tree_insert_offset(&cluster->root, entry->offset, |
2015 | &entry->offset_index, 1); | 2105 | &entry->offset_index, 1); |
2016 | BUG_ON(ret); | 2106 | BUG_ON(ret); |
@@ -2025,6 +2115,7 @@ static int setup_cluster_no_bitmap(struct btrfs_block_group_cache *block_group, | |||
2025 | struct btrfs_free_cluster *cluster, | 2115 | struct btrfs_free_cluster *cluster, |
2026 | u64 offset, u64 bytes, u64 min_bytes) | 2116 | u64 offset, u64 bytes, u64 min_bytes) |
2027 | { | 2117 | { |
2118 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2028 | struct btrfs_free_space *first = NULL; | 2119 | struct btrfs_free_space *first = NULL; |
2029 | struct btrfs_free_space *entry = NULL; | 2120 | struct btrfs_free_space *entry = NULL; |
2030 | struct btrfs_free_space *prev = NULL; | 2121 | struct btrfs_free_space *prev = NULL; |
@@ -2035,7 +2126,7 @@ static int setup_cluster_no_bitmap(struct btrfs_block_group_cache *block_group, | |||
2035 | u64 max_extent; | 2126 | u64 max_extent; |
2036 | u64 max_gap = 128 * 1024; | 2127 | u64 max_gap = 128 * 1024; |
2037 | 2128 | ||
2038 | entry = tree_search_offset(block_group, offset, 0, 1); | 2129 | entry = tree_search_offset(ctl, offset, 0, 1); |
2039 | if (!entry) | 2130 | if (!entry) |
2040 | return -ENOSPC; | 2131 | return -ENOSPC; |
2041 | 2132 | ||
@@ -2101,7 +2192,7 @@ static int setup_cluster_no_bitmap(struct btrfs_block_group_cache *block_group, | |||
2101 | if (entry->bitmap) | 2192 | if (entry->bitmap) |
2102 | continue; | 2193 | continue; |
2103 | 2194 | ||
2104 | rb_erase(&entry->offset_index, &block_group->free_space_offset); | 2195 | rb_erase(&entry->offset_index, &ctl->free_space_offset); |
2105 | ret = tree_insert_offset(&cluster->root, entry->offset, | 2196 | ret = tree_insert_offset(&cluster->root, entry->offset, |
2106 | &entry->offset_index, 0); | 2197 | &entry->offset_index, 0); |
2107 | BUG_ON(ret); | 2198 | BUG_ON(ret); |
@@ -2120,16 +2211,15 @@ static int setup_cluster_bitmap(struct btrfs_block_group_cache *block_group, | |||
2120 | struct btrfs_free_cluster *cluster, | 2211 | struct btrfs_free_cluster *cluster, |
2121 | u64 offset, u64 bytes, u64 min_bytes) | 2212 | u64 offset, u64 bytes, u64 min_bytes) |
2122 | { | 2213 | { |
2214 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2123 | struct btrfs_free_space *entry; | 2215 | struct btrfs_free_space *entry; |
2124 | struct rb_node *node; | 2216 | struct rb_node *node; |
2125 | int ret = -ENOSPC; | 2217 | int ret = -ENOSPC; |
2126 | 2218 | ||
2127 | if (block_group->total_bitmaps == 0) | 2219 | if (ctl->total_bitmaps == 0) |
2128 | return -ENOSPC; | 2220 | return -ENOSPC; |
2129 | 2221 | ||
2130 | entry = tree_search_offset(block_group, | 2222 | entry = tree_search_offset(ctl, offset_to_bitmap(ctl, offset), 0, 1); |
2131 | offset_to_bitmap(block_group, offset), | ||
2132 | 0, 1); | ||
2133 | if (!entry) | 2223 | if (!entry) |
2134 | return -ENOSPC; | 2224 | return -ENOSPC; |
2135 | 2225 | ||
@@ -2162,6 +2252,7 @@ int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | |||
2162 | struct btrfs_free_cluster *cluster, | 2252 | struct btrfs_free_cluster *cluster, |
2163 | u64 offset, u64 bytes, u64 empty_size) | 2253 | u64 offset, u64 bytes, u64 empty_size) |
2164 | { | 2254 | { |
2255 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2165 | u64 min_bytes; | 2256 | u64 min_bytes; |
2166 | int ret; | 2257 | int ret; |
2167 | 2258 | ||
@@ -2181,14 +2272,14 @@ int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | |||
2181 | } else | 2272 | } else |
2182 | min_bytes = max(bytes, (bytes + empty_size) >> 2); | 2273 | min_bytes = max(bytes, (bytes + empty_size) >> 2); |
2183 | 2274 | ||
2184 | spin_lock(&block_group->tree_lock); | 2275 | spin_lock(&ctl->tree_lock); |
2185 | 2276 | ||
2186 | /* | 2277 | /* |
2187 | * If we know we don't have enough space to make a cluster don't even | 2278 | * If we know we don't have enough space to make a cluster don't even |
2188 | * bother doing all the work to try and find one. | 2279 | * bother doing all the work to try and find one. |
2189 | */ | 2280 | */ |
2190 | if (block_group->free_space < min_bytes) { | 2281 | if (ctl->free_space < min_bytes) { |
2191 | spin_unlock(&block_group->tree_lock); | 2282 | spin_unlock(&ctl->tree_lock); |
2192 | return -ENOSPC; | 2283 | return -ENOSPC; |
2193 | } | 2284 | } |
2194 | 2285 | ||
@@ -2214,7 +2305,7 @@ int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | |||
2214 | } | 2305 | } |
2215 | out: | 2306 | out: |
2216 | spin_unlock(&cluster->lock); | 2307 | spin_unlock(&cluster->lock); |
2217 | spin_unlock(&block_group->tree_lock); | 2308 | spin_unlock(&ctl->tree_lock); |
2218 | 2309 | ||
2219 | return ret; | 2310 | return ret; |
2220 | } | 2311 | } |
@@ -2235,6 +2326,7 @@ void btrfs_init_free_cluster(struct btrfs_free_cluster *cluster) | |||
2235 | int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | 2326 | int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, |
2236 | u64 *trimmed, u64 start, u64 end, u64 minlen) | 2327 | u64 *trimmed, u64 start, u64 end, u64 minlen) |
2237 | { | 2328 | { |
2329 | struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl; | ||
2238 | struct btrfs_free_space *entry = NULL; | 2330 | struct btrfs_free_space *entry = NULL; |
2239 | struct btrfs_fs_info *fs_info = block_group->fs_info; | 2331 | struct btrfs_fs_info *fs_info = block_group->fs_info; |
2240 | u64 bytes = 0; | 2332 | u64 bytes = 0; |
@@ -2244,52 +2336,50 @@ int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | |||
2244 | *trimmed = 0; | 2336 | *trimmed = 0; |
2245 | 2337 | ||
2246 | while (start < end) { | 2338 | while (start < end) { |
2247 | spin_lock(&block_group->tree_lock); | 2339 | spin_lock(&ctl->tree_lock); |
2248 | 2340 | ||
2249 | if (block_group->free_space < minlen) { | 2341 | if (ctl->free_space < minlen) { |
2250 | spin_unlock(&block_group->tree_lock); | 2342 | spin_unlock(&ctl->tree_lock); |
2251 | break; | 2343 | break; |
2252 | } | 2344 | } |
2253 | 2345 | ||
2254 | entry = tree_search_offset(block_group, start, 0, 1); | 2346 | entry = tree_search_offset(ctl, start, 0, 1); |
2255 | if (!entry) | 2347 | if (!entry) |
2256 | entry = tree_search_offset(block_group, | 2348 | entry = tree_search_offset(ctl, |
2257 | offset_to_bitmap(block_group, | 2349 | offset_to_bitmap(ctl, start), |
2258 | start), | ||
2259 | 1, 1); | 2350 | 1, 1); |
2260 | 2351 | ||
2261 | if (!entry || entry->offset >= end) { | 2352 | if (!entry || entry->offset >= end) { |
2262 | spin_unlock(&block_group->tree_lock); | 2353 | spin_unlock(&ctl->tree_lock); |
2263 | break; | 2354 | break; |
2264 | } | 2355 | } |
2265 | 2356 | ||
2266 | if (entry->bitmap) { | 2357 | if (entry->bitmap) { |
2267 | ret = search_bitmap(block_group, entry, &start, &bytes); | 2358 | ret = search_bitmap(ctl, entry, &start, &bytes); |
2268 | if (!ret) { | 2359 | if (!ret) { |
2269 | if (start >= end) { | 2360 | if (start >= end) { |
2270 | spin_unlock(&block_group->tree_lock); | 2361 | spin_unlock(&ctl->tree_lock); |
2271 | break; | 2362 | break; |
2272 | } | 2363 | } |
2273 | bytes = min(bytes, end - start); | 2364 | bytes = min(bytes, end - start); |
2274 | bitmap_clear_bits(block_group, entry, | 2365 | bitmap_clear_bits(ctl, entry, start, bytes); |
2275 | start, bytes); | ||
2276 | if (entry->bytes == 0) | 2366 | if (entry->bytes == 0) |
2277 | free_bitmap(block_group, entry); | 2367 | free_bitmap(ctl, entry); |
2278 | } else { | 2368 | } else { |
2279 | start = entry->offset + BITS_PER_BITMAP * | 2369 | start = entry->offset + BITS_PER_BITMAP * |
2280 | block_group->sectorsize; | 2370 | block_group->sectorsize; |
2281 | spin_unlock(&block_group->tree_lock); | 2371 | spin_unlock(&ctl->tree_lock); |
2282 | ret = 0; | 2372 | ret = 0; |
2283 | continue; | 2373 | continue; |
2284 | } | 2374 | } |
2285 | } else { | 2375 | } else { |
2286 | start = entry->offset; | 2376 | start = entry->offset; |
2287 | bytes = min(entry->bytes, end - start); | 2377 | bytes = min(entry->bytes, end - start); |
2288 | unlink_free_space(block_group, entry); | 2378 | unlink_free_space(ctl, entry); |
2289 | kmem_cache_free(btrfs_free_space_cachep, entry); | 2379 | kmem_cache_free(btrfs_free_space_cachep, entry); |
2290 | } | 2380 | } |
2291 | 2381 | ||
2292 | spin_unlock(&block_group->tree_lock); | 2382 | spin_unlock(&ctl->tree_lock); |
2293 | 2383 | ||
2294 | if (bytes >= minlen) { | 2384 | if (bytes >= minlen) { |
2295 | int update_ret; | 2385 | int update_ret; |
@@ -2301,8 +2391,7 @@ int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | |||
2301 | bytes, | 2391 | bytes, |
2302 | &actually_trimmed); | 2392 | &actually_trimmed); |
2303 | 2393 | ||
2304 | btrfs_add_free_space(block_group, | 2394 | btrfs_add_free_space(block_group, start, bytes); |
2305 | start, bytes); | ||
2306 | if (!update_ret) | 2395 | if (!update_ret) |
2307 | btrfs_update_reserved_bytes(block_group, | 2396 | btrfs_update_reserved_bytes(block_group, |
2308 | bytes, 0, 1); | 2397 | bytes, 0, 1); |
@@ -2324,3 +2413,145 @@ int btrfs_trim_block_group(struct btrfs_block_group_cache *block_group, | |||
2324 | 2413 | ||
2325 | return ret; | 2414 | return ret; |
2326 | } | 2415 | } |
2416 | |||
2417 | /* | ||
2418 | * Find the left-most item in the cache tree, and then return the | ||
2419 | * smallest inode number in the item. | ||
2420 | * | ||
2421 | * Note: the returned inode number may not be the smallest one in | ||
2422 | * the tree, if the left-most item is a bitmap. | ||
2423 | */ | ||
2424 | u64 btrfs_find_ino_for_alloc(struct btrfs_root *fs_root) | ||
2425 | { | ||
2426 | struct btrfs_free_space_ctl *ctl = fs_root->free_ino_ctl; | ||
2427 | struct btrfs_free_space *entry = NULL; | ||
2428 | u64 ino = 0; | ||
2429 | |||
2430 | spin_lock(&ctl->tree_lock); | ||
2431 | |||
2432 | if (RB_EMPTY_ROOT(&ctl->free_space_offset)) | ||
2433 | goto out; | ||
2434 | |||
2435 | entry = rb_entry(rb_first(&ctl->free_space_offset), | ||
2436 | struct btrfs_free_space, offset_index); | ||
2437 | |||
2438 | if (!entry->bitmap) { | ||
2439 | ino = entry->offset; | ||
2440 | |||
2441 | unlink_free_space(ctl, entry); | ||
2442 | entry->offset++; | ||
2443 | entry->bytes--; | ||
2444 | if (!entry->bytes) | ||
2445 | kmem_cache_free(btrfs_free_space_cachep, entry); | ||
2446 | else | ||
2447 | link_free_space(ctl, entry); | ||
2448 | } else { | ||
2449 | u64 offset = 0; | ||
2450 | u64 count = 1; | ||
2451 | int ret; | ||
2452 | |||
2453 | ret = search_bitmap(ctl, entry, &offset, &count); | ||
2454 | BUG_ON(ret); | ||
2455 | |||
2456 | ino = offset; | ||
2457 | bitmap_clear_bits(ctl, entry, offset, 1); | ||
2458 | if (entry->bytes == 0) | ||
2459 | free_bitmap(ctl, entry); | ||
2460 | } | ||
2461 | out: | ||
2462 | spin_unlock(&ctl->tree_lock); | ||
2463 | |||
2464 | return ino; | ||
2465 | } | ||
2466 | |||
2467 | struct inode *lookup_free_ino_inode(struct btrfs_root *root, | ||
2468 | struct btrfs_path *path) | ||
2469 | { | ||
2470 | struct inode *inode = NULL; | ||
2471 | |||
2472 | spin_lock(&root->cache_lock); | ||
2473 | if (root->cache_inode) | ||
2474 | inode = igrab(root->cache_inode); | ||
2475 | spin_unlock(&root->cache_lock); | ||
2476 | if (inode) | ||
2477 | return inode; | ||
2478 | |||
2479 | inode = __lookup_free_space_inode(root, path, 0); | ||
2480 | if (IS_ERR(inode)) | ||
2481 | return inode; | ||
2482 | |||
2483 | spin_lock(&root->cache_lock); | ||
2484 | if (!root->fs_info->closing) | ||
2485 | root->cache_inode = igrab(inode); | ||
2486 | spin_unlock(&root->cache_lock); | ||
2487 | |||
2488 | return inode; | ||
2489 | } | ||
2490 | |||
2491 | int create_free_ino_inode(struct btrfs_root *root, | ||
2492 | struct btrfs_trans_handle *trans, | ||
2493 | struct btrfs_path *path) | ||
2494 | { | ||
2495 | return __create_free_space_inode(root, trans, path, | ||
2496 | BTRFS_FREE_INO_OBJECTID, 0); | ||
2497 | } | ||
2498 | |||
2499 | int load_free_ino_cache(struct btrfs_fs_info *fs_info, struct btrfs_root *root) | ||
2500 | { | ||
2501 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
2502 | struct btrfs_path *path; | ||
2503 | struct inode *inode; | ||
2504 | int ret = 0; | ||
2505 | u64 root_gen = btrfs_root_generation(&root->root_item); | ||
2506 | |||
2507 | /* | ||
2508 | * If we're unmounting then just return, since this does a search on the | ||
2509 | * normal root and not the commit root and we could deadlock. | ||
2510 | */ | ||
2511 | smp_mb(); | ||
2512 | if (fs_info->closing) | ||
2513 | return 0; | ||
2514 | |||
2515 | path = btrfs_alloc_path(); | ||
2516 | if (!path) | ||
2517 | return 0; | ||
2518 | |||
2519 | inode = lookup_free_ino_inode(root, path); | ||
2520 | if (IS_ERR(inode)) | ||
2521 | goto out; | ||
2522 | |||
2523 | if (root_gen != BTRFS_I(inode)->generation) | ||
2524 | goto out_put; | ||
2525 | |||
2526 | ret = __load_free_space_cache(root, inode, ctl, path, 0); | ||
2527 | |||
2528 | if (ret < 0) | ||
2529 | printk(KERN_ERR "btrfs: failed to load free ino cache for " | ||
2530 | "root %llu\n", root->root_key.objectid); | ||
2531 | out_put: | ||
2532 | iput(inode); | ||
2533 | out: | ||
2534 | btrfs_free_path(path); | ||
2535 | return ret; | ||
2536 | } | ||
2537 | |||
2538 | int btrfs_write_out_ino_cache(struct btrfs_root *root, | ||
2539 | struct btrfs_trans_handle *trans, | ||
2540 | struct btrfs_path *path) | ||
2541 | { | ||
2542 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
2543 | struct inode *inode; | ||
2544 | int ret; | ||
2545 | |||
2546 | inode = lookup_free_ino_inode(root, path); | ||
2547 | if (IS_ERR(inode)) | ||
2548 | return 0; | ||
2549 | |||
2550 | ret = __btrfs_write_out_cache(root, inode, ctl, NULL, trans, path, 0); | ||
2551 | if (ret < 0) | ||
2552 | printk(KERN_ERR "btrfs: failed to write free ino cache " | ||
2553 | "for root %llu\n", root->root_key.objectid); | ||
2554 | |||
2555 | iput(inode); | ||
2556 | return ret; | ||
2557 | } | ||
diff --git a/fs/btrfs/free-space-cache.h b/fs/btrfs/free-space-cache.h index 12b2b5165f8a..8f2613f779ed 100644 --- a/fs/btrfs/free-space-cache.h +++ b/fs/btrfs/free-space-cache.h | |||
@@ -27,6 +27,25 @@ struct btrfs_free_space { | |||
27 | struct list_head list; | 27 | struct list_head list; |
28 | }; | 28 | }; |
29 | 29 | ||
30 | struct btrfs_free_space_ctl { | ||
31 | spinlock_t tree_lock; | ||
32 | struct rb_root free_space_offset; | ||
33 | u64 free_space; | ||
34 | int extents_thresh; | ||
35 | int free_extents; | ||
36 | int total_bitmaps; | ||
37 | int unit; | ||
38 | u64 start; | ||
39 | struct btrfs_free_space_op *op; | ||
40 | void *private; | ||
41 | }; | ||
42 | |||
43 | struct btrfs_free_space_op { | ||
44 | void (*recalc_thresholds)(struct btrfs_free_space_ctl *ctl); | ||
45 | bool (*use_bitmap)(struct btrfs_free_space_ctl *ctl, | ||
46 | struct btrfs_free_space *info); | ||
47 | }; | ||
48 | |||
30 | struct inode *lookup_free_space_inode(struct btrfs_root *root, | 49 | struct inode *lookup_free_space_inode(struct btrfs_root *root, |
31 | struct btrfs_block_group_cache | 50 | struct btrfs_block_group_cache |
32 | *block_group, struct btrfs_path *path); | 51 | *block_group, struct btrfs_path *path); |
@@ -45,14 +64,36 @@ int btrfs_write_out_cache(struct btrfs_root *root, | |||
45 | struct btrfs_trans_handle *trans, | 64 | struct btrfs_trans_handle *trans, |
46 | struct btrfs_block_group_cache *block_group, | 65 | struct btrfs_block_group_cache *block_group, |
47 | struct btrfs_path *path); | 66 | struct btrfs_path *path); |
48 | int btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | 67 | |
49 | u64 bytenr, u64 size); | 68 | struct inode *lookup_free_ino_inode(struct btrfs_root *root, |
69 | struct btrfs_path *path); | ||
70 | int create_free_ino_inode(struct btrfs_root *root, | ||
71 | struct btrfs_trans_handle *trans, | ||
72 | struct btrfs_path *path); | ||
73 | int load_free_ino_cache(struct btrfs_fs_info *fs_info, | ||
74 | struct btrfs_root *root); | ||
75 | int btrfs_write_out_ino_cache(struct btrfs_root *root, | ||
76 | struct btrfs_trans_handle *trans, | ||
77 | struct btrfs_path *path); | ||
78 | |||
79 | void btrfs_init_free_space_ctl(struct btrfs_block_group_cache *block_group); | ||
80 | int __btrfs_add_free_space(struct btrfs_free_space_ctl *ctl, | ||
81 | u64 bytenr, u64 size); | ||
82 | static inline int | ||
83 | btrfs_add_free_space(struct btrfs_block_group_cache *block_group, | ||
84 | u64 bytenr, u64 size) | ||
85 | { | ||
86 | return __btrfs_add_free_space(block_group->free_space_ctl, | ||
87 | bytenr, size); | ||
88 | } | ||
50 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, | 89 | int btrfs_remove_free_space(struct btrfs_block_group_cache *block_group, |
51 | u64 bytenr, u64 size); | 90 | u64 bytenr, u64 size); |
91 | void __btrfs_remove_free_space_cache(struct btrfs_free_space_ctl *ctl); | ||
52 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache | 92 | void btrfs_remove_free_space_cache(struct btrfs_block_group_cache |
53 | *block_group); | 93 | *block_group); |
54 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, | 94 | u64 btrfs_find_space_for_alloc(struct btrfs_block_group_cache *block_group, |
55 | u64 offset, u64 bytes, u64 empty_size); | 95 | u64 offset, u64 bytes, u64 empty_size); |
96 | u64 btrfs_find_ino_for_alloc(struct btrfs_root *fs_root); | ||
56 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, | 97 | void btrfs_dump_free_space(struct btrfs_block_group_cache *block_group, |
57 | u64 bytes); | 98 | u64 bytes); |
58 | int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, | 99 | int btrfs_find_space_cluster(struct btrfs_trans_handle *trans, |
diff --git a/fs/btrfs/inode-map.c b/fs/btrfs/inode-map.c index c05a08f4c411..000970512624 100644 --- a/fs/btrfs/inode-map.c +++ b/fs/btrfs/inode-map.c | |||
@@ -16,11 +16,430 @@ | |||
16 | * Boston, MA 021110-1307, USA. | 16 | * Boston, MA 021110-1307, USA. |
17 | */ | 17 | */ |
18 | 18 | ||
19 | #include <linux/delay.h> | ||
20 | #include <linux/kthread.h> | ||
21 | #include <linux/pagemap.h> | ||
22 | |||
19 | #include "ctree.h" | 23 | #include "ctree.h" |
20 | #include "disk-io.h" | 24 | #include "disk-io.h" |
25 | #include "free-space-cache.h" | ||
26 | #include "inode-map.h" | ||
21 | #include "transaction.h" | 27 | #include "transaction.h" |
22 | 28 | ||
23 | int btrfs_find_highest_inode(struct btrfs_root *root, u64 *objectid) | 29 | static int caching_kthread(void *data) |
30 | { | ||
31 | struct btrfs_root *root = data; | ||
32 | struct btrfs_fs_info *fs_info = root->fs_info; | ||
33 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
34 | struct btrfs_key key; | ||
35 | struct btrfs_path *path; | ||
36 | struct extent_buffer *leaf; | ||
37 | u64 last = (u64)-1; | ||
38 | int slot; | ||
39 | int ret; | ||
40 | |||
41 | path = btrfs_alloc_path(); | ||
42 | if (!path) | ||
43 | return -ENOMEM; | ||
44 | |||
45 | /* Since the commit root is read-only, we can safely skip locking. */ | ||
46 | path->skip_locking = 1; | ||
47 | path->search_commit_root = 1; | ||
48 | path->reada = 2; | ||
49 | |||
50 | key.objectid = BTRFS_FIRST_FREE_OBJECTID; | ||
51 | key.offset = 0; | ||
52 | key.type = BTRFS_INODE_ITEM_KEY; | ||
53 | again: | ||
54 | /* need to make sure the commit_root doesn't disappear */ | ||
55 | mutex_lock(&root->fs_commit_mutex); | ||
56 | |||
57 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | ||
58 | if (ret < 0) | ||
59 | goto out; | ||
60 | |||
61 | while (1) { | ||
62 | smp_mb(); | ||
63 | if (fs_info->closing > 1) | ||
64 | goto out; | ||
65 | |||
66 | leaf = path->nodes[0]; | ||
67 | slot = path->slots[0]; | ||
68 | if (path->slots[0] >= btrfs_header_nritems(leaf)) { | ||
69 | ret = btrfs_next_leaf(root, path); | ||
70 | if (ret < 0) | ||
71 | goto out; | ||
72 | else if (ret > 0) | ||
73 | break; | ||
74 | |||
75 | if (need_resched() || | ||
76 | btrfs_transaction_in_commit(fs_info)) { | ||
77 | leaf = path->nodes[0]; | ||
78 | |||
79 | if (btrfs_header_nritems(leaf) == 0) { | ||
80 | WARN_ON(1); | ||
81 | break; | ||
82 | } | ||
83 | |||
84 | /* | ||
85 | * Save the key so we can advances forward | ||
86 | * in the next search. | ||
87 | */ | ||
88 | btrfs_item_key_to_cpu(leaf, &key, 0); | ||
89 | btrfs_release_path(path); | ||
90 | root->cache_progress = last; | ||
91 | mutex_unlock(&root->fs_commit_mutex); | ||
92 | schedule_timeout(1); | ||
93 | goto again; | ||
94 | } else | ||
95 | continue; | ||
96 | } | ||
97 | |||
98 | btrfs_item_key_to_cpu(leaf, &key, slot); | ||
99 | |||
100 | if (key.type != BTRFS_INODE_ITEM_KEY) | ||
101 | goto next; | ||
102 | |||
103 | if (key.objectid >= BTRFS_LAST_FREE_OBJECTID) | ||
104 | break; | ||
105 | |||
106 | if (last != (u64)-1 && last + 1 != key.objectid) { | ||
107 | __btrfs_add_free_space(ctl, last + 1, | ||
108 | key.objectid - last - 1); | ||
109 | wake_up(&root->cache_wait); | ||
110 | } | ||
111 | |||
112 | last = key.objectid; | ||
113 | next: | ||
114 | path->slots[0]++; | ||
115 | } | ||
116 | |||
117 | if (last < BTRFS_LAST_FREE_OBJECTID - 1) { | ||
118 | __btrfs_add_free_space(ctl, last + 1, | ||
119 | BTRFS_LAST_FREE_OBJECTID - last - 1); | ||
120 | } | ||
121 | |||
122 | spin_lock(&root->cache_lock); | ||
123 | root->cached = BTRFS_CACHE_FINISHED; | ||
124 | spin_unlock(&root->cache_lock); | ||
125 | |||
126 | root->cache_progress = (u64)-1; | ||
127 | btrfs_unpin_free_ino(root); | ||
128 | out: | ||
129 | wake_up(&root->cache_wait); | ||
130 | mutex_unlock(&root->fs_commit_mutex); | ||
131 | |||
132 | btrfs_free_path(path); | ||
133 | |||
134 | return ret; | ||
135 | } | ||
136 | |||
137 | static void start_caching(struct btrfs_root *root) | ||
138 | { | ||
139 | struct task_struct *tsk; | ||
140 | int ret; | ||
141 | |||
142 | spin_lock(&root->cache_lock); | ||
143 | if (root->cached != BTRFS_CACHE_NO) { | ||
144 | spin_unlock(&root->cache_lock); | ||
145 | return; | ||
146 | } | ||
147 | |||
148 | root->cached = BTRFS_CACHE_STARTED; | ||
149 | spin_unlock(&root->cache_lock); | ||
150 | |||
151 | ret = load_free_ino_cache(root->fs_info, root); | ||
152 | if (ret == 1) { | ||
153 | spin_lock(&root->cache_lock); | ||
154 | root->cached = BTRFS_CACHE_FINISHED; | ||
155 | spin_unlock(&root->cache_lock); | ||
156 | return; | ||
157 | } | ||
158 | |||
159 | tsk = kthread_run(caching_kthread, root, "btrfs-ino-cache-%llu\n", | ||
160 | root->root_key.objectid); | ||
161 | BUG_ON(IS_ERR(tsk)); | ||
162 | } | ||
163 | |||
164 | int btrfs_find_free_ino(struct btrfs_root *root, u64 *objectid) | ||
165 | { | ||
166 | again: | ||
167 | *objectid = btrfs_find_ino_for_alloc(root); | ||
168 | |||
169 | if (*objectid != 0) | ||
170 | return 0; | ||
171 | |||
172 | start_caching(root); | ||
173 | |||
174 | wait_event(root->cache_wait, | ||
175 | root->cached == BTRFS_CACHE_FINISHED || | ||
176 | root->free_ino_ctl->free_space > 0); | ||
177 | |||
178 | if (root->cached == BTRFS_CACHE_FINISHED && | ||
179 | root->free_ino_ctl->free_space == 0) | ||
180 | return -ENOSPC; | ||
181 | else | ||
182 | goto again; | ||
183 | } | ||
184 | |||
185 | void btrfs_return_ino(struct btrfs_root *root, u64 objectid) | ||
186 | { | ||
187 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
188 | struct btrfs_free_space_ctl *pinned = root->free_ino_pinned; | ||
189 | again: | ||
190 | if (root->cached == BTRFS_CACHE_FINISHED) { | ||
191 | __btrfs_add_free_space(ctl, objectid, 1); | ||
192 | } else { | ||
193 | /* | ||
194 | * If we are in the process of caching free ino chunks, | ||
195 | * to avoid adding the same inode number to the free_ino | ||
196 | * tree twice due to cross transaction, we'll leave it | ||
197 | * in the pinned tree until a transaction is committed | ||
198 | * or the caching work is done. | ||
199 | */ | ||
200 | |||
201 | mutex_lock(&root->fs_commit_mutex); | ||
202 | spin_lock(&root->cache_lock); | ||
203 | if (root->cached == BTRFS_CACHE_FINISHED) { | ||
204 | spin_unlock(&root->cache_lock); | ||
205 | mutex_unlock(&root->fs_commit_mutex); | ||
206 | goto again; | ||
207 | } | ||
208 | spin_unlock(&root->cache_lock); | ||
209 | |||
210 | start_caching(root); | ||
211 | |||
212 | if (objectid <= root->cache_progress) | ||
213 | __btrfs_add_free_space(ctl, objectid, 1); | ||
214 | else | ||
215 | __btrfs_add_free_space(pinned, objectid, 1); | ||
216 | |||
217 | mutex_unlock(&root->fs_commit_mutex); | ||
218 | } | ||
219 | } | ||
220 | |||
221 | /* | ||
222 | * When a transaction is committed, we'll move those inode numbers which | ||
223 | * are smaller than root->cache_progress from pinned tree to free_ino tree, | ||
224 | * and others will just be dropped, because the commit root we were | ||
225 | * searching has changed. | ||
226 | * | ||
227 | * Must be called with root->fs_commit_mutex held | ||
228 | */ | ||
229 | void btrfs_unpin_free_ino(struct btrfs_root *root) | ||
230 | { | ||
231 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
232 | struct rb_root *rbroot = &root->free_ino_pinned->free_space_offset; | ||
233 | struct btrfs_free_space *info; | ||
234 | struct rb_node *n; | ||
235 | u64 count; | ||
236 | |||
237 | while (1) { | ||
238 | n = rb_first(rbroot); | ||
239 | if (!n) | ||
240 | break; | ||
241 | |||
242 | info = rb_entry(n, struct btrfs_free_space, offset_index); | ||
243 | BUG_ON(info->bitmap); | ||
244 | |||
245 | if (info->offset > root->cache_progress) | ||
246 | goto free; | ||
247 | else if (info->offset + info->bytes > root->cache_progress) | ||
248 | count = root->cache_progress - info->offset + 1; | ||
249 | else | ||
250 | count = info->bytes; | ||
251 | |||
252 | __btrfs_add_free_space(ctl, info->offset, count); | ||
253 | free: | ||
254 | rb_erase(&info->offset_index, rbroot); | ||
255 | kfree(info); | ||
256 | } | ||
257 | } | ||
258 | |||
259 | #define INIT_THRESHOLD (((1024 * 32) / 2) / sizeof(struct btrfs_free_space)) | ||
260 | #define INODES_PER_BITMAP (PAGE_CACHE_SIZE * 8) | ||
261 | |||
262 | /* | ||
263 | * The goal is to keep the memory used by the free_ino tree won't | ||
264 | * exceed the memory if we use bitmaps only. | ||
265 | */ | ||
266 | static void recalculate_thresholds(struct btrfs_free_space_ctl *ctl) | ||
267 | { | ||
268 | struct btrfs_free_space *info; | ||
269 | struct rb_node *n; | ||
270 | int max_ino; | ||
271 | int max_bitmaps; | ||
272 | |||
273 | n = rb_last(&ctl->free_space_offset); | ||
274 | if (!n) { | ||
275 | ctl->extents_thresh = INIT_THRESHOLD; | ||
276 | return; | ||
277 | } | ||
278 | info = rb_entry(n, struct btrfs_free_space, offset_index); | ||
279 | |||
280 | /* | ||
281 | * Find the maximum inode number in the filesystem. Note we | ||
282 | * ignore the fact that this can be a bitmap, because we are | ||
283 | * not doing precise calculation. | ||
284 | */ | ||
285 | max_ino = info->bytes - 1; | ||
286 | |||
287 | max_bitmaps = ALIGN(max_ino, INODES_PER_BITMAP) / INODES_PER_BITMAP; | ||
288 | if (max_bitmaps <= ctl->total_bitmaps) { | ||
289 | ctl->extents_thresh = 0; | ||
290 | return; | ||
291 | } | ||
292 | |||
293 | ctl->extents_thresh = (max_bitmaps - ctl->total_bitmaps) * | ||
294 | PAGE_CACHE_SIZE / sizeof(*info); | ||
295 | } | ||
296 | |||
297 | /* | ||
298 | * We don't fall back to bitmap, if we are below the extents threshold | ||
299 | * or this chunk of inode numbers is a big one. | ||
300 | */ | ||
301 | static bool use_bitmap(struct btrfs_free_space_ctl *ctl, | ||
302 | struct btrfs_free_space *info) | ||
303 | { | ||
304 | if (ctl->free_extents < ctl->extents_thresh || | ||
305 | info->bytes > INODES_PER_BITMAP / 10) | ||
306 | return false; | ||
307 | |||
308 | return true; | ||
309 | } | ||
310 | |||
311 | static struct btrfs_free_space_op free_ino_op = { | ||
312 | .recalc_thresholds = recalculate_thresholds, | ||
313 | .use_bitmap = use_bitmap, | ||
314 | }; | ||
315 | |||
316 | static void pinned_recalc_thresholds(struct btrfs_free_space_ctl *ctl) | ||
317 | { | ||
318 | } | ||
319 | |||
320 | static bool pinned_use_bitmap(struct btrfs_free_space_ctl *ctl, | ||
321 | struct btrfs_free_space *info) | ||
322 | { | ||
323 | /* | ||
324 | * We always use extents for two reasons: | ||
325 | * | ||
326 | * - The pinned tree is only used during the process of caching | ||
327 | * work. | ||
328 | * - Make code simpler. See btrfs_unpin_free_ino(). | ||
329 | */ | ||
330 | return false; | ||
331 | } | ||
332 | |||
333 | static struct btrfs_free_space_op pinned_free_ino_op = { | ||
334 | .recalc_thresholds = pinned_recalc_thresholds, | ||
335 | .use_bitmap = pinned_use_bitmap, | ||
336 | }; | ||
337 | |||
338 | void btrfs_init_free_ino_ctl(struct btrfs_root *root) | ||
339 | { | ||
340 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
341 | struct btrfs_free_space_ctl *pinned = root->free_ino_pinned; | ||
342 | |||
343 | spin_lock_init(&ctl->tree_lock); | ||
344 | ctl->unit = 1; | ||
345 | ctl->start = 0; | ||
346 | ctl->private = NULL; | ||
347 | ctl->op = &free_ino_op; | ||
348 | |||
349 | /* | ||
350 | * Initially we allow to use 16K of ram to cache chunks of | ||
351 | * inode numbers before we resort to bitmaps. This is somewhat | ||
352 | * arbitrary, but it will be adjusted in runtime. | ||
353 | */ | ||
354 | ctl->extents_thresh = INIT_THRESHOLD; | ||
355 | |||
356 | spin_lock_init(&pinned->tree_lock); | ||
357 | pinned->unit = 1; | ||
358 | pinned->start = 0; | ||
359 | pinned->private = NULL; | ||
360 | pinned->extents_thresh = 0; | ||
361 | pinned->op = &pinned_free_ino_op; | ||
362 | } | ||
363 | |||
364 | int btrfs_save_ino_cache(struct btrfs_root *root, | ||
365 | struct btrfs_trans_handle *trans) | ||
366 | { | ||
367 | struct btrfs_free_space_ctl *ctl = root->free_ino_ctl; | ||
368 | struct btrfs_path *path; | ||
369 | struct inode *inode; | ||
370 | u64 alloc_hint = 0; | ||
371 | int ret; | ||
372 | int prealloc; | ||
373 | bool retry = false; | ||
374 | |||
375 | path = btrfs_alloc_path(); | ||
376 | if (!path) | ||
377 | return -ENOMEM; | ||
378 | again: | ||
379 | inode = lookup_free_ino_inode(root, path); | ||
380 | if (IS_ERR(inode) && PTR_ERR(inode) != -ENOENT) { | ||
381 | ret = PTR_ERR(inode); | ||
382 | goto out; | ||
383 | } | ||
384 | |||
385 | if (IS_ERR(inode)) { | ||
386 | BUG_ON(retry); | ||
387 | retry = true; | ||
388 | |||
389 | ret = create_free_ino_inode(root, trans, path); | ||
390 | if (ret) | ||
391 | goto out; | ||
392 | goto again; | ||
393 | } | ||
394 | |||
395 | BTRFS_I(inode)->generation = 0; | ||
396 | ret = btrfs_update_inode(trans, root, inode); | ||
397 | WARN_ON(ret); | ||
398 | |||
399 | if (i_size_read(inode) > 0) { | ||
400 | ret = btrfs_truncate_free_space_cache(root, trans, path, inode); | ||
401 | if (ret) | ||
402 | goto out_put; | ||
403 | } | ||
404 | |||
405 | spin_lock(&root->cache_lock); | ||
406 | if (root->cached != BTRFS_CACHE_FINISHED) { | ||
407 | ret = -1; | ||
408 | spin_unlock(&root->cache_lock); | ||
409 | goto out_put; | ||
410 | } | ||
411 | spin_unlock(&root->cache_lock); | ||
412 | |||
413 | spin_lock(&ctl->tree_lock); | ||
414 | prealloc = sizeof(struct btrfs_free_space) * ctl->free_extents; | ||
415 | prealloc = ALIGN(prealloc, PAGE_CACHE_SIZE); | ||
416 | prealloc += ctl->total_bitmaps * PAGE_CACHE_SIZE; | ||
417 | spin_unlock(&ctl->tree_lock); | ||
418 | |||
419 | /* Just to make sure we have enough space */ | ||
420 | prealloc += 8 * PAGE_CACHE_SIZE; | ||
421 | |||
422 | ret = btrfs_check_data_free_space(inode, prealloc); | ||
423 | if (ret) | ||
424 | goto out_put; | ||
425 | |||
426 | ret = btrfs_prealloc_file_range_trans(inode, trans, 0, 0, prealloc, | ||
427 | prealloc, prealloc, &alloc_hint); | ||
428 | if (ret) | ||
429 | goto out_put; | ||
430 | btrfs_free_reserved_data_space(inode, prealloc); | ||
431 | |||
432 | out_put: | ||
433 | iput(inode); | ||
434 | out: | ||
435 | if (ret == 0) | ||
436 | ret = btrfs_write_out_ino_cache(root, trans, path); | ||
437 | |||
438 | btrfs_free_path(path); | ||
439 | return ret; | ||
440 | } | ||
441 | |||
442 | static int btrfs_find_highest_objectid(struct btrfs_root *root, u64 *objectid) | ||
24 | { | 443 | { |
25 | struct btrfs_path *path; | 444 | struct btrfs_path *path; |
26 | int ret; | 445 | int ret; |
@@ -55,15 +474,14 @@ error: | |||
55 | return ret; | 474 | return ret; |
56 | } | 475 | } |
57 | 476 | ||
58 | int btrfs_find_free_objectid(struct btrfs_trans_handle *trans, | 477 | int btrfs_find_free_objectid(struct btrfs_root *root, u64 *objectid) |
59 | struct btrfs_root *root, | ||
60 | u64 dirid, u64 *objectid) | ||
61 | { | 478 | { |
62 | int ret; | 479 | int ret; |
63 | mutex_lock(&root->objectid_mutex); | 480 | mutex_lock(&root->objectid_mutex); |
64 | 481 | ||
65 | if (unlikely(root->highest_objectid < BTRFS_FIRST_FREE_OBJECTID)) { | 482 | if (unlikely(root->highest_objectid < BTRFS_FIRST_FREE_OBJECTID)) { |
66 | ret = btrfs_find_highest_inode(root, &root->highest_objectid); | 483 | ret = btrfs_find_highest_objectid(root, |
484 | &root->highest_objectid); | ||
67 | if (ret) | 485 | if (ret) |
68 | goto out; | 486 | goto out; |
69 | } | 487 | } |
diff --git a/fs/btrfs/inode-map.h b/fs/btrfs/inode-map.h new file mode 100644 index 000000000000..ddb347bfee23 --- /dev/null +++ b/fs/btrfs/inode-map.h | |||
@@ -0,0 +1,13 @@ | |||
1 | #ifndef __BTRFS_INODE_MAP | ||
2 | #define __BTRFS_INODE_MAP | ||
3 | |||
4 | void btrfs_init_free_ino_ctl(struct btrfs_root *root); | ||
5 | void btrfs_unpin_free_ino(struct btrfs_root *root); | ||
6 | void btrfs_return_ino(struct btrfs_root *root, u64 objectid); | ||
7 | int btrfs_find_free_ino(struct btrfs_root *root, u64 *objectid); | ||
8 | int btrfs_save_ino_cache(struct btrfs_root *root, | ||
9 | struct btrfs_trans_handle *trans); | ||
10 | |||
11 | int btrfs_find_free_objectid(struct btrfs_root *root, u64 *objectid); | ||
12 | |||
13 | #endif | ||
diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c index 1d1017f91558..8ae72c3eedb1 100644 --- a/fs/btrfs/inode.c +++ b/fs/btrfs/inode.c | |||
@@ -52,6 +52,7 @@ | |||
52 | #include "compression.h" | 52 | #include "compression.h" |
53 | #include "locking.h" | 53 | #include "locking.h" |
54 | #include "free-space-cache.h" | 54 | #include "free-space-cache.h" |
55 | #include "inode-map.h" | ||
55 | 56 | ||
56 | struct btrfs_iget_args { | 57 | struct btrfs_iget_args { |
57 | u64 ino; | 58 | u64 ino; |
@@ -139,7 +140,7 @@ static noinline int insert_inline_extent(struct btrfs_trans_handle *trans, | |||
139 | path->leave_spinning = 1; | 140 | path->leave_spinning = 1; |
140 | btrfs_set_trans_block_group(trans, inode); | 141 | btrfs_set_trans_block_group(trans, inode); |
141 | 142 | ||
142 | key.objectid = inode->i_ino; | 143 | key.objectid = btrfs_ino(inode); |
143 | key.offset = start; | 144 | key.offset = start; |
144 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); | 145 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); |
145 | datasize = btrfs_file_extent_calc_inline_size(cur_size); | 146 | datasize = btrfs_file_extent_calc_inline_size(cur_size); |
@@ -746,6 +747,15 @@ static u64 get_extent_allocation_hint(struct inode *inode, u64 start, | |||
746 | return alloc_hint; | 747 | return alloc_hint; |
747 | } | 748 | } |
748 | 749 | ||
750 | static inline bool is_free_space_inode(struct btrfs_root *root, | ||
751 | struct inode *inode) | ||
752 | { | ||
753 | if (root == root->fs_info->tree_root || | ||
754 | BTRFS_I(inode)->location.objectid == BTRFS_FREE_INO_OBJECTID) | ||
755 | return true; | ||
756 | return false; | ||
757 | } | ||
758 | |||
749 | /* | 759 | /* |
750 | * when extent_io.c finds a delayed allocation range in the file, | 760 | * when extent_io.c finds a delayed allocation range in the file, |
751 | * the call backs end up in this code. The basic idea is to | 761 | * the call backs end up in this code. The basic idea is to |
@@ -778,7 +788,7 @@ static noinline int cow_file_range(struct inode *inode, | |||
778 | struct extent_map_tree *em_tree = &BTRFS_I(inode)->extent_tree; | 788 | struct extent_map_tree *em_tree = &BTRFS_I(inode)->extent_tree; |
779 | int ret = 0; | 789 | int ret = 0; |
780 | 790 | ||
781 | BUG_ON(root == root->fs_info->tree_root); | 791 | BUG_ON(is_free_space_inode(root, inode)); |
782 | trans = btrfs_join_transaction(root, 1); | 792 | trans = btrfs_join_transaction(root, 1); |
783 | BUG_ON(IS_ERR(trans)); | 793 | BUG_ON(IS_ERR(trans)); |
784 | btrfs_set_trans_block_group(trans, inode); | 794 | btrfs_set_trans_block_group(trans, inode); |
@@ -1050,29 +1060,31 @@ static noinline int run_delalloc_nocow(struct inode *inode, | |||
1050 | int type; | 1060 | int type; |
1051 | int nocow; | 1061 | int nocow; |
1052 | int check_prev = 1; | 1062 | int check_prev = 1; |
1053 | bool nolock = false; | 1063 | bool nolock; |
1064 | u64 ino = btrfs_ino(inode); | ||
1054 | 1065 | ||
1055 | path = btrfs_alloc_path(); | 1066 | path = btrfs_alloc_path(); |
1056 | BUG_ON(!path); | 1067 | BUG_ON(!path); |
1057 | if (root == root->fs_info->tree_root) { | 1068 | |
1058 | nolock = true; | 1069 | nolock = is_free_space_inode(root, inode); |
1070 | |||
1071 | if (nolock) | ||
1059 | trans = btrfs_join_transaction_nolock(root, 1); | 1072 | trans = btrfs_join_transaction_nolock(root, 1); |
1060 | } else { | 1073 | else |
1061 | trans = btrfs_join_transaction(root, 1); | 1074 | trans = btrfs_join_transaction(root, 1); |
1062 | } | ||
1063 | BUG_ON(IS_ERR(trans)); | 1075 | BUG_ON(IS_ERR(trans)); |
1064 | 1076 | ||
1065 | cow_start = (u64)-1; | 1077 | cow_start = (u64)-1; |
1066 | cur_offset = start; | 1078 | cur_offset = start; |
1067 | while (1) { | 1079 | while (1) { |
1068 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 1080 | ret = btrfs_lookup_file_extent(trans, root, path, ino, |
1069 | cur_offset, 0); | 1081 | cur_offset, 0); |
1070 | BUG_ON(ret < 0); | 1082 | BUG_ON(ret < 0); |
1071 | if (ret > 0 && path->slots[0] > 0 && check_prev) { | 1083 | if (ret > 0 && path->slots[0] > 0 && check_prev) { |
1072 | leaf = path->nodes[0]; | 1084 | leaf = path->nodes[0]; |
1073 | btrfs_item_key_to_cpu(leaf, &found_key, | 1085 | btrfs_item_key_to_cpu(leaf, &found_key, |
1074 | path->slots[0] - 1); | 1086 | path->slots[0] - 1); |
1075 | if (found_key.objectid == inode->i_ino && | 1087 | if (found_key.objectid == ino && |
1076 | found_key.type == BTRFS_EXTENT_DATA_KEY) | 1088 | found_key.type == BTRFS_EXTENT_DATA_KEY) |
1077 | path->slots[0]--; | 1089 | path->slots[0]--; |
1078 | } | 1090 | } |
@@ -1093,7 +1105,7 @@ next_slot: | |||
1093 | num_bytes = 0; | 1105 | num_bytes = 0; |
1094 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); | 1106 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); |
1095 | 1107 | ||
1096 | if (found_key.objectid > inode->i_ino || | 1108 | if (found_key.objectid > ino || |
1097 | found_key.type > BTRFS_EXTENT_DATA_KEY || | 1109 | found_key.type > BTRFS_EXTENT_DATA_KEY || |
1098 | found_key.offset > end) | 1110 | found_key.offset > end) |
1099 | break; | 1111 | break; |
@@ -1128,7 +1140,7 @@ next_slot: | |||
1128 | goto out_check; | 1140 | goto out_check; |
1129 | if (btrfs_extent_readonly(root, disk_bytenr)) | 1141 | if (btrfs_extent_readonly(root, disk_bytenr)) |
1130 | goto out_check; | 1142 | goto out_check; |
1131 | if (btrfs_cross_ref_exist(trans, root, inode->i_ino, | 1143 | if (btrfs_cross_ref_exist(trans, root, ino, |
1132 | found_key.offset - | 1144 | found_key.offset - |
1133 | extent_offset, disk_bytenr)) | 1145 | extent_offset, disk_bytenr)) |
1134 | goto out_check; | 1146 | goto out_check; |
@@ -1317,8 +1329,7 @@ static int btrfs_set_bit_hook(struct inode *inode, | |||
1317 | if (!(state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { | 1329 | if (!(state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { |
1318 | struct btrfs_root *root = BTRFS_I(inode)->root; | 1330 | struct btrfs_root *root = BTRFS_I(inode)->root; |
1319 | u64 len = state->end + 1 - state->start; | 1331 | u64 len = state->end + 1 - state->start; |
1320 | int do_list = (root->root_key.objectid != | 1332 | bool do_list = !is_free_space_inode(root, inode); |
1321 | BTRFS_ROOT_TREE_OBJECTID); | ||
1322 | 1333 | ||
1323 | if (*bits & EXTENT_FIRST_DELALLOC) | 1334 | if (*bits & EXTENT_FIRST_DELALLOC) |
1324 | *bits &= ~EXTENT_FIRST_DELALLOC; | 1335 | *bits &= ~EXTENT_FIRST_DELALLOC; |
@@ -1351,8 +1362,7 @@ static int btrfs_clear_bit_hook(struct inode *inode, | |||
1351 | if ((state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { | 1362 | if ((state->state & EXTENT_DELALLOC) && (*bits & EXTENT_DELALLOC)) { |
1352 | struct btrfs_root *root = BTRFS_I(inode)->root; | 1363 | struct btrfs_root *root = BTRFS_I(inode)->root; |
1353 | u64 len = state->end + 1 - state->start; | 1364 | u64 len = state->end + 1 - state->start; |
1354 | int do_list = (root->root_key.objectid != | 1365 | bool do_list = !is_free_space_inode(root, inode); |
1355 | BTRFS_ROOT_TREE_OBJECTID); | ||
1356 | 1366 | ||
1357 | if (*bits & EXTENT_FIRST_DELALLOC) | 1367 | if (*bits & EXTENT_FIRST_DELALLOC) |
1358 | *bits &= ~EXTENT_FIRST_DELALLOC; | 1368 | *bits &= ~EXTENT_FIRST_DELALLOC; |
@@ -1459,7 +1469,7 @@ static int btrfs_submit_bio_hook(struct inode *inode, int rw, struct bio *bio, | |||
1459 | 1469 | ||
1460 | skip_sum = BTRFS_I(inode)->flags & BTRFS_INODE_NODATASUM; | 1470 | skip_sum = BTRFS_I(inode)->flags & BTRFS_INODE_NODATASUM; |
1461 | 1471 | ||
1462 | if (root == root->fs_info->tree_root) | 1472 | if (is_free_space_inode(root, inode)) |
1463 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 2); | 1473 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 2); |
1464 | else | 1474 | else |
1465 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 0); | 1475 | ret = btrfs_bio_wq_end_io(root->fs_info, bio, 0); |
@@ -1645,7 +1655,7 @@ static int insert_reserved_file_extent(struct btrfs_trans_handle *trans, | |||
1645 | &hint, 0); | 1655 | &hint, 0); |
1646 | BUG_ON(ret); | 1656 | BUG_ON(ret); |
1647 | 1657 | ||
1648 | ins.objectid = inode->i_ino; | 1658 | ins.objectid = btrfs_ino(inode); |
1649 | ins.offset = file_pos; | 1659 | ins.offset = file_pos; |
1650 | ins.type = BTRFS_EXTENT_DATA_KEY; | 1660 | ins.type = BTRFS_EXTENT_DATA_KEY; |
1651 | ret = btrfs_insert_empty_item(trans, root, path, &ins, sizeof(*fi)); | 1661 | ret = btrfs_insert_empty_item(trans, root, path, &ins, sizeof(*fi)); |
@@ -1676,7 +1686,7 @@ static int insert_reserved_file_extent(struct btrfs_trans_handle *trans, | |||
1676 | ins.type = BTRFS_EXTENT_ITEM_KEY; | 1686 | ins.type = BTRFS_EXTENT_ITEM_KEY; |
1677 | ret = btrfs_alloc_reserved_file_extent(trans, root, | 1687 | ret = btrfs_alloc_reserved_file_extent(trans, root, |
1678 | root->root_key.objectid, | 1688 | root->root_key.objectid, |
1679 | inode->i_ino, file_pos, &ins); | 1689 | btrfs_ino(inode), file_pos, &ins); |
1680 | BUG_ON(ret); | 1690 | BUG_ON(ret); |
1681 | btrfs_free_path(path); | 1691 | btrfs_free_path(path); |
1682 | 1692 | ||
@@ -1702,7 +1712,7 @@ static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end) | |||
1702 | struct extent_state *cached_state = NULL; | 1712 | struct extent_state *cached_state = NULL; |
1703 | int compress_type = 0; | 1713 | int compress_type = 0; |
1704 | int ret; | 1714 | int ret; |
1705 | bool nolock = false; | 1715 | bool nolock; |
1706 | 1716 | ||
1707 | ret = btrfs_dec_test_ordered_pending(inode, &ordered_extent, start, | 1717 | ret = btrfs_dec_test_ordered_pending(inode, &ordered_extent, start, |
1708 | end - start + 1); | 1718 | end - start + 1); |
@@ -1710,7 +1720,7 @@ static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end) | |||
1710 | return 0; | 1720 | return 0; |
1711 | BUG_ON(!ordered_extent); | 1721 | BUG_ON(!ordered_extent); |
1712 | 1722 | ||
1713 | nolock = (root == root->fs_info->tree_root); | 1723 | nolock = is_free_space_inode(root, inode); |
1714 | 1724 | ||
1715 | if (test_bit(BTRFS_ORDERED_NOCOW, &ordered_extent->flags)) { | 1725 | if (test_bit(BTRFS_ORDERED_NOCOW, &ordered_extent->flags)) { |
1716 | BUG_ON(!list_empty(&ordered_extent->list)); | 1726 | BUG_ON(!list_empty(&ordered_extent->list)); |
@@ -2005,8 +2015,9 @@ good: | |||
2005 | return 0; | 2015 | return 0; |
2006 | 2016 | ||
2007 | zeroit: | 2017 | zeroit: |
2008 | printk_ratelimited(KERN_INFO "btrfs csum failed ino %lu off %llu csum %u " | 2018 | printk_ratelimited(KERN_INFO "btrfs csum failed ino %llu off %llu csum %u " |
2009 | "private %llu\n", page->mapping->host->i_ino, | 2019 | "private %llu\n", |
2020 | (unsigned long long)btrfs_ino(page->mapping->host), | ||
2010 | (unsigned long long)start, csum, | 2021 | (unsigned long long)start, csum, |
2011 | (unsigned long long)private); | 2022 | (unsigned long long)private); |
2012 | memset(kaddr + offset, 1, end - start + 1); | 2023 | memset(kaddr + offset, 1, end - start + 1); |
@@ -2243,7 +2254,7 @@ int btrfs_orphan_add(struct btrfs_trans_handle *trans, struct inode *inode) | |||
2243 | 2254 | ||
2244 | /* insert an orphan item to track this unlinked/truncated file */ | 2255 | /* insert an orphan item to track this unlinked/truncated file */ |
2245 | if (insert >= 1) { | 2256 | if (insert >= 1) { |
2246 | ret = btrfs_insert_orphan_item(trans, root, inode->i_ino); | 2257 | ret = btrfs_insert_orphan_item(trans, root, btrfs_ino(inode)); |
2247 | BUG_ON(ret); | 2258 | BUG_ON(ret); |
2248 | } | 2259 | } |
2249 | 2260 | ||
@@ -2280,7 +2291,7 @@ int btrfs_orphan_del(struct btrfs_trans_handle *trans, struct inode *inode) | |||
2280 | spin_unlock(&root->orphan_lock); | 2291 | spin_unlock(&root->orphan_lock); |
2281 | 2292 | ||
2282 | if (trans && delete_item) { | 2293 | if (trans && delete_item) { |
2283 | ret = btrfs_del_orphan_item(trans, root, inode->i_ino); | 2294 | ret = btrfs_del_orphan_item(trans, root, btrfs_ino(inode)); |
2284 | BUG_ON(ret); | 2295 | BUG_ON(ret); |
2285 | } | 2296 | } |
2286 | 2297 | ||
@@ -2542,7 +2553,8 @@ static void btrfs_read_locked_inode(struct inode *inode) | |||
2542 | * try to precache a NULL acl entry for files that don't have | 2553 | * try to precache a NULL acl entry for files that don't have |
2543 | * any xattrs or acls | 2554 | * any xattrs or acls |
2544 | */ | 2555 | */ |
2545 | maybe_acls = acls_after_inode_item(leaf, path->slots[0], inode->i_ino); | 2556 | maybe_acls = acls_after_inode_item(leaf, path->slots[0], |
2557 | btrfs_ino(inode)); | ||
2546 | if (!maybe_acls) | 2558 | if (!maybe_acls) |
2547 | cache_no_acl(inode); | 2559 | cache_no_acl(inode); |
2548 | 2560 | ||
@@ -2646,11 +2658,26 @@ noinline int btrfs_update_inode(struct btrfs_trans_handle *trans, | |||
2646 | struct extent_buffer *leaf; | 2658 | struct extent_buffer *leaf; |
2647 | int ret; | 2659 | int ret; |
2648 | 2660 | ||
2661 | /* | ||
2662 | * If root is tree root, it means this inode is used to | ||
2663 | * store free space information. And these inodes are updated | ||
2664 | * when committing the transaction, so they needn't delaye to | ||
2665 | * be updated, or deadlock will occured. | ||
2666 | */ | ||
2667 | if (!is_free_space_inode(root, inode)) { | ||
2668 | ret = btrfs_delayed_update_inode(trans, root, inode); | ||
2669 | if (!ret) | ||
2670 | btrfs_set_inode_last_trans(trans, inode); | ||
2671 | return ret; | ||
2672 | } | ||
2673 | |||
2649 | path = btrfs_alloc_path(); | 2674 | path = btrfs_alloc_path(); |
2650 | BUG_ON(!path); | 2675 | if (!path) |
2676 | return -ENOMEM; | ||
2677 | |||
2651 | path->leave_spinning = 1; | 2678 | path->leave_spinning = 1; |
2652 | ret = btrfs_lookup_inode(trans, root, path, | 2679 | ret = btrfs_lookup_inode(trans, root, path, &BTRFS_I(inode)->location, |
2653 | &BTRFS_I(inode)->location, 1); | 2680 | 1); |
2654 | if (ret) { | 2681 | if (ret) { |
2655 | if (ret > 0) | 2682 | if (ret > 0) |
2656 | ret = -ENOENT; | 2683 | ret = -ENOENT; |
@@ -2660,7 +2687,7 @@ noinline int btrfs_update_inode(struct btrfs_trans_handle *trans, | |||
2660 | btrfs_unlock_up_safe(path, 1); | 2687 | btrfs_unlock_up_safe(path, 1); |
2661 | leaf = path->nodes[0]; | 2688 | leaf = path->nodes[0]; |
2662 | inode_item = btrfs_item_ptr(leaf, path->slots[0], | 2689 | inode_item = btrfs_item_ptr(leaf, path->slots[0], |
2663 | struct btrfs_inode_item); | 2690 | struct btrfs_inode_item); |
2664 | 2691 | ||
2665 | fill_inode_item(trans, leaf, inode_item, inode); | 2692 | fill_inode_item(trans, leaf, inode_item, inode); |
2666 | btrfs_mark_buffer_dirty(leaf); | 2693 | btrfs_mark_buffer_dirty(leaf); |
@@ -2671,7 +2698,6 @@ failed: | |||
2671 | return ret; | 2698 | return ret; |
2672 | } | 2699 | } |
2673 | 2700 | ||
2674 | |||
2675 | /* | 2701 | /* |
2676 | * unlink helper that gets used here in inode.c and in the tree logging | 2702 | * unlink helper that gets used here in inode.c and in the tree logging |
2677 | * recovery code. It remove a link in a directory with a given name, and | 2703 | * recovery code. It remove a link in a directory with a given name, and |
@@ -2688,6 +2714,8 @@ static int __btrfs_unlink_inode(struct btrfs_trans_handle *trans, | |||
2688 | struct btrfs_dir_item *di; | 2714 | struct btrfs_dir_item *di; |
2689 | struct btrfs_key key; | 2715 | struct btrfs_key key; |
2690 | u64 index; | 2716 | u64 index; |
2717 | u64 ino = btrfs_ino(inode); | ||
2718 | u64 dir_ino = btrfs_ino(dir); | ||
2691 | 2719 | ||
2692 | path = btrfs_alloc_path(); | 2720 | path = btrfs_alloc_path(); |
2693 | if (!path) { | 2721 | if (!path) { |
@@ -2696,7 +2724,7 @@ static int __btrfs_unlink_inode(struct btrfs_trans_handle *trans, | |||
2696 | } | 2724 | } |
2697 | 2725 | ||
2698 | path->leave_spinning = 1; | 2726 | path->leave_spinning = 1; |
2699 | di = btrfs_lookup_dir_item(trans, root, path, dir->i_ino, | 2727 | di = btrfs_lookup_dir_item(trans, root, path, dir_ino, |
2700 | name, name_len, -1); | 2728 | name, name_len, -1); |
2701 | if (IS_ERR(di)) { | 2729 | if (IS_ERR(di)) { |
2702 | ret = PTR_ERR(di); | 2730 | ret = PTR_ERR(di); |
@@ -2713,31 +2741,21 @@ static int __btrfs_unlink_inode(struct btrfs_trans_handle *trans, | |||
2713 | goto err; | 2741 | goto err; |
2714 | btrfs_release_path(path); | 2742 | btrfs_release_path(path); |
2715 | 2743 | ||
2716 | ret = btrfs_del_inode_ref(trans, root, name, name_len, | 2744 | ret = btrfs_del_inode_ref(trans, root, name, name_len, ino, |
2717 | inode->i_ino, | 2745 | dir_ino, &index); |
2718 | dir->i_ino, &index); | ||
2719 | if (ret) { | 2746 | if (ret) { |
2720 | printk(KERN_INFO "btrfs failed to delete reference to %.*s, " | 2747 | printk(KERN_INFO "btrfs failed to delete reference to %.*s, " |
2721 | "inode %lu parent %lu\n", name_len, name, | 2748 | "inode %llu parent %llu\n", name_len, name, |
2722 | inode->i_ino, dir->i_ino); | 2749 | (unsigned long long)ino, (unsigned long long)dir_ino); |
2723 | goto err; | 2750 | goto err; |
2724 | } | 2751 | } |
2725 | 2752 | ||
2726 | di = btrfs_lookup_dir_index_item(trans, root, path, dir->i_ino, | 2753 | ret = btrfs_delete_delayed_dir_index(trans, root, dir, index); |
2727 | index, name, name_len, -1); | 2754 | if (ret) |
2728 | if (IS_ERR(di)) { | ||
2729 | ret = PTR_ERR(di); | ||
2730 | goto err; | ||
2731 | } | ||
2732 | if (!di) { | ||
2733 | ret = -ENOENT; | ||
2734 | goto err; | 2755 | goto err; |
2735 | } | ||
2736 | ret = btrfs_delete_one_dir_name(trans, root, path, di); | ||
2737 | btrfs_release_path(path); | ||
2738 | 2756 | ||
2739 | ret = btrfs_del_inode_ref_in_log(trans, root, name, name_len, | 2757 | ret = btrfs_del_inode_ref_in_log(trans, root, name, name_len, |
2740 | inode, dir->i_ino); | 2758 | inode, dir_ino); |
2741 | BUG_ON(ret != 0 && ret != -ENOENT); | 2759 | BUG_ON(ret != 0 && ret != -ENOENT); |
2742 | 2760 | ||
2743 | ret = btrfs_del_dir_entries_in_log(trans, root, name, name_len, | 2761 | ret = btrfs_del_dir_entries_in_log(trans, root, name, name_len, |
@@ -2815,12 +2833,14 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2815 | int check_link = 1; | 2833 | int check_link = 1; |
2816 | int err = -ENOSPC; | 2834 | int err = -ENOSPC; |
2817 | int ret; | 2835 | int ret; |
2836 | u64 ino = btrfs_ino(inode); | ||
2837 | u64 dir_ino = btrfs_ino(dir); | ||
2818 | 2838 | ||
2819 | trans = btrfs_start_transaction(root, 10); | 2839 | trans = btrfs_start_transaction(root, 10); |
2820 | if (!IS_ERR(trans) || PTR_ERR(trans) != -ENOSPC) | 2840 | if (!IS_ERR(trans) || PTR_ERR(trans) != -ENOSPC) |
2821 | return trans; | 2841 | return trans; |
2822 | 2842 | ||
2823 | if (inode->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) | 2843 | if (ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) |
2824 | return ERR_PTR(-ENOSPC); | 2844 | return ERR_PTR(-ENOSPC); |
2825 | 2845 | ||
2826 | /* check if there is someone else holds reference */ | 2846 | /* check if there is someone else holds reference */ |
@@ -2879,7 +2899,7 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2879 | 2899 | ||
2880 | if (ret == 0 && S_ISREG(inode->i_mode)) { | 2900 | if (ret == 0 && S_ISREG(inode->i_mode)) { |
2881 | ret = btrfs_lookup_file_extent(trans, root, path, | 2901 | ret = btrfs_lookup_file_extent(trans, root, path, |
2882 | inode->i_ino, (u64)-1, 0); | 2902 | ino, (u64)-1, 0); |
2883 | if (ret < 0) { | 2903 | if (ret < 0) { |
2884 | err = ret; | 2904 | err = ret; |
2885 | goto out; | 2905 | goto out; |
@@ -2895,7 +2915,7 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2895 | goto out; | 2915 | goto out; |
2896 | } | 2916 | } |
2897 | 2917 | ||
2898 | di = btrfs_lookup_dir_item(trans, root, path, dir->i_ino, | 2918 | di = btrfs_lookup_dir_item(trans, root, path, dir_ino, |
2899 | dentry->d_name.name, dentry->d_name.len, 0); | 2919 | dentry->d_name.name, dentry->d_name.len, 0); |
2900 | if (IS_ERR(di)) { | 2920 | if (IS_ERR(di)) { |
2901 | err = PTR_ERR(di); | 2921 | err = PTR_ERR(di); |
@@ -2912,7 +2932,7 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2912 | 2932 | ||
2913 | ref = btrfs_lookup_inode_ref(trans, root, path, | 2933 | ref = btrfs_lookup_inode_ref(trans, root, path, |
2914 | dentry->d_name.name, dentry->d_name.len, | 2934 | dentry->d_name.name, dentry->d_name.len, |
2915 | inode->i_ino, dir->i_ino, 0); | 2935 | ino, dir_ino, 0); |
2916 | if (IS_ERR(ref)) { | 2936 | if (IS_ERR(ref)) { |
2917 | err = PTR_ERR(ref); | 2937 | err = PTR_ERR(ref); |
2918 | goto out; | 2938 | goto out; |
@@ -2923,7 +2943,15 @@ static struct btrfs_trans_handle *__unlink_start_trans(struct inode *dir, | |||
2923 | index = btrfs_inode_ref_index(path->nodes[0], ref); | 2943 | index = btrfs_inode_ref_index(path->nodes[0], ref); |
2924 | btrfs_release_path(path); | 2944 | btrfs_release_path(path); |
2925 | 2945 | ||
2926 | di = btrfs_lookup_dir_index_item(trans, root, path, dir->i_ino, index, | 2946 | /* |
2947 | * This is a commit root search, if we can lookup inode item and other | ||
2948 | * relative items in the commit root, it means the transaction of | ||
2949 | * dir/file creation has been committed, and the dir index item that we | ||
2950 | * delay to insert has also been inserted into the commit root. So | ||
2951 | * we needn't worry about the delayed insertion of the dir index item | ||
2952 | * here. | ||
2953 | */ | ||
2954 | di = btrfs_lookup_dir_index_item(trans, root, path, dir_ino, index, | ||
2927 | dentry->d_name.name, dentry->d_name.len, 0); | 2955 | dentry->d_name.name, dentry->d_name.len, 0); |
2928 | if (IS_ERR(di)) { | 2956 | if (IS_ERR(di)) { |
2929 | err = PTR_ERR(di); | 2957 | err = PTR_ERR(di); |
@@ -2998,12 +3026,13 @@ int btrfs_unlink_subvol(struct btrfs_trans_handle *trans, | |||
2998 | struct btrfs_key key; | 3026 | struct btrfs_key key; |
2999 | u64 index; | 3027 | u64 index; |
3000 | int ret; | 3028 | int ret; |
3029 | u64 dir_ino = btrfs_ino(dir); | ||
3001 | 3030 | ||
3002 | path = btrfs_alloc_path(); | 3031 | path = btrfs_alloc_path(); |
3003 | if (!path) | 3032 | if (!path) |
3004 | return -ENOMEM; | 3033 | return -ENOMEM; |
3005 | 3034 | ||
3006 | di = btrfs_lookup_dir_item(trans, root, path, dir->i_ino, | 3035 | di = btrfs_lookup_dir_item(trans, root, path, dir_ino, |
3007 | name, name_len, -1); | 3036 | name, name_len, -1); |
3008 | BUG_ON(IS_ERR_OR_NULL(di)); | 3037 | BUG_ON(IS_ERR_OR_NULL(di)); |
3009 | 3038 | ||
@@ -3016,10 +3045,10 @@ int btrfs_unlink_subvol(struct btrfs_trans_handle *trans, | |||
3016 | 3045 | ||
3017 | ret = btrfs_del_root_ref(trans, root->fs_info->tree_root, | 3046 | ret = btrfs_del_root_ref(trans, root->fs_info->tree_root, |
3018 | objectid, root->root_key.objectid, | 3047 | objectid, root->root_key.objectid, |
3019 | dir->i_ino, &index, name, name_len); | 3048 | dir_ino, &index, name, name_len); |
3020 | if (ret < 0) { | 3049 | if (ret < 0) { |
3021 | BUG_ON(ret != -ENOENT); | 3050 | BUG_ON(ret != -ENOENT); |
3022 | di = btrfs_search_dir_index_item(root, path, dir->i_ino, | 3051 | di = btrfs_search_dir_index_item(root, path, dir_ino, |
3023 | name, name_len); | 3052 | name, name_len); |
3024 | BUG_ON(IS_ERR_OR_NULL(di)); | 3053 | BUG_ON(IS_ERR_OR_NULL(di)); |
3025 | 3054 | ||
@@ -3028,24 +3057,16 @@ int btrfs_unlink_subvol(struct btrfs_trans_handle *trans, | |||
3028 | btrfs_release_path(path); | 3057 | btrfs_release_path(path); |
3029 | index = key.offset; | 3058 | index = key.offset; |
3030 | } | 3059 | } |
3060 | btrfs_release_path(path); | ||
3031 | 3061 | ||
3032 | di = btrfs_lookup_dir_index_item(trans, root, path, dir->i_ino, | 3062 | ret = btrfs_delete_delayed_dir_index(trans, root, dir, index); |
3033 | index, name, name_len, -1); | ||
3034 | BUG_ON(IS_ERR_OR_NULL(di)); | ||
3035 | |||
3036 | leaf = path->nodes[0]; | ||
3037 | btrfs_dir_item_key_to_cpu(leaf, di, &key); | ||
3038 | WARN_ON(key.type != BTRFS_ROOT_ITEM_KEY || key.objectid != objectid); | ||
3039 | ret = btrfs_delete_one_dir_name(trans, root, path, di); | ||
3040 | BUG_ON(ret); | 3063 | BUG_ON(ret); |
3041 | btrfs_release_path(path); | ||
3042 | 3064 | ||
3043 | btrfs_i_size_write(dir, dir->i_size - name_len * 2); | 3065 | btrfs_i_size_write(dir, dir->i_size - name_len * 2); |
3044 | dir->i_mtime = dir->i_ctime = CURRENT_TIME; | 3066 | dir->i_mtime = dir->i_ctime = CURRENT_TIME; |
3045 | ret = btrfs_update_inode(trans, root, dir); | 3067 | ret = btrfs_update_inode(trans, root, dir); |
3046 | BUG_ON(ret); | 3068 | BUG_ON(ret); |
3047 | 3069 | ||
3048 | btrfs_free_path(path); | ||
3049 | return 0; | 3070 | return 0; |
3050 | } | 3071 | } |
3051 | 3072 | ||
@@ -3058,7 +3079,7 @@ static int btrfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
3058 | unsigned long nr = 0; | 3079 | unsigned long nr = 0; |
3059 | 3080 | ||
3060 | if (inode->i_size > BTRFS_EMPTY_DIR_SIZE || | 3081 | if (inode->i_size > BTRFS_EMPTY_DIR_SIZE || |
3061 | inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) | 3082 | btrfs_ino(inode) == BTRFS_FIRST_FREE_OBJECTID) |
3062 | return -ENOTEMPTY; | 3083 | return -ENOTEMPTY; |
3063 | 3084 | ||
3064 | trans = __unlink_start_trans(dir, dentry); | 3085 | trans = __unlink_start_trans(dir, dentry); |
@@ -3067,7 +3088,7 @@ static int btrfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
3067 | 3088 | ||
3068 | btrfs_set_trans_block_group(trans, dir); | 3089 | btrfs_set_trans_block_group(trans, dir); |
3069 | 3090 | ||
3070 | if (unlikely(inode->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { | 3091 | if (unlikely(btrfs_ino(inode) == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { |
3071 | err = btrfs_unlink_subvol(trans, root, dir, | 3092 | err = btrfs_unlink_subvol(trans, root, dir, |
3072 | BTRFS_I(inode)->location.objectid, | 3093 | BTRFS_I(inode)->location.objectid, |
3073 | dentry->d_name.name, | 3094 | dentry->d_name.name, |
@@ -3127,17 +3148,27 @@ int btrfs_truncate_inode_items(struct btrfs_trans_handle *trans, | |||
3127 | int encoding; | 3148 | int encoding; |
3128 | int ret; | 3149 | int ret; |
3129 | int err = 0; | 3150 | int err = 0; |
3151 | u64 ino = btrfs_ino(inode); | ||
3130 | 3152 | ||
3131 | BUG_ON(new_size > 0 && min_type != BTRFS_EXTENT_DATA_KEY); | 3153 | BUG_ON(new_size > 0 && min_type != BTRFS_EXTENT_DATA_KEY); |
3132 | 3154 | ||
3133 | if (root->ref_cows || root == root->fs_info->tree_root) | 3155 | if (root->ref_cows || root == root->fs_info->tree_root) |
3134 | btrfs_drop_extent_cache(inode, new_size & (~mask), (u64)-1, 0); | 3156 | btrfs_drop_extent_cache(inode, new_size & (~mask), (u64)-1, 0); |
3135 | 3157 | ||
3158 | /* | ||
3159 | * This function is also used to drop the items in the log tree before | ||
3160 | * we relog the inode, so if root != BTRFS_I(inode)->root, it means | ||
3161 | * it is used to drop the loged items. So we shouldn't kill the delayed | ||
3162 | * items. | ||
3163 | */ | ||
3164 | if (min_type == 0 && root == BTRFS_I(inode)->root) | ||
3165 | btrfs_kill_delayed_inode_items(inode); | ||
3166 | |||
3136 | path = btrfs_alloc_path(); | 3167 | path = btrfs_alloc_path(); |
3137 | BUG_ON(!path); | 3168 | BUG_ON(!path); |
3138 | path->reada = -1; | 3169 | path->reada = -1; |
3139 | 3170 | ||
3140 | key.objectid = inode->i_ino; | 3171 | key.objectid = ino; |
3141 | key.offset = (u64)-1; | 3172 | key.offset = (u64)-1; |
3142 | key.type = (u8)-1; | 3173 | key.type = (u8)-1; |
3143 | 3174 | ||
@@ -3165,7 +3196,7 @@ search_again: | |||
3165 | found_type = btrfs_key_type(&found_key); | 3196 | found_type = btrfs_key_type(&found_key); |
3166 | encoding = 0; | 3197 | encoding = 0; |
3167 | 3198 | ||
3168 | if (found_key.objectid != inode->i_ino) | 3199 | if (found_key.objectid != ino) |
3169 | break; | 3200 | break; |
3170 | 3201 | ||
3171 | if (found_type < min_type) | 3202 | if (found_type < min_type) |
@@ -3284,7 +3315,7 @@ delete: | |||
3284 | ret = btrfs_free_extent(trans, root, extent_start, | 3315 | ret = btrfs_free_extent(trans, root, extent_start, |
3285 | extent_num_bytes, 0, | 3316 | extent_num_bytes, 0, |
3286 | btrfs_header_owner(leaf), | 3317 | btrfs_header_owner(leaf), |
3287 | inode->i_ino, extent_offset); | 3318 | ino, extent_offset); |
3288 | BUG_ON(ret); | 3319 | BUG_ON(ret); |
3289 | } | 3320 | } |
3290 | 3321 | ||
@@ -3293,7 +3324,9 @@ delete: | |||
3293 | 3324 | ||
3294 | if (path->slots[0] == 0 || | 3325 | if (path->slots[0] == 0 || |
3295 | path->slots[0] != pending_del_slot) { | 3326 | path->slots[0] != pending_del_slot) { |
3296 | if (root->ref_cows) { | 3327 | if (root->ref_cows && |
3328 | BTRFS_I(inode)->location.objectid != | ||
3329 | BTRFS_FREE_INO_OBJECTID) { | ||
3297 | err = -EAGAIN; | 3330 | err = -EAGAIN; |
3298 | goto out; | 3331 | goto out; |
3299 | } | 3332 | } |
@@ -3483,7 +3516,7 @@ int btrfs_cont_expand(struct inode *inode, loff_t oldsize, loff_t size) | |||
3483 | break; | 3516 | break; |
3484 | 3517 | ||
3485 | err = btrfs_insert_file_extent(trans, root, | 3518 | err = btrfs_insert_file_extent(trans, root, |
3486 | inode->i_ino, cur_offset, 0, | 3519 | btrfs_ino(inode), cur_offset, 0, |
3487 | 0, hole_size, 0, hole_size, | 3520 | 0, hole_size, 0, hole_size, |
3488 | 0, 0, 0); | 3521 | 0, 0, 0); |
3489 | if (err) | 3522 | if (err) |
@@ -3585,7 +3618,7 @@ void btrfs_evict_inode(struct inode *inode) | |||
3585 | 3618 | ||
3586 | truncate_inode_pages(&inode->i_data, 0); | 3619 | truncate_inode_pages(&inode->i_data, 0); |
3587 | if (inode->i_nlink && (btrfs_root_refs(&root->root_item) != 0 || | 3620 | if (inode->i_nlink && (btrfs_root_refs(&root->root_item) != 0 || |
3588 | root == root->fs_info->tree_root)) | 3621 | is_free_space_inode(root, inode))) |
3589 | goto no_delete; | 3622 | goto no_delete; |
3590 | 3623 | ||
3591 | if (is_bad_inode(inode)) { | 3624 | if (is_bad_inode(inode)) { |
@@ -3638,6 +3671,10 @@ void btrfs_evict_inode(struct inode *inode) | |||
3638 | BUG_ON(ret); | 3671 | BUG_ON(ret); |
3639 | } | 3672 | } |
3640 | 3673 | ||
3674 | if (!(root == root->fs_info->tree_root || | ||
3675 | root->root_key.objectid == BTRFS_TREE_RELOC_OBJECTID)) | ||
3676 | btrfs_return_ino(root, btrfs_ino(inode)); | ||
3677 | |||
3641 | nr = trans->blocks_used; | 3678 | nr = trans->blocks_used; |
3642 | btrfs_end_transaction(trans, root); | 3679 | btrfs_end_transaction(trans, root); |
3643 | btrfs_btree_balance_dirty(root, nr); | 3680 | btrfs_btree_balance_dirty(root, nr); |
@@ -3663,7 +3700,7 @@ static int btrfs_inode_by_name(struct inode *dir, struct dentry *dentry, | |||
3663 | path = btrfs_alloc_path(); | 3700 | path = btrfs_alloc_path(); |
3664 | BUG_ON(!path); | 3701 | BUG_ON(!path); |
3665 | 3702 | ||
3666 | di = btrfs_lookup_dir_item(NULL, root, path, dir->i_ino, name, | 3703 | di = btrfs_lookup_dir_item(NULL, root, path, btrfs_ino(dir), name, |
3667 | namelen, 0); | 3704 | namelen, 0); |
3668 | if (IS_ERR(di)) | 3705 | if (IS_ERR(di)) |
3669 | ret = PTR_ERR(di); | 3706 | ret = PTR_ERR(di); |
@@ -3716,7 +3753,7 @@ static int fixup_tree_root_location(struct btrfs_root *root, | |||
3716 | 3753 | ||
3717 | leaf = path->nodes[0]; | 3754 | leaf = path->nodes[0]; |
3718 | ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref); | 3755 | ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref); |
3719 | if (btrfs_root_ref_dirid(leaf, ref) != dir->i_ino || | 3756 | if (btrfs_root_ref_dirid(leaf, ref) != btrfs_ino(dir) || |
3720 | btrfs_root_ref_name_len(leaf, ref) != dentry->d_name.len) | 3757 | btrfs_root_ref_name_len(leaf, ref) != dentry->d_name.len) |
3721 | goto out; | 3758 | goto out; |
3722 | 3759 | ||
@@ -3755,6 +3792,7 @@ static void inode_tree_add(struct inode *inode) | |||
3755 | struct btrfs_inode *entry; | 3792 | struct btrfs_inode *entry; |
3756 | struct rb_node **p; | 3793 | struct rb_node **p; |
3757 | struct rb_node *parent; | 3794 | struct rb_node *parent; |
3795 | u64 ino = btrfs_ino(inode); | ||
3758 | again: | 3796 | again: |
3759 | p = &root->inode_tree.rb_node; | 3797 | p = &root->inode_tree.rb_node; |
3760 | parent = NULL; | 3798 | parent = NULL; |
@@ -3767,9 +3805,9 @@ again: | |||
3767 | parent = *p; | 3805 | parent = *p; |
3768 | entry = rb_entry(parent, struct btrfs_inode, rb_node); | 3806 | entry = rb_entry(parent, struct btrfs_inode, rb_node); |
3769 | 3807 | ||
3770 | if (inode->i_ino < entry->vfs_inode.i_ino) | 3808 | if (ino < btrfs_ino(&entry->vfs_inode)) |
3771 | p = &parent->rb_left; | 3809 | p = &parent->rb_left; |
3772 | else if (inode->i_ino > entry->vfs_inode.i_ino) | 3810 | else if (ino > btrfs_ino(&entry->vfs_inode)) |
3773 | p = &parent->rb_right; | 3811 | p = &parent->rb_right; |
3774 | else { | 3812 | else { |
3775 | WARN_ON(!(entry->vfs_inode.i_state & | 3813 | WARN_ON(!(entry->vfs_inode.i_state & |
@@ -3833,9 +3871,9 @@ again: | |||
3833 | prev = node; | 3871 | prev = node; |
3834 | entry = rb_entry(node, struct btrfs_inode, rb_node); | 3872 | entry = rb_entry(node, struct btrfs_inode, rb_node); |
3835 | 3873 | ||
3836 | if (objectid < entry->vfs_inode.i_ino) | 3874 | if (objectid < btrfs_ino(&entry->vfs_inode)) |
3837 | node = node->rb_left; | 3875 | node = node->rb_left; |
3838 | else if (objectid > entry->vfs_inode.i_ino) | 3876 | else if (objectid > btrfs_ino(&entry->vfs_inode)) |
3839 | node = node->rb_right; | 3877 | node = node->rb_right; |
3840 | else | 3878 | else |
3841 | break; | 3879 | break; |
@@ -3843,7 +3881,7 @@ again: | |||
3843 | if (!node) { | 3881 | if (!node) { |
3844 | while (prev) { | 3882 | while (prev) { |
3845 | entry = rb_entry(prev, struct btrfs_inode, rb_node); | 3883 | entry = rb_entry(prev, struct btrfs_inode, rb_node); |
3846 | if (objectid <= entry->vfs_inode.i_ino) { | 3884 | if (objectid <= btrfs_ino(&entry->vfs_inode)) { |
3847 | node = prev; | 3885 | node = prev; |
3848 | break; | 3886 | break; |
3849 | } | 3887 | } |
@@ -3852,7 +3890,7 @@ again: | |||
3852 | } | 3890 | } |
3853 | while (node) { | 3891 | while (node) { |
3854 | entry = rb_entry(node, struct btrfs_inode, rb_node); | 3892 | entry = rb_entry(node, struct btrfs_inode, rb_node); |
3855 | objectid = entry->vfs_inode.i_ino + 1; | 3893 | objectid = btrfs_ino(&entry->vfs_inode) + 1; |
3856 | inode = igrab(&entry->vfs_inode); | 3894 | inode = igrab(&entry->vfs_inode); |
3857 | if (inode) { | 3895 | if (inode) { |
3858 | spin_unlock(&root->inode_lock); | 3896 | spin_unlock(&root->inode_lock); |
@@ -3890,7 +3928,7 @@ static int btrfs_init_locked_inode(struct inode *inode, void *p) | |||
3890 | static int btrfs_find_actor(struct inode *inode, void *opaque) | 3928 | static int btrfs_find_actor(struct inode *inode, void *opaque) |
3891 | { | 3929 | { |
3892 | struct btrfs_iget_args *args = opaque; | 3930 | struct btrfs_iget_args *args = opaque; |
3893 | return args->ino == inode->i_ino && | 3931 | return args->ino == btrfs_ino(inode) && |
3894 | args->root == BTRFS_I(inode)->root; | 3932 | args->root == BTRFS_I(inode)->root; |
3895 | } | 3933 | } |
3896 | 3934 | ||
@@ -4035,7 +4073,7 @@ static struct dentry *btrfs_lookup(struct inode *dir, struct dentry *dentry, | |||
4035 | return d_splice_alias(inode, dentry); | 4073 | return d_splice_alias(inode, dentry); |
4036 | } | 4074 | } |
4037 | 4075 | ||
4038 | static unsigned char btrfs_filetype_table[] = { | 4076 | unsigned char btrfs_filetype_table[] = { |
4039 | DT_UNKNOWN, DT_REG, DT_DIR, DT_CHR, DT_BLK, DT_FIFO, DT_SOCK, DT_LNK | 4077 | DT_UNKNOWN, DT_REG, DT_DIR, DT_CHR, DT_BLK, DT_FIFO, DT_SOCK, DT_LNK |
4040 | }; | 4078 | }; |
4041 | 4079 | ||
@@ -4049,6 +4087,8 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4049 | struct btrfs_key key; | 4087 | struct btrfs_key key; |
4050 | struct btrfs_key found_key; | 4088 | struct btrfs_key found_key; |
4051 | struct btrfs_path *path; | 4089 | struct btrfs_path *path; |
4090 | struct list_head ins_list; | ||
4091 | struct list_head del_list; | ||
4052 | int ret; | 4092 | int ret; |
4053 | struct extent_buffer *leaf; | 4093 | struct extent_buffer *leaf; |
4054 | int slot; | 4094 | int slot; |
@@ -4061,6 +4101,7 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4061 | char tmp_name[32]; | 4101 | char tmp_name[32]; |
4062 | char *name_ptr; | 4102 | char *name_ptr; |
4063 | int name_len; | 4103 | int name_len; |
4104 | int is_curr = 0; /* filp->f_pos points to the current index? */ | ||
4064 | 4105 | ||
4065 | /* FIXME, use a real flag for deciding about the key type */ | 4106 | /* FIXME, use a real flag for deciding about the key type */ |
4066 | if (root->fs_info->tree_root == root) | 4107 | if (root->fs_info->tree_root == root) |
@@ -4068,9 +4109,7 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4068 | 4109 | ||
4069 | /* special case for "." */ | 4110 | /* special case for "." */ |
4070 | if (filp->f_pos == 0) { | 4111 | if (filp->f_pos == 0) { |
4071 | over = filldir(dirent, ".", 1, | 4112 | over = filldir(dirent, ".", 1, 1, btrfs_ino(inode), DT_DIR); |
4072 | 1, inode->i_ino, | ||
4073 | DT_DIR); | ||
4074 | if (over) | 4113 | if (over) |
4075 | return 0; | 4114 | return 0; |
4076 | filp->f_pos = 1; | 4115 | filp->f_pos = 1; |
@@ -4085,11 +4124,19 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4085 | filp->f_pos = 2; | 4124 | filp->f_pos = 2; |
4086 | } | 4125 | } |
4087 | path = btrfs_alloc_path(); | 4126 | path = btrfs_alloc_path(); |
4127 | if (!path) | ||
4128 | return -ENOMEM; | ||
4088 | path->reada = 2; | 4129 | path->reada = 2; |
4089 | 4130 | ||
4131 | if (key_type == BTRFS_DIR_INDEX_KEY) { | ||
4132 | INIT_LIST_HEAD(&ins_list); | ||
4133 | INIT_LIST_HEAD(&del_list); | ||
4134 | btrfs_get_delayed_items(inode, &ins_list, &del_list); | ||
4135 | } | ||
4136 | |||
4090 | btrfs_set_key_type(&key, key_type); | 4137 | btrfs_set_key_type(&key, key_type); |
4091 | key.offset = filp->f_pos; | 4138 | key.offset = filp->f_pos; |
4092 | key.objectid = inode->i_ino; | 4139 | key.objectid = btrfs_ino(inode); |
4093 | 4140 | ||
4094 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); | 4141 | ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); |
4095 | if (ret < 0) | 4142 | if (ret < 0) |
@@ -4116,8 +4163,13 @@ static int btrfs_real_readdir(struct file *filp, void *dirent, | |||
4116 | break; | 4163 | break; |
4117 | if (found_key.offset < filp->f_pos) | 4164 | if (found_key.offset < filp->f_pos) |
4118 | goto next; | 4165 | goto next; |
4166 | if (key_type == BTRFS_DIR_INDEX_KEY && | ||
4167 | btrfs_should_delete_dir_index(&del_list, | ||
4168 | found_key.offset)) | ||
4169 | goto next; | ||
4119 | 4170 | ||
4120 | filp->f_pos = found_key.offset; | 4171 | filp->f_pos = found_key.offset; |
4172 | is_curr = 1; | ||
4121 | 4173 | ||
4122 | di = btrfs_item_ptr(leaf, slot, struct btrfs_dir_item); | 4174 | di = btrfs_item_ptr(leaf, slot, struct btrfs_dir_item); |
4123 | di_cur = 0; | 4175 | di_cur = 0; |
@@ -4172,6 +4224,15 @@ next: | |||
4172 | path->slots[0]++; | 4224 | path->slots[0]++; |
4173 | } | 4225 | } |
4174 | 4226 | ||
4227 | if (key_type == BTRFS_DIR_INDEX_KEY) { | ||
4228 | if (is_curr) | ||
4229 | filp->f_pos++; | ||
4230 | ret = btrfs_readdir_delayed_dir_index(filp, dirent, filldir, | ||
4231 | &ins_list); | ||
4232 | if (ret) | ||
4233 | goto nopos; | ||
4234 | } | ||
4235 | |||
4175 | /* Reached end of directory/root. Bump pos past the last item. */ | 4236 | /* Reached end of directory/root. Bump pos past the last item. */ |
4176 | if (key_type == BTRFS_DIR_INDEX_KEY) | 4237 | if (key_type == BTRFS_DIR_INDEX_KEY) |
4177 | /* | 4238 | /* |
@@ -4184,6 +4245,8 @@ next: | |||
4184 | nopos: | 4245 | nopos: |
4185 | ret = 0; | 4246 | ret = 0; |
4186 | err: | 4247 | err: |
4248 | if (key_type == BTRFS_DIR_INDEX_KEY) | ||
4249 | btrfs_put_delayed_items(&ins_list, &del_list); | ||
4187 | btrfs_free_path(path); | 4250 | btrfs_free_path(path); |
4188 | return ret; | 4251 | return ret; |
4189 | } | 4252 | } |
@@ -4199,7 +4262,8 @@ int btrfs_write_inode(struct inode *inode, struct writeback_control *wbc) | |||
4199 | return 0; | 4262 | return 0; |
4200 | 4263 | ||
4201 | smp_mb(); | 4264 | smp_mb(); |
4202 | nolock = (root->fs_info->closing && root == root->fs_info->tree_root); | 4265 | if (root->fs_info->closing && is_free_space_inode(root, inode)) |
4266 | nolock = true; | ||
4203 | 4267 | ||
4204 | if (wbc->sync_mode == WB_SYNC_ALL) { | 4268 | if (wbc->sync_mode == WB_SYNC_ALL) { |
4205 | if (nolock) | 4269 | if (nolock) |
@@ -4243,8 +4307,9 @@ void btrfs_dirty_inode(struct inode *inode) | |||
4243 | trans = btrfs_start_transaction(root, 1); | 4307 | trans = btrfs_start_transaction(root, 1); |
4244 | if (IS_ERR(trans)) { | 4308 | if (IS_ERR(trans)) { |
4245 | printk_ratelimited(KERN_ERR "btrfs: fail to " | 4309 | printk_ratelimited(KERN_ERR "btrfs: fail to " |
4246 | "dirty inode %lu error %ld\n", | 4310 | "dirty inode %llu error %ld\n", |
4247 | inode->i_ino, PTR_ERR(trans)); | 4311 | (unsigned long long)btrfs_ino(inode), |
4312 | PTR_ERR(trans)); | ||
4248 | return; | 4313 | return; |
4249 | } | 4314 | } |
4250 | btrfs_set_trans_block_group(trans, inode); | 4315 | btrfs_set_trans_block_group(trans, inode); |
@@ -4252,11 +4317,14 @@ void btrfs_dirty_inode(struct inode *inode) | |||
4252 | ret = btrfs_update_inode(trans, root, inode); | 4317 | ret = btrfs_update_inode(trans, root, inode); |
4253 | if (ret) { | 4318 | if (ret) { |
4254 | printk_ratelimited(KERN_ERR "btrfs: fail to " | 4319 | printk_ratelimited(KERN_ERR "btrfs: fail to " |
4255 | "dirty inode %lu error %d\n", | 4320 | "dirty inode %llu error %d\n", |
4256 | inode->i_ino, ret); | 4321 | (unsigned long long)btrfs_ino(inode), |
4322 | ret); | ||
4257 | } | 4323 | } |
4258 | } | 4324 | } |
4259 | btrfs_end_transaction(trans, root); | 4325 | btrfs_end_transaction(trans, root); |
4326 | if (BTRFS_I(inode)->delayed_node) | ||
4327 | btrfs_balance_delayed_items(root); | ||
4260 | } | 4328 | } |
4261 | 4329 | ||
4262 | /* | 4330 | /* |
@@ -4272,7 +4340,7 @@ static int btrfs_set_inode_index_count(struct inode *inode) | |||
4272 | struct extent_buffer *leaf; | 4340 | struct extent_buffer *leaf; |
4273 | int ret; | 4341 | int ret; |
4274 | 4342 | ||
4275 | key.objectid = inode->i_ino; | 4343 | key.objectid = btrfs_ino(inode); |
4276 | btrfs_set_key_type(&key, BTRFS_DIR_INDEX_KEY); | 4344 | btrfs_set_key_type(&key, BTRFS_DIR_INDEX_KEY); |
4277 | key.offset = (u64)-1; | 4345 | key.offset = (u64)-1; |
4278 | 4346 | ||
@@ -4304,7 +4372,7 @@ static int btrfs_set_inode_index_count(struct inode *inode) | |||
4304 | leaf = path->nodes[0]; | 4372 | leaf = path->nodes[0]; |
4305 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); | 4373 | btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); |
4306 | 4374 | ||
4307 | if (found_key.objectid != inode->i_ino || | 4375 | if (found_key.objectid != btrfs_ino(inode) || |
4308 | btrfs_key_type(&found_key) != BTRFS_DIR_INDEX_KEY) { | 4376 | btrfs_key_type(&found_key) != BTRFS_DIR_INDEX_KEY) { |
4309 | BTRFS_I(inode)->index_cnt = 2; | 4377 | BTRFS_I(inode)->index_cnt = 2; |
4310 | goto out; | 4378 | goto out; |
@@ -4325,9 +4393,12 @@ int btrfs_set_inode_index(struct inode *dir, u64 *index) | |||
4325 | int ret = 0; | 4393 | int ret = 0; |
4326 | 4394 | ||
4327 | if (BTRFS_I(dir)->index_cnt == (u64)-1) { | 4395 | if (BTRFS_I(dir)->index_cnt == (u64)-1) { |
4328 | ret = btrfs_set_inode_index_count(dir); | 4396 | ret = btrfs_inode_delayed_dir_index_count(dir); |
4329 | if (ret) | 4397 | if (ret) { |
4330 | return ret; | 4398 | ret = btrfs_set_inode_index_count(dir); |
4399 | if (ret) | ||
4400 | return ret; | ||
4401 | } | ||
4331 | } | 4402 | } |
4332 | 4403 | ||
4333 | *index = BTRFS_I(dir)->index_cnt; | 4404 | *index = BTRFS_I(dir)->index_cnt; |
@@ -4363,6 +4434,12 @@ static struct inode *btrfs_new_inode(struct btrfs_trans_handle *trans, | |||
4363 | return ERR_PTR(-ENOMEM); | 4434 | return ERR_PTR(-ENOMEM); |
4364 | } | 4435 | } |
4365 | 4436 | ||
4437 | /* | ||
4438 | * we have to initialize this early, so we can reclaim the inode | ||
4439 | * number if we fail afterwards in this function. | ||
4440 | */ | ||
4441 | inode->i_ino = objectid; | ||
4442 | |||
4366 | if (dir) { | 4443 | if (dir) { |
4367 | trace_btrfs_inode_request(dir); | 4444 | trace_btrfs_inode_request(dir); |
4368 | 4445 | ||
@@ -4408,7 +4485,6 @@ static struct inode *btrfs_new_inode(struct btrfs_trans_handle *trans, | |||
4408 | goto fail; | 4485 | goto fail; |
4409 | 4486 | ||
4410 | inode_init_owner(inode, dir, mode); | 4487 | inode_init_owner(inode, dir, mode); |
4411 | inode->i_ino = objectid; | ||
4412 | inode_set_bytes(inode, 0); | 4488 | inode_set_bytes(inode, 0); |
4413 | inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME; | 4489 | inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME; |
4414 | inode_item = btrfs_item_ptr(path->nodes[0], path->slots[0], | 4490 | inode_item = btrfs_item_ptr(path->nodes[0], path->slots[0], |
@@ -4472,29 +4548,29 @@ int btrfs_add_link(struct btrfs_trans_handle *trans, | |||
4472 | int ret = 0; | 4548 | int ret = 0; |
4473 | struct btrfs_key key; | 4549 | struct btrfs_key key; |
4474 | struct btrfs_root *root = BTRFS_I(parent_inode)->root; | 4550 | struct btrfs_root *root = BTRFS_I(parent_inode)->root; |
4551 | u64 ino = btrfs_ino(inode); | ||
4552 | u64 parent_ino = btrfs_ino(parent_inode); | ||
4475 | 4553 | ||
4476 | if (unlikely(inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 4554 | if (unlikely(ino == BTRFS_FIRST_FREE_OBJECTID)) { |
4477 | memcpy(&key, &BTRFS_I(inode)->root->root_key, sizeof(key)); | 4555 | memcpy(&key, &BTRFS_I(inode)->root->root_key, sizeof(key)); |
4478 | } else { | 4556 | } else { |
4479 | key.objectid = inode->i_ino; | 4557 | key.objectid = ino; |
4480 | btrfs_set_key_type(&key, BTRFS_INODE_ITEM_KEY); | 4558 | btrfs_set_key_type(&key, BTRFS_INODE_ITEM_KEY); |
4481 | key.offset = 0; | 4559 | key.offset = 0; |
4482 | } | 4560 | } |
4483 | 4561 | ||
4484 | if (unlikely(inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 4562 | if (unlikely(ino == BTRFS_FIRST_FREE_OBJECTID)) { |
4485 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, | 4563 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, |
4486 | key.objectid, root->root_key.objectid, | 4564 | key.objectid, root->root_key.objectid, |
4487 | parent_inode->i_ino, | 4565 | parent_ino, index, name, name_len); |
4488 | index, name, name_len); | ||
4489 | } else if (add_backref) { | 4566 | } else if (add_backref) { |
4490 | ret = btrfs_insert_inode_ref(trans, root, | 4567 | ret = btrfs_insert_inode_ref(trans, root, name, name_len, ino, |
4491 | name, name_len, inode->i_ino, | 4568 | parent_ino, index); |
4492 | parent_inode->i_ino, index); | ||
4493 | } | 4569 | } |
4494 | 4570 | ||
4495 | if (ret == 0) { | 4571 | if (ret == 0) { |
4496 | ret = btrfs_insert_dir_item(trans, root, name, name_len, | 4572 | ret = btrfs_insert_dir_item(trans, root, name, name_len, |
4497 | parent_inode->i_ino, &key, | 4573 | parent_inode, &key, |
4498 | btrfs_inode_type(inode), index); | 4574 | btrfs_inode_type(inode), index); |
4499 | BUG_ON(ret); | 4575 | BUG_ON(ret); |
4500 | 4576 | ||
@@ -4537,10 +4613,6 @@ static int btrfs_mknod(struct inode *dir, struct dentry *dentry, | |||
4537 | if (!new_valid_dev(rdev)) | 4613 | if (!new_valid_dev(rdev)) |
4538 | return -EINVAL; | 4614 | return -EINVAL; |
4539 | 4615 | ||
4540 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
4541 | if (err) | ||
4542 | return err; | ||
4543 | |||
4544 | /* | 4616 | /* |
4545 | * 2 for inode item and ref | 4617 | * 2 for inode item and ref |
4546 | * 2 for dir items | 4618 | * 2 for dir items |
@@ -4552,8 +4624,12 @@ static int btrfs_mknod(struct inode *dir, struct dentry *dentry, | |||
4552 | 4624 | ||
4553 | btrfs_set_trans_block_group(trans, dir); | 4625 | btrfs_set_trans_block_group(trans, dir); |
4554 | 4626 | ||
4627 | err = btrfs_find_free_ino(root, &objectid); | ||
4628 | if (err) | ||
4629 | goto out_unlock; | ||
4630 | |||
4555 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 4631 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
4556 | dentry->d_name.len, dir->i_ino, objectid, | 4632 | dentry->d_name.len, btrfs_ino(dir), objectid, |
4557 | BTRFS_I(dir)->block_group, mode, &index); | 4633 | BTRFS_I(dir)->block_group, mode, &index); |
4558 | if (IS_ERR(inode)) { | 4634 | if (IS_ERR(inode)) { |
4559 | err = PTR_ERR(inode); | 4635 | err = PTR_ERR(inode); |
@@ -4600,9 +4676,6 @@ static int btrfs_create(struct inode *dir, struct dentry *dentry, | |||
4600 | u64 objectid; | 4676 | u64 objectid; |
4601 | u64 index = 0; | 4677 | u64 index = 0; |
4602 | 4678 | ||
4603 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
4604 | if (err) | ||
4605 | return err; | ||
4606 | /* | 4679 | /* |
4607 | * 2 for inode item and ref | 4680 | * 2 for inode item and ref |
4608 | * 2 for dir items | 4681 | * 2 for dir items |
@@ -4614,8 +4687,12 @@ static int btrfs_create(struct inode *dir, struct dentry *dentry, | |||
4614 | 4687 | ||
4615 | btrfs_set_trans_block_group(trans, dir); | 4688 | btrfs_set_trans_block_group(trans, dir); |
4616 | 4689 | ||
4690 | err = btrfs_find_free_ino(root, &objectid); | ||
4691 | if (err) | ||
4692 | goto out_unlock; | ||
4693 | |||
4617 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 4694 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
4618 | dentry->d_name.len, dir->i_ino, objectid, | 4695 | dentry->d_name.len, btrfs_ino(dir), objectid, |
4619 | BTRFS_I(dir)->block_group, mode, &index); | 4696 | BTRFS_I(dir)->block_group, mode, &index); |
4620 | if (IS_ERR(inode)) { | 4697 | if (IS_ERR(inode)) { |
4621 | err = PTR_ERR(inode); | 4698 | err = PTR_ERR(inode); |
@@ -4726,10 +4803,6 @@ static int btrfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
4726 | u64 index = 0; | 4803 | u64 index = 0; |
4727 | unsigned long nr = 1; | 4804 | unsigned long nr = 1; |
4728 | 4805 | ||
4729 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
4730 | if (err) | ||
4731 | return err; | ||
4732 | |||
4733 | /* | 4806 | /* |
4734 | * 2 items for inode and ref | 4807 | * 2 items for inode and ref |
4735 | * 2 items for dir items | 4808 | * 2 items for dir items |
@@ -4740,8 +4813,12 @@ static int btrfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
4740 | return PTR_ERR(trans); | 4813 | return PTR_ERR(trans); |
4741 | btrfs_set_trans_block_group(trans, dir); | 4814 | btrfs_set_trans_block_group(trans, dir); |
4742 | 4815 | ||
4816 | err = btrfs_find_free_ino(root, &objectid); | ||
4817 | if (err) | ||
4818 | goto out_fail; | ||
4819 | |||
4743 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 4820 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
4744 | dentry->d_name.len, dir->i_ino, objectid, | 4821 | dentry->d_name.len, btrfs_ino(dir), objectid, |
4745 | BTRFS_I(dir)->block_group, S_IFDIR | mode, | 4822 | BTRFS_I(dir)->block_group, S_IFDIR | mode, |
4746 | &index); | 4823 | &index); |
4747 | if (IS_ERR(inode)) { | 4824 | if (IS_ERR(inode)) { |
@@ -4864,7 +4941,7 @@ struct extent_map *btrfs_get_extent(struct inode *inode, struct page *page, | |||
4864 | u64 bytenr; | 4941 | u64 bytenr; |
4865 | u64 extent_start = 0; | 4942 | u64 extent_start = 0; |
4866 | u64 extent_end = 0; | 4943 | u64 extent_end = 0; |
4867 | u64 objectid = inode->i_ino; | 4944 | u64 objectid = btrfs_ino(inode); |
4868 | u32 found_type; | 4945 | u32 found_type; |
4869 | struct btrfs_path *path = NULL; | 4946 | struct btrfs_path *path = NULL; |
4870 | struct btrfs_root *root = BTRFS_I(inode)->root; | 4947 | struct btrfs_root *root = BTRFS_I(inode)->root; |
@@ -5372,7 +5449,7 @@ static noinline int can_nocow_odirect(struct btrfs_trans_handle *trans, | |||
5372 | if (!path) | 5449 | if (!path) |
5373 | return -ENOMEM; | 5450 | return -ENOMEM; |
5374 | 5451 | ||
5375 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 5452 | ret = btrfs_lookup_file_extent(trans, root, path, btrfs_ino(inode), |
5376 | offset, 0); | 5453 | offset, 0); |
5377 | if (ret < 0) | 5454 | if (ret < 0) |
5378 | goto out; | 5455 | goto out; |
@@ -5389,7 +5466,7 @@ static noinline int can_nocow_odirect(struct btrfs_trans_handle *trans, | |||
5389 | ret = 0; | 5466 | ret = 0; |
5390 | leaf = path->nodes[0]; | 5467 | leaf = path->nodes[0]; |
5391 | btrfs_item_key_to_cpu(leaf, &key, slot); | 5468 | btrfs_item_key_to_cpu(leaf, &key, slot); |
5392 | if (key.objectid != inode->i_ino || | 5469 | if (key.objectid != btrfs_ino(inode) || |
5393 | key.type != BTRFS_EXTENT_DATA_KEY) { | 5470 | key.type != BTRFS_EXTENT_DATA_KEY) { |
5394 | /* not our file or wrong item type, must cow */ | 5471 | /* not our file or wrong item type, must cow */ |
5395 | goto out; | 5472 | goto out; |
@@ -5423,7 +5500,7 @@ static noinline int can_nocow_odirect(struct btrfs_trans_handle *trans, | |||
5423 | * look for other files referencing this extent, if we | 5500 | * look for other files referencing this extent, if we |
5424 | * find any we must cow | 5501 | * find any we must cow |
5425 | */ | 5502 | */ |
5426 | if (btrfs_cross_ref_exist(trans, root, inode->i_ino, | 5503 | if (btrfs_cross_ref_exist(trans, root, btrfs_ino(inode), |
5427 | key.offset - backref_offset, disk_bytenr)) | 5504 | key.offset - backref_offset, disk_bytenr)) |
5428 | goto out; | 5505 | goto out; |
5429 | 5506 | ||
@@ -5613,9 +5690,10 @@ static void btrfs_endio_direct_read(struct bio *bio, int err) | |||
5613 | 5690 | ||
5614 | flush_dcache_page(bvec->bv_page); | 5691 | flush_dcache_page(bvec->bv_page); |
5615 | if (csum != *private) { | 5692 | if (csum != *private) { |
5616 | printk(KERN_ERR "btrfs csum failed ino %lu off" | 5693 | printk(KERN_ERR "btrfs csum failed ino %llu off" |
5617 | " %llu csum %u private %u\n", | 5694 | " %llu csum %u private %u\n", |
5618 | inode->i_ino, (unsigned long long)start, | 5695 | (unsigned long long)btrfs_ino(inode), |
5696 | (unsigned long long)start, | ||
5619 | csum, *private); | 5697 | csum, *private); |
5620 | err = -EIO; | 5698 | err = -EIO; |
5621 | } | 5699 | } |
@@ -5762,9 +5840,9 @@ static void btrfs_end_dio_bio(struct bio *bio, int err) | |||
5762 | struct btrfs_dio_private *dip = bio->bi_private; | 5840 | struct btrfs_dio_private *dip = bio->bi_private; |
5763 | 5841 | ||
5764 | if (err) { | 5842 | if (err) { |
5765 | printk(KERN_ERR "btrfs direct IO failed ino %lu rw %lu " | 5843 | printk(KERN_ERR "btrfs direct IO failed ino %llu rw %lu " |
5766 | "sector %#Lx len %u err no %d\n", | 5844 | "sector %#Lx len %u err no %d\n", |
5767 | dip->inode->i_ino, bio->bi_rw, | 5845 | (unsigned long long)btrfs_ino(dip->inode), bio->bi_rw, |
5768 | (unsigned long long)bio->bi_sector, bio->bi_size, err); | 5846 | (unsigned long long)bio->bi_sector, bio->bi_size, err); |
5769 | dip->errors = 1; | 5847 | dip->errors = 1; |
5770 | 5848 | ||
@@ -6607,6 +6685,8 @@ struct inode *btrfs_alloc_inode(struct super_block *sb) | |||
6607 | ei->dummy_inode = 0; | 6685 | ei->dummy_inode = 0; |
6608 | ei->force_compress = BTRFS_COMPRESS_NONE; | 6686 | ei->force_compress = BTRFS_COMPRESS_NONE; |
6609 | 6687 | ||
6688 | ei->delayed_node = NULL; | ||
6689 | |||
6610 | inode = &ei->vfs_inode; | 6690 | inode = &ei->vfs_inode; |
6611 | extent_map_tree_init(&ei->extent_tree); | 6691 | extent_map_tree_init(&ei->extent_tree); |
6612 | extent_io_tree_init(&ei->io_tree, &inode->i_data); | 6692 | extent_io_tree_init(&ei->io_tree, &inode->i_data); |
@@ -6674,8 +6754,8 @@ void btrfs_destroy_inode(struct inode *inode) | |||
6674 | 6754 | ||
6675 | spin_lock(&root->orphan_lock); | 6755 | spin_lock(&root->orphan_lock); |
6676 | if (!list_empty(&BTRFS_I(inode)->i_orphan)) { | 6756 | if (!list_empty(&BTRFS_I(inode)->i_orphan)) { |
6677 | printk(KERN_INFO "BTRFS: inode %lu still on the orphan list\n", | 6757 | printk(KERN_INFO "BTRFS: inode %llu still on the orphan list\n", |
6678 | inode->i_ino); | 6758 | (unsigned long long)btrfs_ino(inode)); |
6679 | list_del_init(&BTRFS_I(inode)->i_orphan); | 6759 | list_del_init(&BTRFS_I(inode)->i_orphan); |
6680 | } | 6760 | } |
6681 | spin_unlock(&root->orphan_lock); | 6761 | spin_unlock(&root->orphan_lock); |
@@ -6697,6 +6777,7 @@ void btrfs_destroy_inode(struct inode *inode) | |||
6697 | inode_tree_del(inode); | 6777 | inode_tree_del(inode); |
6698 | btrfs_drop_extent_cache(inode, 0, (u64)-1, 0); | 6778 | btrfs_drop_extent_cache(inode, 0, (u64)-1, 0); |
6699 | free: | 6779 | free: |
6780 | btrfs_remove_delayed_node(inode); | ||
6700 | call_rcu(&inode->i_rcu, btrfs_i_callback); | 6781 | call_rcu(&inode->i_rcu, btrfs_i_callback); |
6701 | } | 6782 | } |
6702 | 6783 | ||
@@ -6705,7 +6786,7 @@ int btrfs_drop_inode(struct inode *inode) | |||
6705 | struct btrfs_root *root = BTRFS_I(inode)->root; | 6786 | struct btrfs_root *root = BTRFS_I(inode)->root; |
6706 | 6787 | ||
6707 | if (btrfs_root_refs(&root->root_item) == 0 && | 6788 | if (btrfs_root_refs(&root->root_item) == 0 && |
6708 | root != root->fs_info->tree_root) | 6789 | !is_free_space_inode(root, inode)) |
6709 | return 1; | 6790 | return 1; |
6710 | else | 6791 | else |
6711 | return generic_drop_inode(inode); | 6792 | return generic_drop_inode(inode); |
@@ -6808,38 +6889,39 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6808 | struct btrfs_trans_handle *trans; | 6889 | struct btrfs_trans_handle *trans; |
6809 | struct btrfs_root *root = BTRFS_I(old_dir)->root; | 6890 | struct btrfs_root *root = BTRFS_I(old_dir)->root; |
6810 | struct btrfs_root *dest = BTRFS_I(new_dir)->root; | 6891 | struct btrfs_root *dest = BTRFS_I(new_dir)->root; |
6811 | struct inode *newinode = new_dentry->d_inode; | 6892 | struct inode *new_inode = new_dentry->d_inode; |
6812 | struct inode *old_inode = old_dentry->d_inode; | 6893 | struct inode *old_inode = old_dentry->d_inode; |
6813 | struct timespec ctime = CURRENT_TIME; | 6894 | struct timespec ctime = CURRENT_TIME; |
6814 | u64 index = 0; | 6895 | u64 index = 0; |
6815 | u64 root_objectid; | 6896 | u64 root_objectid; |
6816 | int ret; | 6897 | int ret; |
6898 | u64 old_ino = btrfs_ino(old_inode); | ||
6817 | 6899 | ||
6818 | if (new_dir->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) | 6900 | if (btrfs_ino(new_dir) == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID) |
6819 | return -EPERM; | 6901 | return -EPERM; |
6820 | 6902 | ||
6821 | /* we only allow rename subvolume link between subvolumes */ | 6903 | /* we only allow rename subvolume link between subvolumes */ |
6822 | if (old_inode->i_ino != BTRFS_FIRST_FREE_OBJECTID && root != dest) | 6904 | if (old_ino != BTRFS_FIRST_FREE_OBJECTID && root != dest) |
6823 | return -EXDEV; | 6905 | return -EXDEV; |
6824 | 6906 | ||
6825 | if (old_inode->i_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID || | 6907 | if (old_ino == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID || |
6826 | (newinode && newinode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) | 6908 | (new_inode && btrfs_ino(new_inode) == BTRFS_FIRST_FREE_OBJECTID)) |
6827 | return -ENOTEMPTY; | 6909 | return -ENOTEMPTY; |
6828 | 6910 | ||
6829 | if (S_ISDIR(old_inode->i_mode) && newinode && | 6911 | if (S_ISDIR(old_inode->i_mode) && new_inode && |
6830 | newinode->i_size > BTRFS_EMPTY_DIR_SIZE) | 6912 | new_inode->i_size > BTRFS_EMPTY_DIR_SIZE) |
6831 | return -ENOTEMPTY; | 6913 | return -ENOTEMPTY; |
6832 | /* | 6914 | /* |
6833 | * we're using rename to replace one file with another. | 6915 | * we're using rename to replace one file with another. |
6834 | * and the replacement file is large. Start IO on it now so | 6916 | * and the replacement file is large. Start IO on it now so |
6835 | * we don't add too much work to the end of the transaction | 6917 | * we don't add too much work to the end of the transaction |
6836 | */ | 6918 | */ |
6837 | if (newinode && S_ISREG(old_inode->i_mode) && newinode->i_size && | 6919 | if (new_inode && S_ISREG(old_inode->i_mode) && new_inode->i_size && |
6838 | old_inode->i_size > BTRFS_ORDERED_OPERATIONS_FLUSH_LIMIT) | 6920 | old_inode->i_size > BTRFS_ORDERED_OPERATIONS_FLUSH_LIMIT) |
6839 | filemap_flush(old_inode->i_mapping); | 6921 | filemap_flush(old_inode->i_mapping); |
6840 | 6922 | ||
6841 | /* close the racy window with snapshot create/destroy ioctl */ | 6923 | /* close the racy window with snapshot create/destroy ioctl */ |
6842 | if (old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) | 6924 | if (old_ino == BTRFS_FIRST_FREE_OBJECTID) |
6843 | down_read(&root->fs_info->subvol_sem); | 6925 | down_read(&root->fs_info->subvol_sem); |
6844 | /* | 6926 | /* |
6845 | * We want to reserve the absolute worst case amount of items. So if | 6927 | * We want to reserve the absolute worst case amount of items. So if |
@@ -6864,15 +6946,15 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6864 | if (ret) | 6946 | if (ret) |
6865 | goto out_fail; | 6947 | goto out_fail; |
6866 | 6948 | ||
6867 | if (unlikely(old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 6949 | if (unlikely(old_ino == BTRFS_FIRST_FREE_OBJECTID)) { |
6868 | /* force full log commit if subvolume involved. */ | 6950 | /* force full log commit if subvolume involved. */ |
6869 | root->fs_info->last_trans_log_full_commit = trans->transid; | 6951 | root->fs_info->last_trans_log_full_commit = trans->transid; |
6870 | } else { | 6952 | } else { |
6871 | ret = btrfs_insert_inode_ref(trans, dest, | 6953 | ret = btrfs_insert_inode_ref(trans, dest, |
6872 | new_dentry->d_name.name, | 6954 | new_dentry->d_name.name, |
6873 | new_dentry->d_name.len, | 6955 | new_dentry->d_name.len, |
6874 | old_inode->i_ino, | 6956 | old_ino, |
6875 | new_dir->i_ino, index); | 6957 | btrfs_ino(new_dir), index); |
6876 | if (ret) | 6958 | if (ret) |
6877 | goto out_fail; | 6959 | goto out_fail; |
6878 | /* | 6960 | /* |
@@ -6888,10 +6970,8 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6888 | * make sure the inode gets flushed if it is replacing | 6970 | * make sure the inode gets flushed if it is replacing |
6889 | * something. | 6971 | * something. |
6890 | */ | 6972 | */ |
6891 | if (newinode && newinode->i_size && | 6973 | if (new_inode && new_inode->i_size && S_ISREG(old_inode->i_mode)) |
6892 | old_inode && S_ISREG(old_inode->i_mode)) { | ||
6893 | btrfs_add_ordered_operation(trans, root, old_inode); | 6974 | btrfs_add_ordered_operation(trans, root, old_inode); |
6894 | } | ||
6895 | 6975 | ||
6896 | old_dir->i_ctime = old_dir->i_mtime = ctime; | 6976 | old_dir->i_ctime = old_dir->i_mtime = ctime; |
6897 | new_dir->i_ctime = new_dir->i_mtime = ctime; | 6977 | new_dir->i_ctime = new_dir->i_mtime = ctime; |
@@ -6900,7 +6980,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6900 | if (old_dentry->d_parent != new_dentry->d_parent) | 6980 | if (old_dentry->d_parent != new_dentry->d_parent) |
6901 | btrfs_record_unlink_dir(trans, old_dir, old_inode, 1); | 6981 | btrfs_record_unlink_dir(trans, old_dir, old_inode, 1); |
6902 | 6982 | ||
6903 | if (unlikely(old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID)) { | 6983 | if (unlikely(old_ino == BTRFS_FIRST_FREE_OBJECTID)) { |
6904 | root_objectid = BTRFS_I(old_inode)->root->root_key.objectid; | 6984 | root_objectid = BTRFS_I(old_inode)->root->root_key.objectid; |
6905 | ret = btrfs_unlink_subvol(trans, root, old_dir, root_objectid, | 6985 | ret = btrfs_unlink_subvol(trans, root, old_dir, root_objectid, |
6906 | old_dentry->d_name.name, | 6986 | old_dentry->d_name.name, |
@@ -6915,16 +6995,16 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6915 | } | 6995 | } |
6916 | BUG_ON(ret); | 6996 | BUG_ON(ret); |
6917 | 6997 | ||
6918 | if (newinode) { | 6998 | if (new_inode) { |
6919 | newinode->i_ctime = CURRENT_TIME; | 6999 | new_inode->i_ctime = CURRENT_TIME; |
6920 | if (unlikely(newinode->i_ino == | 7000 | if (unlikely(btrfs_ino(new_inode) == |
6921 | BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { | 7001 | BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)) { |
6922 | root_objectid = BTRFS_I(newinode)->location.objectid; | 7002 | root_objectid = BTRFS_I(new_inode)->location.objectid; |
6923 | ret = btrfs_unlink_subvol(trans, dest, new_dir, | 7003 | ret = btrfs_unlink_subvol(trans, dest, new_dir, |
6924 | root_objectid, | 7004 | root_objectid, |
6925 | new_dentry->d_name.name, | 7005 | new_dentry->d_name.name, |
6926 | new_dentry->d_name.len); | 7006 | new_dentry->d_name.len); |
6927 | BUG_ON(newinode->i_nlink == 0); | 7007 | BUG_ON(new_inode->i_nlink == 0); |
6928 | } else { | 7008 | } else { |
6929 | ret = btrfs_unlink_inode(trans, dest, new_dir, | 7009 | ret = btrfs_unlink_inode(trans, dest, new_dir, |
6930 | new_dentry->d_inode, | 7010 | new_dentry->d_inode, |
@@ -6932,7 +7012,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6932 | new_dentry->d_name.len); | 7012 | new_dentry->d_name.len); |
6933 | } | 7013 | } |
6934 | BUG_ON(ret); | 7014 | BUG_ON(ret); |
6935 | if (newinode->i_nlink == 0) { | 7015 | if (new_inode->i_nlink == 0) { |
6936 | ret = btrfs_orphan_add(trans, new_dentry->d_inode); | 7016 | ret = btrfs_orphan_add(trans, new_dentry->d_inode); |
6937 | BUG_ON(ret); | 7017 | BUG_ON(ret); |
6938 | } | 7018 | } |
@@ -6945,7 +7025,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6945 | new_dentry->d_name.len, 0, index); | 7025 | new_dentry->d_name.len, 0, index); |
6946 | BUG_ON(ret); | 7026 | BUG_ON(ret); |
6947 | 7027 | ||
6948 | if (old_inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) { | 7028 | if (old_ino != BTRFS_FIRST_FREE_OBJECTID) { |
6949 | struct dentry *parent = dget_parent(new_dentry); | 7029 | struct dentry *parent = dget_parent(new_dentry); |
6950 | btrfs_log_new_name(trans, old_inode, old_dir, parent); | 7030 | btrfs_log_new_name(trans, old_inode, old_dir, parent); |
6951 | dput(parent); | 7031 | dput(parent); |
@@ -6954,7 +7034,7 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
6954 | out_fail: | 7034 | out_fail: |
6955 | btrfs_end_transaction_throttle(trans, root); | 7035 | btrfs_end_transaction_throttle(trans, root); |
6956 | out_notrans: | 7036 | out_notrans: |
6957 | if (old_inode->i_ino == BTRFS_FIRST_FREE_OBJECTID) | 7037 | if (old_ino == BTRFS_FIRST_FREE_OBJECTID) |
6958 | up_read(&root->fs_info->subvol_sem); | 7038 | up_read(&root->fs_info->subvol_sem); |
6959 | 7039 | ||
6960 | return ret; | 7040 | return ret; |
@@ -7031,9 +7111,6 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
7031 | if (name_len > BTRFS_MAX_INLINE_DATA_SIZE(root)) | 7111 | if (name_len > BTRFS_MAX_INLINE_DATA_SIZE(root)) |
7032 | return -ENAMETOOLONG; | 7112 | return -ENAMETOOLONG; |
7033 | 7113 | ||
7034 | err = btrfs_find_free_objectid(NULL, root, dir->i_ino, &objectid); | ||
7035 | if (err) | ||
7036 | return err; | ||
7037 | /* | 7114 | /* |
7038 | * 2 items for inode item and ref | 7115 | * 2 items for inode item and ref |
7039 | * 2 items for dir items | 7116 | * 2 items for dir items |
@@ -7045,8 +7122,12 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
7045 | 7122 | ||
7046 | btrfs_set_trans_block_group(trans, dir); | 7123 | btrfs_set_trans_block_group(trans, dir); |
7047 | 7124 | ||
7125 | err = btrfs_find_free_ino(root, &objectid); | ||
7126 | if (err) | ||
7127 | goto out_unlock; | ||
7128 | |||
7048 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, | 7129 | inode = btrfs_new_inode(trans, root, dir, dentry->d_name.name, |
7049 | dentry->d_name.len, dir->i_ino, objectid, | 7130 | dentry->d_name.len, btrfs_ino(dir), objectid, |
7050 | BTRFS_I(dir)->block_group, S_IFLNK|S_IRWXUGO, | 7131 | BTRFS_I(dir)->block_group, S_IFLNK|S_IRWXUGO, |
7051 | &index); | 7132 | &index); |
7052 | if (IS_ERR(inode)) { | 7133 | if (IS_ERR(inode)) { |
@@ -7078,7 +7159,7 @@ static int btrfs_symlink(struct inode *dir, struct dentry *dentry, | |||
7078 | 7159 | ||
7079 | path = btrfs_alloc_path(); | 7160 | path = btrfs_alloc_path(); |
7080 | BUG_ON(!path); | 7161 | BUG_ON(!path); |
7081 | key.objectid = inode->i_ino; | 7162 | key.objectid = btrfs_ino(inode); |
7082 | key.offset = 0; | 7163 | key.offset = 0; |
7083 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); | 7164 | btrfs_set_key_type(&key, BTRFS_EXTENT_DATA_KEY); |
7084 | datasize = btrfs_file_extent_calc_inline_size(name_len); | 7165 | datasize = btrfs_file_extent_calc_inline_size(name_len); |
diff --git a/fs/btrfs/ioctl.c b/fs/btrfs/ioctl.c index d11fc6548e15..ed8c055ab70f 100644 --- a/fs/btrfs/ioctl.c +++ b/fs/btrfs/ioctl.c | |||
@@ -50,6 +50,7 @@ | |||
50 | #include "print-tree.h" | 50 | #include "print-tree.h" |
51 | #include "volumes.h" | 51 | #include "volumes.h" |
52 | #include "locking.h" | 52 | #include "locking.h" |
53 | #include "inode-map.h" | ||
53 | 54 | ||
54 | /* Mask out flags that are inappropriate for the given type of inode. */ | 55 | /* Mask out flags that are inappropriate for the given type of inode. */ |
55 | static inline __u32 btrfs_mask_flags(umode_t mode, __u32 flags) | 56 | static inline __u32 btrfs_mask_flags(umode_t mode, __u32 flags) |
@@ -81,6 +82,13 @@ static unsigned int btrfs_flags_to_ioctl(unsigned int flags) | |||
81 | iflags |= FS_NOATIME_FL; | 82 | iflags |= FS_NOATIME_FL; |
82 | if (flags & BTRFS_INODE_DIRSYNC) | 83 | if (flags & BTRFS_INODE_DIRSYNC) |
83 | iflags |= FS_DIRSYNC_FL; | 84 | iflags |= FS_DIRSYNC_FL; |
85 | if (flags & BTRFS_INODE_NODATACOW) | ||
86 | iflags |= FS_NOCOW_FL; | ||
87 | |||
88 | if ((flags & BTRFS_INODE_COMPRESS) && !(flags & BTRFS_INODE_NOCOMPRESS)) | ||
89 | iflags |= FS_COMPR_FL; | ||
90 | else if (flags & BTRFS_INODE_NOCOMPRESS) | ||
91 | iflags |= FS_NOCOMP_FL; | ||
84 | 92 | ||
85 | return iflags; | 93 | return iflags; |
86 | } | 94 | } |
@@ -144,16 +152,13 @@ static int check_flags(unsigned int flags) | |||
144 | if (flags & ~(FS_IMMUTABLE_FL | FS_APPEND_FL | \ | 152 | if (flags & ~(FS_IMMUTABLE_FL | FS_APPEND_FL | \ |
145 | FS_NOATIME_FL | FS_NODUMP_FL | \ | 153 | FS_NOATIME_FL | FS_NODUMP_FL | \ |
146 | FS_SYNC_FL | FS_DIRSYNC_FL | \ | 154 | FS_SYNC_FL | FS_DIRSYNC_FL | \ |
147 | FS_NOCOMP_FL | FS_COMPR_FL | \ | 155 | FS_NOCOMP_FL | FS_COMPR_FL | |
148 | FS_NOCOW_FL | FS_COW_FL)) | 156 | FS_NOCOW_FL)) |
149 | return -EOPNOTSUPP; | 157 | return -EOPNOTSUPP; |
150 | 158 | ||
151 | if ((flags & FS_NOCOMP_FL) && (flags & FS_COMPR_FL)) | 159 | if ((flags & FS_NOCOMP_FL) && (flags & FS_COMPR_FL)) |
152 | return -EINVAL; | 160 | return -EINVAL; |
153 | 161 | ||
154 | if ((flags & FS_NOCOW_FL) && (flags & FS_COW_FL)) | ||
155 | return -EINVAL; | ||
156 | |||
157 | return 0; | 162 | return 0; |
158 | } | 163 | } |
159 | 164 | ||
@@ -218,6 +223,10 @@ static int btrfs_ioctl_setflags(struct file *file, void __user *arg) | |||
218 | ip->flags |= BTRFS_INODE_DIRSYNC; | 223 | ip->flags |= BTRFS_INODE_DIRSYNC; |
219 | else | 224 | else |
220 | ip->flags &= ~BTRFS_INODE_DIRSYNC; | 225 | ip->flags &= ~BTRFS_INODE_DIRSYNC; |
226 | if (flags & FS_NOCOW_FL) | ||
227 | ip->flags |= BTRFS_INODE_NODATACOW; | ||
228 | else | ||
229 | ip->flags &= ~BTRFS_INODE_NODATACOW; | ||
221 | 230 | ||
222 | /* | 231 | /* |
223 | * The COMPRESS flag can only be changed by users, while the NOCOMPRESS | 232 | * The COMPRESS flag can only be changed by users, while the NOCOMPRESS |
@@ -230,11 +239,9 @@ static int btrfs_ioctl_setflags(struct file *file, void __user *arg) | |||
230 | } else if (flags & FS_COMPR_FL) { | 239 | } else if (flags & FS_COMPR_FL) { |
231 | ip->flags |= BTRFS_INODE_COMPRESS; | 240 | ip->flags |= BTRFS_INODE_COMPRESS; |
232 | ip->flags &= ~BTRFS_INODE_NOCOMPRESS; | 241 | ip->flags &= ~BTRFS_INODE_NOCOMPRESS; |
242 | } else { | ||
243 | ip->flags &= ~(BTRFS_INODE_COMPRESS | BTRFS_INODE_NOCOMPRESS); | ||
233 | } | 244 | } |
234 | if (flags & FS_NOCOW_FL) | ||
235 | ip->flags |= BTRFS_INODE_NODATACOW; | ||
236 | else if (flags & FS_COW_FL) | ||
237 | ip->flags &= ~BTRFS_INODE_NODATACOW; | ||
238 | 245 | ||
239 | trans = btrfs_join_transaction(root, 1); | 246 | trans = btrfs_join_transaction(root, 1); |
240 | BUG_ON(IS_ERR(trans)); | 247 | BUG_ON(IS_ERR(trans)); |
@@ -323,8 +330,7 @@ static noinline int create_subvol(struct btrfs_root *root, | |||
323 | u64 new_dirid = BTRFS_FIRST_FREE_OBJECTID; | 330 | u64 new_dirid = BTRFS_FIRST_FREE_OBJECTID; |
324 | u64 index = 0; | 331 | u64 index = 0; |
325 | 332 | ||
326 | ret = btrfs_find_free_objectid(NULL, root->fs_info->tree_root, | 333 | ret = btrfs_find_free_objectid(root->fs_info->tree_root, &objectid); |
327 | 0, &objectid); | ||
328 | if (ret) { | 334 | if (ret) { |
329 | dput(parent); | 335 | dput(parent); |
330 | return ret; | 336 | return ret; |
@@ -416,7 +422,7 @@ static noinline int create_subvol(struct btrfs_root *root, | |||
416 | BUG_ON(ret); | 422 | BUG_ON(ret); |
417 | 423 | ||
418 | ret = btrfs_insert_dir_item(trans, root, | 424 | ret = btrfs_insert_dir_item(trans, root, |
419 | name, namelen, dir->i_ino, &key, | 425 | name, namelen, dir, &key, |
420 | BTRFS_FT_DIR, index); | 426 | BTRFS_FT_DIR, index); |
421 | if (ret) | 427 | if (ret) |
422 | goto fail; | 428 | goto fail; |
@@ -427,7 +433,7 @@ static noinline int create_subvol(struct btrfs_root *root, | |||
427 | 433 | ||
428 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, | 434 | ret = btrfs_add_root_ref(trans, root->fs_info->tree_root, |
429 | objectid, root->root_key.objectid, | 435 | objectid, root->root_key.objectid, |
430 | dir->i_ino, index, name, namelen); | 436 | btrfs_ino(dir), index, name, namelen); |
431 | 437 | ||
432 | BUG_ON(ret); | 438 | BUG_ON(ret); |
433 | 439 | ||
@@ -1123,7 +1129,7 @@ static noinline int btrfs_ioctl_subvol_getflags(struct file *file, | |||
1123 | int ret = 0; | 1129 | int ret = 0; |
1124 | u64 flags = 0; | 1130 | u64 flags = 0; |
1125 | 1131 | ||
1126 | if (inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) | 1132 | if (btrfs_ino(inode) != BTRFS_FIRST_FREE_OBJECTID) |
1127 | return -EINVAL; | 1133 | return -EINVAL; |
1128 | 1134 | ||
1129 | down_read(&root->fs_info->subvol_sem); | 1135 | down_read(&root->fs_info->subvol_sem); |
@@ -1150,7 +1156,7 @@ static noinline int btrfs_ioctl_subvol_setflags(struct file *file, | |||
1150 | if (root->fs_info->sb->s_flags & MS_RDONLY) | 1156 | if (root->fs_info->sb->s_flags & MS_RDONLY) |
1151 | return -EROFS; | 1157 | return -EROFS; |
1152 | 1158 | ||
1153 | if (inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) | 1159 | if (btrfs_ino(inode) != BTRFS_FIRST_FREE_OBJECTID) |
1154 | return -EINVAL; | 1160 | return -EINVAL; |
1155 | 1161 | ||
1156 | if (copy_from_user(&flags, arg, sizeof(flags))) | 1162 | if (copy_from_user(&flags, arg, sizeof(flags))) |
@@ -1633,7 +1639,7 @@ static noinline int btrfs_ioctl_snap_destroy(struct file *file, | |||
1633 | goto out_dput; | 1639 | goto out_dput; |
1634 | } | 1640 | } |
1635 | 1641 | ||
1636 | if (inode->i_ino != BTRFS_FIRST_FREE_OBJECTID) { | 1642 | if (btrfs_ino(inode) != BTRFS_FIRST_FREE_OBJECTID) { |
1637 | err = -EINVAL; | 1643 | err = -EINVAL; |
1638 | goto out_dput; | 1644 | goto out_dput; |
1639 | } | 1645 | } |
@@ -1919,7 +1925,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
1919 | } | 1925 | } |
1920 | 1926 | ||
1921 | /* clone data */ | 1927 | /* clone data */ |
1922 | key.objectid = src->i_ino; | 1928 | key.objectid = btrfs_ino(src); |
1923 | key.type = BTRFS_EXTENT_DATA_KEY; | 1929 | key.type = BTRFS_EXTENT_DATA_KEY; |
1924 | key.offset = 0; | 1930 | key.offset = 0; |
1925 | 1931 | ||
@@ -1946,7 +1952,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
1946 | 1952 | ||
1947 | btrfs_item_key_to_cpu(leaf, &key, slot); | 1953 | btrfs_item_key_to_cpu(leaf, &key, slot); |
1948 | if (btrfs_key_type(&key) > BTRFS_EXTENT_DATA_KEY || | 1954 | if (btrfs_key_type(&key) > BTRFS_EXTENT_DATA_KEY || |
1949 | key.objectid != src->i_ino) | 1955 | key.objectid != btrfs_ino(src)) |
1950 | break; | 1956 | break; |
1951 | 1957 | ||
1952 | if (btrfs_key_type(&key) == BTRFS_EXTENT_DATA_KEY) { | 1958 | if (btrfs_key_type(&key) == BTRFS_EXTENT_DATA_KEY) { |
@@ -1989,7 +1995,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
1989 | goto next; | 1995 | goto next; |
1990 | 1996 | ||
1991 | memcpy(&new_key, &key, sizeof(new_key)); | 1997 | memcpy(&new_key, &key, sizeof(new_key)); |
1992 | new_key.objectid = inode->i_ino; | 1998 | new_key.objectid = btrfs_ino(inode); |
1993 | if (off <= key.offset) | 1999 | if (off <= key.offset) |
1994 | new_key.offset = key.offset + destoff - off; | 2000 | new_key.offset = key.offset + destoff - off; |
1995 | else | 2001 | else |
@@ -2043,7 +2049,7 @@ static noinline long btrfs_ioctl_clone(struct file *file, unsigned long srcfd, | |||
2043 | ret = btrfs_inc_extent_ref(trans, root, | 2049 | ret = btrfs_inc_extent_ref(trans, root, |
2044 | disko, diskl, 0, | 2050 | disko, diskl, 0, |
2045 | root->root_key.objectid, | 2051 | root->root_key.objectid, |
2046 | inode->i_ino, | 2052 | btrfs_ino(inode), |
2047 | new_key.offset - datao); | 2053 | new_key.offset - datao); |
2048 | BUG_ON(ret); | 2054 | BUG_ON(ret); |
2049 | } | 2055 | } |
diff --git a/fs/btrfs/relocation.c b/fs/btrfs/relocation.c index f726e72dd362..051992c7fcc9 100644 --- a/fs/btrfs/relocation.c +++ b/fs/btrfs/relocation.c | |||
@@ -30,6 +30,7 @@ | |||
30 | #include "btrfs_inode.h" | 30 | #include "btrfs_inode.h" |
31 | #include "async-thread.h" | 31 | #include "async-thread.h" |
32 | #include "free-space-cache.h" | 32 | #include "free-space-cache.h" |
33 | #include "inode-map.h" | ||
33 | 34 | ||
34 | /* | 35 | /* |
35 | * backref_node, mapping_node and tree_block start with this | 36 | * backref_node, mapping_node and tree_block start with this |
@@ -1409,9 +1410,9 @@ again: | |||
1409 | prev = node; | 1410 | prev = node; |
1410 | entry = rb_entry(node, struct btrfs_inode, rb_node); | 1411 | entry = rb_entry(node, struct btrfs_inode, rb_node); |
1411 | 1412 | ||
1412 | if (objectid < entry->vfs_inode.i_ino) | 1413 | if (objectid < btrfs_ino(&entry->vfs_inode)) |
1413 | node = node->rb_left; | 1414 | node = node->rb_left; |
1414 | else if (objectid > entry->vfs_inode.i_ino) | 1415 | else if (objectid > btrfs_ino(&entry->vfs_inode)) |
1415 | node = node->rb_right; | 1416 | node = node->rb_right; |
1416 | else | 1417 | else |
1417 | break; | 1418 | break; |
@@ -1419,7 +1420,7 @@ again: | |||
1419 | if (!node) { | 1420 | if (!node) { |
1420 | while (prev) { | 1421 | while (prev) { |
1421 | entry = rb_entry(prev, struct btrfs_inode, rb_node); | 1422 | entry = rb_entry(prev, struct btrfs_inode, rb_node); |
1422 | if (objectid <= entry->vfs_inode.i_ino) { | 1423 | if (objectid <= btrfs_ino(&entry->vfs_inode)) { |
1423 | node = prev; | 1424 | node = prev; |
1424 | break; | 1425 | break; |
1425 | } | 1426 | } |
@@ -1434,7 +1435,7 @@ again: | |||
1434 | return inode; | 1435 | return inode; |
1435 | } | 1436 | } |
1436 | 1437 | ||
1437 | objectid = entry->vfs_inode.i_ino + 1; | 1438 | objectid = btrfs_ino(&entry->vfs_inode) + 1; |
1438 | if (cond_resched_lock(&root->inode_lock)) | 1439 | if (cond_resched_lock(&root->inode_lock)) |
1439 | goto again; | 1440 | goto again; |
1440 | 1441 | ||
@@ -1470,7 +1471,7 @@ static int get_new_location(struct inode *reloc_inode, u64 *new_bytenr, | |||
1470 | return -ENOMEM; | 1471 | return -ENOMEM; |
1471 | 1472 | ||
1472 | bytenr -= BTRFS_I(reloc_inode)->index_cnt; | 1473 | bytenr -= BTRFS_I(reloc_inode)->index_cnt; |
1473 | ret = btrfs_lookup_file_extent(NULL, root, path, reloc_inode->i_ino, | 1474 | ret = btrfs_lookup_file_extent(NULL, root, path, btrfs_ino(reloc_inode), |
1474 | bytenr, 0); | 1475 | bytenr, 0); |
1475 | if (ret < 0) | 1476 | if (ret < 0) |
1476 | goto out; | 1477 | goto out; |
@@ -1558,11 +1559,11 @@ int replace_file_extents(struct btrfs_trans_handle *trans, | |||
1558 | if (first) { | 1559 | if (first) { |
1559 | inode = find_next_inode(root, key.objectid); | 1560 | inode = find_next_inode(root, key.objectid); |
1560 | first = 0; | 1561 | first = 0; |
1561 | } else if (inode && inode->i_ino < key.objectid) { | 1562 | } else if (inode && btrfs_ino(inode) < key.objectid) { |
1562 | btrfs_add_delayed_iput(inode); | 1563 | btrfs_add_delayed_iput(inode); |
1563 | inode = find_next_inode(root, key.objectid); | 1564 | inode = find_next_inode(root, key.objectid); |
1564 | } | 1565 | } |
1565 | if (inode && inode->i_ino == key.objectid) { | 1566 | if (inode && btrfs_ino(inode) == key.objectid) { |
1566 | end = key.offset + | 1567 | end = key.offset + |
1567 | btrfs_file_extent_num_bytes(leaf, fi); | 1568 | btrfs_file_extent_num_bytes(leaf, fi); |
1568 | WARN_ON(!IS_ALIGNED(key.offset, | 1569 | WARN_ON(!IS_ALIGNED(key.offset, |
@@ -1893,6 +1894,7 @@ static int invalidate_extent_cache(struct btrfs_root *root, | |||
1893 | struct inode *inode = NULL; | 1894 | struct inode *inode = NULL; |
1894 | u64 objectid; | 1895 | u64 objectid; |
1895 | u64 start, end; | 1896 | u64 start, end; |
1897 | u64 ino; | ||
1896 | 1898 | ||
1897 | objectid = min_key->objectid; | 1899 | objectid = min_key->objectid; |
1898 | while (1) { | 1900 | while (1) { |
@@ -1905,17 +1907,18 @@ static int invalidate_extent_cache(struct btrfs_root *root, | |||
1905 | inode = find_next_inode(root, objectid); | 1907 | inode = find_next_inode(root, objectid); |
1906 | if (!inode) | 1908 | if (!inode) |
1907 | break; | 1909 | break; |
1910 | ino = btrfs_ino(inode); | ||
1908 | 1911 | ||
1909 | if (inode->i_ino > max_key->objectid) { | 1912 | if (ino > max_key->objectid) { |
1910 | iput(inode); | 1913 | iput(inode); |
1911 | break; | 1914 | break; |
1912 | } | 1915 | } |
1913 | 1916 | ||
1914 | objectid = inode->i_ino + 1; | 1917 | objectid = ino + 1; |
1915 | if (!S_ISREG(inode->i_mode)) | 1918 | if (!S_ISREG(inode->i_mode)) |
1916 | continue; | 1919 | continue; |
1917 | 1920 | ||
1918 | if (unlikely(min_key->objectid == inode->i_ino)) { | 1921 | if (unlikely(min_key->objectid == ino)) { |
1919 | if (min_key->type > BTRFS_EXTENT_DATA_KEY) | 1922 | if (min_key->type > BTRFS_EXTENT_DATA_KEY) |
1920 | continue; | 1923 | continue; |
1921 | if (min_key->type < BTRFS_EXTENT_DATA_KEY) | 1924 | if (min_key->type < BTRFS_EXTENT_DATA_KEY) |
@@ -1928,7 +1931,7 @@ static int invalidate_extent_cache(struct btrfs_root *root, | |||
1928 | start = 0; | 1931 | start = 0; |
1929 | } | 1932 | } |
1930 | 1933 | ||
1931 | if (unlikely(max_key->objectid == inode->i_ino)) { | 1934 | if (unlikely(max_key->objectid == ino)) { |
1932 | if (max_key->type < BTRFS_EXTENT_DATA_KEY) | 1935 | if (max_key->type < BTRFS_EXTENT_DATA_KEY) |
1933 | continue; | 1936 | continue; |
1934 | if (max_key->type > BTRFS_EXTENT_DATA_KEY) { | 1937 | if (max_key->type > BTRFS_EXTENT_DATA_KEY) { |
@@ -3897,7 +3900,7 @@ struct inode *create_reloc_inode(struct btrfs_fs_info *fs_info, | |||
3897 | if (IS_ERR(trans)) | 3900 | if (IS_ERR(trans)) |
3898 | return ERR_CAST(trans); | 3901 | return ERR_CAST(trans); |
3899 | 3902 | ||
3900 | err = btrfs_find_free_objectid(trans, root, objectid, &objectid); | 3903 | err = btrfs_find_free_objectid(root, &objectid); |
3901 | if (err) | 3904 | if (err) |
3902 | goto out; | 3905 | goto out; |
3903 | 3906 | ||
diff --git a/fs/btrfs/super.c b/fs/btrfs/super.c index 3e28521643fb..fb72e2bea882 100644 --- a/fs/btrfs/super.c +++ b/fs/btrfs/super.c | |||
@@ -40,6 +40,7 @@ | |||
40 | #include <linux/magic.h> | 40 | #include <linux/magic.h> |
41 | #include <linux/slab.h> | 41 | #include <linux/slab.h> |
42 | #include "compat.h" | 42 | #include "compat.h" |
43 | #include "delayed-inode.h" | ||
43 | #include "ctree.h" | 44 | #include "ctree.h" |
44 | #include "disk-io.h" | 45 | #include "disk-io.h" |
45 | #include "transaction.h" | 46 | #include "transaction.h" |
@@ -1206,10 +1207,14 @@ static int __init init_btrfs_fs(void) | |||
1206 | if (err) | 1207 | if (err) |
1207 | goto free_extent_io; | 1208 | goto free_extent_io; |
1208 | 1209 | ||
1209 | err = btrfs_interface_init(); | 1210 | err = btrfs_delayed_inode_init(); |
1210 | if (err) | 1211 | if (err) |
1211 | goto free_extent_map; | 1212 | goto free_extent_map; |
1212 | 1213 | ||
1214 | err = btrfs_interface_init(); | ||
1215 | if (err) | ||
1216 | goto free_delayed_inode; | ||
1217 | |||
1213 | err = register_filesystem(&btrfs_fs_type); | 1218 | err = register_filesystem(&btrfs_fs_type); |
1214 | if (err) | 1219 | if (err) |
1215 | goto unregister_ioctl; | 1220 | goto unregister_ioctl; |
@@ -1219,6 +1224,8 @@ static int __init init_btrfs_fs(void) | |||
1219 | 1224 | ||
1220 | unregister_ioctl: | 1225 | unregister_ioctl: |
1221 | btrfs_interface_exit(); | 1226 | btrfs_interface_exit(); |
1227 | free_delayed_inode: | ||
1228 | btrfs_delayed_inode_exit(); | ||
1222 | free_extent_map: | 1229 | free_extent_map: |
1223 | extent_map_exit(); | 1230 | extent_map_exit(); |
1224 | free_extent_io: | 1231 | free_extent_io: |
@@ -1235,6 +1242,7 @@ free_sysfs: | |||
1235 | static void __exit exit_btrfs_fs(void) | 1242 | static void __exit exit_btrfs_fs(void) |
1236 | { | 1243 | { |
1237 | btrfs_destroy_cachep(); | 1244 | btrfs_destroy_cachep(); |
1245 | btrfs_delayed_inode_exit(); | ||
1238 | extent_map_exit(); | 1246 | extent_map_exit(); |
1239 | extent_io_exit(); | 1247 | extent_io_exit(); |
1240 | btrfs_interface_exit(); | 1248 | btrfs_interface_exit(); |
diff --git a/fs/btrfs/sysfs.c b/fs/btrfs/sysfs.c index ab9633fd72a4..c3c223ae6691 100644 --- a/fs/btrfs/sysfs.c +++ b/fs/btrfs/sysfs.c | |||
@@ -174,18 +174,6 @@ static const struct sysfs_ops btrfs_root_attr_ops = { | |||
174 | .store = btrfs_root_attr_store, | 174 | .store = btrfs_root_attr_store, |
175 | }; | 175 | }; |
176 | 176 | ||
177 | static struct kobj_type btrfs_root_ktype = { | ||
178 | .default_attrs = btrfs_root_attrs, | ||
179 | .sysfs_ops = &btrfs_root_attr_ops, | ||
180 | .release = btrfs_root_release, | ||
181 | }; | ||
182 | |||
183 | static struct kobj_type btrfs_super_ktype = { | ||
184 | .default_attrs = btrfs_super_attrs, | ||
185 | .sysfs_ops = &btrfs_super_attr_ops, | ||
186 | .release = btrfs_super_release, | ||
187 | }; | ||
188 | |||
189 | /* /sys/fs/btrfs/ entry */ | 177 | /* /sys/fs/btrfs/ entry */ |
190 | static struct kset *btrfs_kset; | 178 | static struct kset *btrfs_kset; |
191 | 179 | ||
diff --git a/fs/btrfs/transaction.c b/fs/btrfs/transaction.c index 211aceeb9ea0..33679fc710c6 100644 --- a/fs/btrfs/transaction.c +++ b/fs/btrfs/transaction.c | |||
@@ -27,6 +27,7 @@ | |||
27 | #include "transaction.h" | 27 | #include "transaction.h" |
28 | #include "locking.h" | 28 | #include "locking.h" |
29 | #include "tree-log.h" | 29 | #include "tree-log.h" |
30 | #include "inode-map.h" | ||
30 | 31 | ||
31 | #define BTRFS_ROOT_TRANS_TAG 0 | 32 | #define BTRFS_ROOT_TRANS_TAG 0 |
32 | 33 | ||
@@ -443,19 +444,40 @@ static int __btrfs_end_transaction(struct btrfs_trans_handle *trans, | |||
443 | int btrfs_end_transaction(struct btrfs_trans_handle *trans, | 444 | int btrfs_end_transaction(struct btrfs_trans_handle *trans, |
444 | struct btrfs_root *root) | 445 | struct btrfs_root *root) |
445 | { | 446 | { |
446 | return __btrfs_end_transaction(trans, root, 0, 1); | 447 | int ret; |
448 | |||
449 | ret = __btrfs_end_transaction(trans, root, 0, 1); | ||
450 | if (ret) | ||
451 | return ret; | ||
452 | return 0; | ||
447 | } | 453 | } |
448 | 454 | ||
449 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, | 455 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, |
450 | struct btrfs_root *root) | 456 | struct btrfs_root *root) |
451 | { | 457 | { |
452 | return __btrfs_end_transaction(trans, root, 1, 1); | 458 | int ret; |
459 | |||
460 | ret = __btrfs_end_transaction(trans, root, 1, 1); | ||
461 | if (ret) | ||
462 | return ret; | ||
463 | return 0; | ||
453 | } | 464 | } |
454 | 465 | ||
455 | int btrfs_end_transaction_nolock(struct btrfs_trans_handle *trans, | 466 | int btrfs_end_transaction_nolock(struct btrfs_trans_handle *trans, |
456 | struct btrfs_root *root) | 467 | struct btrfs_root *root) |
457 | { | 468 | { |
458 | return __btrfs_end_transaction(trans, root, 0, 0); | 469 | int ret; |
470 | |||
471 | ret = __btrfs_end_transaction(trans, root, 0, 0); | ||
472 | if (ret) | ||
473 | return ret; | ||
474 | return 0; | ||
475 | } | ||
476 | |||
477 | int btrfs_end_transaction_dmeta(struct btrfs_trans_handle *trans, | ||
478 | struct btrfs_root *root) | ||
479 | { | ||
480 | return __btrfs_end_transaction(trans, root, 1, 1); | ||
459 | } | 481 | } |
460 | 482 | ||
461 | /* | 483 | /* |
@@ -716,8 +738,14 @@ static noinline int commit_fs_roots(struct btrfs_trans_handle *trans, | |||
716 | btrfs_update_reloc_root(trans, root); | 738 | btrfs_update_reloc_root(trans, root); |
717 | btrfs_orphan_commit_root(trans, root); | 739 | btrfs_orphan_commit_root(trans, root); |
718 | 740 | ||
741 | btrfs_save_ino_cache(root, trans); | ||
742 | |||
719 | if (root->commit_root != root->node) { | 743 | if (root->commit_root != root->node) { |
744 | mutex_lock(&root->fs_commit_mutex); | ||
720 | switch_commit_root(root); | 745 | switch_commit_root(root); |
746 | btrfs_unpin_free_ino(root); | ||
747 | mutex_unlock(&root->fs_commit_mutex); | ||
748 | |||
721 | btrfs_set_root_node(&root->root_item, | 749 | btrfs_set_root_node(&root->root_item, |
722 | root->node); | 750 | root->node); |
723 | } | 751 | } |
@@ -795,7 +823,7 @@ static noinline int create_pending_snapshot(struct btrfs_trans_handle *trans, | |||
795 | goto fail; | 823 | goto fail; |
796 | } | 824 | } |
797 | 825 | ||
798 | ret = btrfs_find_free_objectid(trans, tree_root, 0, &objectid); | 826 | ret = btrfs_find_free_objectid(tree_root, &objectid); |
799 | if (ret) { | 827 | if (ret) { |
800 | pending->error = ret; | 828 | pending->error = ret; |
801 | goto fail; | 829 | goto fail; |
@@ -832,7 +860,7 @@ static noinline int create_pending_snapshot(struct btrfs_trans_handle *trans, | |||
832 | BUG_ON(ret); | 860 | BUG_ON(ret); |
833 | ret = btrfs_insert_dir_item(trans, parent_root, | 861 | ret = btrfs_insert_dir_item(trans, parent_root, |
834 | dentry->d_name.name, dentry->d_name.len, | 862 | dentry->d_name.name, dentry->d_name.len, |
835 | parent_inode->i_ino, &key, | 863 | parent_inode, &key, |
836 | BTRFS_FT_DIR, index); | 864 | BTRFS_FT_DIR, index); |
837 | BUG_ON(ret); | 865 | BUG_ON(ret); |
838 | 866 | ||
@@ -874,7 +902,7 @@ static noinline int create_pending_snapshot(struct btrfs_trans_handle *trans, | |||
874 | */ | 902 | */ |
875 | ret = btrfs_add_root_ref(trans, tree_root, objectid, | 903 | ret = btrfs_add_root_ref(trans, tree_root, objectid, |
876 | parent_root->root_key.objectid, | 904 | parent_root->root_key.objectid, |
877 | parent_inode->i_ino, index, | 905 | btrfs_ino(parent_inode), index, |
878 | dentry->d_name.name, dentry->d_name.len); | 906 | dentry->d_name.name, dentry->d_name.len); |
879 | BUG_ON(ret); | 907 | BUG_ON(ret); |
880 | dput(parent); | 908 | dput(parent); |
@@ -902,6 +930,14 @@ static noinline int create_pending_snapshots(struct btrfs_trans_handle *trans, | |||
902 | int ret; | 930 | int ret; |
903 | 931 | ||
904 | list_for_each_entry(pending, head, list) { | 932 | list_for_each_entry(pending, head, list) { |
933 | /* | ||
934 | * We must deal with the delayed items before creating | ||
935 | * snapshots, or we will create a snapthot with inconsistent | ||
936 | * information. | ||
937 | */ | ||
938 | ret = btrfs_run_delayed_items(trans, fs_info->fs_root); | ||
939 | BUG_ON(ret); | ||
940 | |||
905 | ret = create_pending_snapshot(trans, fs_info, pending); | 941 | ret = create_pending_snapshot(trans, fs_info, pending); |
906 | BUG_ON(ret); | 942 | BUG_ON(ret); |
907 | } | 943 | } |
@@ -1155,6 +1191,9 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, | |||
1155 | BUG_ON(ret); | 1191 | BUG_ON(ret); |
1156 | } | 1192 | } |
1157 | 1193 | ||
1194 | ret = btrfs_run_delayed_items(trans, root); | ||
1195 | BUG_ON(ret); | ||
1196 | |||
1158 | /* | 1197 | /* |
1159 | * rename don't use btrfs_join_transaction, so, once we | 1198 | * rename don't use btrfs_join_transaction, so, once we |
1160 | * set the transaction to blocked above, we aren't going | 1199 | * set the transaction to blocked above, we aren't going |
@@ -1181,6 +1220,9 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, | |||
1181 | ret = create_pending_snapshots(trans, root->fs_info); | 1220 | ret = create_pending_snapshots(trans, root->fs_info); |
1182 | BUG_ON(ret); | 1221 | BUG_ON(ret); |
1183 | 1222 | ||
1223 | ret = btrfs_run_delayed_items(trans, root); | ||
1224 | BUG_ON(ret); | ||
1225 | |||
1184 | ret = btrfs_run_delayed_refs(trans, root, (unsigned long)-1); | 1226 | ret = btrfs_run_delayed_refs(trans, root, (unsigned long)-1); |
1185 | BUG_ON(ret); | 1227 | BUG_ON(ret); |
1186 | 1228 | ||
@@ -1297,6 +1339,8 @@ int btrfs_clean_old_snapshots(struct btrfs_root *root) | |||
1297 | root = list_entry(list.next, struct btrfs_root, root_list); | 1339 | root = list_entry(list.next, struct btrfs_root, root_list); |
1298 | list_del(&root->root_list); | 1340 | list_del(&root->root_list); |
1299 | 1341 | ||
1342 | btrfs_kill_all_delayed_nodes(root); | ||
1343 | |||
1300 | if (btrfs_header_backref_rev(root->node) < | 1344 | if (btrfs_header_backref_rev(root->node) < |
1301 | BTRFS_MIXED_BACKREF_REV) | 1345 | BTRFS_MIXED_BACKREF_REV) |
1302 | btrfs_drop_snapshot(root, NULL, 0); | 1346 | btrfs_drop_snapshot(root, NULL, 0); |
diff --git a/fs/btrfs/transaction.h b/fs/btrfs/transaction.h index 000a41008c3b..804c88639e5d 100644 --- a/fs/btrfs/transaction.h +++ b/fs/btrfs/transaction.h | |||
@@ -112,6 +112,8 @@ int btrfs_commit_transaction_async(struct btrfs_trans_handle *trans, | |||
112 | int wait_for_unblock); | 112 | int wait_for_unblock); |
113 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, | 113 | int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans, |
114 | struct btrfs_root *root); | 114 | struct btrfs_root *root); |
115 | int btrfs_end_transaction_dmeta(struct btrfs_trans_handle *trans, | ||
116 | struct btrfs_root *root); | ||
115 | int btrfs_should_end_transaction(struct btrfs_trans_handle *trans, | 117 | int btrfs_should_end_transaction(struct btrfs_trans_handle *trans, |
116 | struct btrfs_root *root); | 118 | struct btrfs_root *root); |
117 | void btrfs_throttle(struct btrfs_root *root); | 119 | void btrfs_throttle(struct btrfs_root *root); |
diff --git a/fs/btrfs/tree-log.c b/fs/btrfs/tree-log.c index c599e8c2a53c..a794b9f60138 100644 --- a/fs/btrfs/tree-log.c +++ b/fs/btrfs/tree-log.c | |||
@@ -519,7 +519,7 @@ static noinline int replay_one_extent(struct btrfs_trans_handle *trans, | |||
519 | * file. This must be done before the btrfs_drop_extents run | 519 | * file. This must be done before the btrfs_drop_extents run |
520 | * so we don't try to drop this extent. | 520 | * so we don't try to drop this extent. |
521 | */ | 521 | */ |
522 | ret = btrfs_lookup_file_extent(trans, root, path, inode->i_ino, | 522 | ret = btrfs_lookup_file_extent(trans, root, path, btrfs_ino(inode), |
523 | start, 0); | 523 | start, 0); |
524 | 524 | ||
525 | if (ret == 0 && | 525 | if (ret == 0 && |
@@ -832,7 +832,7 @@ again: | |||
832 | read_extent_buffer(eb, name, (unsigned long)(ref + 1), namelen); | 832 | read_extent_buffer(eb, name, (unsigned long)(ref + 1), namelen); |
833 | 833 | ||
834 | /* if we already have a perfect match, we're done */ | 834 | /* if we already have a perfect match, we're done */ |
835 | if (inode_in_dir(root, path, dir->i_ino, inode->i_ino, | 835 | if (inode_in_dir(root, path, btrfs_ino(dir), btrfs_ino(inode), |
836 | btrfs_inode_ref_index(eb, ref), | 836 | btrfs_inode_ref_index(eb, ref), |
837 | name, namelen)) { | 837 | name, namelen)) { |
838 | goto out; | 838 | goto out; |
@@ -960,8 +960,9 @@ static noinline int fixup_inode_link_count(struct btrfs_trans_handle *trans, | |||
960 | unsigned long ptr; | 960 | unsigned long ptr; |
961 | unsigned long ptr_end; | 961 | unsigned long ptr_end; |
962 | int name_len; | 962 | int name_len; |
963 | u64 ino = btrfs_ino(inode); | ||
963 | 964 | ||
964 | key.objectid = inode->i_ino; | 965 | key.objectid = ino; |
965 | key.type = BTRFS_INODE_REF_KEY; | 966 | key.type = BTRFS_INODE_REF_KEY; |
966 | key.offset = (u64)-1; | 967 | key.offset = (u64)-1; |
967 | 968 | ||
@@ -980,7 +981,7 @@ static noinline int fixup_inode_link_count(struct btrfs_trans_handle *trans, | |||
980 | } | 981 | } |
981 | btrfs_item_key_to_cpu(path->nodes[0], &key, | 982 | btrfs_item_key_to_cpu(path->nodes[0], &key, |
982 | path->slots[0]); | 983 | path->slots[0]); |
983 | if (key.objectid != inode->i_ino || | 984 | if (key.objectid != ino || |
984 | key.type != BTRFS_INODE_REF_KEY) | 985 | key.type != BTRFS_INODE_REF_KEY) |
985 | break; | 986 | break; |
986 | ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); | 987 | ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); |
@@ -1011,10 +1012,10 @@ static noinline int fixup_inode_link_count(struct btrfs_trans_handle *trans, | |||
1011 | if (inode->i_nlink == 0) { | 1012 | if (inode->i_nlink == 0) { |
1012 | if (S_ISDIR(inode->i_mode)) { | 1013 | if (S_ISDIR(inode->i_mode)) { |
1013 | ret = replay_dir_deletes(trans, root, NULL, path, | 1014 | ret = replay_dir_deletes(trans, root, NULL, path, |
1014 | inode->i_ino, 1); | 1015 | ino, 1); |
1015 | BUG_ON(ret); | 1016 | BUG_ON(ret); |
1016 | } | 1017 | } |
1017 | ret = insert_orphan_item(trans, root, inode->i_ino); | 1018 | ret = insert_orphan_item(trans, root, ino); |
1018 | BUG_ON(ret); | 1019 | BUG_ON(ret); |
1019 | } | 1020 | } |
1020 | btrfs_free_path(path); | 1021 | btrfs_free_path(path); |
@@ -2197,6 +2198,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2197 | int ret; | 2198 | int ret; |
2198 | int err = 0; | 2199 | int err = 0; |
2199 | int bytes_del = 0; | 2200 | int bytes_del = 0; |
2201 | u64 dir_ino = btrfs_ino(dir); | ||
2200 | 2202 | ||
2201 | if (BTRFS_I(dir)->logged_trans < trans->transid) | 2203 | if (BTRFS_I(dir)->logged_trans < trans->transid) |
2202 | return 0; | 2204 | return 0; |
@@ -2214,7 +2216,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2214 | goto out_unlock; | 2216 | goto out_unlock; |
2215 | } | 2217 | } |
2216 | 2218 | ||
2217 | di = btrfs_lookup_dir_item(trans, log, path, dir->i_ino, | 2219 | di = btrfs_lookup_dir_item(trans, log, path, dir_ino, |
2218 | name, name_len, -1); | 2220 | name, name_len, -1); |
2219 | if (IS_ERR(di)) { | 2221 | if (IS_ERR(di)) { |
2220 | err = PTR_ERR(di); | 2222 | err = PTR_ERR(di); |
@@ -2226,7 +2228,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2226 | BUG_ON(ret); | 2228 | BUG_ON(ret); |
2227 | } | 2229 | } |
2228 | btrfs_release_path(path); | 2230 | btrfs_release_path(path); |
2229 | di = btrfs_lookup_dir_index_item(trans, log, path, dir->i_ino, | 2231 | di = btrfs_lookup_dir_index_item(trans, log, path, dir_ino, |
2230 | index, name, name_len, -1); | 2232 | index, name, name_len, -1); |
2231 | if (IS_ERR(di)) { | 2233 | if (IS_ERR(di)) { |
2232 | err = PTR_ERR(di); | 2234 | err = PTR_ERR(di); |
@@ -2244,7 +2246,7 @@ int btrfs_del_dir_entries_in_log(struct btrfs_trans_handle *trans, | |||
2244 | if (bytes_del) { | 2246 | if (bytes_del) { |
2245 | struct btrfs_key key; | 2247 | struct btrfs_key key; |
2246 | 2248 | ||
2247 | key.objectid = dir->i_ino; | 2249 | key.objectid = dir_ino; |
2248 | key.offset = 0; | 2250 | key.offset = 0; |
2249 | key.type = BTRFS_INODE_ITEM_KEY; | 2251 | key.type = BTRFS_INODE_ITEM_KEY; |
2250 | btrfs_release_path(path); | 2252 | btrfs_release_path(path); |
@@ -2303,7 +2305,7 @@ int btrfs_del_inode_ref_in_log(struct btrfs_trans_handle *trans, | |||
2303 | log = root->log_root; | 2305 | log = root->log_root; |
2304 | mutex_lock(&BTRFS_I(inode)->log_mutex); | 2306 | mutex_lock(&BTRFS_I(inode)->log_mutex); |
2305 | 2307 | ||
2306 | ret = btrfs_del_inode_ref(trans, log, name, name_len, inode->i_ino, | 2308 | ret = btrfs_del_inode_ref(trans, log, name, name_len, btrfs_ino(inode), |
2307 | dirid, &index); | 2309 | dirid, &index); |
2308 | mutex_unlock(&BTRFS_I(inode)->log_mutex); | 2310 | mutex_unlock(&BTRFS_I(inode)->log_mutex); |
2309 | if (ret == -ENOSPC) { | 2311 | if (ret == -ENOSPC) { |
@@ -2369,13 +2371,14 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2369 | int nritems; | 2371 | int nritems; |
2370 | u64 first_offset = min_offset; | 2372 | u64 first_offset = min_offset; |
2371 | u64 last_offset = (u64)-1; | 2373 | u64 last_offset = (u64)-1; |
2374 | u64 ino = btrfs_ino(inode); | ||
2372 | 2375 | ||
2373 | log = root->log_root; | 2376 | log = root->log_root; |
2374 | max_key.objectid = inode->i_ino; | 2377 | max_key.objectid = ino; |
2375 | max_key.offset = (u64)-1; | 2378 | max_key.offset = (u64)-1; |
2376 | max_key.type = key_type; | 2379 | max_key.type = key_type; |
2377 | 2380 | ||
2378 | min_key.objectid = inode->i_ino; | 2381 | min_key.objectid = ino; |
2379 | min_key.type = key_type; | 2382 | min_key.type = key_type; |
2380 | min_key.offset = min_offset; | 2383 | min_key.offset = min_offset; |
2381 | 2384 | ||
@@ -2388,9 +2391,8 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2388 | * we didn't find anything from this transaction, see if there | 2391 | * we didn't find anything from this transaction, see if there |
2389 | * is anything at all | 2392 | * is anything at all |
2390 | */ | 2393 | */ |
2391 | if (ret != 0 || min_key.objectid != inode->i_ino || | 2394 | if (ret != 0 || min_key.objectid != ino || min_key.type != key_type) { |
2392 | min_key.type != key_type) { | 2395 | min_key.objectid = ino; |
2393 | min_key.objectid = inode->i_ino; | ||
2394 | min_key.type = key_type; | 2396 | min_key.type = key_type; |
2395 | min_key.offset = (u64)-1; | 2397 | min_key.offset = (u64)-1; |
2396 | btrfs_release_path(path); | 2398 | btrfs_release_path(path); |
@@ -2399,7 +2401,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2399 | btrfs_release_path(path); | 2401 | btrfs_release_path(path); |
2400 | return ret; | 2402 | return ret; |
2401 | } | 2403 | } |
2402 | ret = btrfs_previous_item(root, path, inode->i_ino, key_type); | 2404 | ret = btrfs_previous_item(root, path, ino, key_type); |
2403 | 2405 | ||
2404 | /* if ret == 0 there are items for this type, | 2406 | /* if ret == 0 there are items for this type, |
2405 | * create a range to tell us the last key of this type. | 2407 | * create a range to tell us the last key of this type. |
@@ -2417,7 +2419,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2417 | } | 2419 | } |
2418 | 2420 | ||
2419 | /* go backward to find any previous key */ | 2421 | /* go backward to find any previous key */ |
2420 | ret = btrfs_previous_item(root, path, inode->i_ino, key_type); | 2422 | ret = btrfs_previous_item(root, path, ino, key_type); |
2421 | if (ret == 0) { | 2423 | if (ret == 0) { |
2422 | struct btrfs_key tmp; | 2424 | struct btrfs_key tmp; |
2423 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); | 2425 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); |
@@ -2452,8 +2454,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2452 | for (i = path->slots[0]; i < nritems; i++) { | 2454 | for (i = path->slots[0]; i < nritems; i++) { |
2453 | btrfs_item_key_to_cpu(src, &min_key, i); | 2455 | btrfs_item_key_to_cpu(src, &min_key, i); |
2454 | 2456 | ||
2455 | if (min_key.objectid != inode->i_ino || | 2457 | if (min_key.objectid != ino || min_key.type != key_type) |
2456 | min_key.type != key_type) | ||
2457 | goto done; | 2458 | goto done; |
2458 | ret = overwrite_item(trans, log, dst_path, src, i, | 2459 | ret = overwrite_item(trans, log, dst_path, src, i, |
2459 | &min_key); | 2460 | &min_key); |
@@ -2474,7 +2475,7 @@ static noinline int log_dir_items(struct btrfs_trans_handle *trans, | |||
2474 | goto done; | 2475 | goto done; |
2475 | } | 2476 | } |
2476 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); | 2477 | btrfs_item_key_to_cpu(path->nodes[0], &tmp, path->slots[0]); |
2477 | if (tmp.objectid != inode->i_ino || tmp.type != key_type) { | 2478 | if (tmp.objectid != ino || tmp.type != key_type) { |
2478 | last_offset = (u64)-1; | 2479 | last_offset = (u64)-1; |
2479 | goto done; | 2480 | goto done; |
2480 | } | 2481 | } |
@@ -2500,8 +2501,7 @@ done: | |||
2500 | * is valid | 2501 | * is valid |
2501 | */ | 2502 | */ |
2502 | ret = insert_dir_log_key(trans, log, path, key_type, | 2503 | ret = insert_dir_log_key(trans, log, path, key_type, |
2503 | inode->i_ino, first_offset, | 2504 | ino, first_offset, last_offset); |
2504 | last_offset); | ||
2505 | if (ret) | 2505 | if (ret) |
2506 | err = ret; | 2506 | err = ret; |
2507 | } | 2507 | } |
@@ -2745,6 +2745,7 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2745 | int nritems; | 2745 | int nritems; |
2746 | int ins_start_slot = 0; | 2746 | int ins_start_slot = 0; |
2747 | int ins_nr; | 2747 | int ins_nr; |
2748 | u64 ino = btrfs_ino(inode); | ||
2748 | 2749 | ||
2749 | log = root->log_root; | 2750 | log = root->log_root; |
2750 | 2751 | ||
@@ -2757,11 +2758,11 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2757 | return -ENOMEM; | 2758 | return -ENOMEM; |
2758 | } | 2759 | } |
2759 | 2760 | ||
2760 | min_key.objectid = inode->i_ino; | 2761 | min_key.objectid = ino; |
2761 | min_key.type = BTRFS_INODE_ITEM_KEY; | 2762 | min_key.type = BTRFS_INODE_ITEM_KEY; |
2762 | min_key.offset = 0; | 2763 | min_key.offset = 0; |
2763 | 2764 | ||
2764 | max_key.objectid = inode->i_ino; | 2765 | max_key.objectid = ino; |
2765 | 2766 | ||
2766 | /* today the code can only do partial logging of directories */ | 2767 | /* today the code can only do partial logging of directories */ |
2767 | if (!S_ISDIR(inode->i_mode)) | 2768 | if (!S_ISDIR(inode->i_mode)) |
@@ -2773,6 +2774,13 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2773 | max_key.type = (u8)-1; | 2774 | max_key.type = (u8)-1; |
2774 | max_key.offset = (u64)-1; | 2775 | max_key.offset = (u64)-1; |
2775 | 2776 | ||
2777 | ret = btrfs_commit_inode_delayed_items(trans, inode); | ||
2778 | if (ret) { | ||
2779 | btrfs_free_path(path); | ||
2780 | btrfs_free_path(dst_path); | ||
2781 | return ret; | ||
2782 | } | ||
2783 | |||
2776 | mutex_lock(&BTRFS_I(inode)->log_mutex); | 2784 | mutex_lock(&BTRFS_I(inode)->log_mutex); |
2777 | 2785 | ||
2778 | /* | 2786 | /* |
@@ -2784,8 +2792,7 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2784 | 2792 | ||
2785 | if (inode_only == LOG_INODE_EXISTS) | 2793 | if (inode_only == LOG_INODE_EXISTS) |
2786 | max_key_type = BTRFS_XATTR_ITEM_KEY; | 2794 | max_key_type = BTRFS_XATTR_ITEM_KEY; |
2787 | ret = drop_objectid_items(trans, log, path, | 2795 | ret = drop_objectid_items(trans, log, path, ino, max_key_type); |
2788 | inode->i_ino, max_key_type); | ||
2789 | } else { | 2796 | } else { |
2790 | ret = btrfs_truncate_inode_items(trans, log, inode, 0, 0); | 2797 | ret = btrfs_truncate_inode_items(trans, log, inode, 0, 0); |
2791 | } | 2798 | } |
@@ -2803,7 +2810,7 @@ static int btrfs_log_inode(struct btrfs_trans_handle *trans, | |||
2803 | break; | 2810 | break; |
2804 | again: | 2811 | again: |
2805 | /* note, ins_nr might be > 0 here, cleanup outside the loop */ | 2812 | /* note, ins_nr might be > 0 here, cleanup outside the loop */ |
2806 | if (min_key.objectid != inode->i_ino) | 2813 | if (min_key.objectid != ino) |
2807 | break; | 2814 | break; |
2808 | if (min_key.type > max_key.type) | 2815 | if (min_key.type > max_key.type) |
2809 | break; | 2816 | break; |
diff --git a/fs/btrfs/xattr.c b/fs/btrfs/xattr.c index 4ca88d1e18e2..f3107e4b4d56 100644 --- a/fs/btrfs/xattr.c +++ b/fs/btrfs/xattr.c | |||
@@ -44,7 +44,7 @@ ssize_t __btrfs_getxattr(struct inode *inode, const char *name, | |||
44 | return -ENOMEM; | 44 | return -ENOMEM; |
45 | 45 | ||
46 | /* lookup the xattr by name */ | 46 | /* lookup the xattr by name */ |
47 | di = btrfs_lookup_xattr(NULL, root, path, inode->i_ino, name, | 47 | di = btrfs_lookup_xattr(NULL, root, path, btrfs_ino(inode), name, |
48 | strlen(name), 0); | 48 | strlen(name), 0); |
49 | if (!di) { | 49 | if (!di) { |
50 | ret = -ENODATA; | 50 | ret = -ENODATA; |
@@ -103,7 +103,7 @@ static int do_setxattr(struct btrfs_trans_handle *trans, | |||
103 | return -ENOMEM; | 103 | return -ENOMEM; |
104 | 104 | ||
105 | /* first lets see if we already have this xattr */ | 105 | /* first lets see if we already have this xattr */ |
106 | di = btrfs_lookup_xattr(trans, root, path, inode->i_ino, name, | 106 | di = btrfs_lookup_xattr(trans, root, path, btrfs_ino(inode), name, |
107 | strlen(name), -1); | 107 | strlen(name), -1); |
108 | if (IS_ERR(di)) { | 108 | if (IS_ERR(di)) { |
109 | ret = PTR_ERR(di); | 109 | ret = PTR_ERR(di); |
@@ -136,7 +136,7 @@ static int do_setxattr(struct btrfs_trans_handle *trans, | |||
136 | } | 136 | } |
137 | 137 | ||
138 | /* ok we have to create a completely new xattr */ | 138 | /* ok we have to create a completely new xattr */ |
139 | ret = btrfs_insert_xattr_item(trans, root, path, inode->i_ino, | 139 | ret = btrfs_insert_xattr_item(trans, root, path, btrfs_ino(inode), |
140 | name, name_len, value, size); | 140 | name, name_len, value, size); |
141 | BUG_ON(ret); | 141 | BUG_ON(ret); |
142 | out: | 142 | out: |
@@ -190,7 +190,7 @@ ssize_t btrfs_listxattr(struct dentry *dentry, char *buffer, size_t size) | |||
190 | * NOTE: we set key.offset = 0; because we want to start with the | 190 | * NOTE: we set key.offset = 0; because we want to start with the |
191 | * first xattr that we find and walk forward | 191 | * first xattr that we find and walk forward |
192 | */ | 192 | */ |
193 | key.objectid = inode->i_ino; | 193 | key.objectid = btrfs_ino(inode); |
194 | btrfs_set_key_type(&key, BTRFS_XATTR_ITEM_KEY); | 194 | btrfs_set_key_type(&key, BTRFS_XATTR_ITEM_KEY); |
195 | key.offset = 0; | 195 | key.offset = 0; |
196 | 196 | ||
diff --git a/fs/ceph/addr.c b/fs/ceph/addr.c index e159c529fd2b..38b8ab554924 100644 --- a/fs/ceph/addr.c +++ b/fs/ceph/addr.c | |||
@@ -775,6 +775,13 @@ get_more_pages: | |||
775 | ci->i_truncate_seq, | 775 | ci->i_truncate_seq, |
776 | ci->i_truncate_size, | 776 | ci->i_truncate_size, |
777 | &inode->i_mtime, true, 1, 0); | 777 | &inode->i_mtime, true, 1, 0); |
778 | |||
779 | if (!req) { | ||
780 | rc = -ENOMEM; | ||
781 | unlock_page(page); | ||
782 | break; | ||
783 | } | ||
784 | |||
778 | max_pages = req->r_num_pages; | 785 | max_pages = req->r_num_pages; |
779 | 786 | ||
780 | alloc_page_vec(fsc, req); | 787 | alloc_page_vec(fsc, req); |
diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c index 5323c330bbf3..2a5404c1c42f 100644 --- a/fs/ceph/caps.c +++ b/fs/ceph/caps.c | |||
@@ -819,7 +819,7 @@ int __ceph_caps_used(struct ceph_inode_info *ci) | |||
819 | used |= CEPH_CAP_FILE_CACHE; | 819 | used |= CEPH_CAP_FILE_CACHE; |
820 | if (ci->i_wr_ref) | 820 | if (ci->i_wr_ref) |
821 | used |= CEPH_CAP_FILE_WR; | 821 | used |= CEPH_CAP_FILE_WR; |
822 | if (ci->i_wrbuffer_ref) | 822 | if (ci->i_wb_ref || ci->i_wrbuffer_ref) |
823 | used |= CEPH_CAP_FILE_BUFFER; | 823 | used |= CEPH_CAP_FILE_BUFFER; |
824 | return used; | 824 | return used; |
825 | } | 825 | } |
@@ -1331,10 +1331,11 @@ static void ceph_flush_snaps(struct ceph_inode_info *ci) | |||
1331 | } | 1331 | } |
1332 | 1332 | ||
1333 | /* | 1333 | /* |
1334 | * Mark caps dirty. If inode is newly dirty, add to the global dirty | 1334 | * Mark caps dirty. If inode is newly dirty, return the dirty flags. |
1335 | * list. | 1335 | * Caller is then responsible for calling __mark_inode_dirty with the |
1336 | * returned flags value. | ||
1336 | */ | 1337 | */ |
1337 | void __ceph_mark_dirty_caps(struct ceph_inode_info *ci, int mask) | 1338 | int __ceph_mark_dirty_caps(struct ceph_inode_info *ci, int mask) |
1338 | { | 1339 | { |
1339 | struct ceph_mds_client *mdsc = | 1340 | struct ceph_mds_client *mdsc = |
1340 | ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc; | 1341 | ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc; |
@@ -1357,7 +1358,7 @@ void __ceph_mark_dirty_caps(struct ceph_inode_info *ci, int mask) | |||
1357 | list_add(&ci->i_dirty_item, &mdsc->cap_dirty); | 1358 | list_add(&ci->i_dirty_item, &mdsc->cap_dirty); |
1358 | spin_unlock(&mdsc->cap_dirty_lock); | 1359 | spin_unlock(&mdsc->cap_dirty_lock); |
1359 | if (ci->i_flushing_caps == 0) { | 1360 | if (ci->i_flushing_caps == 0) { |
1360 | igrab(inode); | 1361 | ihold(inode); |
1361 | dirty |= I_DIRTY_SYNC; | 1362 | dirty |= I_DIRTY_SYNC; |
1362 | } | 1363 | } |
1363 | } | 1364 | } |
@@ -1365,9 +1366,8 @@ void __ceph_mark_dirty_caps(struct ceph_inode_info *ci, int mask) | |||
1365 | if (((was | ci->i_flushing_caps) & CEPH_CAP_FILE_BUFFER) && | 1366 | if (((was | ci->i_flushing_caps) & CEPH_CAP_FILE_BUFFER) && |
1366 | (mask & CEPH_CAP_FILE_BUFFER)) | 1367 | (mask & CEPH_CAP_FILE_BUFFER)) |
1367 | dirty |= I_DIRTY_DATASYNC; | 1368 | dirty |= I_DIRTY_DATASYNC; |
1368 | if (dirty) | ||
1369 | __mark_inode_dirty(inode, dirty); | ||
1370 | __cap_delay_requeue(mdsc, ci); | 1369 | __cap_delay_requeue(mdsc, ci); |
1370 | return dirty; | ||
1371 | } | 1371 | } |
1372 | 1372 | ||
1373 | /* | 1373 | /* |
@@ -1990,11 +1990,11 @@ static void __take_cap_refs(struct ceph_inode_info *ci, int got) | |||
1990 | if (got & CEPH_CAP_FILE_WR) | 1990 | if (got & CEPH_CAP_FILE_WR) |
1991 | ci->i_wr_ref++; | 1991 | ci->i_wr_ref++; |
1992 | if (got & CEPH_CAP_FILE_BUFFER) { | 1992 | if (got & CEPH_CAP_FILE_BUFFER) { |
1993 | if (ci->i_wrbuffer_ref == 0) | 1993 | if (ci->i_wb_ref == 0) |
1994 | igrab(&ci->vfs_inode); | 1994 | ihold(&ci->vfs_inode); |
1995 | ci->i_wrbuffer_ref++; | 1995 | ci->i_wb_ref++; |
1996 | dout("__take_cap_refs %p wrbuffer %d -> %d (?)\n", | 1996 | dout("__take_cap_refs %p wb %d -> %d (?)\n", |
1997 | &ci->vfs_inode, ci->i_wrbuffer_ref-1, ci->i_wrbuffer_ref); | 1997 | &ci->vfs_inode, ci->i_wb_ref-1, ci->i_wb_ref); |
1998 | } | 1998 | } |
1999 | } | 1999 | } |
2000 | 2000 | ||
@@ -2169,12 +2169,12 @@ void ceph_put_cap_refs(struct ceph_inode_info *ci, int had) | |||
2169 | if (--ci->i_rdcache_ref == 0) | 2169 | if (--ci->i_rdcache_ref == 0) |
2170 | last++; | 2170 | last++; |
2171 | if (had & CEPH_CAP_FILE_BUFFER) { | 2171 | if (had & CEPH_CAP_FILE_BUFFER) { |
2172 | if (--ci->i_wrbuffer_ref == 0) { | 2172 | if (--ci->i_wb_ref == 0) { |
2173 | last++; | 2173 | last++; |
2174 | put++; | 2174 | put++; |
2175 | } | 2175 | } |
2176 | dout("put_cap_refs %p wrbuffer %d -> %d (?)\n", | 2176 | dout("put_cap_refs %p wb %d -> %d (?)\n", |
2177 | inode, ci->i_wrbuffer_ref+1, ci->i_wrbuffer_ref); | 2177 | inode, ci->i_wb_ref+1, ci->i_wb_ref); |
2178 | } | 2178 | } |
2179 | if (had & CEPH_CAP_FILE_WR) | 2179 | if (had & CEPH_CAP_FILE_WR) |
2180 | if (--ci->i_wr_ref == 0) { | 2180 | if (--ci->i_wr_ref == 0) { |
diff --git a/fs/ceph/file.c b/fs/ceph/file.c index 159b512d5a27..203252d88d9f 100644 --- a/fs/ceph/file.c +++ b/fs/ceph/file.c | |||
@@ -734,9 +734,12 @@ retry_snap: | |||
734 | } | 734 | } |
735 | } | 735 | } |
736 | if (ret >= 0) { | 736 | if (ret >= 0) { |
737 | int dirty; | ||
737 | spin_lock(&inode->i_lock); | 738 | spin_lock(&inode->i_lock); |
738 | __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR); | 739 | dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR); |
739 | spin_unlock(&inode->i_lock); | 740 | spin_unlock(&inode->i_lock); |
741 | if (dirty) | ||
742 | __mark_inode_dirty(inode, dirty); | ||
740 | } | 743 | } |
741 | 744 | ||
742 | out: | 745 | out: |
diff --git a/fs/ceph/inode.c b/fs/ceph/inode.c index b54c97da1c43..70b6a4839c38 100644 --- a/fs/ceph/inode.c +++ b/fs/ceph/inode.c | |||
@@ -355,6 +355,7 @@ struct inode *ceph_alloc_inode(struct super_block *sb) | |||
355 | ci->i_rd_ref = 0; | 355 | ci->i_rd_ref = 0; |
356 | ci->i_rdcache_ref = 0; | 356 | ci->i_rdcache_ref = 0; |
357 | ci->i_wr_ref = 0; | 357 | ci->i_wr_ref = 0; |
358 | ci->i_wb_ref = 0; | ||
358 | ci->i_wrbuffer_ref = 0; | 359 | ci->i_wrbuffer_ref = 0; |
359 | ci->i_wrbuffer_ref_head = 0; | 360 | ci->i_wrbuffer_ref_head = 0; |
360 | ci->i_shared_gen = 0; | 361 | ci->i_shared_gen = 0; |
@@ -1567,6 +1568,7 @@ int ceph_setattr(struct dentry *dentry, struct iattr *attr) | |||
1567 | int release = 0, dirtied = 0; | 1568 | int release = 0, dirtied = 0; |
1568 | int mask = 0; | 1569 | int mask = 0; |
1569 | int err = 0; | 1570 | int err = 0; |
1571 | int inode_dirty_flags = 0; | ||
1570 | 1572 | ||
1571 | if (ceph_snap(inode) != CEPH_NOSNAP) | 1573 | if (ceph_snap(inode) != CEPH_NOSNAP) |
1572 | return -EROFS; | 1574 | return -EROFS; |
@@ -1725,13 +1727,16 @@ int ceph_setattr(struct dentry *dentry, struct iattr *attr) | |||
1725 | dout("setattr %p ATTR_FILE ... hrm!\n", inode); | 1727 | dout("setattr %p ATTR_FILE ... hrm!\n", inode); |
1726 | 1728 | ||
1727 | if (dirtied) { | 1729 | if (dirtied) { |
1728 | __ceph_mark_dirty_caps(ci, dirtied); | 1730 | inode_dirty_flags = __ceph_mark_dirty_caps(ci, dirtied); |
1729 | inode->i_ctime = CURRENT_TIME; | 1731 | inode->i_ctime = CURRENT_TIME; |
1730 | } | 1732 | } |
1731 | 1733 | ||
1732 | release &= issued; | 1734 | release &= issued; |
1733 | spin_unlock(&inode->i_lock); | 1735 | spin_unlock(&inode->i_lock); |
1734 | 1736 | ||
1737 | if (inode_dirty_flags) | ||
1738 | __mark_inode_dirty(inode, inode_dirty_flags); | ||
1739 | |||
1735 | if (mask) { | 1740 | if (mask) { |
1736 | req->r_inode = igrab(inode); | 1741 | req->r_inode = igrab(inode); |
1737 | req->r_inode_drop = release; | 1742 | req->r_inode_drop = release; |
diff --git a/fs/ceph/mds_client.c b/fs/ceph/mds_client.c index f60b07b0feb0..d0fae4ce9ba5 100644 --- a/fs/ceph/mds_client.c +++ b/fs/ceph/mds_client.c | |||
@@ -3304,8 +3304,8 @@ static void con_put(struct ceph_connection *con) | |||
3304 | { | 3304 | { |
3305 | struct ceph_mds_session *s = con->private; | 3305 | struct ceph_mds_session *s = con->private; |
3306 | 3306 | ||
3307 | dout("mdsc con_put %p (%d)\n", s, atomic_read(&s->s_ref) - 1); | ||
3307 | ceph_put_mds_session(s); | 3308 | ceph_put_mds_session(s); |
3308 | dout("mdsc con_put %p (%d)\n", s, atomic_read(&s->s_ref)); | ||
3309 | } | 3309 | } |
3310 | 3310 | ||
3311 | /* | 3311 | /* |
diff --git a/fs/ceph/snap.c b/fs/ceph/snap.c index e86ec1155f8f..24067d68a554 100644 --- a/fs/ceph/snap.c +++ b/fs/ceph/snap.c | |||
@@ -206,7 +206,7 @@ void ceph_put_snap_realm(struct ceph_mds_client *mdsc, | |||
206 | up_write(&mdsc->snap_rwsem); | 206 | up_write(&mdsc->snap_rwsem); |
207 | } else { | 207 | } else { |
208 | spin_lock(&mdsc->snap_empty_lock); | 208 | spin_lock(&mdsc->snap_empty_lock); |
209 | list_add(&mdsc->snap_empty, &realm->empty_item); | 209 | list_add(&realm->empty_item, &mdsc->snap_empty); |
210 | spin_unlock(&mdsc->snap_empty_lock); | 210 | spin_unlock(&mdsc->snap_empty_lock); |
211 | } | 211 | } |
212 | } | 212 | } |
diff --git a/fs/ceph/super.h b/fs/ceph/super.h index 619fe719968f..f5cabefa98dc 100644 --- a/fs/ceph/super.h +++ b/fs/ceph/super.h | |||
@@ -293,7 +293,7 @@ struct ceph_inode_info { | |||
293 | 293 | ||
294 | /* held references to caps */ | 294 | /* held references to caps */ |
295 | int i_pin_ref; | 295 | int i_pin_ref; |
296 | int i_rd_ref, i_rdcache_ref, i_wr_ref; | 296 | int i_rd_ref, i_rdcache_ref, i_wr_ref, i_wb_ref; |
297 | int i_wrbuffer_ref, i_wrbuffer_ref_head; | 297 | int i_wrbuffer_ref, i_wrbuffer_ref_head; |
298 | u32 i_shared_gen; /* increment each time we get FILE_SHARED */ | 298 | u32 i_shared_gen; /* increment each time we get FILE_SHARED */ |
299 | u32 i_rdcache_gen; /* incremented each time we get FILE_CACHE. */ | 299 | u32 i_rdcache_gen; /* incremented each time we get FILE_CACHE. */ |
@@ -506,7 +506,7 @@ static inline int __ceph_caps_dirty(struct ceph_inode_info *ci) | |||
506 | { | 506 | { |
507 | return ci->i_dirty_caps | ci->i_flushing_caps; | 507 | return ci->i_dirty_caps | ci->i_flushing_caps; |
508 | } | 508 | } |
509 | extern void __ceph_mark_dirty_caps(struct ceph_inode_info *ci, int mask); | 509 | extern int __ceph_mark_dirty_caps(struct ceph_inode_info *ci, int mask); |
510 | 510 | ||
511 | extern int ceph_caps_revoking(struct ceph_inode_info *ci, int mask); | 511 | extern int ceph_caps_revoking(struct ceph_inode_info *ci, int mask); |
512 | extern int __ceph_caps_used(struct ceph_inode_info *ci); | 512 | extern int __ceph_caps_used(struct ceph_inode_info *ci); |
diff --git a/fs/ceph/xattr.c b/fs/ceph/xattr.c index 8c9eba6ef9df..f2b628696180 100644 --- a/fs/ceph/xattr.c +++ b/fs/ceph/xattr.c | |||
@@ -703,6 +703,7 @@ int ceph_setxattr(struct dentry *dentry, const char *name, | |||
703 | struct ceph_inode_xattr *xattr = NULL; | 703 | struct ceph_inode_xattr *xattr = NULL; |
704 | int issued; | 704 | int issued; |
705 | int required_blob_size; | 705 | int required_blob_size; |
706 | int dirty; | ||
706 | 707 | ||
707 | if (ceph_snap(inode) != CEPH_NOSNAP) | 708 | if (ceph_snap(inode) != CEPH_NOSNAP) |
708 | return -EROFS; | 709 | return -EROFS; |
@@ -763,11 +764,12 @@ retry: | |||
763 | dout("setxattr %p issued %s\n", inode, ceph_cap_string(issued)); | 764 | dout("setxattr %p issued %s\n", inode, ceph_cap_string(issued)); |
764 | err = __set_xattr(ci, newname, name_len, newval, | 765 | err = __set_xattr(ci, newname, name_len, newval, |
765 | val_len, 1, 1, 1, &xattr); | 766 | val_len, 1, 1, 1, &xattr); |
766 | __ceph_mark_dirty_caps(ci, CEPH_CAP_XATTR_EXCL); | 767 | dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_XATTR_EXCL); |
767 | ci->i_xattrs.dirty = true; | 768 | ci->i_xattrs.dirty = true; |
768 | inode->i_ctime = CURRENT_TIME; | 769 | inode->i_ctime = CURRENT_TIME; |
769 | spin_unlock(&inode->i_lock); | 770 | spin_unlock(&inode->i_lock); |
770 | 771 | if (dirty) | |
772 | __mark_inode_dirty(inode, dirty); | ||
771 | return err; | 773 | return err; |
772 | 774 | ||
773 | do_sync: | 775 | do_sync: |
@@ -810,6 +812,7 @@ int ceph_removexattr(struct dentry *dentry, const char *name) | |||
810 | struct ceph_vxattr_cb *vxattrs = ceph_inode_vxattrs(inode); | 812 | struct ceph_vxattr_cb *vxattrs = ceph_inode_vxattrs(inode); |
811 | int issued; | 813 | int issued; |
812 | int err; | 814 | int err; |
815 | int dirty; | ||
813 | 816 | ||
814 | if (ceph_snap(inode) != CEPH_NOSNAP) | 817 | if (ceph_snap(inode) != CEPH_NOSNAP) |
815 | return -EROFS; | 818 | return -EROFS; |
@@ -833,12 +836,13 @@ int ceph_removexattr(struct dentry *dentry, const char *name) | |||
833 | goto do_sync; | 836 | goto do_sync; |
834 | 837 | ||
835 | err = __remove_xattr_by_name(ceph_inode(inode), name); | 838 | err = __remove_xattr_by_name(ceph_inode(inode), name); |
836 | __ceph_mark_dirty_caps(ci, CEPH_CAP_XATTR_EXCL); | 839 | dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_XATTR_EXCL); |
837 | ci->i_xattrs.dirty = true; | 840 | ci->i_xattrs.dirty = true; |
838 | inode->i_ctime = CURRENT_TIME; | 841 | inode->i_ctime = CURRENT_TIME; |
839 | 842 | ||
840 | spin_unlock(&inode->i_lock); | 843 | spin_unlock(&inode->i_lock); |
841 | 844 | if (dirty) | |
845 | __mark_inode_dirty(inode, dirty); | ||
842 | return err; | 846 | return err; |
843 | do_sync: | 847 | do_sync: |
844 | spin_unlock(&inode->i_lock); | 848 | spin_unlock(&inode->i_lock); |
diff --git a/fs/cifs/cifs_unicode.c b/fs/cifs/cifs_unicode.c index 23d43cde4306..1b2e180b018d 100644 --- a/fs/cifs/cifs_unicode.c +++ b/fs/cifs/cifs_unicode.c | |||
@@ -277,6 +277,7 @@ cifsConvertToUCS(__le16 *target, const char *source, int srclen, | |||
277 | 277 | ||
278 | for (i = 0, j = 0; i < srclen; j++) { | 278 | for (i = 0, j = 0; i < srclen; j++) { |
279 | src_char = source[i]; | 279 | src_char = source[i]; |
280 | charlen = 1; | ||
280 | switch (src_char) { | 281 | switch (src_char) { |
281 | case 0: | 282 | case 0: |
282 | put_unaligned(0, &target[j]); | 283 | put_unaligned(0, &target[j]); |
@@ -316,16 +317,13 @@ cifsConvertToUCS(__le16 *target, const char *source, int srclen, | |||
316 | dst_char = cpu_to_le16(0x003f); | 317 | dst_char = cpu_to_le16(0x003f); |
317 | charlen = 1; | 318 | charlen = 1; |
318 | } | 319 | } |
319 | /* | ||
320 | * character may take more than one byte in the source | ||
321 | * string, but will take exactly two bytes in the | ||
322 | * target string | ||
323 | */ | ||
324 | i += charlen; | ||
325 | continue; | ||
326 | } | 320 | } |
321 | /* | ||
322 | * character may take more than one byte in the source string, | ||
323 | * but will take exactly two bytes in the target string | ||
324 | */ | ||
325 | i += charlen; | ||
327 | put_unaligned(dst_char, &target[j]); | 326 | put_unaligned(dst_char, &target[j]); |
328 | i++; /* move to next char in source string */ | ||
329 | } | 327 | } |
330 | 328 | ||
331 | ctoUCS_out: | 329 | ctoUCS_out: |
diff --git a/fs/cifs/connect.c b/fs/cifs/connect.c index 4bc862a80efa..277262a8e82f 100644 --- a/fs/cifs/connect.c +++ b/fs/cifs/connect.c | |||
@@ -274,7 +274,8 @@ static int coalesce_t2(struct smb_hdr *psecond, struct smb_hdr *pTargetSMB) | |||
274 | char *data_area_of_target; | 274 | char *data_area_of_target; |
275 | char *data_area_of_buf2; | 275 | char *data_area_of_buf2; |
276 | int remaining; | 276 | int remaining; |
277 | __u16 byte_count, total_data_size, total_in_buf, total_in_buf2; | 277 | unsigned int byte_count, total_in_buf; |
278 | __u16 total_data_size, total_in_buf2; | ||
278 | 279 | ||
279 | total_data_size = get_unaligned_le16(&pSMBt->t2_rsp.TotalDataCount); | 280 | total_data_size = get_unaligned_le16(&pSMBt->t2_rsp.TotalDataCount); |
280 | 281 | ||
@@ -287,7 +288,7 @@ static int coalesce_t2(struct smb_hdr *psecond, struct smb_hdr *pTargetSMB) | |||
287 | remaining = total_data_size - total_in_buf; | 288 | remaining = total_data_size - total_in_buf; |
288 | 289 | ||
289 | if (remaining < 0) | 290 | if (remaining < 0) |
290 | return -EINVAL; | 291 | return -EPROTO; |
291 | 292 | ||
292 | if (remaining == 0) /* nothing to do, ignore */ | 293 | if (remaining == 0) /* nothing to do, ignore */ |
293 | return 0; | 294 | return 0; |
@@ -308,20 +309,29 @@ static int coalesce_t2(struct smb_hdr *psecond, struct smb_hdr *pTargetSMB) | |||
308 | data_area_of_target += total_in_buf; | 309 | data_area_of_target += total_in_buf; |
309 | 310 | ||
310 | /* copy second buffer into end of first buffer */ | 311 | /* copy second buffer into end of first buffer */ |
311 | memcpy(data_area_of_target, data_area_of_buf2, total_in_buf2); | ||
312 | total_in_buf += total_in_buf2; | 312 | total_in_buf += total_in_buf2; |
313 | /* is the result too big for the field? */ | ||
314 | if (total_in_buf > USHRT_MAX) | ||
315 | return -EPROTO; | ||
313 | put_unaligned_le16(total_in_buf, &pSMBt->t2_rsp.DataCount); | 316 | put_unaligned_le16(total_in_buf, &pSMBt->t2_rsp.DataCount); |
317 | |||
318 | /* fix up the BCC */ | ||
314 | byte_count = get_bcc_le(pTargetSMB); | 319 | byte_count = get_bcc_le(pTargetSMB); |
315 | byte_count += total_in_buf2; | 320 | byte_count += total_in_buf2; |
321 | /* is the result too big for the field? */ | ||
322 | if (byte_count > USHRT_MAX) | ||
323 | return -EPROTO; | ||
316 | put_bcc_le(byte_count, pTargetSMB); | 324 | put_bcc_le(byte_count, pTargetSMB); |
317 | 325 | ||
318 | byte_count = pTargetSMB->smb_buf_length; | 326 | byte_count = pTargetSMB->smb_buf_length; |
319 | byte_count += total_in_buf2; | 327 | byte_count += total_in_buf2; |
320 | 328 | /* don't allow buffer to overflow */ | |
321 | /* BB also add check that we are not beyond maximum buffer size */ | 329 | if (byte_count > CIFSMaxBufSize) |
322 | 330 | return -ENOBUFS; | |
323 | pTargetSMB->smb_buf_length = byte_count; | 331 | pTargetSMB->smb_buf_length = byte_count; |
324 | 332 | ||
333 | memcpy(data_area_of_target, data_area_of_buf2, total_in_buf2); | ||
334 | |||
325 | if (remaining == total_in_buf2) { | 335 | if (remaining == total_in_buf2) { |
326 | cFYI(1, "found the last secondary response"); | 336 | cFYI(1, "found the last secondary response"); |
327 | return 0; /* we are done */ | 337 | return 0; /* we are done */ |
@@ -607,59 +617,63 @@ incomplete_rcv: | |||
607 | list_for_each_safe(tmp, tmp2, &server->pending_mid_q) { | 617 | list_for_each_safe(tmp, tmp2, &server->pending_mid_q) { |
608 | mid_entry = list_entry(tmp, struct mid_q_entry, qhead); | 618 | mid_entry = list_entry(tmp, struct mid_q_entry, qhead); |
609 | 619 | ||
610 | if ((mid_entry->mid == smb_buffer->Mid) && | 620 | if (mid_entry->mid != smb_buffer->Mid || |
611 | (mid_entry->midState == MID_REQUEST_SUBMITTED) && | 621 | mid_entry->midState != MID_REQUEST_SUBMITTED || |
612 | (mid_entry->command == smb_buffer->Command)) { | 622 | mid_entry->command != smb_buffer->Command) { |
613 | if (length == 0 && | 623 | mid_entry = NULL; |
614 | check2ndT2(smb_buffer, server->maxBuf) > 0) { | 624 | continue; |
615 | /* We have a multipart transact2 resp */ | 625 | } |
616 | isMultiRsp = true; | 626 | |
617 | if (mid_entry->resp_buf) { | 627 | if (length == 0 && |
618 | /* merge response - fix up 1st*/ | 628 | check2ndT2(smb_buffer, server->maxBuf) > 0) { |
619 | if (coalesce_t2(smb_buffer, | 629 | /* We have a multipart transact2 resp */ |
620 | mid_entry->resp_buf)) { | 630 | isMultiRsp = true; |
621 | mid_entry->multiRsp = | 631 | if (mid_entry->resp_buf) { |
622 | true; | 632 | /* merge response - fix up 1st*/ |
623 | break; | 633 | length = coalesce_t2(smb_buffer, |
624 | } else { | 634 | mid_entry->resp_buf); |
625 | /* all parts received */ | 635 | if (length > 0) { |
626 | mid_entry->multiEnd = | 636 | length = 0; |
627 | true; | 637 | mid_entry->multiRsp = true; |
628 | goto multi_t2_fnd; | 638 | break; |
629 | } | ||
630 | } else { | 639 | } else { |
631 | if (!isLargeBuf) { | 640 | /* all parts received or |
632 | cERROR(1, "1st trans2 resp needs bigbuf"); | 641 | * packet is malformed |
633 | /* BB maybe we can fix this up, switch | 642 | */ |
634 | to already allocated large buffer? */ | 643 | mid_entry->multiEnd = true; |
635 | } else { | 644 | goto multi_t2_fnd; |
636 | /* Have first buffer */ | 645 | } |
637 | mid_entry->resp_buf = | 646 | } else { |
638 | smb_buffer; | 647 | if (!isLargeBuf) { |
639 | mid_entry->largeBuf = | 648 | /* |
640 | true; | 649 | * FIXME: switch to already |
641 | bigbuf = NULL; | 650 | * allocated largebuf? |
642 | } | 651 | */ |
652 | cERROR(1, "1st trans2 resp " | ||
653 | "needs bigbuf"); | ||
654 | } else { | ||
655 | /* Have first buffer */ | ||
656 | mid_entry->resp_buf = | ||
657 | smb_buffer; | ||
658 | mid_entry->largeBuf = true; | ||
659 | bigbuf = NULL; | ||
643 | } | 660 | } |
644 | break; | ||
645 | } | 661 | } |
646 | mid_entry->resp_buf = smb_buffer; | 662 | break; |
647 | mid_entry->largeBuf = isLargeBuf; | 663 | } |
664 | mid_entry->resp_buf = smb_buffer; | ||
665 | mid_entry->largeBuf = isLargeBuf; | ||
648 | multi_t2_fnd: | 666 | multi_t2_fnd: |
649 | if (length == 0) | 667 | if (length == 0) |
650 | mid_entry->midState = | 668 | mid_entry->midState = MID_RESPONSE_RECEIVED; |
651 | MID_RESPONSE_RECEIVED; | 669 | else |
652 | else | 670 | mid_entry->midState = MID_RESPONSE_MALFORMED; |
653 | mid_entry->midState = | ||
654 | MID_RESPONSE_MALFORMED; | ||
655 | #ifdef CONFIG_CIFS_STATS2 | 671 | #ifdef CONFIG_CIFS_STATS2 |
656 | mid_entry->when_received = jiffies; | 672 | mid_entry->when_received = jiffies; |
657 | #endif | 673 | #endif |
658 | list_del_init(&mid_entry->qhead); | 674 | list_del_init(&mid_entry->qhead); |
659 | mid_entry->callback(mid_entry); | 675 | mid_entry->callback(mid_entry); |
660 | break; | 676 | break; |
661 | } | ||
662 | mid_entry = NULL; | ||
663 | } | 677 | } |
664 | spin_unlock(&GlobalMid_Lock); | 678 | spin_unlock(&GlobalMid_Lock); |
665 | 679 | ||
@@ -2659,6 +2673,11 @@ is_path_accessible(int xid, struct cifsTconInfo *tcon, | |||
2659 | 0 /* not legacy */, cifs_sb->local_nls, | 2673 | 0 /* not legacy */, cifs_sb->local_nls, |
2660 | cifs_sb->mnt_cifs_flags & | 2674 | cifs_sb->mnt_cifs_flags & |
2661 | CIFS_MOUNT_MAP_SPECIAL_CHR); | 2675 | CIFS_MOUNT_MAP_SPECIAL_CHR); |
2676 | |||
2677 | if (rc == -EOPNOTSUPP || rc == -EINVAL) | ||
2678 | rc = SMBQueryInformation(xid, tcon, full_path, pfile_info, | ||
2679 | cifs_sb->local_nls, cifs_sb->mnt_cifs_flags & | ||
2680 | CIFS_MOUNT_MAP_SPECIAL_CHR); | ||
2662 | kfree(pfile_info); | 2681 | kfree(pfile_info); |
2663 | return rc; | 2682 | return rc; |
2664 | } | 2683 | } |
diff --git a/fs/cifs/sess.c b/fs/cifs/sess.c index f6728eb6f4b9..645114ad0a10 100644 --- a/fs/cifs/sess.c +++ b/fs/cifs/sess.c | |||
@@ -276,7 +276,7 @@ static void ascii_ssetup_strings(char **pbcc_area, struct cifsSesInfo *ses, | |||
276 | } | 276 | } |
277 | 277 | ||
278 | static void | 278 | static void |
279 | decode_unicode_ssetup(char **pbcc_area, __u16 bleft, struct cifsSesInfo *ses, | 279 | decode_unicode_ssetup(char **pbcc_area, int bleft, struct cifsSesInfo *ses, |
280 | const struct nls_table *nls_cp) | 280 | const struct nls_table *nls_cp) |
281 | { | 281 | { |
282 | int len; | 282 | int len; |
@@ -284,19 +284,6 @@ decode_unicode_ssetup(char **pbcc_area, __u16 bleft, struct cifsSesInfo *ses, | |||
284 | 284 | ||
285 | cFYI(1, "bleft %d", bleft); | 285 | cFYI(1, "bleft %d", bleft); |
286 | 286 | ||
287 | /* | ||
288 | * Windows servers do not always double null terminate their final | ||
289 | * Unicode string. Check to see if there are an uneven number of bytes | ||
290 | * left. If so, then add an extra NULL pad byte to the end of the | ||
291 | * response. | ||
292 | * | ||
293 | * See section 2.7.2 in "Implementing CIFS" for details | ||
294 | */ | ||
295 | if (bleft % 2) { | ||
296 | data[bleft] = 0; | ||
297 | ++bleft; | ||
298 | } | ||
299 | |||
300 | kfree(ses->serverOS); | 287 | kfree(ses->serverOS); |
301 | ses->serverOS = cifs_strndup_from_ucs(data, bleft, true, nls_cp); | 288 | ses->serverOS = cifs_strndup_from_ucs(data, bleft, true, nls_cp); |
302 | cFYI(1, "serverOS=%s", ses->serverOS); | 289 | cFYI(1, "serverOS=%s", ses->serverOS); |
@@ -929,7 +916,9 @@ ssetup_ntlmssp_authenticate: | |||
929 | } | 916 | } |
930 | 917 | ||
931 | /* BB check if Unicode and decode strings */ | 918 | /* BB check if Unicode and decode strings */ |
932 | if (smb_buf->Flags2 & SMBFLG2_UNICODE) { | 919 | if (bytes_remaining == 0) { |
920 | /* no string area to decode, do nothing */ | ||
921 | } else if (smb_buf->Flags2 & SMBFLG2_UNICODE) { | ||
933 | /* unicode string area must be word-aligned */ | 922 | /* unicode string area must be word-aligned */ |
934 | if (((unsigned long) bcc_ptr - (unsigned long) smb_buf) % 2) { | 923 | if (((unsigned long) bcc_ptr - (unsigned long) smb_buf) % 2) { |
935 | ++bcc_ptr; | 924 | ++bcc_ptr; |
diff --git a/fs/configfs/dir.c b/fs/configfs/dir.c index 3313dd19f543..9a37a9b6de3a 100644 --- a/fs/configfs/dir.c +++ b/fs/configfs/dir.c | |||
@@ -53,11 +53,14 @@ DEFINE_SPINLOCK(configfs_dirent_lock); | |||
53 | static void configfs_d_iput(struct dentry * dentry, | 53 | static void configfs_d_iput(struct dentry * dentry, |
54 | struct inode * inode) | 54 | struct inode * inode) |
55 | { | 55 | { |
56 | struct configfs_dirent * sd = dentry->d_fsdata; | 56 | struct configfs_dirent *sd = dentry->d_fsdata; |
57 | 57 | ||
58 | if (sd) { | 58 | if (sd) { |
59 | BUG_ON(sd->s_dentry != dentry); | 59 | BUG_ON(sd->s_dentry != dentry); |
60 | /* Coordinate with configfs_readdir */ | ||
61 | spin_lock(&configfs_dirent_lock); | ||
60 | sd->s_dentry = NULL; | 62 | sd->s_dentry = NULL; |
63 | spin_unlock(&configfs_dirent_lock); | ||
61 | configfs_put(sd); | 64 | configfs_put(sd); |
62 | } | 65 | } |
63 | iput(inode); | 66 | iput(inode); |
@@ -689,7 +692,8 @@ static int create_default_group(struct config_group *parent_group, | |||
689 | sd = child->d_fsdata; | 692 | sd = child->d_fsdata; |
690 | sd->s_type |= CONFIGFS_USET_DEFAULT; | 693 | sd->s_type |= CONFIGFS_USET_DEFAULT; |
691 | } else { | 694 | } else { |
692 | d_delete(child); | 695 | BUG_ON(child->d_inode); |
696 | d_drop(child); | ||
693 | dput(child); | 697 | dput(child); |
694 | } | 698 | } |
695 | } | 699 | } |
@@ -1545,7 +1549,7 @@ static int configfs_readdir(struct file * filp, void * dirent, filldir_t filldir | |||
1545 | struct configfs_dirent * parent_sd = dentry->d_fsdata; | 1549 | struct configfs_dirent * parent_sd = dentry->d_fsdata; |
1546 | struct configfs_dirent *cursor = filp->private_data; | 1550 | struct configfs_dirent *cursor = filp->private_data; |
1547 | struct list_head *p, *q = &cursor->s_sibling; | 1551 | struct list_head *p, *q = &cursor->s_sibling; |
1548 | ino_t ino; | 1552 | ino_t ino = 0; |
1549 | int i = filp->f_pos; | 1553 | int i = filp->f_pos; |
1550 | 1554 | ||
1551 | switch (i) { | 1555 | switch (i) { |
@@ -1573,6 +1577,7 @@ static int configfs_readdir(struct file * filp, void * dirent, filldir_t filldir | |||
1573 | struct configfs_dirent *next; | 1577 | struct configfs_dirent *next; |
1574 | const char * name; | 1578 | const char * name; |
1575 | int len; | 1579 | int len; |
1580 | struct inode *inode = NULL; | ||
1576 | 1581 | ||
1577 | next = list_entry(p, struct configfs_dirent, | 1582 | next = list_entry(p, struct configfs_dirent, |
1578 | s_sibling); | 1583 | s_sibling); |
@@ -1581,9 +1586,28 @@ static int configfs_readdir(struct file * filp, void * dirent, filldir_t filldir | |||
1581 | 1586 | ||
1582 | name = configfs_get_name(next); | 1587 | name = configfs_get_name(next); |
1583 | len = strlen(name); | 1588 | len = strlen(name); |
1584 | if (next->s_dentry) | 1589 | |
1585 | ino = next->s_dentry->d_inode->i_ino; | 1590 | /* |
1586 | else | 1591 | * We'll have a dentry and an inode for |
1592 | * PINNED items and for open attribute | ||
1593 | * files. We lock here to prevent a race | ||
1594 | * with configfs_d_iput() clearing | ||
1595 | * s_dentry before calling iput(). | ||
1596 | * | ||
1597 | * Why do we go to the trouble? If | ||
1598 | * someone has an attribute file open, | ||
1599 | * the inode number should match until | ||
1600 | * they close it. Beyond that, we don't | ||
1601 | * care. | ||
1602 | */ | ||
1603 | spin_lock(&configfs_dirent_lock); | ||
1604 | dentry = next->s_dentry; | ||
1605 | if (dentry) | ||
1606 | inode = dentry->d_inode; | ||
1607 | if (inode) | ||
1608 | ino = inode->i_ino; | ||
1609 | spin_unlock(&configfs_dirent_lock); | ||
1610 | if (!inode) | ||
1587 | ino = iunique(configfs_sb, 2); | 1611 | ino = iunique(configfs_sb, 2); |
1588 | 1612 | ||
1589 | if (filldir(dirent, name, len, filp->f_pos, ino, | 1613 | if (filldir(dirent, name, len, filp->f_pos, ino, |
@@ -1683,7 +1707,8 @@ int configfs_register_subsystem(struct configfs_subsystem *subsys) | |||
1683 | err = configfs_attach_group(sd->s_element, &group->cg_item, | 1707 | err = configfs_attach_group(sd->s_element, &group->cg_item, |
1684 | dentry); | 1708 | dentry); |
1685 | if (err) { | 1709 | if (err) { |
1686 | d_delete(dentry); | 1710 | BUG_ON(dentry->d_inode); |
1711 | d_drop(dentry); | ||
1687 | dput(dentry); | 1712 | dput(dentry); |
1688 | } else { | 1713 | } else { |
1689 | spin_lock(&configfs_dirent_lock); | 1714 | spin_lock(&configfs_dirent_lock); |
diff --git a/fs/fuse/dir.c b/fs/fuse/dir.c index c6ba49bd95b3..b32eb29a4e6f 100644 --- a/fs/fuse/dir.c +++ b/fs/fuse/dir.c | |||
@@ -174,7 +174,7 @@ static int fuse_dentry_revalidate(struct dentry *entry, struct nameidata *nd) | |||
174 | if (!inode) | 174 | if (!inode) |
175 | return 0; | 175 | return 0; |
176 | 176 | ||
177 | if (nd->flags & LOOKUP_RCU) | 177 | if (nd && (nd->flags & LOOKUP_RCU)) |
178 | return -ECHILD; | 178 | return -ECHILD; |
179 | 179 | ||
180 | fc = get_fuse_conn(inode); | 180 | fc = get_fuse_conn(inode); |
diff --git a/fs/hpfs/Kconfig b/fs/hpfs/Kconfig index 0c39dc3ef7d7..56bd15c5bf6c 100644 --- a/fs/hpfs/Kconfig +++ b/fs/hpfs/Kconfig | |||
@@ -1,7 +1,6 @@ | |||
1 | config HPFS_FS | 1 | config HPFS_FS |
2 | tristate "OS/2 HPFS file system support" | 2 | tristate "OS/2 HPFS file system support" |
3 | depends on BLOCK | 3 | depends on BLOCK |
4 | depends on BROKEN || !PREEMPT | ||
5 | help | 4 | help |
6 | OS/2 is IBM's operating system for PC's, the same as Warp, and HPFS | 5 | OS/2 is IBM's operating system for PC's, the same as Warp, and HPFS |
7 | is the file system used for organizing files on OS/2 hard disk | 6 | is the file system used for organizing files on OS/2 hard disk |
diff --git a/fs/hpfs/alloc.c b/fs/hpfs/alloc.c index 5503e2c28910..7a5eb2c718c8 100644 --- a/fs/hpfs/alloc.c +++ b/fs/hpfs/alloc.c | |||
@@ -8,8 +8,6 @@ | |||
8 | 8 | ||
9 | #include "hpfs_fn.h" | 9 | #include "hpfs_fn.h" |
10 | 10 | ||
11 | static int hpfs_alloc_if_possible_nolock(struct super_block *s, secno sec); | ||
12 | |||
13 | /* | 11 | /* |
14 | * Check if a sector is allocated in bitmap | 12 | * Check if a sector is allocated in bitmap |
15 | * This is really slow. Turned on only if chk==2 | 13 | * This is really slow. Turned on only if chk==2 |
@@ -18,9 +16,9 @@ static int hpfs_alloc_if_possible_nolock(struct super_block *s, secno sec); | |||
18 | static int chk_if_allocated(struct super_block *s, secno sec, char *msg) | 16 | static int chk_if_allocated(struct super_block *s, secno sec, char *msg) |
19 | { | 17 | { |
20 | struct quad_buffer_head qbh; | 18 | struct quad_buffer_head qbh; |
21 | unsigned *bmp; | 19 | u32 *bmp; |
22 | if (!(bmp = hpfs_map_bitmap(s, sec >> 14, &qbh, "chk"))) goto fail; | 20 | if (!(bmp = hpfs_map_bitmap(s, sec >> 14, &qbh, "chk"))) goto fail; |
23 | if ((bmp[(sec & 0x3fff) >> 5] >> (sec & 0x1f)) & 1) { | 21 | if ((cpu_to_le32(bmp[(sec & 0x3fff) >> 5]) >> (sec & 0x1f)) & 1) { |
24 | hpfs_error(s, "sector '%s' - %08x not allocated in bitmap", msg, sec); | 22 | hpfs_error(s, "sector '%s' - %08x not allocated in bitmap", msg, sec); |
25 | goto fail1; | 23 | goto fail1; |
26 | } | 24 | } |
@@ -28,7 +26,7 @@ static int chk_if_allocated(struct super_block *s, secno sec, char *msg) | |||
28 | if (sec >= hpfs_sb(s)->sb_dirband_start && sec < hpfs_sb(s)->sb_dirband_start + hpfs_sb(s)->sb_dirband_size) { | 26 | if (sec >= hpfs_sb(s)->sb_dirband_start && sec < hpfs_sb(s)->sb_dirband_start + hpfs_sb(s)->sb_dirband_size) { |
29 | unsigned ssec = (sec - hpfs_sb(s)->sb_dirband_start) / 4; | 27 | unsigned ssec = (sec - hpfs_sb(s)->sb_dirband_start) / 4; |
30 | if (!(bmp = hpfs_map_dnode_bitmap(s, &qbh))) goto fail; | 28 | if (!(bmp = hpfs_map_dnode_bitmap(s, &qbh))) goto fail; |
31 | if ((bmp[ssec >> 5] >> (ssec & 0x1f)) & 1) { | 29 | if ((le32_to_cpu(bmp[ssec >> 5]) >> (ssec & 0x1f)) & 1) { |
32 | hpfs_error(s, "sector '%s' - %08x not allocated in directory bitmap", msg, sec); | 30 | hpfs_error(s, "sector '%s' - %08x not allocated in directory bitmap", msg, sec); |
33 | goto fail1; | 31 | goto fail1; |
34 | } | 32 | } |
@@ -75,7 +73,6 @@ static secno alloc_in_bmp(struct super_block *s, secno near, unsigned n, unsigne | |||
75 | hpfs_error(s, "Bad allocation size: %d", n); | 73 | hpfs_error(s, "Bad allocation size: %d", n); |
76 | return 0; | 74 | return 0; |
77 | } | 75 | } |
78 | lock_super(s); | ||
79 | if (bs != ~0x3fff) { | 76 | if (bs != ~0x3fff) { |
80 | if (!(bmp = hpfs_map_bitmap(s, near >> 14, &qbh, "aib"))) goto uls; | 77 | if (!(bmp = hpfs_map_bitmap(s, near >> 14, &qbh, "aib"))) goto uls; |
81 | } else { | 78 | } else { |
@@ -85,10 +82,6 @@ static secno alloc_in_bmp(struct super_block *s, secno near, unsigned n, unsigne | |||
85 | ret = bs + nr; | 82 | ret = bs + nr; |
86 | goto rt; | 83 | goto rt; |
87 | } | 84 | } |
88 | /*if (!tstbits(bmp, nr + n, n + forward)) { | ||
89 | ret = bs + nr + n; | ||
90 | goto rt; | ||
91 | }*/ | ||
92 | q = nr + n; b = 0; | 85 | q = nr + n; b = 0; |
93 | while ((a = tstbits(bmp, q, n + forward)) != 0) { | 86 | while ((a = tstbits(bmp, q, n + forward)) != 0) { |
94 | q += a; | 87 | q += a; |
@@ -105,14 +98,14 @@ static secno alloc_in_bmp(struct super_block *s, secno near, unsigned n, unsigne | |||
105 | goto rt; | 98 | goto rt; |
106 | } | 99 | } |
107 | nr >>= 5; | 100 | nr >>= 5; |
108 | /*for (i = nr + 1; i != nr; i++, i &= 0x1ff) {*/ | 101 | /*for (i = nr + 1; i != nr; i++, i &= 0x1ff) */ |
109 | i = nr; | 102 | i = nr; |
110 | do { | 103 | do { |
111 | if (!bmp[i]) goto cont; | 104 | if (!le32_to_cpu(bmp[i])) goto cont; |
112 | if (n + forward >= 0x3f && bmp[i] != -1) goto cont; | 105 | if (n + forward >= 0x3f && le32_to_cpu(bmp[i]) != 0xffffffff) goto cont; |
113 | q = i<<5; | 106 | q = i<<5; |
114 | if (i > 0) { | 107 | if (i > 0) { |
115 | unsigned k = bmp[i-1]; | 108 | unsigned k = le32_to_cpu(bmp[i-1]); |
116 | while (k & 0x80000000) { | 109 | while (k & 0x80000000) { |
117 | q--; k <<= 1; | 110 | q--; k <<= 1; |
118 | } | 111 | } |
@@ -132,18 +125,17 @@ static secno alloc_in_bmp(struct super_block *s, secno near, unsigned n, unsigne | |||
132 | } while (i != nr); | 125 | } while (i != nr); |
133 | rt: | 126 | rt: |
134 | if (ret) { | 127 | if (ret) { |
135 | if (hpfs_sb(s)->sb_chk && ((ret >> 14) != (bs >> 14) || (bmp[(ret & 0x3fff) >> 5] | ~(((1 << n) - 1) << (ret & 0x1f))) != 0xffffffff)) { | 128 | if (hpfs_sb(s)->sb_chk && ((ret >> 14) != (bs >> 14) || (le32_to_cpu(bmp[(ret & 0x3fff) >> 5]) | ~(((1 << n) - 1) << (ret & 0x1f))) != 0xffffffff)) { |
136 | hpfs_error(s, "Allocation doesn't work! Wanted %d, allocated at %08x", n, ret); | 129 | hpfs_error(s, "Allocation doesn't work! Wanted %d, allocated at %08x", n, ret); |
137 | ret = 0; | 130 | ret = 0; |
138 | goto b; | 131 | goto b; |
139 | } | 132 | } |
140 | bmp[(ret & 0x3fff) >> 5] &= ~(((1 << n) - 1) << (ret & 0x1f)); | 133 | bmp[(ret & 0x3fff) >> 5] &= cpu_to_le32(~(((1 << n) - 1) << (ret & 0x1f))); |
141 | hpfs_mark_4buffers_dirty(&qbh); | 134 | hpfs_mark_4buffers_dirty(&qbh); |
142 | } | 135 | } |
143 | b: | 136 | b: |
144 | hpfs_brelse4(&qbh); | 137 | hpfs_brelse4(&qbh); |
145 | uls: | 138 | uls: |
146 | unlock_super(s); | ||
147 | return ret; | 139 | return ret; |
148 | } | 140 | } |
149 | 141 | ||
@@ -155,7 +147,7 @@ static secno alloc_in_bmp(struct super_block *s, secno near, unsigned n, unsigne | |||
155 | * sectors | 147 | * sectors |
156 | */ | 148 | */ |
157 | 149 | ||
158 | secno hpfs_alloc_sector(struct super_block *s, secno near, unsigned n, int forward, int lock) | 150 | secno hpfs_alloc_sector(struct super_block *s, secno near, unsigned n, int forward) |
159 | { | 151 | { |
160 | secno sec; | 152 | secno sec; |
161 | int i; | 153 | int i; |
@@ -167,7 +159,6 @@ secno hpfs_alloc_sector(struct super_block *s, secno near, unsigned n, int forwa | |||
167 | forward = -forward; | 159 | forward = -forward; |
168 | f_p = 1; | 160 | f_p = 1; |
169 | } | 161 | } |
170 | if (lock) hpfs_lock_creation(s); | ||
171 | n_bmps = (sbi->sb_fs_size + 0x4000 - 1) >> 14; | 162 | n_bmps = (sbi->sb_fs_size + 0x4000 - 1) >> 14; |
172 | if (near && near < sbi->sb_fs_size) { | 163 | if (near && near < sbi->sb_fs_size) { |
173 | if ((sec = alloc_in_bmp(s, near, n, f_p ? forward : forward/4))) goto ret; | 164 | if ((sec = alloc_in_bmp(s, near, n, f_p ? forward : forward/4))) goto ret; |
@@ -214,18 +205,17 @@ secno hpfs_alloc_sector(struct super_block *s, secno near, unsigned n, int forwa | |||
214 | ret: | 205 | ret: |
215 | if (sec && f_p) { | 206 | if (sec && f_p) { |
216 | for (i = 0; i < forward; i++) { | 207 | for (i = 0; i < forward; i++) { |
217 | if (!hpfs_alloc_if_possible_nolock(s, sec + i + 1)) { | 208 | if (!hpfs_alloc_if_possible(s, sec + i + 1)) { |
218 | hpfs_error(s, "Prealloc doesn't work! Wanted %d, allocated at %08x, can't allocate %d", forward, sec, i); | 209 | hpfs_error(s, "Prealloc doesn't work! Wanted %d, allocated at %08x, can't allocate %d", forward, sec, i); |
219 | sec = 0; | 210 | sec = 0; |
220 | break; | 211 | break; |
221 | } | 212 | } |
222 | } | 213 | } |
223 | } | 214 | } |
224 | if (lock) hpfs_unlock_creation(s); | ||
225 | return sec; | 215 | return sec; |
226 | } | 216 | } |
227 | 217 | ||
228 | static secno alloc_in_dirband(struct super_block *s, secno near, int lock) | 218 | static secno alloc_in_dirband(struct super_block *s, secno near) |
229 | { | 219 | { |
230 | unsigned nr = near; | 220 | unsigned nr = near; |
231 | secno sec; | 221 | secno sec; |
@@ -236,49 +226,35 @@ static secno alloc_in_dirband(struct super_block *s, secno near, int lock) | |||
236 | nr = sbi->sb_dirband_start + sbi->sb_dirband_size - 4; | 226 | nr = sbi->sb_dirband_start + sbi->sb_dirband_size - 4; |
237 | nr -= sbi->sb_dirband_start; | 227 | nr -= sbi->sb_dirband_start; |
238 | nr >>= 2; | 228 | nr >>= 2; |
239 | if (lock) hpfs_lock_creation(s); | ||
240 | sec = alloc_in_bmp(s, (~0x3fff) | nr, 1, 0); | 229 | sec = alloc_in_bmp(s, (~0x3fff) | nr, 1, 0); |
241 | if (lock) hpfs_unlock_creation(s); | ||
242 | if (!sec) return 0; | 230 | if (!sec) return 0; |
243 | return ((sec & 0x3fff) << 2) + sbi->sb_dirband_start; | 231 | return ((sec & 0x3fff) << 2) + sbi->sb_dirband_start; |
244 | } | 232 | } |
245 | 233 | ||
246 | /* Alloc sector if it's free */ | 234 | /* Alloc sector if it's free */ |
247 | 235 | ||
248 | static int hpfs_alloc_if_possible_nolock(struct super_block *s, secno sec) | 236 | int hpfs_alloc_if_possible(struct super_block *s, secno sec) |
249 | { | 237 | { |
250 | struct quad_buffer_head qbh; | 238 | struct quad_buffer_head qbh; |
251 | unsigned *bmp; | 239 | u32 *bmp; |
252 | lock_super(s); | ||
253 | if (!(bmp = hpfs_map_bitmap(s, sec >> 14, &qbh, "aip"))) goto end; | 240 | if (!(bmp = hpfs_map_bitmap(s, sec >> 14, &qbh, "aip"))) goto end; |
254 | if (bmp[(sec & 0x3fff) >> 5] & (1 << (sec & 0x1f))) { | 241 | if (le32_to_cpu(bmp[(sec & 0x3fff) >> 5]) & (1 << (sec & 0x1f))) { |
255 | bmp[(sec & 0x3fff) >> 5] &= ~(1 << (sec & 0x1f)); | 242 | bmp[(sec & 0x3fff) >> 5] &= cpu_to_le32(~(1 << (sec & 0x1f))); |
256 | hpfs_mark_4buffers_dirty(&qbh); | 243 | hpfs_mark_4buffers_dirty(&qbh); |
257 | hpfs_brelse4(&qbh); | 244 | hpfs_brelse4(&qbh); |
258 | unlock_super(s); | ||
259 | return 1; | 245 | return 1; |
260 | } | 246 | } |
261 | hpfs_brelse4(&qbh); | 247 | hpfs_brelse4(&qbh); |
262 | end: | 248 | end: |
263 | unlock_super(s); | ||
264 | return 0; | 249 | return 0; |
265 | } | 250 | } |
266 | 251 | ||
267 | int hpfs_alloc_if_possible(struct super_block *s, secno sec) | ||
268 | { | ||
269 | int r; | ||
270 | hpfs_lock_creation(s); | ||
271 | r = hpfs_alloc_if_possible_nolock(s, sec); | ||
272 | hpfs_unlock_creation(s); | ||
273 | return r; | ||
274 | } | ||
275 | |||
276 | /* Free sectors in bitmaps */ | 252 | /* Free sectors in bitmaps */ |
277 | 253 | ||
278 | void hpfs_free_sectors(struct super_block *s, secno sec, unsigned n) | 254 | void hpfs_free_sectors(struct super_block *s, secno sec, unsigned n) |
279 | { | 255 | { |
280 | struct quad_buffer_head qbh; | 256 | struct quad_buffer_head qbh; |
281 | unsigned *bmp; | 257 | u32 *bmp; |
282 | struct hpfs_sb_info *sbi = hpfs_sb(s); | 258 | struct hpfs_sb_info *sbi = hpfs_sb(s); |
283 | /*printk("2 - ");*/ | 259 | /*printk("2 - ");*/ |
284 | if (!n) return; | 260 | if (!n) return; |
@@ -286,26 +262,22 @@ void hpfs_free_sectors(struct super_block *s, secno sec, unsigned n) | |||
286 | hpfs_error(s, "Trying to free reserved sector %08x", sec); | 262 | hpfs_error(s, "Trying to free reserved sector %08x", sec); |
287 | return; | 263 | return; |
288 | } | 264 | } |
289 | lock_super(s); | ||
290 | sbi->sb_max_fwd_alloc += n > 0xffff ? 0xffff : n; | 265 | sbi->sb_max_fwd_alloc += n > 0xffff ? 0xffff : n; |
291 | if (sbi->sb_max_fwd_alloc > 0xffffff) sbi->sb_max_fwd_alloc = 0xffffff; | 266 | if (sbi->sb_max_fwd_alloc > 0xffffff) sbi->sb_max_fwd_alloc = 0xffffff; |
292 | new_map: | 267 | new_map: |
293 | if (!(bmp = hpfs_map_bitmap(s, sec >> 14, &qbh, "free"))) { | 268 | if (!(bmp = hpfs_map_bitmap(s, sec >> 14, &qbh, "free"))) { |
294 | unlock_super(s); | ||
295 | return; | 269 | return; |
296 | } | 270 | } |
297 | new_tst: | 271 | new_tst: |
298 | if ((bmp[(sec & 0x3fff) >> 5] >> (sec & 0x1f) & 1)) { | 272 | if ((le32_to_cpu(bmp[(sec & 0x3fff) >> 5]) >> (sec & 0x1f) & 1)) { |
299 | hpfs_error(s, "sector %08x not allocated", sec); | 273 | hpfs_error(s, "sector %08x not allocated", sec); |
300 | hpfs_brelse4(&qbh); | 274 | hpfs_brelse4(&qbh); |
301 | unlock_super(s); | ||
302 | return; | 275 | return; |
303 | } | 276 | } |
304 | bmp[(sec & 0x3fff) >> 5] |= 1 << (sec & 0x1f); | 277 | bmp[(sec & 0x3fff) >> 5] |= cpu_to_le32(1 << (sec & 0x1f)); |
305 | if (!--n) { | 278 | if (!--n) { |
306 | hpfs_mark_4buffers_dirty(&qbh); | 279 | hpfs_mark_4buffers_dirty(&qbh); |
307 | hpfs_brelse4(&qbh); | 280 | hpfs_brelse4(&qbh); |
308 | unlock_super(s); | ||
309 | return; | 281 | return; |
310 | } | 282 | } |
311 | if (!(++sec & 0x3fff)) { | 283 | if (!(++sec & 0x3fff)) { |
@@ -327,13 +299,13 @@ int hpfs_check_free_dnodes(struct super_block *s, int n) | |||
327 | int n_bmps = (hpfs_sb(s)->sb_fs_size + 0x4000 - 1) >> 14; | 299 | int n_bmps = (hpfs_sb(s)->sb_fs_size + 0x4000 - 1) >> 14; |
328 | int b = hpfs_sb(s)->sb_c_bitmap & 0x0fffffff; | 300 | int b = hpfs_sb(s)->sb_c_bitmap & 0x0fffffff; |
329 | int i, j; | 301 | int i, j; |
330 | unsigned *bmp; | 302 | u32 *bmp; |
331 | struct quad_buffer_head qbh; | 303 | struct quad_buffer_head qbh; |
332 | if ((bmp = hpfs_map_dnode_bitmap(s, &qbh))) { | 304 | if ((bmp = hpfs_map_dnode_bitmap(s, &qbh))) { |
333 | for (j = 0; j < 512; j++) { | 305 | for (j = 0; j < 512; j++) { |
334 | unsigned k; | 306 | unsigned k; |
335 | if (!bmp[j]) continue; | 307 | if (!le32_to_cpu(bmp[j])) continue; |
336 | for (k = bmp[j]; k; k >>= 1) if (k & 1) if (!--n) { | 308 | for (k = le32_to_cpu(bmp[j]); k; k >>= 1) if (k & 1) if (!--n) { |
337 | hpfs_brelse4(&qbh); | 309 | hpfs_brelse4(&qbh); |
338 | return 0; | 310 | return 0; |
339 | } | 311 | } |
@@ -352,10 +324,10 @@ int hpfs_check_free_dnodes(struct super_block *s, int n) | |||
352 | chk_bmp: | 324 | chk_bmp: |
353 | if (bmp) { | 325 | if (bmp) { |
354 | for (j = 0; j < 512; j++) { | 326 | for (j = 0; j < 512; j++) { |
355 | unsigned k; | 327 | u32 k; |
356 | if (!bmp[j]) continue; | 328 | if (!le32_to_cpu(bmp[j])) continue; |
357 | for (k = 0xf; k; k <<= 4) | 329 | for (k = 0xf; k; k <<= 4) |
358 | if ((bmp[j] & k) == k) { | 330 | if ((le32_to_cpu(bmp[j]) & k) == k) { |
359 | if (!--n) { | 331 | if (!--n) { |
360 | hpfs_brelse4(&qbh); | 332 | hpfs_brelse4(&qbh); |
361 | return 0; | 333 | return 0; |
@@ -379,44 +351,40 @@ void hpfs_free_dnode(struct super_block *s, dnode_secno dno) | |||
379 | hpfs_free_sectors(s, dno, 4); | 351 | hpfs_free_sectors(s, dno, 4); |
380 | } else { | 352 | } else { |
381 | struct quad_buffer_head qbh; | 353 | struct quad_buffer_head qbh; |
382 | unsigned *bmp; | 354 | u32 *bmp; |
383 | unsigned ssec = (dno - hpfs_sb(s)->sb_dirband_start) / 4; | 355 | unsigned ssec = (dno - hpfs_sb(s)->sb_dirband_start) / 4; |
384 | lock_super(s); | ||
385 | if (!(bmp = hpfs_map_dnode_bitmap(s, &qbh))) { | 356 | if (!(bmp = hpfs_map_dnode_bitmap(s, &qbh))) { |
386 | unlock_super(s); | ||
387 | return; | 357 | return; |
388 | } | 358 | } |
389 | bmp[ssec >> 5] |= 1 << (ssec & 0x1f); | 359 | bmp[ssec >> 5] |= cpu_to_le32(1 << (ssec & 0x1f)); |
390 | hpfs_mark_4buffers_dirty(&qbh); | 360 | hpfs_mark_4buffers_dirty(&qbh); |
391 | hpfs_brelse4(&qbh); | 361 | hpfs_brelse4(&qbh); |
392 | unlock_super(s); | ||
393 | } | 362 | } |
394 | } | 363 | } |
395 | 364 | ||
396 | struct dnode *hpfs_alloc_dnode(struct super_block *s, secno near, | 365 | struct dnode *hpfs_alloc_dnode(struct super_block *s, secno near, |
397 | dnode_secno *dno, struct quad_buffer_head *qbh, | 366 | dnode_secno *dno, struct quad_buffer_head *qbh) |
398 | int lock) | ||
399 | { | 367 | { |
400 | struct dnode *d; | 368 | struct dnode *d; |
401 | if (hpfs_count_one_bitmap(s, hpfs_sb(s)->sb_dmap) > FREE_DNODES_ADD) { | 369 | if (hpfs_count_one_bitmap(s, hpfs_sb(s)->sb_dmap) > FREE_DNODES_ADD) { |
402 | if (!(*dno = alloc_in_dirband(s, near, lock))) | 370 | if (!(*dno = alloc_in_dirband(s, near))) |
403 | if (!(*dno = hpfs_alloc_sector(s, near, 4, 0, lock))) return NULL; | 371 | if (!(*dno = hpfs_alloc_sector(s, near, 4, 0))) return NULL; |
404 | } else { | 372 | } else { |
405 | if (!(*dno = hpfs_alloc_sector(s, near, 4, 0, lock))) | 373 | if (!(*dno = hpfs_alloc_sector(s, near, 4, 0))) |
406 | if (!(*dno = alloc_in_dirband(s, near, lock))) return NULL; | 374 | if (!(*dno = alloc_in_dirband(s, near))) return NULL; |
407 | } | 375 | } |
408 | if (!(d = hpfs_get_4sectors(s, *dno, qbh))) { | 376 | if (!(d = hpfs_get_4sectors(s, *dno, qbh))) { |
409 | hpfs_free_dnode(s, *dno); | 377 | hpfs_free_dnode(s, *dno); |
410 | return NULL; | 378 | return NULL; |
411 | } | 379 | } |
412 | memset(d, 0, 2048); | 380 | memset(d, 0, 2048); |
413 | d->magic = DNODE_MAGIC; | 381 | d->magic = cpu_to_le32(DNODE_MAGIC); |
414 | d->first_free = 52; | 382 | d->first_free = cpu_to_le32(52); |
415 | d->dirent[0] = 32; | 383 | d->dirent[0] = 32; |
416 | d->dirent[2] = 8; | 384 | d->dirent[2] = 8; |
417 | d->dirent[30] = 1; | 385 | d->dirent[30] = 1; |
418 | d->dirent[31] = 255; | 386 | d->dirent[31] = 255; |
419 | d->self = *dno; | 387 | d->self = cpu_to_le32(*dno); |
420 | return d; | 388 | return d; |
421 | } | 389 | } |
422 | 390 | ||
@@ -424,16 +392,16 @@ struct fnode *hpfs_alloc_fnode(struct super_block *s, secno near, fnode_secno *f | |||
424 | struct buffer_head **bh) | 392 | struct buffer_head **bh) |
425 | { | 393 | { |
426 | struct fnode *f; | 394 | struct fnode *f; |
427 | if (!(*fno = hpfs_alloc_sector(s, near, 1, FNODE_ALLOC_FWD, 1))) return NULL; | 395 | if (!(*fno = hpfs_alloc_sector(s, near, 1, FNODE_ALLOC_FWD))) return NULL; |
428 | if (!(f = hpfs_get_sector(s, *fno, bh))) { | 396 | if (!(f = hpfs_get_sector(s, *fno, bh))) { |
429 | hpfs_free_sectors(s, *fno, 1); | 397 | hpfs_free_sectors(s, *fno, 1); |
430 | return NULL; | 398 | return NULL; |
431 | } | 399 | } |
432 | memset(f, 0, 512); | 400 | memset(f, 0, 512); |
433 | f->magic = FNODE_MAGIC; | 401 | f->magic = cpu_to_le32(FNODE_MAGIC); |
434 | f->ea_offs = 0xc4; | 402 | f->ea_offs = cpu_to_le16(0xc4); |
435 | f->btree.n_free_nodes = 8; | 403 | f->btree.n_free_nodes = 8; |
436 | f->btree.first_free = 8; | 404 | f->btree.first_free = cpu_to_le16(8); |
437 | return f; | 405 | return f; |
438 | } | 406 | } |
439 | 407 | ||
@@ -441,16 +409,16 @@ struct anode *hpfs_alloc_anode(struct super_block *s, secno near, anode_secno *a | |||
441 | struct buffer_head **bh) | 409 | struct buffer_head **bh) |
442 | { | 410 | { |
443 | struct anode *a; | 411 | struct anode *a; |
444 | if (!(*ano = hpfs_alloc_sector(s, near, 1, ANODE_ALLOC_FWD, 1))) return NULL; | 412 | if (!(*ano = hpfs_alloc_sector(s, near, 1, ANODE_ALLOC_FWD))) return NULL; |
445 | if (!(a = hpfs_get_sector(s, *ano, bh))) { | 413 | if (!(a = hpfs_get_sector(s, *ano, bh))) { |
446 | hpfs_free_sectors(s, *ano, 1); | 414 | hpfs_free_sectors(s, *ano, 1); |
447 | return NULL; | 415 | return NULL; |
448 | } | 416 | } |
449 | memset(a, 0, 512); | 417 | memset(a, 0, 512); |
450 | a->magic = ANODE_MAGIC; | 418 | a->magic = cpu_to_le32(ANODE_MAGIC); |
451 | a->self = *ano; | 419 | a->self = cpu_to_le32(*ano); |
452 | a->btree.n_free_nodes = 40; | 420 | a->btree.n_free_nodes = 40; |
453 | a->btree.n_used_nodes = 0; | 421 | a->btree.n_used_nodes = 0; |
454 | a->btree.first_free = 8; | 422 | a->btree.first_free = cpu_to_le16(8); |
455 | return a; | 423 | return a; |
456 | } | 424 | } |
diff --git a/fs/hpfs/anode.c b/fs/hpfs/anode.c index 6a2f04bf3df0..08b503e8ed29 100644 --- a/fs/hpfs/anode.c +++ b/fs/hpfs/anode.c | |||
@@ -22,8 +22,8 @@ secno hpfs_bplus_lookup(struct super_block *s, struct inode *inode, | |||
22 | if (hpfs_sb(s)->sb_chk) if (hpfs_stop_cycles(s, a, &c1, &c2, "hpfs_bplus_lookup")) return -1; | 22 | if (hpfs_sb(s)->sb_chk) if (hpfs_stop_cycles(s, a, &c1, &c2, "hpfs_bplus_lookup")) return -1; |
23 | if (btree->internal) { | 23 | if (btree->internal) { |
24 | for (i = 0; i < btree->n_used_nodes; i++) | 24 | for (i = 0; i < btree->n_used_nodes; i++) |
25 | if (btree->u.internal[i].file_secno > sec) { | 25 | if (le32_to_cpu(btree->u.internal[i].file_secno) > sec) { |
26 | a = btree->u.internal[i].down; | 26 | a = le32_to_cpu(btree->u.internal[i].down); |
27 | brelse(bh); | 27 | brelse(bh); |
28 | if (!(anode = hpfs_map_anode(s, a, &bh))) return -1; | 28 | if (!(anode = hpfs_map_anode(s, a, &bh))) return -1; |
29 | btree = &anode->btree; | 29 | btree = &anode->btree; |
@@ -34,18 +34,18 @@ secno hpfs_bplus_lookup(struct super_block *s, struct inode *inode, | |||
34 | return -1; | 34 | return -1; |
35 | } | 35 | } |
36 | for (i = 0; i < btree->n_used_nodes; i++) | 36 | for (i = 0; i < btree->n_used_nodes; i++) |
37 | if (btree->u.external[i].file_secno <= sec && | 37 | if (le32_to_cpu(btree->u.external[i].file_secno) <= sec && |
38 | btree->u.external[i].file_secno + btree->u.external[i].length > sec) { | 38 | le32_to_cpu(btree->u.external[i].file_secno) + le32_to_cpu(btree->u.external[i].length) > sec) { |
39 | a = btree->u.external[i].disk_secno + sec - btree->u.external[i].file_secno; | 39 | a = le32_to_cpu(btree->u.external[i].disk_secno) + sec - le32_to_cpu(btree->u.external[i].file_secno); |
40 | if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, a, 1, "data")) { | 40 | if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, a, 1, "data")) { |
41 | brelse(bh); | 41 | brelse(bh); |
42 | return -1; | 42 | return -1; |
43 | } | 43 | } |
44 | if (inode) { | 44 | if (inode) { |
45 | struct hpfs_inode_info *hpfs_inode = hpfs_i(inode); | 45 | struct hpfs_inode_info *hpfs_inode = hpfs_i(inode); |
46 | hpfs_inode->i_file_sec = btree->u.external[i].file_secno; | 46 | hpfs_inode->i_file_sec = le32_to_cpu(btree->u.external[i].file_secno); |
47 | hpfs_inode->i_disk_sec = btree->u.external[i].disk_secno; | 47 | hpfs_inode->i_disk_sec = le32_to_cpu(btree->u.external[i].disk_secno); |
48 | hpfs_inode->i_n_secs = btree->u.external[i].length; | 48 | hpfs_inode->i_n_secs = le32_to_cpu(btree->u.external[i].length); |
49 | } | 49 | } |
50 | brelse(bh); | 50 | brelse(bh); |
51 | return a; | 51 | return a; |
@@ -83,8 +83,8 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
83 | return -1; | 83 | return -1; |
84 | } | 84 | } |
85 | if (btree->internal) { | 85 | if (btree->internal) { |
86 | a = btree->u.internal[n].down; | 86 | a = le32_to_cpu(btree->u.internal[n].down); |
87 | btree->u.internal[n].file_secno = -1; | 87 | btree->u.internal[n].file_secno = cpu_to_le32(-1); |
88 | mark_buffer_dirty(bh); | 88 | mark_buffer_dirty(bh); |
89 | brelse(bh); | 89 | brelse(bh); |
90 | if (hpfs_sb(s)->sb_chk) | 90 | if (hpfs_sb(s)->sb_chk) |
@@ -94,15 +94,15 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
94 | goto go_down; | 94 | goto go_down; |
95 | } | 95 | } |
96 | if (n >= 0) { | 96 | if (n >= 0) { |
97 | if (btree->u.external[n].file_secno + btree->u.external[n].length != fsecno) { | 97 | if (le32_to_cpu(btree->u.external[n].file_secno) + le32_to_cpu(btree->u.external[n].length) != fsecno) { |
98 | hpfs_error(s, "allocated size %08x, trying to add sector %08x, %cnode %08x", | 98 | hpfs_error(s, "allocated size %08x, trying to add sector %08x, %cnode %08x", |
99 | btree->u.external[n].file_secno + btree->u.external[n].length, fsecno, | 99 | le32_to_cpu(btree->u.external[n].file_secno) + le32_to_cpu(btree->u.external[n].length), fsecno, |
100 | fnod?'f':'a', node); | 100 | fnod?'f':'a', node); |
101 | brelse(bh); | 101 | brelse(bh); |
102 | return -1; | 102 | return -1; |
103 | } | 103 | } |
104 | if (hpfs_alloc_if_possible(s, se = btree->u.external[n].disk_secno + btree->u.external[n].length)) { | 104 | if (hpfs_alloc_if_possible(s, se = le32_to_cpu(btree->u.external[n].disk_secno) + le32_to_cpu(btree->u.external[n].length))) { |
105 | btree->u.external[n].length++; | 105 | btree->u.external[n].length = cpu_to_le32(le32_to_cpu(btree->u.external[n].length) + 1); |
106 | mark_buffer_dirty(bh); | 106 | mark_buffer_dirty(bh); |
107 | brelse(bh); | 107 | brelse(bh); |
108 | return se; | 108 | return se; |
@@ -115,20 +115,20 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
115 | } | 115 | } |
116 | se = !fnod ? node : (node + 16384) & ~16383; | 116 | se = !fnod ? node : (node + 16384) & ~16383; |
117 | } | 117 | } |
118 | if (!(se = hpfs_alloc_sector(s, se, 1, fsecno*ALLOC_M>ALLOC_FWD_MAX ? ALLOC_FWD_MAX : fsecno*ALLOC_M<ALLOC_FWD_MIN ? ALLOC_FWD_MIN : fsecno*ALLOC_M, 1))) { | 118 | if (!(se = hpfs_alloc_sector(s, se, 1, fsecno*ALLOC_M>ALLOC_FWD_MAX ? ALLOC_FWD_MAX : fsecno*ALLOC_M<ALLOC_FWD_MIN ? ALLOC_FWD_MIN : fsecno*ALLOC_M))) { |
119 | brelse(bh); | 119 | brelse(bh); |
120 | return -1; | 120 | return -1; |
121 | } | 121 | } |
122 | fs = n < 0 ? 0 : btree->u.external[n].file_secno + btree->u.external[n].length; | 122 | fs = n < 0 ? 0 : le32_to_cpu(btree->u.external[n].file_secno) + le32_to_cpu(btree->u.external[n].length); |
123 | if (!btree->n_free_nodes) { | 123 | if (!btree->n_free_nodes) { |
124 | up = a != node ? anode->up : -1; | 124 | up = a != node ? le32_to_cpu(anode->up) : -1; |
125 | if (!(anode = hpfs_alloc_anode(s, a, &na, &bh1))) { | 125 | if (!(anode = hpfs_alloc_anode(s, a, &na, &bh1))) { |
126 | brelse(bh); | 126 | brelse(bh); |
127 | hpfs_free_sectors(s, se, 1); | 127 | hpfs_free_sectors(s, se, 1); |
128 | return -1; | 128 | return -1; |
129 | } | 129 | } |
130 | if (a == node && fnod) { | 130 | if (a == node && fnod) { |
131 | anode->up = node; | 131 | anode->up = cpu_to_le32(node); |
132 | anode->btree.fnode_parent = 1; | 132 | anode->btree.fnode_parent = 1; |
133 | anode->btree.n_used_nodes = btree->n_used_nodes; | 133 | anode->btree.n_used_nodes = btree->n_used_nodes; |
134 | anode->btree.first_free = btree->first_free; | 134 | anode->btree.first_free = btree->first_free; |
@@ -137,9 +137,9 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
137 | btree->internal = 1; | 137 | btree->internal = 1; |
138 | btree->n_free_nodes = 11; | 138 | btree->n_free_nodes = 11; |
139 | btree->n_used_nodes = 1; | 139 | btree->n_used_nodes = 1; |
140 | btree->first_free = (char *)&(btree->u.internal[1]) - (char *)btree; | 140 | btree->first_free = cpu_to_le16((char *)&(btree->u.internal[1]) - (char *)btree); |
141 | btree->u.internal[0].file_secno = -1; | 141 | btree->u.internal[0].file_secno = cpu_to_le32(-1); |
142 | btree->u.internal[0].down = na; | 142 | btree->u.internal[0].down = cpu_to_le32(na); |
143 | mark_buffer_dirty(bh); | 143 | mark_buffer_dirty(bh); |
144 | } else if (!(ranode = hpfs_alloc_anode(s, /*a*/0, &ra, &bh2))) { | 144 | } else if (!(ranode = hpfs_alloc_anode(s, /*a*/0, &ra, &bh2))) { |
145 | brelse(bh); | 145 | brelse(bh); |
@@ -153,15 +153,15 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
153 | btree = &anode->btree; | 153 | btree = &anode->btree; |
154 | } | 154 | } |
155 | btree->n_free_nodes--; n = btree->n_used_nodes++; | 155 | btree->n_free_nodes--; n = btree->n_used_nodes++; |
156 | btree->first_free += 12; | 156 | btree->first_free = cpu_to_le16(le16_to_cpu(btree->first_free) + 12); |
157 | btree->u.external[n].disk_secno = se; | 157 | btree->u.external[n].disk_secno = cpu_to_le32(se); |
158 | btree->u.external[n].file_secno = fs; | 158 | btree->u.external[n].file_secno = cpu_to_le32(fs); |
159 | btree->u.external[n].length = 1; | 159 | btree->u.external[n].length = cpu_to_le32(1); |
160 | mark_buffer_dirty(bh); | 160 | mark_buffer_dirty(bh); |
161 | brelse(bh); | 161 | brelse(bh); |
162 | if ((a == node && fnod) || na == -1) return se; | 162 | if ((a == node && fnod) || na == -1) return se; |
163 | c2 = 0; | 163 | c2 = 0; |
164 | while (up != -1) { | 164 | while (up != (anode_secno)-1) { |
165 | struct anode *new_anode; | 165 | struct anode *new_anode; |
166 | if (hpfs_sb(s)->sb_chk) | 166 | if (hpfs_sb(s)->sb_chk) |
167 | if (hpfs_stop_cycles(s, up, &c1, &c2, "hpfs_add_sector_to_btree #2")) return -1; | 167 | if (hpfs_stop_cycles(s, up, &c1, &c2, "hpfs_add_sector_to_btree #2")) return -1; |
@@ -174,47 +174,47 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
174 | } | 174 | } |
175 | if (btree->n_free_nodes) { | 175 | if (btree->n_free_nodes) { |
176 | btree->n_free_nodes--; n = btree->n_used_nodes++; | 176 | btree->n_free_nodes--; n = btree->n_used_nodes++; |
177 | btree->first_free += 8; | 177 | btree->first_free = cpu_to_le16(le16_to_cpu(btree->first_free) + 8); |
178 | btree->u.internal[n].file_secno = -1; | 178 | btree->u.internal[n].file_secno = cpu_to_le32(-1); |
179 | btree->u.internal[n].down = na; | 179 | btree->u.internal[n].down = cpu_to_le32(na); |
180 | btree->u.internal[n-1].file_secno = fs; | 180 | btree->u.internal[n-1].file_secno = cpu_to_le32(fs); |
181 | mark_buffer_dirty(bh); | 181 | mark_buffer_dirty(bh); |
182 | brelse(bh); | 182 | brelse(bh); |
183 | brelse(bh2); | 183 | brelse(bh2); |
184 | hpfs_free_sectors(s, ra, 1); | 184 | hpfs_free_sectors(s, ra, 1); |
185 | if ((anode = hpfs_map_anode(s, na, &bh))) { | 185 | if ((anode = hpfs_map_anode(s, na, &bh))) { |
186 | anode->up = up; | 186 | anode->up = cpu_to_le32(up); |
187 | anode->btree.fnode_parent = up == node && fnod; | 187 | anode->btree.fnode_parent = up == node && fnod; |
188 | mark_buffer_dirty(bh); | 188 | mark_buffer_dirty(bh); |
189 | brelse(bh); | 189 | brelse(bh); |
190 | } | 190 | } |
191 | return se; | 191 | return se; |
192 | } | 192 | } |
193 | up = up != node ? anode->up : -1; | 193 | up = up != node ? le32_to_cpu(anode->up) : -1; |
194 | btree->u.internal[btree->n_used_nodes - 1].file_secno = /*fs*/-1; | 194 | btree->u.internal[btree->n_used_nodes - 1].file_secno = cpu_to_le32(/*fs*/-1); |
195 | mark_buffer_dirty(bh); | 195 | mark_buffer_dirty(bh); |
196 | brelse(bh); | 196 | brelse(bh); |
197 | a = na; | 197 | a = na; |
198 | if ((new_anode = hpfs_alloc_anode(s, a, &na, &bh))) { | 198 | if ((new_anode = hpfs_alloc_anode(s, a, &na, &bh))) { |
199 | anode = new_anode; | 199 | anode = new_anode; |
200 | /*anode->up = up != -1 ? up : ra;*/ | 200 | /*anode->up = cpu_to_le32(up != -1 ? up : ra);*/ |
201 | anode->btree.internal = 1; | 201 | anode->btree.internal = 1; |
202 | anode->btree.n_used_nodes = 1; | 202 | anode->btree.n_used_nodes = 1; |
203 | anode->btree.n_free_nodes = 59; | 203 | anode->btree.n_free_nodes = 59; |
204 | anode->btree.first_free = 16; | 204 | anode->btree.first_free = cpu_to_le16(16); |
205 | anode->btree.u.internal[0].down = a; | 205 | anode->btree.u.internal[0].down = cpu_to_le32(a); |
206 | anode->btree.u.internal[0].file_secno = -1; | 206 | anode->btree.u.internal[0].file_secno = cpu_to_le32(-1); |
207 | mark_buffer_dirty(bh); | 207 | mark_buffer_dirty(bh); |
208 | brelse(bh); | 208 | brelse(bh); |
209 | if ((anode = hpfs_map_anode(s, a, &bh))) { | 209 | if ((anode = hpfs_map_anode(s, a, &bh))) { |
210 | anode->up = na; | 210 | anode->up = cpu_to_le32(na); |
211 | mark_buffer_dirty(bh); | 211 | mark_buffer_dirty(bh); |
212 | brelse(bh); | 212 | brelse(bh); |
213 | } | 213 | } |
214 | } else na = a; | 214 | } else na = a; |
215 | } | 215 | } |
216 | if ((anode = hpfs_map_anode(s, na, &bh))) { | 216 | if ((anode = hpfs_map_anode(s, na, &bh))) { |
217 | anode->up = node; | 217 | anode->up = cpu_to_le32(node); |
218 | if (fnod) anode->btree.fnode_parent = 1; | 218 | if (fnod) anode->btree.fnode_parent = 1; |
219 | mark_buffer_dirty(bh); | 219 | mark_buffer_dirty(bh); |
220 | brelse(bh); | 220 | brelse(bh); |
@@ -232,14 +232,14 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
232 | } | 232 | } |
233 | btree = &fnode->btree; | 233 | btree = &fnode->btree; |
234 | } | 234 | } |
235 | ranode->up = node; | 235 | ranode->up = cpu_to_le32(node); |
236 | memcpy(&ranode->btree, btree, btree->first_free); | 236 | memcpy(&ranode->btree, btree, le16_to_cpu(btree->first_free)); |
237 | if (fnod) ranode->btree.fnode_parent = 1; | 237 | if (fnod) ranode->btree.fnode_parent = 1; |
238 | ranode->btree.n_free_nodes = (ranode->btree.internal ? 60 : 40) - ranode->btree.n_used_nodes; | 238 | ranode->btree.n_free_nodes = (ranode->btree.internal ? 60 : 40) - ranode->btree.n_used_nodes; |
239 | if (ranode->btree.internal) for (n = 0; n < ranode->btree.n_used_nodes; n++) { | 239 | if (ranode->btree.internal) for (n = 0; n < ranode->btree.n_used_nodes; n++) { |
240 | struct anode *unode; | 240 | struct anode *unode; |
241 | if ((unode = hpfs_map_anode(s, ranode->u.internal[n].down, &bh1))) { | 241 | if ((unode = hpfs_map_anode(s, le32_to_cpu(ranode->u.internal[n].down), &bh1))) { |
242 | unode->up = ra; | 242 | unode->up = cpu_to_le32(ra); |
243 | unode->btree.fnode_parent = 0; | 243 | unode->btree.fnode_parent = 0; |
244 | mark_buffer_dirty(bh1); | 244 | mark_buffer_dirty(bh1); |
245 | brelse(bh1); | 245 | brelse(bh1); |
@@ -248,11 +248,11 @@ secno hpfs_add_sector_to_btree(struct super_block *s, secno node, int fnod, unsi | |||
248 | btree->internal = 1; | 248 | btree->internal = 1; |
249 | btree->n_free_nodes = fnod ? 10 : 58; | 249 | btree->n_free_nodes = fnod ? 10 : 58; |
250 | btree->n_used_nodes = 2; | 250 | btree->n_used_nodes = 2; |
251 | btree->first_free = (char *)&btree->u.internal[2] - (char *)btree; | 251 | btree->first_free = cpu_to_le16((char *)&btree->u.internal[2] - (char *)btree); |
252 | btree->u.internal[0].file_secno = fs; | 252 | btree->u.internal[0].file_secno = cpu_to_le32(fs); |
253 | btree->u.internal[0].down = ra; | 253 | btree->u.internal[0].down = cpu_to_le32(ra); |
254 | btree->u.internal[1].file_secno = -1; | 254 | btree->u.internal[1].file_secno = cpu_to_le32(-1); |
255 | btree->u.internal[1].down = na; | 255 | btree->u.internal[1].down = cpu_to_le32(na); |
256 | mark_buffer_dirty(bh); | 256 | mark_buffer_dirty(bh); |
257 | brelse(bh); | 257 | brelse(bh); |
258 | mark_buffer_dirty(bh2); | 258 | mark_buffer_dirty(bh2); |
@@ -279,7 +279,7 @@ void hpfs_remove_btree(struct super_block *s, struct bplus_header *btree) | |||
279 | go_down: | 279 | go_down: |
280 | d2 = 0; | 280 | d2 = 0; |
281 | while (btree1->internal) { | 281 | while (btree1->internal) { |
282 | ano = btree1->u.internal[pos].down; | 282 | ano = le32_to_cpu(btree1->u.internal[pos].down); |
283 | if (level) brelse(bh); | 283 | if (level) brelse(bh); |
284 | if (hpfs_sb(s)->sb_chk) | 284 | if (hpfs_sb(s)->sb_chk) |
285 | if (hpfs_stop_cycles(s, ano, &d1, &d2, "hpfs_remove_btree #1")) | 285 | if (hpfs_stop_cycles(s, ano, &d1, &d2, "hpfs_remove_btree #1")) |
@@ -290,7 +290,7 @@ void hpfs_remove_btree(struct super_block *s, struct bplus_header *btree) | |||
290 | pos = 0; | 290 | pos = 0; |
291 | } | 291 | } |
292 | for (i = 0; i < btree1->n_used_nodes; i++) | 292 | for (i = 0; i < btree1->n_used_nodes; i++) |
293 | hpfs_free_sectors(s, btree1->u.external[i].disk_secno, btree1->u.external[i].length); | 293 | hpfs_free_sectors(s, le32_to_cpu(btree1->u.external[i].disk_secno), le32_to_cpu(btree1->u.external[i].length)); |
294 | go_up: | 294 | go_up: |
295 | if (!level) return; | 295 | if (!level) return; |
296 | brelse(bh); | 296 | brelse(bh); |
@@ -298,13 +298,13 @@ void hpfs_remove_btree(struct super_block *s, struct bplus_header *btree) | |||
298 | if (hpfs_stop_cycles(s, ano, &c1, &c2, "hpfs_remove_btree #2")) return; | 298 | if (hpfs_stop_cycles(s, ano, &c1, &c2, "hpfs_remove_btree #2")) return; |
299 | hpfs_free_sectors(s, ano, 1); | 299 | hpfs_free_sectors(s, ano, 1); |
300 | oano = ano; | 300 | oano = ano; |
301 | ano = anode->up; | 301 | ano = le32_to_cpu(anode->up); |
302 | if (--level) { | 302 | if (--level) { |
303 | if (!(anode = hpfs_map_anode(s, ano, &bh))) return; | 303 | if (!(anode = hpfs_map_anode(s, ano, &bh))) return; |
304 | btree1 = &anode->btree; | 304 | btree1 = &anode->btree; |
305 | } else btree1 = btree; | 305 | } else btree1 = btree; |
306 | for (i = 0; i < btree1->n_used_nodes; i++) { | 306 | for (i = 0; i < btree1->n_used_nodes; i++) { |
307 | if (btree1->u.internal[i].down == oano) { | 307 | if (le32_to_cpu(btree1->u.internal[i].down) == oano) { |
308 | if ((pos = i + 1) < btree1->n_used_nodes) | 308 | if ((pos = i + 1) < btree1->n_used_nodes) |
309 | goto go_down; | 309 | goto go_down; |
310 | else | 310 | else |
@@ -411,7 +411,7 @@ void hpfs_truncate_btree(struct super_block *s, secno f, int fno, unsigned secs) | |||
411 | if (fno) { | 411 | if (fno) { |
412 | btree->n_free_nodes = 8; | 412 | btree->n_free_nodes = 8; |
413 | btree->n_used_nodes = 0; | 413 | btree->n_used_nodes = 0; |
414 | btree->first_free = 8; | 414 | btree->first_free = cpu_to_le16(8); |
415 | btree->internal = 0; | 415 | btree->internal = 0; |
416 | mark_buffer_dirty(bh); | 416 | mark_buffer_dirty(bh); |
417 | } else hpfs_free_sectors(s, f, 1); | 417 | } else hpfs_free_sectors(s, f, 1); |
@@ -421,22 +421,22 @@ void hpfs_truncate_btree(struct super_block *s, secno f, int fno, unsigned secs) | |||
421 | while (btree->internal) { | 421 | while (btree->internal) { |
422 | nodes = btree->n_used_nodes + btree->n_free_nodes; | 422 | nodes = btree->n_used_nodes + btree->n_free_nodes; |
423 | for (i = 0; i < btree->n_used_nodes; i++) | 423 | for (i = 0; i < btree->n_used_nodes; i++) |
424 | if (btree->u.internal[i].file_secno >= secs) goto f; | 424 | if (le32_to_cpu(btree->u.internal[i].file_secno) >= secs) goto f; |
425 | brelse(bh); | 425 | brelse(bh); |
426 | hpfs_error(s, "internal btree %08x doesn't end with -1", node); | 426 | hpfs_error(s, "internal btree %08x doesn't end with -1", node); |
427 | return; | 427 | return; |
428 | f: | 428 | f: |
429 | for (j = i + 1; j < btree->n_used_nodes; j++) | 429 | for (j = i + 1; j < btree->n_used_nodes; j++) |
430 | hpfs_ea_remove(s, btree->u.internal[j].down, 1, 0); | 430 | hpfs_ea_remove(s, le32_to_cpu(btree->u.internal[j].down), 1, 0); |
431 | btree->n_used_nodes = i + 1; | 431 | btree->n_used_nodes = i + 1; |
432 | btree->n_free_nodes = nodes - btree->n_used_nodes; | 432 | btree->n_free_nodes = nodes - btree->n_used_nodes; |
433 | btree->first_free = 8 + 8 * btree->n_used_nodes; | 433 | btree->first_free = cpu_to_le16(8 + 8 * btree->n_used_nodes); |
434 | mark_buffer_dirty(bh); | 434 | mark_buffer_dirty(bh); |
435 | if (btree->u.internal[i].file_secno == secs) { | 435 | if (btree->u.internal[i].file_secno == cpu_to_le32(secs)) { |
436 | brelse(bh); | 436 | brelse(bh); |
437 | return; | 437 | return; |
438 | } | 438 | } |
439 | node = btree->u.internal[i].down; | 439 | node = le32_to_cpu(btree->u.internal[i].down); |
440 | brelse(bh); | 440 | brelse(bh); |
441 | if (hpfs_sb(s)->sb_chk) | 441 | if (hpfs_sb(s)->sb_chk) |
442 | if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree")) | 442 | if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree")) |
@@ -446,25 +446,25 @@ void hpfs_truncate_btree(struct super_block *s, secno f, int fno, unsigned secs) | |||
446 | } | 446 | } |
447 | nodes = btree->n_used_nodes + btree->n_free_nodes; | 447 | nodes = btree->n_used_nodes + btree->n_free_nodes; |
448 | for (i = 0; i < btree->n_used_nodes; i++) | 448 | for (i = 0; i < btree->n_used_nodes; i++) |
449 | if (btree->u.external[i].file_secno + btree->u.external[i].length >= secs) goto ff; | 449 | if (le32_to_cpu(btree->u.external[i].file_secno) + le32_to_cpu(btree->u.external[i].length) >= secs) goto ff; |
450 | brelse(bh); | 450 | brelse(bh); |
451 | return; | 451 | return; |
452 | ff: | 452 | ff: |
453 | if (secs <= btree->u.external[i].file_secno) { | 453 | if (secs <= le32_to_cpu(btree->u.external[i].file_secno)) { |
454 | hpfs_error(s, "there is an allocation error in file %08x, sector %08x", f, secs); | 454 | hpfs_error(s, "there is an allocation error in file %08x, sector %08x", f, secs); |
455 | if (i) i--; | 455 | if (i) i--; |
456 | } | 456 | } |
457 | else if (btree->u.external[i].file_secno + btree->u.external[i].length > secs) { | 457 | else if (le32_to_cpu(btree->u.external[i].file_secno) + le32_to_cpu(btree->u.external[i].length) > secs) { |
458 | hpfs_free_sectors(s, btree->u.external[i].disk_secno + secs - | 458 | hpfs_free_sectors(s, le32_to_cpu(btree->u.external[i].disk_secno) + secs - |
459 | btree->u.external[i].file_secno, btree->u.external[i].length | 459 | le32_to_cpu(btree->u.external[i].file_secno), le32_to_cpu(btree->u.external[i].length) |
460 | - secs + btree->u.external[i].file_secno); /* I hope gcc optimizes this :-) */ | 460 | - secs + le32_to_cpu(btree->u.external[i].file_secno)); /* I hope gcc optimizes this :-) */ |
461 | btree->u.external[i].length = secs - btree->u.external[i].file_secno; | 461 | btree->u.external[i].length = cpu_to_le32(secs - le32_to_cpu(btree->u.external[i].file_secno)); |
462 | } | 462 | } |
463 | for (j = i + 1; j < btree->n_used_nodes; j++) | 463 | for (j = i + 1; j < btree->n_used_nodes; j++) |
464 | hpfs_free_sectors(s, btree->u.external[j].disk_secno, btree->u.external[j].length); | 464 | hpfs_free_sectors(s, le32_to_cpu(btree->u.external[j].disk_secno), le32_to_cpu(btree->u.external[j].length)); |
465 | btree->n_used_nodes = i + 1; | 465 | btree->n_used_nodes = i + 1; |
466 | btree->n_free_nodes = nodes - btree->n_used_nodes; | 466 | btree->n_free_nodes = nodes - btree->n_used_nodes; |
467 | btree->first_free = 8 + 12 * btree->n_used_nodes; | 467 | btree->first_free = cpu_to_le16(8 + 12 * btree->n_used_nodes); |
468 | mark_buffer_dirty(bh); | 468 | mark_buffer_dirty(bh); |
469 | brelse(bh); | 469 | brelse(bh); |
470 | } | 470 | } |
@@ -480,12 +480,12 @@ void hpfs_remove_fnode(struct super_block *s, fnode_secno fno) | |||
480 | struct extended_attribute *ea_end; | 480 | struct extended_attribute *ea_end; |
481 | if (!(fnode = hpfs_map_fnode(s, fno, &bh))) return; | 481 | if (!(fnode = hpfs_map_fnode(s, fno, &bh))) return; |
482 | if (!fnode->dirflag) hpfs_remove_btree(s, &fnode->btree); | 482 | if (!fnode->dirflag) hpfs_remove_btree(s, &fnode->btree); |
483 | else hpfs_remove_dtree(s, fnode->u.external[0].disk_secno); | 483 | else hpfs_remove_dtree(s, le32_to_cpu(fnode->u.external[0].disk_secno)); |
484 | ea_end = fnode_end_ea(fnode); | 484 | ea_end = fnode_end_ea(fnode); |
485 | for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) | 485 | for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) |
486 | if (ea->indirect) | 486 | if (ea->indirect) |
487 | hpfs_ea_remove(s, ea_sec(ea), ea->anode, ea_len(ea)); | 487 | hpfs_ea_remove(s, ea_sec(ea), ea->anode, ea_len(ea)); |
488 | hpfs_ea_ext_remove(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l); | 488 | hpfs_ea_ext_remove(s, le32_to_cpu(fnode->ea_secno), fnode->ea_anode, le32_to_cpu(fnode->ea_size_l)); |
489 | brelse(bh); | 489 | brelse(bh); |
490 | hpfs_free_sectors(s, fno, 1); | 490 | hpfs_free_sectors(s, fno, 1); |
491 | } | 491 | } |
diff --git a/fs/hpfs/buffer.c b/fs/hpfs/buffer.c index 793cb9d943d2..9ecde27d1e29 100644 --- a/fs/hpfs/buffer.c +++ b/fs/hpfs/buffer.c | |||
@@ -9,22 +9,6 @@ | |||
9 | #include <linux/slab.h> | 9 | #include <linux/slab.h> |
10 | #include "hpfs_fn.h" | 10 | #include "hpfs_fn.h" |
11 | 11 | ||
12 | void hpfs_lock_creation(struct super_block *s) | ||
13 | { | ||
14 | #ifdef DEBUG_LOCKS | ||
15 | printk("lock creation\n"); | ||
16 | #endif | ||
17 | mutex_lock(&hpfs_sb(s)->hpfs_creation_de); | ||
18 | } | ||
19 | |||
20 | void hpfs_unlock_creation(struct super_block *s) | ||
21 | { | ||
22 | #ifdef DEBUG_LOCKS | ||
23 | printk("unlock creation\n"); | ||
24 | #endif | ||
25 | mutex_unlock(&hpfs_sb(s)->hpfs_creation_de); | ||
26 | } | ||
27 | |||
28 | /* Map a sector into a buffer and return pointers to it and to the buffer. */ | 12 | /* Map a sector into a buffer and return pointers to it and to the buffer. */ |
29 | 13 | ||
30 | void *hpfs_map_sector(struct super_block *s, unsigned secno, struct buffer_head **bhp, | 14 | void *hpfs_map_sector(struct super_block *s, unsigned secno, struct buffer_head **bhp, |
@@ -32,6 +16,8 @@ void *hpfs_map_sector(struct super_block *s, unsigned secno, struct buffer_head | |||
32 | { | 16 | { |
33 | struct buffer_head *bh; | 17 | struct buffer_head *bh; |
34 | 18 | ||
19 | hpfs_lock_assert(s); | ||
20 | |||
35 | cond_resched(); | 21 | cond_resched(); |
36 | 22 | ||
37 | *bhp = bh = sb_bread(s, secno); | 23 | *bhp = bh = sb_bread(s, secno); |
@@ -50,6 +36,8 @@ void *hpfs_get_sector(struct super_block *s, unsigned secno, struct buffer_head | |||
50 | struct buffer_head *bh; | 36 | struct buffer_head *bh; |
51 | /*return hpfs_map_sector(s, secno, bhp, 0);*/ | 37 | /*return hpfs_map_sector(s, secno, bhp, 0);*/ |
52 | 38 | ||
39 | hpfs_lock_assert(s); | ||
40 | |||
53 | cond_resched(); | 41 | cond_resched(); |
54 | 42 | ||
55 | if ((*bhp = bh = sb_getblk(s, secno)) != NULL) { | 43 | if ((*bhp = bh = sb_getblk(s, secno)) != NULL) { |
@@ -70,6 +58,8 @@ void *hpfs_map_4sectors(struct super_block *s, unsigned secno, struct quad_buffe | |||
70 | struct buffer_head *bh; | 58 | struct buffer_head *bh; |
71 | char *data; | 59 | char *data; |
72 | 60 | ||
61 | hpfs_lock_assert(s); | ||
62 | |||
73 | cond_resched(); | 63 | cond_resched(); |
74 | 64 | ||
75 | if (secno & 3) { | 65 | if (secno & 3) { |
@@ -125,6 +115,8 @@ void *hpfs_get_4sectors(struct super_block *s, unsigned secno, | |||
125 | { | 115 | { |
126 | cond_resched(); | 116 | cond_resched(); |
127 | 117 | ||
118 | hpfs_lock_assert(s); | ||
119 | |||
128 | if (secno & 3) { | 120 | if (secno & 3) { |
129 | printk("HPFS: hpfs_get_4sectors: unaligned read\n"); | 121 | printk("HPFS: hpfs_get_4sectors: unaligned read\n"); |
130 | return NULL; | 122 | return NULL; |
diff --git a/fs/hpfs/dir.c b/fs/hpfs/dir.c index b3d7c0ddb609..f46ae025bfb5 100644 --- a/fs/hpfs/dir.c +++ b/fs/hpfs/dir.c | |||
@@ -88,9 +88,9 @@ static int hpfs_readdir(struct file *filp, void *dirent, filldir_t filldir) | |||
88 | hpfs_error(inode->i_sb, "not a directory, fnode %08lx", | 88 | hpfs_error(inode->i_sb, "not a directory, fnode %08lx", |
89 | (unsigned long)inode->i_ino); | 89 | (unsigned long)inode->i_ino); |
90 | } | 90 | } |
91 | if (hpfs_inode->i_dno != fno->u.external[0].disk_secno) { | 91 | if (hpfs_inode->i_dno != le32_to_cpu(fno->u.external[0].disk_secno)) { |
92 | e = 1; | 92 | e = 1; |
93 | hpfs_error(inode->i_sb, "corrupted inode: i_dno == %08x, fnode -> dnode == %08x", hpfs_inode->i_dno, fno->u.external[0].disk_secno); | 93 | hpfs_error(inode->i_sb, "corrupted inode: i_dno == %08x, fnode -> dnode == %08x", hpfs_inode->i_dno, le32_to_cpu(fno->u.external[0].disk_secno)); |
94 | } | 94 | } |
95 | brelse(bh); | 95 | brelse(bh); |
96 | if (e) { | 96 | if (e) { |
@@ -156,7 +156,7 @@ static int hpfs_readdir(struct file *filp, void *dirent, filldir_t filldir) | |||
156 | goto again; | 156 | goto again; |
157 | } | 157 | } |
158 | tempname = hpfs_translate_name(inode->i_sb, de->name, de->namelen, lc, de->not_8x3); | 158 | tempname = hpfs_translate_name(inode->i_sb, de->name, de->namelen, lc, de->not_8x3); |
159 | if (filldir(dirent, tempname, de->namelen, old_pos, de->fnode, DT_UNKNOWN) < 0) { | 159 | if (filldir(dirent, tempname, de->namelen, old_pos, le32_to_cpu(de->fnode), DT_UNKNOWN) < 0) { |
160 | filp->f_pos = old_pos; | 160 | filp->f_pos = old_pos; |
161 | if (tempname != de->name) kfree(tempname); | 161 | if (tempname != de->name) kfree(tempname); |
162 | hpfs_brelse4(&qbh); | 162 | hpfs_brelse4(&qbh); |
@@ -221,7 +221,7 @@ struct dentry *hpfs_lookup(struct inode *dir, struct dentry *dentry, struct name | |||
221 | * Get inode number, what we're after. | 221 | * Get inode number, what we're after. |
222 | */ | 222 | */ |
223 | 223 | ||
224 | ino = de->fnode; | 224 | ino = le32_to_cpu(de->fnode); |
225 | 225 | ||
226 | /* | 226 | /* |
227 | * Go find or make an inode. | 227 | * Go find or make an inode. |
@@ -236,7 +236,7 @@ struct dentry *hpfs_lookup(struct inode *dir, struct dentry *dentry, struct name | |||
236 | hpfs_init_inode(result); | 236 | hpfs_init_inode(result); |
237 | if (de->directory) | 237 | if (de->directory) |
238 | hpfs_read_inode(result); | 238 | hpfs_read_inode(result); |
239 | else if (de->ea_size && hpfs_sb(dir->i_sb)->sb_eas) | 239 | else if (le32_to_cpu(de->ea_size) && hpfs_sb(dir->i_sb)->sb_eas) |
240 | hpfs_read_inode(result); | 240 | hpfs_read_inode(result); |
241 | else { | 241 | else { |
242 | result->i_mode |= S_IFREG; | 242 | result->i_mode |= S_IFREG; |
@@ -250,8 +250,6 @@ struct dentry *hpfs_lookup(struct inode *dir, struct dentry *dentry, struct name | |||
250 | hpfs_result = hpfs_i(result); | 250 | hpfs_result = hpfs_i(result); |
251 | if (!de->directory) hpfs_result->i_parent_dir = dir->i_ino; | 251 | if (!de->directory) hpfs_result->i_parent_dir = dir->i_ino; |
252 | 252 | ||
253 | hpfs_decide_conv(result, name, len); | ||
254 | |||
255 | if (de->has_acl || de->has_xtd_perm) if (!(dir->i_sb->s_flags & MS_RDONLY)) { | 253 | if (de->has_acl || de->has_xtd_perm) if (!(dir->i_sb->s_flags & MS_RDONLY)) { |
256 | hpfs_error(result->i_sb, "ACLs or XPERM found. This is probably HPFS386. This driver doesn't support it now. Send me some info on these structures"); | 254 | hpfs_error(result->i_sb, "ACLs or XPERM found. This is probably HPFS386. This driver doesn't support it now. Send me some info on these structures"); |
257 | goto bail1; | 255 | goto bail1; |
@@ -263,19 +261,19 @@ struct dentry *hpfs_lookup(struct inode *dir, struct dentry *dentry, struct name | |||
263 | */ | 261 | */ |
264 | 262 | ||
265 | if (!result->i_ctime.tv_sec) { | 263 | if (!result->i_ctime.tv_sec) { |
266 | if (!(result->i_ctime.tv_sec = local_to_gmt(dir->i_sb, de->creation_date))) | 264 | if (!(result->i_ctime.tv_sec = local_to_gmt(dir->i_sb, le32_to_cpu(de->creation_date)))) |
267 | result->i_ctime.tv_sec = 1; | 265 | result->i_ctime.tv_sec = 1; |
268 | result->i_ctime.tv_nsec = 0; | 266 | result->i_ctime.tv_nsec = 0; |
269 | result->i_mtime.tv_sec = local_to_gmt(dir->i_sb, de->write_date); | 267 | result->i_mtime.tv_sec = local_to_gmt(dir->i_sb, le32_to_cpu(de->write_date)); |
270 | result->i_mtime.tv_nsec = 0; | 268 | result->i_mtime.tv_nsec = 0; |
271 | result->i_atime.tv_sec = local_to_gmt(dir->i_sb, de->read_date); | 269 | result->i_atime.tv_sec = local_to_gmt(dir->i_sb, le32_to_cpu(de->read_date)); |
272 | result->i_atime.tv_nsec = 0; | 270 | result->i_atime.tv_nsec = 0; |
273 | hpfs_result->i_ea_size = de->ea_size; | 271 | hpfs_result->i_ea_size = le32_to_cpu(de->ea_size); |
274 | if (!hpfs_result->i_ea_mode && de->read_only) | 272 | if (!hpfs_result->i_ea_mode && de->read_only) |
275 | result->i_mode &= ~0222; | 273 | result->i_mode &= ~0222; |
276 | if (!de->directory) { | 274 | if (!de->directory) { |
277 | if (result->i_size == -1) { | 275 | if (result->i_size == -1) { |
278 | result->i_size = de->file_size; | 276 | result->i_size = le32_to_cpu(de->file_size); |
279 | result->i_data.a_ops = &hpfs_aops; | 277 | result->i_data.a_ops = &hpfs_aops; |
280 | hpfs_i(result)->mmu_private = result->i_size; | 278 | hpfs_i(result)->mmu_private = result->i_size; |
281 | /* | 279 | /* |
diff --git a/fs/hpfs/dnode.c b/fs/hpfs/dnode.c index 9b2ffadfc8c4..1e0e2ac30fd3 100644 --- a/fs/hpfs/dnode.c +++ b/fs/hpfs/dnode.c | |||
@@ -14,11 +14,11 @@ static loff_t get_pos(struct dnode *d, struct hpfs_dirent *fde) | |||
14 | struct hpfs_dirent *de_end = dnode_end_de(d); | 14 | struct hpfs_dirent *de_end = dnode_end_de(d); |
15 | int i = 1; | 15 | int i = 1; |
16 | for (de = dnode_first_de(d); de < de_end; de = de_next_de(de)) { | 16 | for (de = dnode_first_de(d); de < de_end; de = de_next_de(de)) { |
17 | if (de == fde) return ((loff_t) d->self << 4) | (loff_t)i; | 17 | if (de == fde) return ((loff_t) le32_to_cpu(d->self) << 4) | (loff_t)i; |
18 | i++; | 18 | i++; |
19 | } | 19 | } |
20 | printk("HPFS: get_pos: not_found\n"); | 20 | printk("HPFS: get_pos: not_found\n"); |
21 | return ((loff_t)d->self << 4) | (loff_t)1; | 21 | return ((loff_t)le32_to_cpu(d->self) << 4) | (loff_t)1; |
22 | } | 22 | } |
23 | 23 | ||
24 | void hpfs_add_pos(struct inode *inode, loff_t *pos) | 24 | void hpfs_add_pos(struct inode *inode, loff_t *pos) |
@@ -130,29 +130,30 @@ static void set_last_pointer(struct super_block *s, struct dnode *d, dnode_secno | |||
130 | { | 130 | { |
131 | struct hpfs_dirent *de; | 131 | struct hpfs_dirent *de; |
132 | if (!(de = dnode_last_de(d))) { | 132 | if (!(de = dnode_last_de(d))) { |
133 | hpfs_error(s, "set_last_pointer: empty dnode %08x", d->self); | 133 | hpfs_error(s, "set_last_pointer: empty dnode %08x", le32_to_cpu(d->self)); |
134 | return; | 134 | return; |
135 | } | 135 | } |
136 | if (hpfs_sb(s)->sb_chk) { | 136 | if (hpfs_sb(s)->sb_chk) { |
137 | if (de->down) { | 137 | if (de->down) { |
138 | hpfs_error(s, "set_last_pointer: dnode %08x has already last pointer %08x", | 138 | hpfs_error(s, "set_last_pointer: dnode %08x has already last pointer %08x", |
139 | d->self, de_down_pointer(de)); | 139 | le32_to_cpu(d->self), de_down_pointer(de)); |
140 | return; | 140 | return; |
141 | } | 141 | } |
142 | if (de->length != 32) { | 142 | if (le16_to_cpu(de->length) != 32) { |
143 | hpfs_error(s, "set_last_pointer: bad last dirent in dnode %08x", d->self); | 143 | hpfs_error(s, "set_last_pointer: bad last dirent in dnode %08x", le32_to_cpu(d->self)); |
144 | return; | 144 | return; |
145 | } | 145 | } |
146 | } | 146 | } |
147 | if (ptr) { | 147 | if (ptr) { |
148 | if ((d->first_free += 4) > 2048) { | 148 | d->first_free = cpu_to_le32(le32_to_cpu(d->first_free) + 4); |
149 | hpfs_error(s,"set_last_pointer: too long dnode %08x", d->self); | 149 | if (le32_to_cpu(d->first_free) > 2048) { |
150 | d->first_free -= 4; | 150 | hpfs_error(s, "set_last_pointer: too long dnode %08x", le32_to_cpu(d->self)); |
151 | d->first_free = cpu_to_le32(le32_to_cpu(d->first_free) - 4); | ||
151 | return; | 152 | return; |
152 | } | 153 | } |
153 | de->length = 36; | 154 | de->length = cpu_to_le16(36); |
154 | de->down = 1; | 155 | de->down = 1; |
155 | *(dnode_secno *)((char *)de + 32) = ptr; | 156 | *(dnode_secno *)((char *)de + 32) = cpu_to_le32(ptr); |
156 | } | 157 | } |
157 | } | 158 | } |
158 | 159 | ||
@@ -168,7 +169,7 @@ struct hpfs_dirent *hpfs_add_de(struct super_block *s, struct dnode *d, | |||
168 | for (de = dnode_first_de(d); de < de_end; de = de_next_de(de)) { | 169 | for (de = dnode_first_de(d); de < de_end; de = de_next_de(de)) { |
169 | int c = hpfs_compare_names(s, name, namelen, de->name, de->namelen, de->last); | 170 | int c = hpfs_compare_names(s, name, namelen, de->name, de->namelen, de->last); |
170 | if (!c) { | 171 | if (!c) { |
171 | hpfs_error(s, "name (%c,%d) already exists in dnode %08x", *name, namelen, d->self); | 172 | hpfs_error(s, "name (%c,%d) already exists in dnode %08x", *name, namelen, le32_to_cpu(d->self)); |
172 | return NULL; | 173 | return NULL; |
173 | } | 174 | } |
174 | if (c < 0) break; | 175 | if (c < 0) break; |
@@ -176,15 +177,14 @@ struct hpfs_dirent *hpfs_add_de(struct super_block *s, struct dnode *d, | |||
176 | memmove((char *)de + d_size, de, (char *)de_end - (char *)de); | 177 | memmove((char *)de + d_size, de, (char *)de_end - (char *)de); |
177 | memset(de, 0, d_size); | 178 | memset(de, 0, d_size); |
178 | if (down_ptr) { | 179 | if (down_ptr) { |
179 | *(int *)((char *)de + d_size - 4) = down_ptr; | 180 | *(dnode_secno *)((char *)de + d_size - 4) = cpu_to_le32(down_ptr); |
180 | de->down = 1; | 181 | de->down = 1; |
181 | } | 182 | } |
182 | de->length = d_size; | 183 | de->length = cpu_to_le16(d_size); |
183 | if (down_ptr) de->down = 1; | ||
184 | de->not_8x3 = hpfs_is_name_long(name, namelen); | 184 | de->not_8x3 = hpfs_is_name_long(name, namelen); |
185 | de->namelen = namelen; | 185 | de->namelen = namelen; |
186 | memcpy(de->name, name, namelen); | 186 | memcpy(de->name, name, namelen); |
187 | d->first_free += d_size; | 187 | d->first_free = cpu_to_le32(le32_to_cpu(d->first_free) + d_size); |
188 | return de; | 188 | return de; |
189 | } | 189 | } |
190 | 190 | ||
@@ -194,25 +194,25 @@ static void hpfs_delete_de(struct super_block *s, struct dnode *d, | |||
194 | struct hpfs_dirent *de) | 194 | struct hpfs_dirent *de) |
195 | { | 195 | { |
196 | if (de->last) { | 196 | if (de->last) { |
197 | hpfs_error(s, "attempt to delete last dirent in dnode %08x", d->self); | 197 | hpfs_error(s, "attempt to delete last dirent in dnode %08x", le32_to_cpu(d->self)); |
198 | return; | 198 | return; |
199 | } | 199 | } |
200 | d->first_free -= de->length; | 200 | d->first_free = cpu_to_le32(le32_to_cpu(d->first_free) - le16_to_cpu(de->length)); |
201 | memmove(de, de_next_de(de), d->first_free + (char *)d - (char *)de); | 201 | memmove(de, de_next_de(de), le32_to_cpu(d->first_free) + (char *)d - (char *)de); |
202 | } | 202 | } |
203 | 203 | ||
204 | static void fix_up_ptrs(struct super_block *s, struct dnode *d) | 204 | static void fix_up_ptrs(struct super_block *s, struct dnode *d) |
205 | { | 205 | { |
206 | struct hpfs_dirent *de; | 206 | struct hpfs_dirent *de; |
207 | struct hpfs_dirent *de_end = dnode_end_de(d); | 207 | struct hpfs_dirent *de_end = dnode_end_de(d); |
208 | dnode_secno dno = d->self; | 208 | dnode_secno dno = le32_to_cpu(d->self); |
209 | for (de = dnode_first_de(d); de < de_end; de = de_next_de(de)) | 209 | for (de = dnode_first_de(d); de < de_end; de = de_next_de(de)) |
210 | if (de->down) { | 210 | if (de->down) { |
211 | struct quad_buffer_head qbh; | 211 | struct quad_buffer_head qbh; |
212 | struct dnode *dd; | 212 | struct dnode *dd; |
213 | if ((dd = hpfs_map_dnode(s, de_down_pointer(de), &qbh))) { | 213 | if ((dd = hpfs_map_dnode(s, de_down_pointer(de), &qbh))) { |
214 | if (dd->up != dno || dd->root_dnode) { | 214 | if (le32_to_cpu(dd->up) != dno || dd->root_dnode) { |
215 | dd->up = dno; | 215 | dd->up = cpu_to_le32(dno); |
216 | dd->root_dnode = 0; | 216 | dd->root_dnode = 0; |
217 | hpfs_mark_4buffers_dirty(&qbh); | 217 | hpfs_mark_4buffers_dirty(&qbh); |
218 | } | 218 | } |
@@ -262,7 +262,7 @@ static int hpfs_add_to_dnode(struct inode *i, dnode_secno dno, | |||
262 | kfree(nname); | 262 | kfree(nname); |
263 | return 1; | 263 | return 1; |
264 | } | 264 | } |
265 | if (d->first_free + de_size(namelen, down_ptr) <= 2048) { | 265 | if (le32_to_cpu(d->first_free) + de_size(namelen, down_ptr) <= 2048) { |
266 | loff_t t; | 266 | loff_t t; |
267 | copy_de(de=hpfs_add_de(i->i_sb, d, name, namelen, down_ptr), new_de); | 267 | copy_de(de=hpfs_add_de(i->i_sb, d, name, namelen, down_ptr), new_de); |
268 | t = get_pos(d, de); | 268 | t = get_pos(d, de); |
@@ -286,11 +286,11 @@ static int hpfs_add_to_dnode(struct inode *i, dnode_secno dno, | |||
286 | kfree(nname); | 286 | kfree(nname); |
287 | return 1; | 287 | return 1; |
288 | } | 288 | } |
289 | memcpy(nd, d, d->first_free); | 289 | memcpy(nd, d, le32_to_cpu(d->first_free)); |
290 | copy_de(de = hpfs_add_de(i->i_sb, nd, name, namelen, down_ptr), new_de); | 290 | copy_de(de = hpfs_add_de(i->i_sb, nd, name, namelen, down_ptr), new_de); |
291 | for_all_poss(i, hpfs_pos_ins, get_pos(nd, de), 1); | 291 | for_all_poss(i, hpfs_pos_ins, get_pos(nd, de), 1); |
292 | h = ((char *)dnode_last_de(nd) - (char *)nd) / 2 + 10; | 292 | h = ((char *)dnode_last_de(nd) - (char *)nd) / 2 + 10; |
293 | if (!(ad = hpfs_alloc_dnode(i->i_sb, d->up, &adno, &qbh1, 0))) { | 293 | if (!(ad = hpfs_alloc_dnode(i->i_sb, le32_to_cpu(d->up), &adno, &qbh1))) { |
294 | hpfs_error(i->i_sb, "unable to alloc dnode - dnode tree will be corrupted"); | 294 | hpfs_error(i->i_sb, "unable to alloc dnode - dnode tree will be corrupted"); |
295 | hpfs_brelse4(&qbh); | 295 | hpfs_brelse4(&qbh); |
296 | kfree(nd); | 296 | kfree(nd); |
@@ -313,20 +313,21 @@ static int hpfs_add_to_dnode(struct inode *i, dnode_secno dno, | |||
313 | down_ptr = adno; | 313 | down_ptr = adno; |
314 | set_last_pointer(i->i_sb, ad, de->down ? de_down_pointer(de) : 0); | 314 | set_last_pointer(i->i_sb, ad, de->down ? de_down_pointer(de) : 0); |
315 | de = de_next_de(de); | 315 | de = de_next_de(de); |
316 | memmove((char *)nd + 20, de, nd->first_free + (char *)nd - (char *)de); | 316 | memmove((char *)nd + 20, de, le32_to_cpu(nd->first_free) + (char *)nd - (char *)de); |
317 | nd->first_free -= (char *)de - (char *)nd - 20; | 317 | nd->first_free = cpu_to_le32(le32_to_cpu(nd->first_free) - ((char *)de - (char *)nd - 20)); |
318 | memcpy(d, nd, nd->first_free); | 318 | memcpy(d, nd, le32_to_cpu(nd->first_free)); |
319 | for_all_poss(i, hpfs_pos_del, (loff_t)dno << 4, pos); | 319 | for_all_poss(i, hpfs_pos_del, (loff_t)dno << 4, pos); |
320 | fix_up_ptrs(i->i_sb, ad); | 320 | fix_up_ptrs(i->i_sb, ad); |
321 | if (!d->root_dnode) { | 321 | if (!d->root_dnode) { |
322 | dno = ad->up = d->up; | 322 | ad->up = d->up; |
323 | dno = le32_to_cpu(ad->up); | ||
323 | hpfs_mark_4buffers_dirty(&qbh); | 324 | hpfs_mark_4buffers_dirty(&qbh); |
324 | hpfs_brelse4(&qbh); | 325 | hpfs_brelse4(&qbh); |
325 | hpfs_mark_4buffers_dirty(&qbh1); | 326 | hpfs_mark_4buffers_dirty(&qbh1); |
326 | hpfs_brelse4(&qbh1); | 327 | hpfs_brelse4(&qbh1); |
327 | goto go_up; | 328 | goto go_up; |
328 | } | 329 | } |
329 | if (!(rd = hpfs_alloc_dnode(i->i_sb, d->up, &rdno, &qbh2, 0))) { | 330 | if (!(rd = hpfs_alloc_dnode(i->i_sb, le32_to_cpu(d->up), &rdno, &qbh2))) { |
330 | hpfs_error(i->i_sb, "unable to alloc dnode - dnode tree will be corrupted"); | 331 | hpfs_error(i->i_sb, "unable to alloc dnode - dnode tree will be corrupted"); |
331 | hpfs_brelse4(&qbh); | 332 | hpfs_brelse4(&qbh); |
332 | hpfs_brelse4(&qbh1); | 333 | hpfs_brelse4(&qbh1); |
@@ -338,7 +339,7 @@ static int hpfs_add_to_dnode(struct inode *i, dnode_secno dno, | |||
338 | i->i_blocks += 4; | 339 | i->i_blocks += 4; |
339 | rd->root_dnode = 1; | 340 | rd->root_dnode = 1; |
340 | rd->up = d->up; | 341 | rd->up = d->up; |
341 | if (!(fnode = hpfs_map_fnode(i->i_sb, d->up, &bh))) { | 342 | if (!(fnode = hpfs_map_fnode(i->i_sb, le32_to_cpu(d->up), &bh))) { |
342 | hpfs_free_dnode(i->i_sb, rdno); | 343 | hpfs_free_dnode(i->i_sb, rdno); |
343 | hpfs_brelse4(&qbh); | 344 | hpfs_brelse4(&qbh); |
344 | hpfs_brelse4(&qbh1); | 345 | hpfs_brelse4(&qbh1); |
@@ -347,10 +348,11 @@ static int hpfs_add_to_dnode(struct inode *i, dnode_secno dno, | |||
347 | kfree(nname); | 348 | kfree(nname); |
348 | return 1; | 349 | return 1; |
349 | } | 350 | } |
350 | fnode->u.external[0].disk_secno = rdno; | 351 | fnode->u.external[0].disk_secno = cpu_to_le32(rdno); |
351 | mark_buffer_dirty(bh); | 352 | mark_buffer_dirty(bh); |
352 | brelse(bh); | 353 | brelse(bh); |
353 | d->up = ad->up = hpfs_i(i)->i_dno = rdno; | 354 | hpfs_i(i)->i_dno = rdno; |
355 | d->up = ad->up = cpu_to_le32(rdno); | ||
354 | d->root_dnode = ad->root_dnode = 0; | 356 | d->root_dnode = ad->root_dnode = 0; |
355 | hpfs_mark_4buffers_dirty(&qbh); | 357 | hpfs_mark_4buffers_dirty(&qbh); |
356 | hpfs_brelse4(&qbh); | 358 | hpfs_brelse4(&qbh); |
@@ -373,7 +375,7 @@ static int hpfs_add_to_dnode(struct inode *i, dnode_secno dno, | |||
373 | 375 | ||
374 | int hpfs_add_dirent(struct inode *i, | 376 | int hpfs_add_dirent(struct inode *i, |
375 | const unsigned char *name, unsigned namelen, | 377 | const unsigned char *name, unsigned namelen, |
376 | struct hpfs_dirent *new_de, int cdepth) | 378 | struct hpfs_dirent *new_de) |
377 | { | 379 | { |
378 | struct hpfs_inode_info *hpfs_inode = hpfs_i(i); | 380 | struct hpfs_inode_info *hpfs_inode = hpfs_i(i); |
379 | struct dnode *d; | 381 | struct dnode *d; |
@@ -403,7 +405,6 @@ int hpfs_add_dirent(struct inode *i, | |||
403 | } | 405 | } |
404 | } | 406 | } |
405 | hpfs_brelse4(&qbh); | 407 | hpfs_brelse4(&qbh); |
406 | if (!cdepth) hpfs_lock_creation(i->i_sb); | ||
407 | if (hpfs_check_free_dnodes(i->i_sb, FREE_DNODES_ADD)) { | 408 | if (hpfs_check_free_dnodes(i->i_sb, FREE_DNODES_ADD)) { |
408 | c = 1; | 409 | c = 1; |
409 | goto ret; | 410 | goto ret; |
@@ -411,7 +412,6 @@ int hpfs_add_dirent(struct inode *i, | |||
411 | i->i_version++; | 412 | i->i_version++; |
412 | c = hpfs_add_to_dnode(i, dno, name, namelen, new_de, 0); | 413 | c = hpfs_add_to_dnode(i, dno, name, namelen, new_de, 0); |
413 | ret: | 414 | ret: |
414 | if (!cdepth) hpfs_unlock_creation(i->i_sb); | ||
415 | return c; | 415 | return c; |
416 | } | 416 | } |
417 | 417 | ||
@@ -437,9 +437,9 @@ static secno move_to_top(struct inode *i, dnode_secno from, dnode_secno to) | |||
437 | return 0; | 437 | return 0; |
438 | if (!(dnode = hpfs_map_dnode(i->i_sb, dno, &qbh))) return 0; | 438 | if (!(dnode = hpfs_map_dnode(i->i_sb, dno, &qbh))) return 0; |
439 | if (hpfs_sb(i->i_sb)->sb_chk) { | 439 | if (hpfs_sb(i->i_sb)->sb_chk) { |
440 | if (dnode->up != chk_up) { | 440 | if (le32_to_cpu(dnode->up) != chk_up) { |
441 | hpfs_error(i->i_sb, "move_to_top: up pointer from %08x should be %08x, is %08x", | 441 | hpfs_error(i->i_sb, "move_to_top: up pointer from %08x should be %08x, is %08x", |
442 | dno, chk_up, dnode->up); | 442 | dno, chk_up, le32_to_cpu(dnode->up)); |
443 | hpfs_brelse4(&qbh); | 443 | hpfs_brelse4(&qbh); |
444 | return 0; | 444 | return 0; |
445 | } | 445 | } |
@@ -455,7 +455,7 @@ static secno move_to_top(struct inode *i, dnode_secno from, dnode_secno to) | |||
455 | hpfs_brelse4(&qbh); | 455 | hpfs_brelse4(&qbh); |
456 | } | 456 | } |
457 | while (!(de = dnode_pre_last_de(dnode))) { | 457 | while (!(de = dnode_pre_last_de(dnode))) { |
458 | dnode_secno up = dnode->up; | 458 | dnode_secno up = le32_to_cpu(dnode->up); |
459 | hpfs_brelse4(&qbh); | 459 | hpfs_brelse4(&qbh); |
460 | hpfs_free_dnode(i->i_sb, dno); | 460 | hpfs_free_dnode(i->i_sb, dno); |
461 | i->i_size -= 2048; | 461 | i->i_size -= 2048; |
@@ -474,8 +474,8 @@ static secno move_to_top(struct inode *i, dnode_secno from, dnode_secno to) | |||
474 | hpfs_brelse4(&qbh); | 474 | hpfs_brelse4(&qbh); |
475 | return 0; | 475 | return 0; |
476 | } | 476 | } |
477 | dnode->first_free -= 4; | 477 | dnode->first_free = cpu_to_le32(le32_to_cpu(dnode->first_free) - 4); |
478 | de->length -= 4; | 478 | de->length = cpu_to_le16(le16_to_cpu(de->length) - 4); |
479 | de->down = 0; | 479 | de->down = 0; |
480 | hpfs_mark_4buffers_dirty(&qbh); | 480 | hpfs_mark_4buffers_dirty(&qbh); |
481 | dno = up; | 481 | dno = up; |
@@ -483,12 +483,12 @@ static secno move_to_top(struct inode *i, dnode_secno from, dnode_secno to) | |||
483 | t = get_pos(dnode, de); | 483 | t = get_pos(dnode, de); |
484 | for_all_poss(i, hpfs_pos_subst, t, 4); | 484 | for_all_poss(i, hpfs_pos_subst, t, 4); |
485 | for_all_poss(i, hpfs_pos_subst, t + 1, 5); | 485 | for_all_poss(i, hpfs_pos_subst, t + 1, 5); |
486 | if (!(nde = kmalloc(de->length, GFP_NOFS))) { | 486 | if (!(nde = kmalloc(le16_to_cpu(de->length), GFP_NOFS))) { |
487 | hpfs_error(i->i_sb, "out of memory for dirent - directory will be corrupted"); | 487 | hpfs_error(i->i_sb, "out of memory for dirent - directory will be corrupted"); |
488 | hpfs_brelse4(&qbh); | 488 | hpfs_brelse4(&qbh); |
489 | return 0; | 489 | return 0; |
490 | } | 490 | } |
491 | memcpy(nde, de, de->length); | 491 | memcpy(nde, de, le16_to_cpu(de->length)); |
492 | ddno = de->down ? de_down_pointer(de) : 0; | 492 | ddno = de->down ? de_down_pointer(de) : 0; |
493 | hpfs_delete_de(i->i_sb, dnode, de); | 493 | hpfs_delete_de(i->i_sb, dnode, de); |
494 | set_last_pointer(i->i_sb, dnode, ddno); | 494 | set_last_pointer(i->i_sb, dnode, ddno); |
@@ -517,11 +517,11 @@ static void delete_empty_dnode(struct inode *i, dnode_secno dno) | |||
517 | try_it_again: | 517 | try_it_again: |
518 | if (hpfs_stop_cycles(i->i_sb, dno, &c1, &c2, "delete_empty_dnode")) return; | 518 | if (hpfs_stop_cycles(i->i_sb, dno, &c1, &c2, "delete_empty_dnode")) return; |
519 | if (!(dnode = hpfs_map_dnode(i->i_sb, dno, &qbh))) return; | 519 | if (!(dnode = hpfs_map_dnode(i->i_sb, dno, &qbh))) return; |
520 | if (dnode->first_free > 56) goto end; | 520 | if (le32_to_cpu(dnode->first_free) > 56) goto end; |
521 | if (dnode->first_free == 52 || dnode->first_free == 56) { | 521 | if (le32_to_cpu(dnode->first_free) == 52 || le32_to_cpu(dnode->first_free) == 56) { |
522 | struct hpfs_dirent *de_end; | 522 | struct hpfs_dirent *de_end; |
523 | int root = dnode->root_dnode; | 523 | int root = dnode->root_dnode; |
524 | up = dnode->up; | 524 | up = le32_to_cpu(dnode->up); |
525 | de = dnode_first_de(dnode); | 525 | de = dnode_first_de(dnode); |
526 | down = de->down ? de_down_pointer(de) : 0; | 526 | down = de->down ? de_down_pointer(de) : 0; |
527 | if (hpfs_sb(i->i_sb)->sb_chk) if (root && !down) { | 527 | if (hpfs_sb(i->i_sb)->sb_chk) if (root && !down) { |
@@ -545,13 +545,13 @@ static void delete_empty_dnode(struct inode *i, dnode_secno dno) | |||
545 | return; | 545 | return; |
546 | } | 546 | } |
547 | if ((d1 = hpfs_map_dnode(i->i_sb, down, &qbh1))) { | 547 | if ((d1 = hpfs_map_dnode(i->i_sb, down, &qbh1))) { |
548 | d1->up = up; | 548 | d1->up = cpu_to_le32(up); |
549 | d1->root_dnode = 1; | 549 | d1->root_dnode = 1; |
550 | hpfs_mark_4buffers_dirty(&qbh1); | 550 | hpfs_mark_4buffers_dirty(&qbh1); |
551 | hpfs_brelse4(&qbh1); | 551 | hpfs_brelse4(&qbh1); |
552 | } | 552 | } |
553 | if ((fnode = hpfs_map_fnode(i->i_sb, up, &bh))) { | 553 | if ((fnode = hpfs_map_fnode(i->i_sb, up, &bh))) { |
554 | fnode->u.external[0].disk_secno = down; | 554 | fnode->u.external[0].disk_secno = cpu_to_le32(down); |
555 | mark_buffer_dirty(bh); | 555 | mark_buffer_dirty(bh); |
556 | brelse(bh); | 556 | brelse(bh); |
557 | } | 557 | } |
@@ -570,22 +570,22 @@ static void delete_empty_dnode(struct inode *i, dnode_secno dno) | |||
570 | for_all_poss(i, hpfs_pos_subst, ((loff_t)dno << 4) | 1, ((loff_t)up << 4) | p); | 570 | for_all_poss(i, hpfs_pos_subst, ((loff_t)dno << 4) | 1, ((loff_t)up << 4) | p); |
571 | if (!down) { | 571 | if (!down) { |
572 | de->down = 0; | 572 | de->down = 0; |
573 | de->length -= 4; | 573 | de->length = cpu_to_le16(le16_to_cpu(de->length) - 4); |
574 | dnode->first_free -= 4; | 574 | dnode->first_free = cpu_to_le32(le32_to_cpu(dnode->first_free) - 4); |
575 | memmove(de_next_de(de), (char *)de_next_de(de) + 4, | 575 | memmove(de_next_de(de), (char *)de_next_de(de) + 4, |
576 | (char *)dnode + dnode->first_free - (char *)de_next_de(de)); | 576 | (char *)dnode + le32_to_cpu(dnode->first_free) - (char *)de_next_de(de)); |
577 | } else { | 577 | } else { |
578 | struct dnode *d1; | 578 | struct dnode *d1; |
579 | struct quad_buffer_head qbh1; | 579 | struct quad_buffer_head qbh1; |
580 | *(dnode_secno *) ((void *) de + de->length - 4) = down; | 580 | *(dnode_secno *) ((void *) de + le16_to_cpu(de->length) - 4) = down; |
581 | if ((d1 = hpfs_map_dnode(i->i_sb, down, &qbh1))) { | 581 | if ((d1 = hpfs_map_dnode(i->i_sb, down, &qbh1))) { |
582 | d1->up = up; | 582 | d1->up = cpu_to_le32(up); |
583 | hpfs_mark_4buffers_dirty(&qbh1); | 583 | hpfs_mark_4buffers_dirty(&qbh1); |
584 | hpfs_brelse4(&qbh1); | 584 | hpfs_brelse4(&qbh1); |
585 | } | 585 | } |
586 | } | 586 | } |
587 | } else { | 587 | } else { |
588 | hpfs_error(i->i_sb, "delete_empty_dnode: dnode %08x, first_free == %03x", dno, dnode->first_free); | 588 | hpfs_error(i->i_sb, "delete_empty_dnode: dnode %08x, first_free == %03x", dno, le32_to_cpu(dnode->first_free)); |
589 | goto end; | 589 | goto end; |
590 | } | 590 | } |
591 | 591 | ||
@@ -596,18 +596,18 @@ static void delete_empty_dnode(struct inode *i, dnode_secno dno) | |||
596 | struct quad_buffer_head qbh1; | 596 | struct quad_buffer_head qbh1; |
597 | if (!de_next->down) goto endm; | 597 | if (!de_next->down) goto endm; |
598 | ndown = de_down_pointer(de_next); | 598 | ndown = de_down_pointer(de_next); |
599 | if (!(de_cp = kmalloc(de->length, GFP_NOFS))) { | 599 | if (!(de_cp = kmalloc(le16_to_cpu(de->length), GFP_NOFS))) { |
600 | printk("HPFS: out of memory for dtree balancing\n"); | 600 | printk("HPFS: out of memory for dtree balancing\n"); |
601 | goto endm; | 601 | goto endm; |
602 | } | 602 | } |
603 | memcpy(de_cp, de, de->length); | 603 | memcpy(de_cp, de, le16_to_cpu(de->length)); |
604 | hpfs_delete_de(i->i_sb, dnode, de); | 604 | hpfs_delete_de(i->i_sb, dnode, de); |
605 | hpfs_mark_4buffers_dirty(&qbh); | 605 | hpfs_mark_4buffers_dirty(&qbh); |
606 | hpfs_brelse4(&qbh); | 606 | hpfs_brelse4(&qbh); |
607 | for_all_poss(i, hpfs_pos_subst, ((loff_t)up << 4) | p, 4); | 607 | for_all_poss(i, hpfs_pos_subst, ((loff_t)up << 4) | p, 4); |
608 | for_all_poss(i, hpfs_pos_del, ((loff_t)up << 4) | p, 1); | 608 | for_all_poss(i, hpfs_pos_del, ((loff_t)up << 4) | p, 1); |
609 | if (de_cp->down) if ((d1 = hpfs_map_dnode(i->i_sb, de_down_pointer(de_cp), &qbh1))) { | 609 | if (de_cp->down) if ((d1 = hpfs_map_dnode(i->i_sb, de_down_pointer(de_cp), &qbh1))) { |
610 | d1->up = ndown; | 610 | d1->up = cpu_to_le32(ndown); |
611 | hpfs_mark_4buffers_dirty(&qbh1); | 611 | hpfs_mark_4buffers_dirty(&qbh1); |
612 | hpfs_brelse4(&qbh1); | 612 | hpfs_brelse4(&qbh1); |
613 | } | 613 | } |
@@ -635,7 +635,7 @@ static void delete_empty_dnode(struct inode *i, dnode_secno dno) | |||
635 | struct hpfs_dirent *del = dnode_last_de(d1); | 635 | struct hpfs_dirent *del = dnode_last_de(d1); |
636 | dlp = del->down ? de_down_pointer(del) : 0; | 636 | dlp = del->down ? de_down_pointer(del) : 0; |
637 | if (!dlp && down) { | 637 | if (!dlp && down) { |
638 | if (d1->first_free > 2044) { | 638 | if (le32_to_cpu(d1->first_free) > 2044) { |
639 | if (hpfs_sb(i->i_sb)->sb_chk >= 2) { | 639 | if (hpfs_sb(i->i_sb)->sb_chk >= 2) { |
640 | printk("HPFS: warning: unbalanced dnode tree, see hpfs.txt 4 more info\n"); | 640 | printk("HPFS: warning: unbalanced dnode tree, see hpfs.txt 4 more info\n"); |
641 | printk("HPFS: warning: terminating balancing operation\n"); | 641 | printk("HPFS: warning: terminating balancing operation\n"); |
@@ -647,38 +647,38 @@ static void delete_empty_dnode(struct inode *i, dnode_secno dno) | |||
647 | printk("HPFS: warning: unbalanced dnode tree, see hpfs.txt 4 more info\n"); | 647 | printk("HPFS: warning: unbalanced dnode tree, see hpfs.txt 4 more info\n"); |
648 | printk("HPFS: warning: goin'on\n"); | 648 | printk("HPFS: warning: goin'on\n"); |
649 | } | 649 | } |
650 | del->length += 4; | 650 | del->length = cpu_to_le16(le16_to_cpu(del->length) + 4); |
651 | del->down = 1; | 651 | del->down = 1; |
652 | d1->first_free += 4; | 652 | d1->first_free = cpu_to_le32(le32_to_cpu(d1->first_free) + 4); |
653 | } | 653 | } |
654 | if (dlp && !down) { | 654 | if (dlp && !down) { |
655 | del->length -= 4; | 655 | del->length = cpu_to_le16(le16_to_cpu(del->length) - 4); |
656 | del->down = 0; | 656 | del->down = 0; |
657 | d1->first_free -= 4; | 657 | d1->first_free = cpu_to_le32(le32_to_cpu(d1->first_free) - 4); |
658 | } else if (down) | 658 | } else if (down) |
659 | *(dnode_secno *) ((void *) del + del->length - 4) = down; | 659 | *(dnode_secno *) ((void *) del + le16_to_cpu(del->length) - 4) = cpu_to_le32(down); |
660 | } else goto endm; | 660 | } else goto endm; |
661 | if (!(de_cp = kmalloc(de_prev->length, GFP_NOFS))) { | 661 | if (!(de_cp = kmalloc(le16_to_cpu(de_prev->length), GFP_NOFS))) { |
662 | printk("HPFS: out of memory for dtree balancing\n"); | 662 | printk("HPFS: out of memory for dtree balancing\n"); |
663 | hpfs_brelse4(&qbh1); | 663 | hpfs_brelse4(&qbh1); |
664 | goto endm; | 664 | goto endm; |
665 | } | 665 | } |
666 | hpfs_mark_4buffers_dirty(&qbh1); | 666 | hpfs_mark_4buffers_dirty(&qbh1); |
667 | hpfs_brelse4(&qbh1); | 667 | hpfs_brelse4(&qbh1); |
668 | memcpy(de_cp, de_prev, de_prev->length); | 668 | memcpy(de_cp, de_prev, le16_to_cpu(de_prev->length)); |
669 | hpfs_delete_de(i->i_sb, dnode, de_prev); | 669 | hpfs_delete_de(i->i_sb, dnode, de_prev); |
670 | if (!de_prev->down) { | 670 | if (!de_prev->down) { |
671 | de_prev->length += 4; | 671 | de_prev->length = cpu_to_le16(le16_to_cpu(de_prev->length) + 4); |
672 | de_prev->down = 1; | 672 | de_prev->down = 1; |
673 | dnode->first_free += 4; | 673 | dnode->first_free = cpu_to_le32(le32_to_cpu(dnode->first_free) + 4); |
674 | } | 674 | } |
675 | *(dnode_secno *) ((void *) de_prev + de_prev->length - 4) = ndown; | 675 | *(dnode_secno *) ((void *) de_prev + le16_to_cpu(de_prev->length) - 4) = cpu_to_le32(ndown); |
676 | hpfs_mark_4buffers_dirty(&qbh); | 676 | hpfs_mark_4buffers_dirty(&qbh); |
677 | hpfs_brelse4(&qbh); | 677 | hpfs_brelse4(&qbh); |
678 | for_all_poss(i, hpfs_pos_subst, ((loff_t)up << 4) | (p - 1), 4); | 678 | for_all_poss(i, hpfs_pos_subst, ((loff_t)up << 4) | (p - 1), 4); |
679 | for_all_poss(i, hpfs_pos_subst, ((loff_t)up << 4) | p, ((loff_t)up << 4) | (p - 1)); | 679 | for_all_poss(i, hpfs_pos_subst, ((loff_t)up << 4) | p, ((loff_t)up << 4) | (p - 1)); |
680 | if (down) if ((d1 = hpfs_map_dnode(i->i_sb, de_down_pointer(de), &qbh1))) { | 680 | if (down) if ((d1 = hpfs_map_dnode(i->i_sb, de_down_pointer(de), &qbh1))) { |
681 | d1->up = ndown; | 681 | d1->up = cpu_to_le32(ndown); |
682 | hpfs_mark_4buffers_dirty(&qbh1); | 682 | hpfs_mark_4buffers_dirty(&qbh1); |
683 | hpfs_brelse4(&qbh1); | 683 | hpfs_brelse4(&qbh1); |
684 | } | 684 | } |
@@ -701,7 +701,6 @@ int hpfs_remove_dirent(struct inode *i, dnode_secno dno, struct hpfs_dirent *de, | |||
701 | { | 701 | { |
702 | struct dnode *dnode = qbh->data; | 702 | struct dnode *dnode = qbh->data; |
703 | dnode_secno down = 0; | 703 | dnode_secno down = 0; |
704 | int lock = 0; | ||
705 | loff_t t; | 704 | loff_t t; |
706 | if (de->first || de->last) { | 705 | if (de->first || de->last) { |
707 | hpfs_error(i->i_sb, "hpfs_remove_dirent: attempt to delete first or last dirent in dnode %08x", dno); | 706 | hpfs_error(i->i_sb, "hpfs_remove_dirent: attempt to delete first or last dirent in dnode %08x", dno); |
@@ -710,11 +709,8 @@ int hpfs_remove_dirent(struct inode *i, dnode_secno dno, struct hpfs_dirent *de, | |||
710 | } | 709 | } |
711 | if (de->down) down = de_down_pointer(de); | 710 | if (de->down) down = de_down_pointer(de); |
712 | if (depth && (de->down || (de == dnode_first_de(dnode) && de_next_de(de)->last))) { | 711 | if (depth && (de->down || (de == dnode_first_de(dnode) && de_next_de(de)->last))) { |
713 | lock = 1; | ||
714 | hpfs_lock_creation(i->i_sb); | ||
715 | if (hpfs_check_free_dnodes(i->i_sb, FREE_DNODES_DEL)) { | 712 | if (hpfs_check_free_dnodes(i->i_sb, FREE_DNODES_DEL)) { |
716 | hpfs_brelse4(qbh); | 713 | hpfs_brelse4(qbh); |
717 | hpfs_unlock_creation(i->i_sb); | ||
718 | return 2; | 714 | return 2; |
719 | } | 715 | } |
720 | } | 716 | } |
@@ -727,11 +723,9 @@ int hpfs_remove_dirent(struct inode *i, dnode_secno dno, struct hpfs_dirent *de, | |||
727 | dnode_secno a = move_to_top(i, down, dno); | 723 | dnode_secno a = move_to_top(i, down, dno); |
728 | for_all_poss(i, hpfs_pos_subst, 5, t); | 724 | for_all_poss(i, hpfs_pos_subst, 5, t); |
729 | if (a) delete_empty_dnode(i, a); | 725 | if (a) delete_empty_dnode(i, a); |
730 | if (lock) hpfs_unlock_creation(i->i_sb); | ||
731 | return !a; | 726 | return !a; |
732 | } | 727 | } |
733 | delete_empty_dnode(i, dno); | 728 | delete_empty_dnode(i, dno); |
734 | if (lock) hpfs_unlock_creation(i->i_sb); | ||
735 | return 0; | 729 | return 0; |
736 | } | 730 | } |
737 | 731 | ||
@@ -751,8 +745,8 @@ void hpfs_count_dnodes(struct super_block *s, dnode_secno dno, int *n_dnodes, | |||
751 | ptr = 0; | 745 | ptr = 0; |
752 | go_up: | 746 | go_up: |
753 | if (!(dnode = hpfs_map_dnode(s, dno, &qbh))) return; | 747 | if (!(dnode = hpfs_map_dnode(s, dno, &qbh))) return; |
754 | if (hpfs_sb(s)->sb_chk) if (odno && odno != -1 && dnode->up != odno) | 748 | if (hpfs_sb(s)->sb_chk) if (odno && odno != -1 && le32_to_cpu(dnode->up) != odno) |
755 | hpfs_error(s, "hpfs_count_dnodes: bad up pointer; dnode %08x, down %08x points to %08x", odno, dno, dnode->up); | 749 | hpfs_error(s, "hpfs_count_dnodes: bad up pointer; dnode %08x, down %08x points to %08x", odno, dno, le32_to_cpu(dnode->up)); |
756 | de = dnode_first_de(dnode); | 750 | de = dnode_first_de(dnode); |
757 | if (ptr) while(1) { | 751 | if (ptr) while(1) { |
758 | if (de->down) if (de_down_pointer(de) == ptr) goto process_de; | 752 | if (de->down) if (de_down_pointer(de) == ptr) goto process_de; |
@@ -776,7 +770,7 @@ void hpfs_count_dnodes(struct super_block *s, dnode_secno dno, int *n_dnodes, | |||
776 | if (!de->first && !de->last && n_items) (*n_items)++; | 770 | if (!de->first && !de->last && n_items) (*n_items)++; |
777 | if ((de = de_next_de(de)) < dnode_end_de(dnode)) goto next_de; | 771 | if ((de = de_next_de(de)) < dnode_end_de(dnode)) goto next_de; |
778 | ptr = dno; | 772 | ptr = dno; |
779 | dno = dnode->up; | 773 | dno = le32_to_cpu(dnode->up); |
780 | if (dnode->root_dnode) { | 774 | if (dnode->root_dnode) { |
781 | hpfs_brelse4(&qbh); | 775 | hpfs_brelse4(&qbh); |
782 | return; | 776 | return; |
@@ -824,8 +818,8 @@ dnode_secno hpfs_de_as_down_as_possible(struct super_block *s, dnode_secno dno) | |||
824 | return d; | 818 | return d; |
825 | if (!(de = map_nth_dirent(s, d, 1, &qbh, NULL))) return dno; | 819 | if (!(de = map_nth_dirent(s, d, 1, &qbh, NULL))) return dno; |
826 | if (hpfs_sb(s)->sb_chk) | 820 | if (hpfs_sb(s)->sb_chk) |
827 | if (up && ((struct dnode *)qbh.data)->up != up) | 821 | if (up && le32_to_cpu(((struct dnode *)qbh.data)->up) != up) |
828 | hpfs_error(s, "hpfs_de_as_down_as_possible: bad up pointer; dnode %08x, down %08x points to %08x", up, d, ((struct dnode *)qbh.data)->up); | 822 | hpfs_error(s, "hpfs_de_as_down_as_possible: bad up pointer; dnode %08x, down %08x points to %08x", up, d, le32_to_cpu(((struct dnode *)qbh.data)->up)); |
829 | if (!de->down) { | 823 | if (!de->down) { |
830 | hpfs_brelse4(&qbh); | 824 | hpfs_brelse4(&qbh); |
831 | return d; | 825 | return d; |
@@ -874,7 +868,7 @@ struct hpfs_dirent *map_pos_dirent(struct inode *inode, loff_t *posp, | |||
874 | /* Going up */ | 868 | /* Going up */ |
875 | if (dnode->root_dnode) goto bail; | 869 | if (dnode->root_dnode) goto bail; |
876 | 870 | ||
877 | if (!(up_dnode = hpfs_map_dnode(inode->i_sb, dnode->up, &qbh0))) | 871 | if (!(up_dnode = hpfs_map_dnode(inode->i_sb, le32_to_cpu(dnode->up), &qbh0))) |
878 | goto bail; | 872 | goto bail; |
879 | 873 | ||
880 | end_up_de = dnode_end_de(up_dnode); | 874 | end_up_de = dnode_end_de(up_dnode); |
@@ -882,16 +876,16 @@ struct hpfs_dirent *map_pos_dirent(struct inode *inode, loff_t *posp, | |||
882 | for (up_de = dnode_first_de(up_dnode); up_de < end_up_de; | 876 | for (up_de = dnode_first_de(up_dnode); up_de < end_up_de; |
883 | up_de = de_next_de(up_de)) { | 877 | up_de = de_next_de(up_de)) { |
884 | if (!(++c & 077)) hpfs_error(inode->i_sb, | 878 | if (!(++c & 077)) hpfs_error(inode->i_sb, |
885 | "map_pos_dirent: pos crossed dnode boundary; dnode = %08x", dnode->up); | 879 | "map_pos_dirent: pos crossed dnode boundary; dnode = %08x", le32_to_cpu(dnode->up)); |
886 | if (up_de->down && de_down_pointer(up_de) == dno) { | 880 | if (up_de->down && de_down_pointer(up_de) == dno) { |
887 | *posp = ((loff_t) dnode->up << 4) + c; | 881 | *posp = ((loff_t) le32_to_cpu(dnode->up) << 4) + c; |
888 | hpfs_brelse4(&qbh0); | 882 | hpfs_brelse4(&qbh0); |
889 | return de; | 883 | return de; |
890 | } | 884 | } |
891 | } | 885 | } |
892 | 886 | ||
893 | hpfs_error(inode->i_sb, "map_pos_dirent: pointer to dnode %08x not found in parent dnode %08x", | 887 | hpfs_error(inode->i_sb, "map_pos_dirent: pointer to dnode %08x not found in parent dnode %08x", |
894 | dno, dnode->up); | 888 | dno, le32_to_cpu(dnode->up)); |
895 | hpfs_brelse4(&qbh0); | 889 | hpfs_brelse4(&qbh0); |
896 | 890 | ||
897 | bail: | 891 | bail: |
@@ -1017,17 +1011,17 @@ struct hpfs_dirent *map_fnode_dirent(struct super_block *s, fnode_secno fno, | |||
1017 | /*name2[15] = 0xff;*/ | 1011 | /*name2[15] = 0xff;*/ |
1018 | name1len = 15; name2len = 256; | 1012 | name1len = 15; name2len = 256; |
1019 | } | 1013 | } |
1020 | if (!(upf = hpfs_map_fnode(s, f->up, &bh))) { | 1014 | if (!(upf = hpfs_map_fnode(s, le32_to_cpu(f->up), &bh))) { |
1021 | kfree(name2); | 1015 | kfree(name2); |
1022 | return NULL; | 1016 | return NULL; |
1023 | } | 1017 | } |
1024 | if (!upf->dirflag) { | 1018 | if (!upf->dirflag) { |
1025 | brelse(bh); | 1019 | brelse(bh); |
1026 | hpfs_error(s, "fnode %08x has non-directory parent %08x", fno, f->up); | 1020 | hpfs_error(s, "fnode %08x has non-directory parent %08x", fno, le32_to_cpu(f->up)); |
1027 | kfree(name2); | 1021 | kfree(name2); |
1028 | return NULL; | 1022 | return NULL; |
1029 | } | 1023 | } |
1030 | dno = upf->u.external[0].disk_secno; | 1024 | dno = le32_to_cpu(upf->u.external[0].disk_secno); |
1031 | brelse(bh); | 1025 | brelse(bh); |
1032 | go_down: | 1026 | go_down: |
1033 | downd = 0; | 1027 | downd = 0; |
@@ -1049,7 +1043,7 @@ struct hpfs_dirent *map_fnode_dirent(struct super_block *s, fnode_secno fno, | |||
1049 | return NULL; | 1043 | return NULL; |
1050 | } | 1044 | } |
1051 | next_de: | 1045 | next_de: |
1052 | if (de->fnode == fno) { | 1046 | if (le32_to_cpu(de->fnode) == fno) { |
1053 | kfree(name2); | 1047 | kfree(name2); |
1054 | return de; | 1048 | return de; |
1055 | } | 1049 | } |
@@ -1065,7 +1059,7 @@ struct hpfs_dirent *map_fnode_dirent(struct super_block *s, fnode_secno fno, | |||
1065 | goto go_down; | 1059 | goto go_down; |
1066 | } | 1060 | } |
1067 | f: | 1061 | f: |
1068 | if (de->fnode == fno) { | 1062 | if (le32_to_cpu(de->fnode) == fno) { |
1069 | kfree(name2); | 1063 | kfree(name2); |
1070 | return de; | 1064 | return de; |
1071 | } | 1065 | } |
@@ -1074,7 +1068,7 @@ struct hpfs_dirent *map_fnode_dirent(struct super_block *s, fnode_secno fno, | |||
1074 | if ((de = de_next_de(de)) < de_end) goto next_de; | 1068 | if ((de = de_next_de(de)) < de_end) goto next_de; |
1075 | if (d->root_dnode) goto not_found; | 1069 | if (d->root_dnode) goto not_found; |
1076 | downd = dno; | 1070 | downd = dno; |
1077 | dno = d->up; | 1071 | dno = le32_to_cpu(d->up); |
1078 | hpfs_brelse4(qbh); | 1072 | hpfs_brelse4(qbh); |
1079 | if (hpfs_sb(s)->sb_chk) | 1073 | if (hpfs_sb(s)->sb_chk) |
1080 | if (hpfs_stop_cycles(s, downd, &d1, &d2, "map_fnode_dirent #2")) { | 1074 | if (hpfs_stop_cycles(s, downd, &d1, &d2, "map_fnode_dirent #2")) { |
diff --git a/fs/hpfs/ea.c b/fs/hpfs/ea.c index 45e53d972b42..d8b84d113c89 100644 --- a/fs/hpfs/ea.c +++ b/fs/hpfs/ea.c | |||
@@ -24,7 +24,7 @@ void hpfs_ea_ext_remove(struct super_block *s, secno a, int ano, unsigned len) | |||
24 | } | 24 | } |
25 | if (hpfs_ea_read(s, a, ano, pos, 4, ex)) return; | 25 | if (hpfs_ea_read(s, a, ano, pos, 4, ex)) return; |
26 | if (ea->indirect) { | 26 | if (ea->indirect) { |
27 | if (ea->valuelen != 8) { | 27 | if (ea_valuelen(ea) != 8) { |
28 | hpfs_error(s, "ea->indirect set while ea->valuelen!=8, %s %08x, pos %08x", | 28 | hpfs_error(s, "ea->indirect set while ea->valuelen!=8, %s %08x, pos %08x", |
29 | ano ? "anode" : "sectors", a, pos); | 29 | ano ? "anode" : "sectors", a, pos); |
30 | return; | 30 | return; |
@@ -33,7 +33,7 @@ void hpfs_ea_ext_remove(struct super_block *s, secno a, int ano, unsigned len) | |||
33 | return; | 33 | return; |
34 | hpfs_ea_remove(s, ea_sec(ea), ea->anode, ea_len(ea)); | 34 | hpfs_ea_remove(s, ea_sec(ea), ea->anode, ea_len(ea)); |
35 | } | 35 | } |
36 | pos += ea->namelen + ea->valuelen + 5; | 36 | pos += ea->namelen + ea_valuelen(ea) + 5; |
37 | } | 37 | } |
38 | if (!ano) hpfs_free_sectors(s, a, (len+511) >> 9); | 38 | if (!ano) hpfs_free_sectors(s, a, (len+511) >> 9); |
39 | else { | 39 | else { |
@@ -76,24 +76,24 @@ int hpfs_read_ea(struct super_block *s, struct fnode *fnode, char *key, | |||
76 | unsigned pos; | 76 | unsigned pos; |
77 | int ano, len; | 77 | int ano, len; |
78 | secno a; | 78 | secno a; |
79 | char ex[4 + 255 + 1 + 8]; | ||
79 | struct extended_attribute *ea; | 80 | struct extended_attribute *ea; |
80 | struct extended_attribute *ea_end = fnode_end_ea(fnode); | 81 | struct extended_attribute *ea_end = fnode_end_ea(fnode); |
81 | for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) | 82 | for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea)) |
82 | if (!strcmp(ea->name, key)) { | 83 | if (!strcmp(ea->name, key)) { |
83 | if (ea->indirect) | 84 | if (ea->indirect) |
84 | goto indirect; | 85 | goto indirect; |
85 | if (ea->valuelen >= size) | 86 | if (ea_valuelen(ea) >= size) |
86 | return -EINVAL; | 87 | return -EINVAL; |
87 | memcpy(buf, ea_data(ea), ea->valuelen); | 88 | memcpy(buf, ea_data(ea), ea_valuelen(ea)); |
88 | buf[ea->valuelen] = 0; | 89 | buf[ea_valuelen(ea)] = 0; |
89 | return 0; | 90 | return 0; |
90 | } | 91 | } |
91 | a = fnode->ea_secno; | 92 | a = le32_to_cpu(fnode->ea_secno); |
92 | len = fnode->ea_size_l; | 93 | len = le32_to_cpu(fnode->ea_size_l); |
93 | ano = fnode->ea_anode; | 94 | ano = fnode->ea_anode; |
94 | pos = 0; | 95 | pos = 0; |
95 | while (pos < len) { | 96 | while (pos < len) { |
96 | char ex[4 + 255 + 1 + 8]; | ||
97 | ea = (struct extended_attribute *)ex; | 97 | ea = (struct extended_attribute *)ex; |
98 | if (pos + 4 > len) { | 98 | if (pos + 4 > len) { |
99 | hpfs_error(s, "EAs don't end correctly, %s %08x, len %08x", | 99 | hpfs_error(s, "EAs don't end correctly, %s %08x, len %08x", |
@@ -106,14 +106,14 @@ int hpfs_read_ea(struct super_block *s, struct fnode *fnode, char *key, | |||
106 | if (!strcmp(ea->name, key)) { | 106 | if (!strcmp(ea->name, key)) { |
107 | if (ea->indirect) | 107 | if (ea->indirect) |
108 | goto indirect; | 108 | goto indirect; |
109 | if (ea->valuelen >= size) | 109 | if (ea_valuelen(ea) >= size) |
110 | return -EINVAL; | 110 | return -EINVAL; |
111 | if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea->valuelen, buf)) | 111 | if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea_valuelen(ea), buf)) |
112 | return -EIO; | 112 | return -EIO; |
113 | buf[ea->valuelen] = 0; | 113 | buf[ea_valuelen(ea)] = 0; |
114 | return 0; | 114 | return 0; |
115 | } | 115 | } |
116 | pos += ea->namelen + ea->valuelen + 5; | 116 | pos += ea->namelen + ea_valuelen(ea) + 5; |
117 | } | 117 | } |
118 | return -ENOENT; | 118 | return -ENOENT; |
119 | indirect: | 119 | indirect: |
@@ -138,16 +138,16 @@ char *hpfs_get_ea(struct super_block *s, struct fnode *fnode, char *key, int *si | |||
138 | if (!strcmp(ea->name, key)) { | 138 | if (!strcmp(ea->name, key)) { |
139 | if (ea->indirect) | 139 | if (ea->indirect) |
140 | return get_indirect_ea(s, ea->anode, ea_sec(ea), *size = ea_len(ea)); | 140 | return get_indirect_ea(s, ea->anode, ea_sec(ea), *size = ea_len(ea)); |
141 | if (!(ret = kmalloc((*size = ea->valuelen) + 1, GFP_NOFS))) { | 141 | if (!(ret = kmalloc((*size = ea_valuelen(ea)) + 1, GFP_NOFS))) { |
142 | printk("HPFS: out of memory for EA\n"); | 142 | printk("HPFS: out of memory for EA\n"); |
143 | return NULL; | 143 | return NULL; |
144 | } | 144 | } |
145 | memcpy(ret, ea_data(ea), ea->valuelen); | 145 | memcpy(ret, ea_data(ea), ea_valuelen(ea)); |
146 | ret[ea->valuelen] = 0; | 146 | ret[ea_valuelen(ea)] = 0; |
147 | return ret; | 147 | return ret; |
148 | } | 148 | } |
149 | a = fnode->ea_secno; | 149 | a = le32_to_cpu(fnode->ea_secno); |
150 | len = fnode->ea_size_l; | 150 | len = le32_to_cpu(fnode->ea_size_l); |
151 | ano = fnode->ea_anode; | 151 | ano = fnode->ea_anode; |
152 | pos = 0; | 152 | pos = 0; |
153 | while (pos < len) { | 153 | while (pos < len) { |
@@ -164,18 +164,18 @@ char *hpfs_get_ea(struct super_block *s, struct fnode *fnode, char *key, int *si | |||
164 | if (!strcmp(ea->name, key)) { | 164 | if (!strcmp(ea->name, key)) { |
165 | if (ea->indirect) | 165 | if (ea->indirect) |
166 | return get_indirect_ea(s, ea->anode, ea_sec(ea), *size = ea_len(ea)); | 166 | return get_indirect_ea(s, ea->anode, ea_sec(ea), *size = ea_len(ea)); |
167 | if (!(ret = kmalloc((*size = ea->valuelen) + 1, GFP_NOFS))) { | 167 | if (!(ret = kmalloc((*size = ea_valuelen(ea)) + 1, GFP_NOFS))) { |
168 | printk("HPFS: out of memory for EA\n"); | 168 | printk("HPFS: out of memory for EA\n"); |
169 | return NULL; | 169 | return NULL; |
170 | } | 170 | } |
171 | if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea->valuelen, ret)) { | 171 | if (hpfs_ea_read(s, a, ano, pos + 4 + ea->namelen + 1, ea_valuelen(ea), ret)) { |
172 | kfree(ret); | 172 | kfree(ret); |
173 | return NULL; | 173 | return NULL; |
174 | } | 174 | } |
175 | ret[ea->valuelen] = 0; | 175 | ret[ea_valuelen(ea)] = 0; |
176 | return ret; | 176 | return ret; |
177 | } | 177 | } |
178 | pos += ea->namelen + ea->valuelen + 5; | 178 | pos += ea->namelen + ea_valuelen(ea) + 5; |
179 | } | 179 | } |
180 | return NULL; | 180 | return NULL; |
181 | } | 181 | } |
@@ -202,13 +202,13 @@ void hpfs_set_ea(struct inode *inode, struct fnode *fnode, const char *key, | |||
202 | if (ea->indirect) { | 202 | if (ea->indirect) { |
203 | if (ea_len(ea) == size) | 203 | if (ea_len(ea) == size) |
204 | set_indirect_ea(s, ea->anode, ea_sec(ea), data, size); | 204 | set_indirect_ea(s, ea->anode, ea_sec(ea), data, size); |
205 | } else if (ea->valuelen == size) { | 205 | } else if (ea_valuelen(ea) == size) { |
206 | memcpy(ea_data(ea), data, size); | 206 | memcpy(ea_data(ea), data, size); |
207 | } | 207 | } |
208 | return; | 208 | return; |
209 | } | 209 | } |
210 | a = fnode->ea_secno; | 210 | a = le32_to_cpu(fnode->ea_secno); |
211 | len = fnode->ea_size_l; | 211 | len = le32_to_cpu(fnode->ea_size_l); |
212 | ano = fnode->ea_anode; | 212 | ano = fnode->ea_anode; |
213 | pos = 0; | 213 | pos = 0; |
214 | while (pos < len) { | 214 | while (pos < len) { |
@@ -228,68 +228,70 @@ void hpfs_set_ea(struct inode *inode, struct fnode *fnode, const char *key, | |||
228 | set_indirect_ea(s, ea->anode, ea_sec(ea), data, size); | 228 | set_indirect_ea(s, ea->anode, ea_sec(ea), data, size); |
229 | } | 229 | } |
230 | else { | 230 | else { |
231 | if (ea->valuelen == size) | 231 | if (ea_valuelen(ea) == size) |
232 | hpfs_ea_write(s, a, ano, pos + 4 + ea->namelen + 1, size, data); | 232 | hpfs_ea_write(s, a, ano, pos + 4 + ea->namelen + 1, size, data); |
233 | } | 233 | } |
234 | return; | 234 | return; |
235 | } | 235 | } |
236 | pos += ea->namelen + ea->valuelen + 5; | 236 | pos += ea->namelen + ea_valuelen(ea) + 5; |
237 | } | 237 | } |
238 | if (!fnode->ea_offs) { | 238 | if (!le16_to_cpu(fnode->ea_offs)) { |
239 | /*if (fnode->ea_size_s) { | 239 | /*if (le16_to_cpu(fnode->ea_size_s)) { |
240 | hpfs_error(s, "fnode %08x: ea_size_s == %03x, ea_offs == 0", | 240 | hpfs_error(s, "fnode %08x: ea_size_s == %03x, ea_offs == 0", |
241 | inode->i_ino, fnode->ea_size_s); | 241 | inode->i_ino, le16_to_cpu(fnode->ea_size_s)); |
242 | return; | 242 | return; |
243 | }*/ | 243 | }*/ |
244 | fnode->ea_offs = 0xc4; | 244 | fnode->ea_offs = cpu_to_le16(0xc4); |
245 | } | 245 | } |
246 | if (fnode->ea_offs < 0xc4 || fnode->ea_offs + fnode->acl_size_s + fnode->ea_size_s > 0x200) { | 246 | if (le16_to_cpu(fnode->ea_offs) < 0xc4 || le16_to_cpu(fnode->ea_offs) + le16_to_cpu(fnode->acl_size_s) + le16_to_cpu(fnode->ea_size_s) > 0x200) { |
247 | hpfs_error(s, "fnode %08lx: ea_offs == %03x, ea_size_s == %03x", | 247 | hpfs_error(s, "fnode %08lx: ea_offs == %03x, ea_size_s == %03x", |
248 | (unsigned long)inode->i_ino, | 248 | (unsigned long)inode->i_ino, |
249 | fnode->ea_offs, fnode->ea_size_s); | 249 | le32_to_cpu(fnode->ea_offs), le16_to_cpu(fnode->ea_size_s)); |
250 | return; | 250 | return; |
251 | } | 251 | } |
252 | if ((fnode->ea_size_s || !fnode->ea_size_l) && | 252 | if ((le16_to_cpu(fnode->ea_size_s) || !le32_to_cpu(fnode->ea_size_l)) && |
253 | fnode->ea_offs + fnode->acl_size_s + fnode->ea_size_s + strlen(key) + size + 5 <= 0x200) { | 253 | le16_to_cpu(fnode->ea_offs) + le16_to_cpu(fnode->acl_size_s) + le16_to_cpu(fnode->ea_size_s) + strlen(key) + size + 5 <= 0x200) { |
254 | ea = fnode_end_ea(fnode); | 254 | ea = fnode_end_ea(fnode); |
255 | *(char *)ea = 0; | 255 | *(char *)ea = 0; |
256 | ea->namelen = strlen(key); | 256 | ea->namelen = strlen(key); |
257 | ea->valuelen = size; | 257 | ea->valuelen_lo = size; |
258 | ea->valuelen_hi = size >> 8; | ||
258 | strcpy(ea->name, key); | 259 | strcpy(ea->name, key); |
259 | memcpy(ea_data(ea), data, size); | 260 | memcpy(ea_data(ea), data, size); |
260 | fnode->ea_size_s += strlen(key) + size + 5; | 261 | fnode->ea_size_s = cpu_to_le16(le16_to_cpu(fnode->ea_size_s) + strlen(key) + size + 5); |
261 | goto ret; | 262 | goto ret; |
262 | } | 263 | } |
263 | /* Most the code here is 99.9993422% unused. I hope there are no bugs. | 264 | /* Most the code here is 99.9993422% unused. I hope there are no bugs. |
264 | But what .. HPFS.IFS has also bugs in ea management. */ | 265 | But what .. HPFS.IFS has also bugs in ea management. */ |
265 | if (fnode->ea_size_s && !fnode->ea_size_l) { | 266 | if (le16_to_cpu(fnode->ea_size_s) && !le32_to_cpu(fnode->ea_size_l)) { |
266 | secno n; | 267 | secno n; |
267 | struct buffer_head *bh; | 268 | struct buffer_head *bh; |
268 | char *data; | 269 | char *data; |
269 | if (!(n = hpfs_alloc_sector(s, fno, 1, 0, 1))) return; | 270 | if (!(n = hpfs_alloc_sector(s, fno, 1, 0))) return; |
270 | if (!(data = hpfs_get_sector(s, n, &bh))) { | 271 | if (!(data = hpfs_get_sector(s, n, &bh))) { |
271 | hpfs_free_sectors(s, n, 1); | 272 | hpfs_free_sectors(s, n, 1); |
272 | return; | 273 | return; |
273 | } | 274 | } |
274 | memcpy(data, fnode_ea(fnode), fnode->ea_size_s); | 275 | memcpy(data, fnode_ea(fnode), le16_to_cpu(fnode->ea_size_s)); |
275 | fnode->ea_size_l = fnode->ea_size_s; | 276 | fnode->ea_size_l = cpu_to_le32(le16_to_cpu(fnode->ea_size_s)); |
276 | fnode->ea_size_s = 0; | 277 | fnode->ea_size_s = cpu_to_le16(0); |
277 | fnode->ea_secno = n; | 278 | fnode->ea_secno = cpu_to_le32(n); |
278 | fnode->ea_anode = 0; | 279 | fnode->ea_anode = cpu_to_le32(0); |
279 | mark_buffer_dirty(bh); | 280 | mark_buffer_dirty(bh); |
280 | brelse(bh); | 281 | brelse(bh); |
281 | } | 282 | } |
282 | pos = fnode->ea_size_l + 5 + strlen(key) + size; | 283 | pos = le32_to_cpu(fnode->ea_size_l) + 5 + strlen(key) + size; |
283 | len = (fnode->ea_size_l + 511) >> 9; | 284 | len = (le32_to_cpu(fnode->ea_size_l) + 511) >> 9; |
284 | if (pos >= 30000) goto bail; | 285 | if (pos >= 30000) goto bail; |
285 | while (((pos + 511) >> 9) > len) { | 286 | while (((pos + 511) >> 9) > len) { |
286 | if (!len) { | 287 | if (!len) { |
287 | if (!(fnode->ea_secno = hpfs_alloc_sector(s, fno, 1, 0, 1))) | 288 | secno q = hpfs_alloc_sector(s, fno, 1, 0); |
288 | goto bail; | 289 | if (!q) goto bail; |
290 | fnode->ea_secno = cpu_to_le32(q); | ||
289 | fnode->ea_anode = 0; | 291 | fnode->ea_anode = 0; |
290 | len++; | 292 | len++; |
291 | } else if (!fnode->ea_anode) { | 293 | } else if (!fnode->ea_anode) { |
292 | if (hpfs_alloc_if_possible(s, fnode->ea_secno + len)) { | 294 | if (hpfs_alloc_if_possible(s, le32_to_cpu(fnode->ea_secno) + len)) { |
293 | len++; | 295 | len++; |
294 | } else { | 296 | } else { |
295 | /* Aargh... don't know how to create ea anodes :-( */ | 297 | /* Aargh... don't know how to create ea anodes :-( */ |
@@ -298,26 +300,26 @@ void hpfs_set_ea(struct inode *inode, struct fnode *fnode, const char *key, | |||
298 | anode_secno a_s; | 300 | anode_secno a_s; |
299 | if (!(anode = hpfs_alloc_anode(s, fno, &a_s, &bh))) | 301 | if (!(anode = hpfs_alloc_anode(s, fno, &a_s, &bh))) |
300 | goto bail; | 302 | goto bail; |
301 | anode->up = fno; | 303 | anode->up = cpu_to_le32(fno); |
302 | anode->btree.fnode_parent = 1; | 304 | anode->btree.fnode_parent = 1; |
303 | anode->btree.n_free_nodes--; | 305 | anode->btree.n_free_nodes--; |
304 | anode->btree.n_used_nodes++; | 306 | anode->btree.n_used_nodes++; |
305 | anode->btree.first_free += 12; | 307 | anode->btree.first_free = cpu_to_le16(le16_to_cpu(anode->btree.first_free) + 12); |
306 | anode->u.external[0].disk_secno = fnode->ea_secno; | 308 | anode->u.external[0].disk_secno = cpu_to_le32(le32_to_cpu(fnode->ea_secno)); |
307 | anode->u.external[0].file_secno = 0; | 309 | anode->u.external[0].file_secno = cpu_to_le32(0); |
308 | anode->u.external[0].length = len; | 310 | anode->u.external[0].length = cpu_to_le32(len); |
309 | mark_buffer_dirty(bh); | 311 | mark_buffer_dirty(bh); |
310 | brelse(bh); | 312 | brelse(bh); |
311 | fnode->ea_anode = 1; | 313 | fnode->ea_anode = 1; |
312 | fnode->ea_secno = a_s;*/ | 314 | fnode->ea_secno = cpu_to_le32(a_s);*/ |
313 | secno new_sec; | 315 | secno new_sec; |
314 | int i; | 316 | int i; |
315 | if (!(new_sec = hpfs_alloc_sector(s, fno, 1, 1 - ((pos + 511) >> 9), 1))) | 317 | if (!(new_sec = hpfs_alloc_sector(s, fno, 1, 1 - ((pos + 511) >> 9)))) |
316 | goto bail; | 318 | goto bail; |
317 | for (i = 0; i < len; i++) { | 319 | for (i = 0; i < len; i++) { |
318 | struct buffer_head *bh1, *bh2; | 320 | struct buffer_head *bh1, *bh2; |
319 | void *b1, *b2; | 321 | void *b1, *b2; |
320 | if (!(b1 = hpfs_map_sector(s, fnode->ea_secno + i, &bh1, len - i - 1))) { | 322 | if (!(b1 = hpfs_map_sector(s, le32_to_cpu(fnode->ea_secno) + i, &bh1, len - i - 1))) { |
321 | hpfs_free_sectors(s, new_sec, (pos + 511) >> 9); | 323 | hpfs_free_sectors(s, new_sec, (pos + 511) >> 9); |
322 | goto bail; | 324 | goto bail; |
323 | } | 325 | } |
@@ -331,13 +333,13 @@ void hpfs_set_ea(struct inode *inode, struct fnode *fnode, const char *key, | |||
331 | mark_buffer_dirty(bh2); | 333 | mark_buffer_dirty(bh2); |
332 | brelse(bh2); | 334 | brelse(bh2); |
333 | } | 335 | } |
334 | hpfs_free_sectors(s, fnode->ea_secno, len); | 336 | hpfs_free_sectors(s, le32_to_cpu(fnode->ea_secno), len); |
335 | fnode->ea_secno = new_sec; | 337 | fnode->ea_secno = cpu_to_le32(new_sec); |
336 | len = (pos + 511) >> 9; | 338 | len = (pos + 511) >> 9; |
337 | } | 339 | } |
338 | } | 340 | } |
339 | if (fnode->ea_anode) { | 341 | if (fnode->ea_anode) { |
340 | if (hpfs_add_sector_to_btree(s, fnode->ea_secno, | 342 | if (hpfs_add_sector_to_btree(s, le32_to_cpu(fnode->ea_secno), |
341 | 0, len) != -1) { | 343 | 0, len) != -1) { |
342 | len++; | 344 | len++; |
343 | } else { | 345 | } else { |
@@ -349,17 +351,17 @@ void hpfs_set_ea(struct inode *inode, struct fnode *fnode, const char *key, | |||
349 | h[1] = strlen(key); | 351 | h[1] = strlen(key); |
350 | h[2] = size & 0xff; | 352 | h[2] = size & 0xff; |
351 | h[3] = size >> 8; | 353 | h[3] = size >> 8; |
352 | if (hpfs_ea_write(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l, 4, h)) goto bail; | 354 | if (hpfs_ea_write(s, le32_to_cpu(fnode->ea_secno), fnode->ea_anode, le32_to_cpu(fnode->ea_size_l), 4, h)) goto bail; |
353 | if (hpfs_ea_write(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l + 4, h[1] + 1, key)) goto bail; | 355 | if (hpfs_ea_write(s, le32_to_cpu(fnode->ea_secno), fnode->ea_anode, le32_to_cpu(fnode->ea_size_l) + 4, h[1] + 1, key)) goto bail; |
354 | if (hpfs_ea_write(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l + 5 + h[1], size, data)) goto bail; | 356 | if (hpfs_ea_write(s, le32_to_cpu(fnode->ea_secno), fnode->ea_anode, le32_to_cpu(fnode->ea_size_l) + 5 + h[1], size, data)) goto bail; |
355 | fnode->ea_size_l = pos; | 357 | fnode->ea_size_l = cpu_to_le32(pos); |
356 | ret: | 358 | ret: |
357 | hpfs_i(inode)->i_ea_size += 5 + strlen(key) + size; | 359 | hpfs_i(inode)->i_ea_size += 5 + strlen(key) + size; |
358 | return; | 360 | return; |
359 | bail: | 361 | bail: |
360 | if (fnode->ea_secno) | 362 | if (le32_to_cpu(fnode->ea_secno)) |
361 | if (fnode->ea_anode) hpfs_truncate_btree(s, fnode->ea_secno, 1, (fnode->ea_size_l + 511) >> 9); | 363 | if (fnode->ea_anode) hpfs_truncate_btree(s, le32_to_cpu(fnode->ea_secno), 1, (le32_to_cpu(fnode->ea_size_l) + 511) >> 9); |
362 | else hpfs_free_sectors(s, fnode->ea_secno + ((fnode->ea_size_l + 511) >> 9), len - ((fnode->ea_size_l + 511) >> 9)); | 364 | else hpfs_free_sectors(s, le32_to_cpu(fnode->ea_secno) + ((le32_to_cpu(fnode->ea_size_l) + 511) >> 9), len - ((le32_to_cpu(fnode->ea_size_l) + 511) >> 9)); |
363 | else fnode->ea_secno = fnode->ea_size_l = 0; | 365 | else fnode->ea_secno = fnode->ea_size_l = cpu_to_le32(0); |
364 | } | 366 | } |
365 | 367 | ||
diff --git a/fs/hpfs/file.c b/fs/hpfs/file.c index 9b9eb6933e43..89c500ee5213 100644 --- a/fs/hpfs/file.c +++ b/fs/hpfs/file.c | |||
@@ -20,8 +20,8 @@ static int hpfs_file_release(struct inode *inode, struct file *file) | |||
20 | 20 | ||
21 | int hpfs_file_fsync(struct file *file, int datasync) | 21 | int hpfs_file_fsync(struct file *file, int datasync) |
22 | { | 22 | { |
23 | /*return file_fsync(file, datasync);*/ | 23 | struct inode *inode = file->f_mapping->host; |
24 | return 0; /* Don't fsync :-) */ | 24 | return sync_blockdev(inode->i_sb->s_bdev); |
25 | } | 25 | } |
26 | 26 | ||
27 | /* | 27 | /* |
@@ -48,38 +48,46 @@ static secno hpfs_bmap(struct inode *inode, unsigned file_secno) | |||
48 | static void hpfs_truncate(struct inode *i) | 48 | static void hpfs_truncate(struct inode *i) |
49 | { | 49 | { |
50 | if (IS_IMMUTABLE(i)) return /*-EPERM*/; | 50 | if (IS_IMMUTABLE(i)) return /*-EPERM*/; |
51 | hpfs_lock(i->i_sb); | 51 | hpfs_lock_assert(i->i_sb); |
52 | |||
52 | hpfs_i(i)->i_n_secs = 0; | 53 | hpfs_i(i)->i_n_secs = 0; |
53 | i->i_blocks = 1 + ((i->i_size + 511) >> 9); | 54 | i->i_blocks = 1 + ((i->i_size + 511) >> 9); |
54 | hpfs_i(i)->mmu_private = i->i_size; | 55 | hpfs_i(i)->mmu_private = i->i_size; |
55 | hpfs_truncate_btree(i->i_sb, i->i_ino, 1, ((i->i_size + 511) >> 9)); | 56 | hpfs_truncate_btree(i->i_sb, i->i_ino, 1, ((i->i_size + 511) >> 9)); |
56 | hpfs_write_inode(i); | 57 | hpfs_write_inode(i); |
57 | hpfs_i(i)->i_n_secs = 0; | 58 | hpfs_i(i)->i_n_secs = 0; |
58 | hpfs_unlock(i->i_sb); | ||
59 | } | 59 | } |
60 | 60 | ||
61 | static int hpfs_get_block(struct inode *inode, sector_t iblock, struct buffer_head *bh_result, int create) | 61 | static int hpfs_get_block(struct inode *inode, sector_t iblock, struct buffer_head *bh_result, int create) |
62 | { | 62 | { |
63 | int r; | ||
63 | secno s; | 64 | secno s; |
65 | hpfs_lock(inode->i_sb); | ||
64 | s = hpfs_bmap(inode, iblock); | 66 | s = hpfs_bmap(inode, iblock); |
65 | if (s) { | 67 | if (s) { |
66 | map_bh(bh_result, inode->i_sb, s); | 68 | map_bh(bh_result, inode->i_sb, s); |
67 | return 0; | 69 | goto ret_0; |
68 | } | 70 | } |
69 | if (!create) return 0; | 71 | if (!create) goto ret_0; |
70 | if (iblock<<9 != hpfs_i(inode)->mmu_private) { | 72 | if (iblock<<9 != hpfs_i(inode)->mmu_private) { |
71 | BUG(); | 73 | BUG(); |
72 | return -EIO; | 74 | r = -EIO; |
75 | goto ret_r; | ||
73 | } | 76 | } |
74 | if ((s = hpfs_add_sector_to_btree(inode->i_sb, inode->i_ino, 1, inode->i_blocks - 1)) == -1) { | 77 | if ((s = hpfs_add_sector_to_btree(inode->i_sb, inode->i_ino, 1, inode->i_blocks - 1)) == -1) { |
75 | hpfs_truncate_btree(inode->i_sb, inode->i_ino, 1, inode->i_blocks - 1); | 78 | hpfs_truncate_btree(inode->i_sb, inode->i_ino, 1, inode->i_blocks - 1); |
76 | return -ENOSPC; | 79 | r = -ENOSPC; |
80 | goto ret_r; | ||
77 | } | 81 | } |
78 | inode->i_blocks++; | 82 | inode->i_blocks++; |
79 | hpfs_i(inode)->mmu_private += 512; | 83 | hpfs_i(inode)->mmu_private += 512; |
80 | set_buffer_new(bh_result); | 84 | set_buffer_new(bh_result); |
81 | map_bh(bh_result, inode->i_sb, s); | 85 | map_bh(bh_result, inode->i_sb, s); |
82 | return 0; | 86 | ret_0: |
87 | r = 0; | ||
88 | ret_r: | ||
89 | hpfs_unlock(inode->i_sb); | ||
90 | return r; | ||
83 | } | 91 | } |
84 | 92 | ||
85 | static int hpfs_writepage(struct page *page, struct writeback_control *wbc) | 93 | static int hpfs_writepage(struct page *page, struct writeback_control *wbc) |
@@ -130,8 +138,11 @@ static ssize_t hpfs_file_write(struct file *file, const char __user *buf, | |||
130 | ssize_t retval; | 138 | ssize_t retval; |
131 | 139 | ||
132 | retval = do_sync_write(file, buf, count, ppos); | 140 | retval = do_sync_write(file, buf, count, ppos); |
133 | if (retval > 0) | 141 | if (retval > 0) { |
142 | hpfs_lock(file->f_path.dentry->d_sb); | ||
134 | hpfs_i(file->f_path.dentry->d_inode)->i_dirty = 1; | 143 | hpfs_i(file->f_path.dentry->d_inode)->i_dirty = 1; |
144 | hpfs_unlock(file->f_path.dentry->d_sb); | ||
145 | } | ||
135 | return retval; | 146 | return retval; |
136 | } | 147 | } |
137 | 148 | ||
diff --git a/fs/hpfs/hpfs.h b/fs/hpfs/hpfs.h index 0e84c73cd9c4..8b0650aae328 100644 --- a/fs/hpfs/hpfs.h +++ b/fs/hpfs/hpfs.h | |||
@@ -19,9 +19,13 @@ | |||
19 | For definitive information on HPFS, ask somebody else -- this is guesswork. | 19 | For definitive information on HPFS, ask somebody else -- this is guesswork. |
20 | There are certain to be many mistakes. */ | 20 | There are certain to be many mistakes. */ |
21 | 21 | ||
22 | #if !defined(__LITTLE_ENDIAN) && !defined(__BIG_ENDIAN) | ||
23 | #error unknown endian | ||
24 | #endif | ||
25 | |||
22 | /* Notation */ | 26 | /* Notation */ |
23 | 27 | ||
24 | typedef unsigned secno; /* sector number, partition relative */ | 28 | typedef u32 secno; /* sector number, partition relative */ |
25 | 29 | ||
26 | typedef secno dnode_secno; /* sector number of a dnode */ | 30 | typedef secno dnode_secno; /* sector number of a dnode */ |
27 | typedef secno fnode_secno; /* sector number of an fnode */ | 31 | typedef secno fnode_secno; /* sector number of an fnode */ |
@@ -38,28 +42,28 @@ typedef u32 time32_t; /* 32-bit time_t type */ | |||
38 | 42 | ||
39 | struct hpfs_boot_block | 43 | struct hpfs_boot_block |
40 | { | 44 | { |
41 | unsigned char jmp[3]; | 45 | u8 jmp[3]; |
42 | unsigned char oem_id[8]; | 46 | u8 oem_id[8]; |
43 | unsigned char bytes_per_sector[2]; /* 512 */ | 47 | u8 bytes_per_sector[2]; /* 512 */ |
44 | unsigned char sectors_per_cluster; | 48 | u8 sectors_per_cluster; |
45 | unsigned char n_reserved_sectors[2]; | 49 | u8 n_reserved_sectors[2]; |
46 | unsigned char n_fats; | 50 | u8 n_fats; |
47 | unsigned char n_rootdir_entries[2]; | 51 | u8 n_rootdir_entries[2]; |
48 | unsigned char n_sectors_s[2]; | 52 | u8 n_sectors_s[2]; |
49 | unsigned char media_byte; | 53 | u8 media_byte; |
50 | unsigned short sectors_per_fat; | 54 | u16 sectors_per_fat; |
51 | unsigned short sectors_per_track; | 55 | u16 sectors_per_track; |
52 | unsigned short heads_per_cyl; | 56 | u16 heads_per_cyl; |
53 | unsigned int n_hidden_sectors; | 57 | u32 n_hidden_sectors; |
54 | unsigned int n_sectors_l; /* size of partition */ | 58 | u32 n_sectors_l; /* size of partition */ |
55 | unsigned char drive_number; | 59 | u8 drive_number; |
56 | unsigned char mbz; | 60 | u8 mbz; |
57 | unsigned char sig_28h; /* 28h */ | 61 | u8 sig_28h; /* 28h */ |
58 | unsigned char vol_serno[4]; | 62 | u8 vol_serno[4]; |
59 | unsigned char vol_label[11]; | 63 | u8 vol_label[11]; |
60 | unsigned char sig_hpfs[8]; /* "HPFS " */ | 64 | u8 sig_hpfs[8]; /* "HPFS " */ |
61 | unsigned char pad[448]; | 65 | u8 pad[448]; |
62 | unsigned short magic; /* aa55 */ | 66 | u16 magic; /* aa55 */ |
63 | }; | 67 | }; |
64 | 68 | ||
65 | 69 | ||
@@ -71,31 +75,29 @@ struct hpfs_boot_block | |||
71 | 75 | ||
72 | struct hpfs_super_block | 76 | struct hpfs_super_block |
73 | { | 77 | { |
74 | unsigned magic; /* f995 e849 */ | 78 | u32 magic; /* f995 e849 */ |
75 | unsigned magic1; /* fa53 e9c5, more magic? */ | 79 | u32 magic1; /* fa53 e9c5, more magic? */ |
76 | /*unsigned huh202;*/ /* ?? 202 = N. of B. in 1.00390625 S.*/ | 80 | u8 version; /* version of a filesystem usually 2 */ |
77 | char version; /* version of a filesystem usually 2 */ | 81 | u8 funcversion; /* functional version - oldest version |
78 | char funcversion; /* functional version - oldest version | ||
79 | of filesystem that can understand | 82 | of filesystem that can understand |
80 | this disk */ | 83 | this disk */ |
81 | unsigned short int zero; /* 0 */ | 84 | u16 zero; /* 0 */ |
82 | fnode_secno root; /* fnode of root directory */ | 85 | fnode_secno root; /* fnode of root directory */ |
83 | secno n_sectors; /* size of filesystem */ | 86 | secno n_sectors; /* size of filesystem */ |
84 | unsigned n_badblocks; /* number of bad blocks */ | 87 | u32 n_badblocks; /* number of bad blocks */ |
85 | secno bitmaps; /* pointers to free space bit maps */ | 88 | secno bitmaps; /* pointers to free space bit maps */ |
86 | unsigned zero1; /* 0 */ | 89 | u32 zero1; /* 0 */ |
87 | secno badblocks; /* bad block list */ | 90 | secno badblocks; /* bad block list */ |
88 | unsigned zero3; /* 0 */ | 91 | u32 zero3; /* 0 */ |
89 | time32_t last_chkdsk; /* date last checked, 0 if never */ | 92 | time32_t last_chkdsk; /* date last checked, 0 if never */ |
90 | /*unsigned zero4;*/ /* 0 */ | 93 | time32_t last_optimize; /* date last optimized, 0 if never */ |
91 | time32_t last_optimize; /* date last optimized, 0 if never */ | ||
92 | secno n_dir_band; /* number of sectors in dir band */ | 94 | secno n_dir_band; /* number of sectors in dir band */ |
93 | secno dir_band_start; /* first sector in dir band */ | 95 | secno dir_band_start; /* first sector in dir band */ |
94 | secno dir_band_end; /* last sector in dir band */ | 96 | secno dir_band_end; /* last sector in dir band */ |
95 | secno dir_band_bitmap; /* free space map, 1 dnode per bit */ | 97 | secno dir_band_bitmap; /* free space map, 1 dnode per bit */ |
96 | char volume_name[32]; /* not used */ | 98 | u8 volume_name[32]; /* not used */ |
97 | secno user_id_table; /* 8 preallocated sectors - user id */ | 99 | secno user_id_table; /* 8 preallocated sectors - user id */ |
98 | unsigned zero6[103]; /* 0 */ | 100 | u32 zero6[103]; /* 0 */ |
99 | }; | 101 | }; |
100 | 102 | ||
101 | 103 | ||
@@ -107,44 +109,65 @@ struct hpfs_super_block | |||
107 | 109 | ||
108 | struct hpfs_spare_block | 110 | struct hpfs_spare_block |
109 | { | 111 | { |
110 | unsigned magic; /* f991 1849 */ | 112 | u32 magic; /* f991 1849 */ |
111 | unsigned magic1; /* fa52 29c5, more magic? */ | 113 | u32 magic1; /* fa52 29c5, more magic? */ |
112 | 114 | ||
113 | unsigned dirty: 1; /* 0 clean, 1 "improperly stopped" */ | 115 | #ifdef __LITTLE_ENDIAN |
114 | /*unsigned flag1234: 4;*/ /* unknown flags */ | 116 | u8 dirty: 1; /* 0 clean, 1 "improperly stopped" */ |
115 | unsigned sparedir_used: 1; /* spare dirblks used */ | 117 | u8 sparedir_used: 1; /* spare dirblks used */ |
116 | unsigned hotfixes_used: 1; /* hotfixes used */ | 118 | u8 hotfixes_used: 1; /* hotfixes used */ |
117 | unsigned bad_sector: 1; /* bad sector, corrupted disk (???) */ | 119 | u8 bad_sector: 1; /* bad sector, corrupted disk (???) */ |
118 | unsigned bad_bitmap: 1; /* bad bitmap */ | 120 | u8 bad_bitmap: 1; /* bad bitmap */ |
119 | unsigned fast: 1; /* partition was fast formatted */ | 121 | u8 fast: 1; /* partition was fast formatted */ |
120 | unsigned old_wrote: 1; /* old version wrote to partion */ | 122 | u8 old_wrote: 1; /* old version wrote to partion */ |
121 | unsigned old_wrote_1: 1; /* old version wrote to partion (?) */ | 123 | u8 old_wrote_1: 1; /* old version wrote to partion (?) */ |
122 | unsigned install_dasd_limits: 1; /* HPFS386 flags */ | 124 | #else |
123 | unsigned resynch_dasd_limits: 1; | 125 | u8 old_wrote_1: 1; /* old version wrote to partion (?) */ |
124 | unsigned dasd_limits_operational: 1; | 126 | u8 old_wrote: 1; /* old version wrote to partion */ |
125 | unsigned multimedia_active: 1; | 127 | u8 fast: 1; /* partition was fast formatted */ |
126 | unsigned dce_acls_active: 1; | 128 | u8 bad_bitmap: 1; /* bad bitmap */ |
127 | unsigned dasd_limits_dirty: 1; | 129 | u8 bad_sector: 1; /* bad sector, corrupted disk (???) */ |
128 | unsigned flag67: 2; | 130 | u8 hotfixes_used: 1; /* hotfixes used */ |
129 | unsigned char mm_contlgulty; | 131 | u8 sparedir_used: 1; /* spare dirblks used */ |
130 | unsigned char unused; | 132 | u8 dirty: 1; /* 0 clean, 1 "improperly stopped" */ |
133 | #endif | ||
134 | |||
135 | #ifdef __LITTLE_ENDIAN | ||
136 | u8 install_dasd_limits: 1; /* HPFS386 flags */ | ||
137 | u8 resynch_dasd_limits: 1; | ||
138 | u8 dasd_limits_operational: 1; | ||
139 | u8 multimedia_active: 1; | ||
140 | u8 dce_acls_active: 1; | ||
141 | u8 dasd_limits_dirty: 1; | ||
142 | u8 flag67: 2; | ||
143 | #else | ||
144 | u8 flag67: 2; | ||
145 | u8 dasd_limits_dirty: 1; | ||
146 | u8 dce_acls_active: 1; | ||
147 | u8 multimedia_active: 1; | ||
148 | u8 dasd_limits_operational: 1; | ||
149 | u8 resynch_dasd_limits: 1; | ||
150 | u8 install_dasd_limits: 1; /* HPFS386 flags */ | ||
151 | #endif | ||
152 | |||
153 | u8 mm_contlgulty; | ||
154 | u8 unused; | ||
131 | 155 | ||
132 | secno hotfix_map; /* info about remapped bad sectors */ | 156 | secno hotfix_map; /* info about remapped bad sectors */ |
133 | unsigned n_spares_used; /* number of hotfixes */ | 157 | u32 n_spares_used; /* number of hotfixes */ |
134 | unsigned n_spares; /* number of spares in hotfix map */ | 158 | u32 n_spares; /* number of spares in hotfix map */ |
135 | unsigned n_dnode_spares_free; /* spare dnodes unused */ | 159 | u32 n_dnode_spares_free; /* spare dnodes unused */ |
136 | unsigned n_dnode_spares; /* length of spare_dnodes[] list, | 160 | u32 n_dnode_spares; /* length of spare_dnodes[] list, |
137 | follows in this block*/ | 161 | follows in this block*/ |
138 | secno code_page_dir; /* code page directory block */ | 162 | secno code_page_dir; /* code page directory block */ |
139 | unsigned n_code_pages; /* number of code pages */ | 163 | u32 n_code_pages; /* number of code pages */ |
140 | /*unsigned large_numbers[2];*/ /* ?? */ | 164 | u32 super_crc; /* on HPFS386 and LAN Server this is |
141 | unsigned super_crc; /* on HPFS386 and LAN Server this is | ||
142 | checksum of superblock, on normal | 165 | checksum of superblock, on normal |
143 | OS/2 unused */ | 166 | OS/2 unused */ |
144 | unsigned spare_crc; /* on HPFS386 checksum of spareblock */ | 167 | u32 spare_crc; /* on HPFS386 checksum of spareblock */ |
145 | unsigned zero1[15]; /* unused */ | 168 | u32 zero1[15]; /* unused */ |
146 | dnode_secno spare_dnodes[100]; /* emergency free dnode list */ | 169 | dnode_secno spare_dnodes[100]; /* emergency free dnode list */ |
147 | unsigned zero2[1]; /* room for more? */ | 170 | u32 zero2[1]; /* room for more? */ |
148 | }; | 171 | }; |
149 | 172 | ||
150 | /* The bad block list is 4 sectors long. The first word must be zero, | 173 | /* The bad block list is 4 sectors long. The first word must be zero, |
@@ -179,18 +202,18 @@ struct hpfs_spare_block | |||
179 | 202 | ||
180 | struct code_page_directory | 203 | struct code_page_directory |
181 | { | 204 | { |
182 | unsigned magic; /* 4945 21f7 */ | 205 | u32 magic; /* 4945 21f7 */ |
183 | unsigned n_code_pages; /* number of pointers following */ | 206 | u32 n_code_pages; /* number of pointers following */ |
184 | unsigned zero1[2]; | 207 | u32 zero1[2]; |
185 | struct { | 208 | struct { |
186 | unsigned short ix; /* index */ | 209 | u16 ix; /* index */ |
187 | unsigned short code_page_number; /* code page number */ | 210 | u16 code_page_number; /* code page number */ |
188 | unsigned bounds; /* matches corresponding word | 211 | u32 bounds; /* matches corresponding word |
189 | in data block */ | 212 | in data block */ |
190 | secno code_page_data; /* sector number of a code_page_data | 213 | secno code_page_data; /* sector number of a code_page_data |
191 | containing c.p. array */ | 214 | containing c.p. array */ |
192 | unsigned short index; /* index in c.p. array in that sector*/ | 215 | u16 index; /* index in c.p. array in that sector*/ |
193 | unsigned short unknown; /* some unknown value; usually 0; | 216 | u16 unknown; /* some unknown value; usually 0; |
194 | 2 in Japanese version */ | 217 | 2 in Japanese version */ |
195 | } array[31]; /* unknown length */ | 218 | } array[31]; /* unknown length */ |
196 | }; | 219 | }; |
@@ -201,21 +224,21 @@ struct code_page_directory | |||
201 | 224 | ||
202 | struct code_page_data | 225 | struct code_page_data |
203 | { | 226 | { |
204 | unsigned magic; /* 8945 21f7 */ | 227 | u32 magic; /* 8945 21f7 */ |
205 | unsigned n_used; /* # elements used in c_p_data[] */ | 228 | u32 n_used; /* # elements used in c_p_data[] */ |
206 | unsigned bounds[3]; /* looks a bit like | 229 | u32 bounds[3]; /* looks a bit like |
207 | (beg1,end1), (beg2,end2) | 230 | (beg1,end1), (beg2,end2) |
208 | one byte each */ | 231 | one byte each */ |
209 | unsigned short offs[3]; /* offsets from start of sector | 232 | u16 offs[3]; /* offsets from start of sector |
210 | to start of c_p_data[ix] */ | 233 | to start of c_p_data[ix] */ |
211 | struct { | 234 | struct { |
212 | unsigned short ix; /* index */ | 235 | u16 ix; /* index */ |
213 | unsigned short code_page_number; /* code page number */ | 236 | u16 code_page_number; /* code page number */ |
214 | unsigned short unknown; /* the same as in cp directory */ | 237 | u16 unknown; /* the same as in cp directory */ |
215 | unsigned char map[128]; /* upcase table for chars 80..ff */ | 238 | u8 map[128]; /* upcase table for chars 80..ff */ |
216 | unsigned short zero2; | 239 | u16 zero2; |
217 | } code_page[3]; | 240 | } code_page[3]; |
218 | unsigned char incognita[78]; | 241 | u8 incognita[78]; |
219 | }; | 242 | }; |
220 | 243 | ||
221 | 244 | ||
@@ -255,50 +278,84 @@ struct code_page_data | |||
255 | #define DNODE_MAGIC 0x77e40aae | 278 | #define DNODE_MAGIC 0x77e40aae |
256 | 279 | ||
257 | struct dnode { | 280 | struct dnode { |
258 | unsigned magic; /* 77e4 0aae */ | 281 | u32 magic; /* 77e4 0aae */ |
259 | unsigned first_free; /* offset from start of dnode to | 282 | u32 first_free; /* offset from start of dnode to |
260 | first free dir entry */ | 283 | first free dir entry */ |
261 | unsigned root_dnode:1; /* Is it root dnode? */ | 284 | #ifdef __LITTLE_ENDIAN |
262 | unsigned increment_me:31; /* some kind of activity counter? | 285 | u8 root_dnode: 1; /* Is it root dnode? */ |
263 | Neither HPFS.IFS nor CHKDSK cares | 286 | u8 increment_me: 7; /* some kind of activity counter? */ |
287 | /* Neither HPFS.IFS nor CHKDSK cares | ||
288 | if you change this word */ | ||
289 | #else | ||
290 | u8 increment_me: 7; /* some kind of activity counter? */ | ||
291 | /* Neither HPFS.IFS nor CHKDSK cares | ||
264 | if you change this word */ | 292 | if you change this word */ |
293 | u8 root_dnode: 1; /* Is it root dnode? */ | ||
294 | #endif | ||
295 | u8 increment_me2[3]; | ||
265 | secno up; /* (root dnode) directory's fnode | 296 | secno up; /* (root dnode) directory's fnode |
266 | (nonroot) parent dnode */ | 297 | (nonroot) parent dnode */ |
267 | dnode_secno self; /* pointer to this dnode */ | 298 | dnode_secno self; /* pointer to this dnode */ |
268 | unsigned char dirent[2028]; /* one or more dirents */ | 299 | u8 dirent[2028]; /* one or more dirents */ |
269 | }; | 300 | }; |
270 | 301 | ||
271 | struct hpfs_dirent { | 302 | struct hpfs_dirent { |
272 | unsigned short length; /* offset to next dirent */ | 303 | u16 length; /* offset to next dirent */ |
273 | unsigned first: 1; /* set on phony ^A^A (".") entry */ | 304 | |
274 | unsigned has_acl: 1; | 305 | #ifdef __LITTLE_ENDIAN |
275 | unsigned down: 1; /* down pointer present (after name) */ | 306 | u8 first: 1; /* set on phony ^A^A (".") entry */ |
276 | unsigned last: 1; /* set on phony \377 entry */ | 307 | u8 has_acl: 1; |
277 | unsigned has_ea: 1; /* entry has EA */ | 308 | u8 down: 1; /* down pointer present (after name) */ |
278 | unsigned has_xtd_perm: 1; /* has extended perm list (???) */ | 309 | u8 last: 1; /* set on phony \377 entry */ |
279 | unsigned has_explicit_acl: 1; | 310 | u8 has_ea: 1; /* entry has EA */ |
280 | unsigned has_needea: 1; /* ?? some EA has NEEDEA set | 311 | u8 has_xtd_perm: 1; /* has extended perm list (???) */ |
312 | u8 has_explicit_acl: 1; | ||
313 | u8 has_needea: 1; /* ?? some EA has NEEDEA set | ||
314 | I have no idea why this is | ||
315 | interesting in a dir entry */ | ||
316 | #else | ||
317 | u8 has_needea: 1; /* ?? some EA has NEEDEA set | ||
281 | I have no idea why this is | 318 | I have no idea why this is |
282 | interesting in a dir entry */ | 319 | interesting in a dir entry */ |
283 | unsigned read_only: 1; /* dos attrib */ | 320 | u8 has_explicit_acl: 1; |
284 | unsigned hidden: 1; /* dos attrib */ | 321 | u8 has_xtd_perm: 1; /* has extended perm list (???) */ |
285 | unsigned system: 1; /* dos attrib */ | 322 | u8 has_ea: 1; /* entry has EA */ |
286 | unsigned flag11: 1; /* would be volume label dos attrib */ | 323 | u8 last: 1; /* set on phony \377 entry */ |
287 | unsigned directory: 1; /* dos attrib */ | 324 | u8 down: 1; /* down pointer present (after name) */ |
288 | unsigned archive: 1; /* dos attrib */ | 325 | u8 has_acl: 1; |
289 | unsigned not_8x3: 1; /* name is not 8.3 */ | 326 | u8 first: 1; /* set on phony ^A^A (".") entry */ |
290 | unsigned flag15: 1; | 327 | #endif |
328 | |||
329 | #ifdef __LITTLE_ENDIAN | ||
330 | u8 read_only: 1; /* dos attrib */ | ||
331 | u8 hidden: 1; /* dos attrib */ | ||
332 | u8 system: 1; /* dos attrib */ | ||
333 | u8 flag11: 1; /* would be volume label dos attrib */ | ||
334 | u8 directory: 1; /* dos attrib */ | ||
335 | u8 archive: 1; /* dos attrib */ | ||
336 | u8 not_8x3: 1; /* name is not 8.3 */ | ||
337 | u8 flag15: 1; | ||
338 | #else | ||
339 | u8 flag15: 1; | ||
340 | u8 not_8x3: 1; /* name is not 8.3 */ | ||
341 | u8 archive: 1; /* dos attrib */ | ||
342 | u8 directory: 1; /* dos attrib */ | ||
343 | u8 flag11: 1; /* would be volume label dos attrib */ | ||
344 | u8 system: 1; /* dos attrib */ | ||
345 | u8 hidden: 1; /* dos attrib */ | ||
346 | u8 read_only: 1; /* dos attrib */ | ||
347 | #endif | ||
348 | |||
291 | fnode_secno fnode; /* fnode giving allocation info */ | 349 | fnode_secno fnode; /* fnode giving allocation info */ |
292 | time32_t write_date; /* mtime */ | 350 | time32_t write_date; /* mtime */ |
293 | unsigned file_size; /* file length, bytes */ | 351 | u32 file_size; /* file length, bytes */ |
294 | time32_t read_date; /* atime */ | 352 | time32_t read_date; /* atime */ |
295 | time32_t creation_date; /* ctime */ | 353 | time32_t creation_date; /* ctime */ |
296 | unsigned ea_size; /* total EA length, bytes */ | 354 | u32 ea_size; /* total EA length, bytes */ |
297 | unsigned char no_of_acls : 3; /* number of ACL's */ | 355 | u8 no_of_acls; /* number of ACL's (low 3 bits) */ |
298 | unsigned char reserver : 5; | 356 | u8 ix; /* code page index (of filename), see |
299 | unsigned char ix; /* code page index (of filename), see | ||
300 | struct code_page_data */ | 357 | struct code_page_data */ |
301 | unsigned char namelen, name[1]; /* file name */ | 358 | u8 namelen, name[1]; /* file name */ |
302 | /* dnode_secno down; btree down pointer, if present, | 359 | /* dnode_secno down; btree down pointer, if present, |
303 | follows name on next word boundary, or maybe it | 360 | follows name on next word boundary, or maybe it |
304 | precedes next dirent, which is on a word boundary. */ | 361 | precedes next dirent, which is on a word boundary. */ |
@@ -318,38 +375,50 @@ struct hpfs_dirent { | |||
318 | 375 | ||
319 | struct bplus_leaf_node | 376 | struct bplus_leaf_node |
320 | { | 377 | { |
321 | unsigned file_secno; /* first file sector in extent */ | 378 | u32 file_secno; /* first file sector in extent */ |
322 | unsigned length; /* length, sectors */ | 379 | u32 length; /* length, sectors */ |
323 | secno disk_secno; /* first corresponding disk sector */ | 380 | secno disk_secno; /* first corresponding disk sector */ |
324 | }; | 381 | }; |
325 | 382 | ||
326 | struct bplus_internal_node | 383 | struct bplus_internal_node |
327 | { | 384 | { |
328 | unsigned file_secno; /* subtree maps sectors < this */ | 385 | u32 file_secno; /* subtree maps sectors < this */ |
329 | anode_secno down; /* pointer to subtree */ | 386 | anode_secno down; /* pointer to subtree */ |
330 | }; | 387 | }; |
331 | 388 | ||
332 | struct bplus_header | 389 | struct bplus_header |
333 | { | 390 | { |
334 | unsigned hbff: 1; /* high bit of first free entry offset */ | 391 | #ifdef __LITTLE_ENDIAN |
335 | unsigned flag1: 1; | 392 | u8 hbff: 1; /* high bit of first free entry offset */ |
336 | unsigned flag2: 1; | 393 | u8 flag1234: 4; |
337 | unsigned flag3: 1; | 394 | u8 fnode_parent: 1; /* ? we're pointed to by an fnode, |
338 | unsigned flag4: 1; | ||
339 | unsigned fnode_parent: 1; /* ? we're pointed to by an fnode, | ||
340 | the data btree or some ea or the | 395 | the data btree or some ea or the |
341 | main ea bootage pointer ea_secno */ | 396 | main ea bootage pointer ea_secno */ |
342 | /* also can get set in fnodes, which | 397 | /* also can get set in fnodes, which |
343 | may be a chkdsk glitch or may mean | 398 | may be a chkdsk glitch or may mean |
344 | this bit is irrelevant in fnodes, | 399 | this bit is irrelevant in fnodes, |
345 | or this interpretation is all wet */ | 400 | or this interpretation is all wet */ |
346 | unsigned binary_search: 1; /* suggest binary search (unused) */ | 401 | u8 binary_search: 1; /* suggest binary search (unused) */ |
347 | unsigned internal: 1; /* 1 -> (internal) tree of anodes | 402 | u8 internal: 1; /* 1 -> (internal) tree of anodes |
403 | 0 -> (leaf) list of extents */ | ||
404 | #else | ||
405 | u8 internal: 1; /* 1 -> (internal) tree of anodes | ||
348 | 0 -> (leaf) list of extents */ | 406 | 0 -> (leaf) list of extents */ |
349 | unsigned char fill[3]; | 407 | u8 binary_search: 1; /* suggest binary search (unused) */ |
350 | unsigned char n_free_nodes; /* free nodes in following array */ | 408 | u8 fnode_parent: 1; /* ? we're pointed to by an fnode, |
351 | unsigned char n_used_nodes; /* used nodes in following array */ | 409 | the data btree or some ea or the |
352 | unsigned short first_free; /* offset from start of header to | 410 | main ea bootage pointer ea_secno */ |
411 | /* also can get set in fnodes, which | ||
412 | may be a chkdsk glitch or may mean | ||
413 | this bit is irrelevant in fnodes, | ||
414 | or this interpretation is all wet */ | ||
415 | u8 flag1234: 4; | ||
416 | u8 hbff: 1; /* high bit of first free entry offset */ | ||
417 | #endif | ||
418 | u8 fill[3]; | ||
419 | u8 n_free_nodes; /* free nodes in following array */ | ||
420 | u8 n_used_nodes; /* used nodes in following array */ | ||
421 | u16 first_free; /* offset from start of header to | ||
353 | first free node in array */ | 422 | first free node in array */ |
354 | union { | 423 | union { |
355 | struct bplus_internal_node internal[0]; /* (internal) 2-word entries giving | 424 | struct bplus_internal_node internal[0]; /* (internal) 2-word entries giving |
@@ -369,37 +438,38 @@ struct bplus_header | |||
369 | 438 | ||
370 | struct fnode | 439 | struct fnode |
371 | { | 440 | { |
372 | unsigned magic; /* f7e4 0aae */ | 441 | u32 magic; /* f7e4 0aae */ |
373 | unsigned zero1[2]; /* read history */ | 442 | u32 zero1[2]; /* read history */ |
374 | unsigned char len, name[15]; /* true length, truncated name */ | 443 | u8 len, name[15]; /* true length, truncated name */ |
375 | fnode_secno up; /* pointer to file's directory fnode */ | 444 | fnode_secno up; /* pointer to file's directory fnode */ |
376 | /*unsigned zero2[3];*/ | ||
377 | secno acl_size_l; | 445 | secno acl_size_l; |
378 | secno acl_secno; | 446 | secno acl_secno; |
379 | unsigned short acl_size_s; | 447 | u16 acl_size_s; |
380 | char acl_anode; | 448 | u8 acl_anode; |
381 | char zero2; /* history bit count */ | 449 | u8 zero2; /* history bit count */ |
382 | unsigned ea_size_l; /* length of disk-resident ea's */ | 450 | u32 ea_size_l; /* length of disk-resident ea's */ |
383 | secno ea_secno; /* first sector of disk-resident ea's*/ | 451 | secno ea_secno; /* first sector of disk-resident ea's*/ |
384 | unsigned short ea_size_s; /* length of fnode-resident ea's */ | 452 | u16 ea_size_s; /* length of fnode-resident ea's */ |
385 | 453 | ||
386 | unsigned flag0: 1; | 454 | #ifdef __LITTLE_ENDIAN |
387 | unsigned ea_anode: 1; /* 1 -> ea_secno is an anode */ | 455 | u8 flag0: 1; |
388 | unsigned flag2: 1; | 456 | u8 ea_anode: 1; /* 1 -> ea_secno is an anode */ |
389 | unsigned flag3: 1; | 457 | u8 flag234567: 6; |
390 | unsigned flag4: 1; | 458 | #else |
391 | unsigned flag5: 1; | 459 | u8 flag234567: 6; |
392 | unsigned flag6: 1; | 460 | u8 ea_anode: 1; /* 1 -> ea_secno is an anode */ |
393 | unsigned flag7: 1; | 461 | u8 flag0: 1; |
394 | unsigned dirflag: 1; /* 1 -> directory. first & only extent | 462 | #endif |
463 | |||
464 | #ifdef __LITTLE_ENDIAN | ||
465 | u8 dirflag: 1; /* 1 -> directory. first & only extent | ||
395 | points to dnode. */ | 466 | points to dnode. */ |
396 | unsigned flag9: 1; | 467 | u8 flag9012345: 7; |
397 | unsigned flag10: 1; | 468 | #else |
398 | unsigned flag11: 1; | 469 | u8 flag9012345: 7; |
399 | unsigned flag12: 1; | 470 | u8 dirflag: 1; /* 1 -> directory. first & only extent |
400 | unsigned flag13: 1; | 471 | points to dnode. */ |
401 | unsigned flag14: 1; | 472 | #endif |
402 | unsigned flag15: 1; | ||
403 | 473 | ||
404 | struct bplus_header btree; /* b+ tree, 8 extents or 12 subtrees */ | 474 | struct bplus_header btree; /* b+ tree, 8 extents or 12 subtrees */ |
405 | union { | 475 | union { |
@@ -407,17 +477,16 @@ struct fnode | |||
407 | struct bplus_internal_node internal[12]; | 477 | struct bplus_internal_node internal[12]; |
408 | } u; | 478 | } u; |
409 | 479 | ||
410 | unsigned file_size; /* file length, bytes */ | 480 | u32 file_size; /* file length, bytes */ |
411 | unsigned n_needea; /* number of EA's with NEEDEA set */ | 481 | u32 n_needea; /* number of EA's with NEEDEA set */ |
412 | char user_id[16]; /* unused */ | 482 | u8 user_id[16]; /* unused */ |
413 | unsigned short ea_offs; /* offset from start of fnode | 483 | u16 ea_offs; /* offset from start of fnode |
414 | to first fnode-resident ea */ | 484 | to first fnode-resident ea */ |
415 | char dasd_limit_treshhold; | 485 | u8 dasd_limit_treshhold; |
416 | char dasd_limit_delta; | 486 | u8 dasd_limit_delta; |
417 | unsigned dasd_limit; | 487 | u32 dasd_limit; |
418 | unsigned dasd_usage; | 488 | u32 dasd_usage; |
419 | /*unsigned zero5[2];*/ | 489 | u8 ea[316]; /* zero or more EA's, packed together |
420 | unsigned char ea[316]; /* zero or more EA's, packed together | ||
421 | with no alignment padding. | 490 | with no alignment padding. |
422 | (Do not use this name, get here | 491 | (Do not use this name, get here |
423 | via fnode + ea_offs. I think.) */ | 492 | via fnode + ea_offs. I think.) */ |
@@ -430,7 +499,7 @@ struct fnode | |||
430 | 499 | ||
431 | struct anode | 500 | struct anode |
432 | { | 501 | { |
433 | unsigned magic; /* 37e4 0aae */ | 502 | u32 magic; /* 37e4 0aae */ |
434 | anode_secno self; /* pointer to this anode */ | 503 | anode_secno self; /* pointer to this anode */ |
435 | secno up; /* parent anode or fnode */ | 504 | secno up; /* parent anode or fnode */ |
436 | 505 | ||
@@ -440,7 +509,7 @@ struct anode | |||
440 | struct bplus_internal_node internal[60]; | 509 | struct bplus_internal_node internal[60]; |
441 | } u; | 510 | } u; |
442 | 511 | ||
443 | unsigned fill[3]; /* unused */ | 512 | u32 fill[3]; /* unused */ |
444 | }; | 513 | }; |
445 | 514 | ||
446 | 515 | ||
@@ -461,25 +530,31 @@ struct anode | |||
461 | 530 | ||
462 | struct extended_attribute | 531 | struct extended_attribute |
463 | { | 532 | { |
464 | unsigned indirect: 1; /* 1 -> value gives sector number | 533 | #ifdef __LITTLE_ENDIAN |
534 | u8 indirect: 1; /* 1 -> value gives sector number | ||
465 | where real value starts */ | 535 | where real value starts */ |
466 | unsigned anode: 1; /* 1 -> sector is an anode | 536 | u8 anode: 1; /* 1 -> sector is an anode |
537 | that points to fragmented value */ | ||
538 | u8 flag23456: 5; | ||
539 | u8 needea: 1; /* required ea */ | ||
540 | #else | ||
541 | u8 needea: 1; /* required ea */ | ||
542 | u8 flag23456: 5; | ||
543 | u8 anode: 1; /* 1 -> sector is an anode | ||
467 | that points to fragmented value */ | 544 | that points to fragmented value */ |
468 | unsigned flag2: 1; | 545 | u8 indirect: 1; /* 1 -> value gives sector number |
469 | unsigned flag3: 1; | 546 | where real value starts */ |
470 | unsigned flag4: 1; | 547 | #endif |
471 | unsigned flag5: 1; | 548 | u8 namelen; /* length of name, bytes */ |
472 | unsigned flag6: 1; | 549 | u8 valuelen_lo; /* length of value, bytes */ |
473 | unsigned needea: 1; /* required ea */ | 550 | u8 valuelen_hi; /* length of value, bytes */ |
474 | unsigned char namelen; /* length of name, bytes */ | 551 | u8 name[0]; |
475 | unsigned short valuelen; /* length of value, bytes */ | ||
476 | unsigned char name[0]; | ||
477 | /* | 552 | /* |
478 | unsigned char name[namelen]; ascii attrib name | 553 | u8 name[namelen]; ascii attrib name |
479 | unsigned char nul; terminating '\0', not counted | 554 | u8 nul; terminating '\0', not counted |
480 | unsigned char value[valuelen]; value, arbitrary | 555 | u8 value[valuelen]; value, arbitrary |
481 | if this.indirect, valuelen is 8 and the value is | 556 | if this.indirect, valuelen is 8 and the value is |
482 | unsigned length; real length of value, bytes | 557 | u32 length; real length of value, bytes |
483 | secno secno; sector address where it starts | 558 | secno secno; sector address where it starts |
484 | if this.anode, the above sector number is the root of an anode tree | 559 | if this.anode, the above sector number is the root of an anode tree |
485 | which points to the value. | 560 | which points to the value. |
diff --git a/fs/hpfs/hpfs_fn.h b/fs/hpfs/hpfs_fn.h index c15adbca07ff..dd552f862c8f 100644 --- a/fs/hpfs/hpfs_fn.h +++ b/fs/hpfs/hpfs_fn.h | |||
@@ -13,6 +13,7 @@ | |||
13 | #include <linux/pagemap.h> | 13 | #include <linux/pagemap.h> |
14 | #include <linux/buffer_head.h> | 14 | #include <linux/buffer_head.h> |
15 | #include <linux/slab.h> | 15 | #include <linux/slab.h> |
16 | #include <asm/unaligned.h> | ||
16 | 17 | ||
17 | #include "hpfs.h" | 18 | #include "hpfs.h" |
18 | 19 | ||
@@ -51,18 +52,16 @@ struct hpfs_inode_info { | |||
51 | unsigned i_disk_sec; /* (files) minimalist cache of alloc info */ | 52 | unsigned i_disk_sec; /* (files) minimalist cache of alloc info */ |
52 | unsigned i_n_secs; /* (files) minimalist cache of alloc info */ | 53 | unsigned i_n_secs; /* (files) minimalist cache of alloc info */ |
53 | unsigned i_ea_size; /* size of extended attributes */ | 54 | unsigned i_ea_size; /* size of extended attributes */ |
54 | unsigned i_conv : 2; /* (files) crlf->newline hackery */ | ||
55 | unsigned i_ea_mode : 1; /* file's permission is stored in ea */ | 55 | unsigned i_ea_mode : 1; /* file's permission is stored in ea */ |
56 | unsigned i_ea_uid : 1; /* file's uid is stored in ea */ | 56 | unsigned i_ea_uid : 1; /* file's uid is stored in ea */ |
57 | unsigned i_ea_gid : 1; /* file's gid is stored in ea */ | 57 | unsigned i_ea_gid : 1; /* file's gid is stored in ea */ |
58 | unsigned i_dirty : 1; | 58 | unsigned i_dirty : 1; |
59 | struct mutex i_mutex; | ||
60 | struct mutex i_parent_mutex; | ||
61 | loff_t **i_rddir_off; | 59 | loff_t **i_rddir_off; |
62 | struct inode vfs_inode; | 60 | struct inode vfs_inode; |
63 | }; | 61 | }; |
64 | 62 | ||
65 | struct hpfs_sb_info { | 63 | struct hpfs_sb_info { |
64 | struct mutex hpfs_mutex; /* global hpfs lock */ | ||
66 | ino_t sb_root; /* inode number of root dir */ | 65 | ino_t sb_root; /* inode number of root dir */ |
67 | unsigned sb_fs_size; /* file system size, sectors */ | 66 | unsigned sb_fs_size; /* file system size, sectors */ |
68 | unsigned sb_bitmaps; /* sector number of bitmap list */ | 67 | unsigned sb_bitmaps; /* sector number of bitmap list */ |
@@ -74,7 +73,6 @@ struct hpfs_sb_info { | |||
74 | uid_t sb_uid; /* uid from mount options */ | 73 | uid_t sb_uid; /* uid from mount options */ |
75 | gid_t sb_gid; /* gid from mount options */ | 74 | gid_t sb_gid; /* gid from mount options */ |
76 | umode_t sb_mode; /* mode from mount options */ | 75 | umode_t sb_mode; /* mode from mount options */ |
77 | unsigned sb_conv : 2; /* crlf->newline hackery */ | ||
78 | unsigned sb_eas : 2; /* eas: 0-ignore, 1-ro, 2-rw */ | 76 | unsigned sb_eas : 2; /* eas: 0-ignore, 1-ro, 2-rw */ |
79 | unsigned sb_err : 2; /* on errs: 0-cont, 1-ro, 2-panic */ | 77 | unsigned sb_err : 2; /* on errs: 0-cont, 1-ro, 2-panic */ |
80 | unsigned sb_chk : 2; /* checks: 0-no, 1-normal, 2-strict */ | 78 | unsigned sb_chk : 2; /* checks: 0-no, 1-normal, 2-strict */ |
@@ -87,20 +85,9 @@ struct hpfs_sb_info { | |||
87 | unsigned *sb_bmp_dir; /* main bitmap directory */ | 85 | unsigned *sb_bmp_dir; /* main bitmap directory */ |
88 | unsigned sb_c_bitmap; /* current bitmap */ | 86 | unsigned sb_c_bitmap; /* current bitmap */ |
89 | unsigned sb_max_fwd_alloc; /* max forwad allocation */ | 87 | unsigned sb_max_fwd_alloc; /* max forwad allocation */ |
90 | struct mutex hpfs_creation_de; /* when creating dirents, nobody else | ||
91 | can alloc blocks */ | ||
92 | /*unsigned sb_mounting : 1;*/ | ||
93 | int sb_timeshift; | 88 | int sb_timeshift; |
94 | }; | 89 | }; |
95 | 90 | ||
96 | /* | ||
97 | * conv= options | ||
98 | */ | ||
99 | |||
100 | #define CONV_BINARY 0 /* no conversion */ | ||
101 | #define CONV_TEXT 1 /* crlf->newline */ | ||
102 | #define CONV_AUTO 2 /* decide based on file contents */ | ||
103 | |||
104 | /* Four 512-byte buffers and the 2k block obtained by concatenating them */ | 91 | /* Four 512-byte buffers and the 2k block obtained by concatenating them */ |
105 | 92 | ||
106 | struct quad_buffer_head { | 93 | struct quad_buffer_head { |
@@ -113,7 +100,7 @@ struct quad_buffer_head { | |||
113 | static inline dnode_secno de_down_pointer (struct hpfs_dirent *de) | 100 | static inline dnode_secno de_down_pointer (struct hpfs_dirent *de) |
114 | { | 101 | { |
115 | CHKCOND(de->down,("HPFS: de_down_pointer: !de->down\n")); | 102 | CHKCOND(de->down,("HPFS: de_down_pointer: !de->down\n")); |
116 | return *(dnode_secno *) ((void *) de + de->length - 4); | 103 | return le32_to_cpu(*(dnode_secno *) ((void *) de + le16_to_cpu(de->length) - 4)); |
117 | } | 104 | } |
118 | 105 | ||
119 | /* The first dir entry in a dnode */ | 106 | /* The first dir entry in a dnode */ |
@@ -127,41 +114,46 @@ static inline struct hpfs_dirent *dnode_first_de (struct dnode *dnode) | |||
127 | 114 | ||
128 | static inline struct hpfs_dirent *dnode_end_de (struct dnode *dnode) | 115 | static inline struct hpfs_dirent *dnode_end_de (struct dnode *dnode) |
129 | { | 116 | { |
130 | CHKCOND(dnode->first_free>=0x14 && dnode->first_free<=0xa00,("HPFS: dnode_end_de: dnode->first_free = %d\n",(int)dnode->first_free)); | 117 | CHKCOND(le32_to_cpu(dnode->first_free)>=0x14 && le32_to_cpu(dnode->first_free)<=0xa00,("HPFS: dnode_end_de: dnode->first_free = %x\n",(unsigned)le32_to_cpu(dnode->first_free))); |
131 | return (void *) dnode + dnode->first_free; | 118 | return (void *) dnode + le32_to_cpu(dnode->first_free); |
132 | } | 119 | } |
133 | 120 | ||
134 | /* The dir entry after dir entry de */ | 121 | /* The dir entry after dir entry de */ |
135 | 122 | ||
136 | static inline struct hpfs_dirent *de_next_de (struct hpfs_dirent *de) | 123 | static inline struct hpfs_dirent *de_next_de (struct hpfs_dirent *de) |
137 | { | 124 | { |
138 | CHKCOND(de->length>=0x20 && de->length<0x800,("HPFS: de_next_de: de->length = %d\n",(int)de->length)); | 125 | CHKCOND(le16_to_cpu(de->length)>=0x20 && le16_to_cpu(de->length)<0x800,("HPFS: de_next_de: de->length = %x\n",(unsigned)le16_to_cpu(de->length))); |
139 | return (void *) de + de->length; | 126 | return (void *) de + le16_to_cpu(de->length); |
140 | } | 127 | } |
141 | 128 | ||
142 | static inline struct extended_attribute *fnode_ea(struct fnode *fnode) | 129 | static inline struct extended_attribute *fnode_ea(struct fnode *fnode) |
143 | { | 130 | { |
144 | return (struct extended_attribute *)((char *)fnode + fnode->ea_offs + fnode->acl_size_s); | 131 | return (struct extended_attribute *)((char *)fnode + le16_to_cpu(fnode->ea_offs) + le16_to_cpu(fnode->acl_size_s)); |
145 | } | 132 | } |
146 | 133 | ||
147 | static inline struct extended_attribute *fnode_end_ea(struct fnode *fnode) | 134 | static inline struct extended_attribute *fnode_end_ea(struct fnode *fnode) |
148 | { | 135 | { |
149 | return (struct extended_attribute *)((char *)fnode + fnode->ea_offs + fnode->acl_size_s + fnode->ea_size_s); | 136 | return (struct extended_attribute *)((char *)fnode + le16_to_cpu(fnode->ea_offs) + le16_to_cpu(fnode->acl_size_s) + le16_to_cpu(fnode->ea_size_s)); |
137 | } | ||
138 | |||
139 | static unsigned ea_valuelen(struct extended_attribute *ea) | ||
140 | { | ||
141 | return ea->valuelen_lo + 256 * ea->valuelen_hi; | ||
150 | } | 142 | } |
151 | 143 | ||
152 | static inline struct extended_attribute *next_ea(struct extended_attribute *ea) | 144 | static inline struct extended_attribute *next_ea(struct extended_attribute *ea) |
153 | { | 145 | { |
154 | return (struct extended_attribute *)((char *)ea + 5 + ea->namelen + ea->valuelen); | 146 | return (struct extended_attribute *)((char *)ea + 5 + ea->namelen + ea_valuelen(ea)); |
155 | } | 147 | } |
156 | 148 | ||
157 | static inline secno ea_sec(struct extended_attribute *ea) | 149 | static inline secno ea_sec(struct extended_attribute *ea) |
158 | { | 150 | { |
159 | return *(secno *)((char *)ea + 9 + ea->namelen); | 151 | return le32_to_cpu(get_unaligned((secno *)((char *)ea + 9 + ea->namelen))); |
160 | } | 152 | } |
161 | 153 | ||
162 | static inline secno ea_len(struct extended_attribute *ea) | 154 | static inline secno ea_len(struct extended_attribute *ea) |
163 | { | 155 | { |
164 | return *(secno *)((char *)ea + 5 + ea->namelen); | 156 | return le32_to_cpu(get_unaligned((secno *)((char *)ea + 5 + ea->namelen))); |
165 | } | 157 | } |
166 | 158 | ||
167 | static inline char *ea_data(struct extended_attribute *ea) | 159 | static inline char *ea_data(struct extended_attribute *ea) |
@@ -186,13 +178,13 @@ static inline void copy_de(struct hpfs_dirent *dst, struct hpfs_dirent *src) | |||
186 | dst->not_8x3 = n; | 178 | dst->not_8x3 = n; |
187 | } | 179 | } |
188 | 180 | ||
189 | static inline unsigned tstbits(unsigned *bmp, unsigned b, unsigned n) | 181 | static inline unsigned tstbits(u32 *bmp, unsigned b, unsigned n) |
190 | { | 182 | { |
191 | int i; | 183 | int i; |
192 | if ((b >= 0x4000) || (b + n - 1 >= 0x4000)) return n; | 184 | if ((b >= 0x4000) || (b + n - 1 >= 0x4000)) return n; |
193 | if (!((bmp[(b & 0x3fff) >> 5] >> (b & 0x1f)) & 1)) return 1; | 185 | if (!((le32_to_cpu(bmp[(b & 0x3fff) >> 5]) >> (b & 0x1f)) & 1)) return 1; |
194 | for (i = 1; i < n; i++) | 186 | for (i = 1; i < n; i++) |
195 | if (/*b+i < 0x4000 &&*/ !((bmp[((b+i) & 0x3fff) >> 5] >> ((b+i) & 0x1f)) & 1)) | 187 | if (!((le32_to_cpu(bmp[((b+i) & 0x3fff) >> 5]) >> ((b+i) & 0x1f)) & 1)) |
196 | return i + 1; | 188 | return i + 1; |
197 | return 0; | 189 | return 0; |
198 | } | 190 | } |
@@ -200,12 +192,12 @@ static inline unsigned tstbits(unsigned *bmp, unsigned b, unsigned n) | |||
200 | /* alloc.c */ | 192 | /* alloc.c */ |
201 | 193 | ||
202 | int hpfs_chk_sectors(struct super_block *, secno, int, char *); | 194 | int hpfs_chk_sectors(struct super_block *, secno, int, char *); |
203 | secno hpfs_alloc_sector(struct super_block *, secno, unsigned, int, int); | 195 | secno hpfs_alloc_sector(struct super_block *, secno, unsigned, int); |
204 | int hpfs_alloc_if_possible(struct super_block *, secno); | 196 | int hpfs_alloc_if_possible(struct super_block *, secno); |
205 | void hpfs_free_sectors(struct super_block *, secno, unsigned); | 197 | void hpfs_free_sectors(struct super_block *, secno, unsigned); |
206 | int hpfs_check_free_dnodes(struct super_block *, int); | 198 | int hpfs_check_free_dnodes(struct super_block *, int); |
207 | void hpfs_free_dnode(struct super_block *, secno); | 199 | void hpfs_free_dnode(struct super_block *, secno); |
208 | struct dnode *hpfs_alloc_dnode(struct super_block *, secno, dnode_secno *, struct quad_buffer_head *, int); | 200 | struct dnode *hpfs_alloc_dnode(struct super_block *, secno, dnode_secno *, struct quad_buffer_head *); |
209 | struct fnode *hpfs_alloc_fnode(struct super_block *, secno, fnode_secno *, struct buffer_head **); | 201 | struct fnode *hpfs_alloc_fnode(struct super_block *, secno, fnode_secno *, struct buffer_head **); |
210 | struct anode *hpfs_alloc_anode(struct super_block *, secno, anode_secno *, struct buffer_head **); | 202 | struct anode *hpfs_alloc_anode(struct super_block *, secno, anode_secno *, struct buffer_head **); |
211 | 203 | ||
@@ -222,8 +214,6 @@ void hpfs_remove_fnode(struct super_block *, fnode_secno fno); | |||
222 | 214 | ||
223 | /* buffer.c */ | 215 | /* buffer.c */ |
224 | 216 | ||
225 | void hpfs_lock_creation(struct super_block *); | ||
226 | void hpfs_unlock_creation(struct super_block *); | ||
227 | void *hpfs_map_sector(struct super_block *, unsigned, struct buffer_head **, int); | 217 | void *hpfs_map_sector(struct super_block *, unsigned, struct buffer_head **, int); |
228 | void *hpfs_get_sector(struct super_block *, unsigned, struct buffer_head **); | 218 | void *hpfs_get_sector(struct super_block *, unsigned, struct buffer_head **); |
229 | void *hpfs_map_4sectors(struct super_block *, unsigned, struct quad_buffer_head *, int); | 219 | void *hpfs_map_4sectors(struct super_block *, unsigned, struct quad_buffer_head *, int); |
@@ -247,7 +237,7 @@ void hpfs_del_pos(struct inode *, loff_t *); | |||
247 | struct hpfs_dirent *hpfs_add_de(struct super_block *, struct dnode *, | 237 | struct hpfs_dirent *hpfs_add_de(struct super_block *, struct dnode *, |
248 | const unsigned char *, unsigned, secno); | 238 | const unsigned char *, unsigned, secno); |
249 | int hpfs_add_dirent(struct inode *, const unsigned char *, unsigned, | 239 | int hpfs_add_dirent(struct inode *, const unsigned char *, unsigned, |
250 | struct hpfs_dirent *, int); | 240 | struct hpfs_dirent *); |
251 | int hpfs_remove_dirent(struct inode *, dnode_secno, struct hpfs_dirent *, struct quad_buffer_head *, int); | 241 | int hpfs_remove_dirent(struct inode *, dnode_secno, struct hpfs_dirent *, struct quad_buffer_head *, int); |
252 | void hpfs_count_dnodes(struct super_block *, dnode_secno, int *, int *, int *); | 242 | void hpfs_count_dnodes(struct super_block *, dnode_secno, int *, int *, int *); |
253 | dnode_secno hpfs_de_as_down_as_possible(struct super_block *, dnode_secno dno); | 243 | dnode_secno hpfs_de_as_down_as_possible(struct super_block *, dnode_secno dno); |
@@ -303,7 +293,6 @@ int hpfs_compare_names(struct super_block *, const unsigned char *, unsigned, | |||
303 | const unsigned char *, unsigned, int); | 293 | const unsigned char *, unsigned, int); |
304 | int hpfs_is_name_long(const unsigned char *, unsigned); | 294 | int hpfs_is_name_long(const unsigned char *, unsigned); |
305 | void hpfs_adjust_length(const unsigned char *, unsigned *); | 295 | void hpfs_adjust_length(const unsigned char *, unsigned *); |
306 | void hpfs_decide_conv(struct inode *, const unsigned char *, unsigned); | ||
307 | 296 | ||
308 | /* namei.c */ | 297 | /* namei.c */ |
309 | 298 | ||
@@ -346,21 +335,26 @@ static inline time32_t gmt_to_local(struct super_block *s, time_t t) | |||
346 | /* | 335 | /* |
347 | * Locking: | 336 | * Locking: |
348 | * | 337 | * |
349 | * hpfs_lock() is a leftover from the big kernel lock. | 338 | * hpfs_lock() locks the whole filesystem. It must be taken |
350 | * Right now, these functions are empty and only left | 339 | * on any method called by the VFS. |
351 | * for documentation purposes. The file system no longer | ||
352 | * works on SMP systems, so the lock is not needed | ||
353 | * any more. | ||
354 | * | 340 | * |
355 | * If someone is interested in making it work again, this | 341 | * We don't do any per-file locking anymore, it is hard to |
356 | * would be the place to start by adding a per-superblock | 342 | * review and HPFS is not performance-sensitive anyway. |
357 | * mutex and fixing all the bugs and performance issues | ||
358 | * caused by that. | ||
359 | */ | 343 | */ |
360 | static inline void hpfs_lock(struct super_block *s) | 344 | static inline void hpfs_lock(struct super_block *s) |
361 | { | 345 | { |
346 | struct hpfs_sb_info *sbi = hpfs_sb(s); | ||
347 | mutex_lock(&sbi->hpfs_mutex); | ||
362 | } | 348 | } |
363 | 349 | ||
364 | static inline void hpfs_unlock(struct super_block *s) | 350 | static inline void hpfs_unlock(struct super_block *s) |
365 | { | 351 | { |
352 | struct hpfs_sb_info *sbi = hpfs_sb(s); | ||
353 | mutex_unlock(&sbi->hpfs_mutex); | ||
354 | } | ||
355 | |||
356 | static inline void hpfs_lock_assert(struct super_block *s) | ||
357 | { | ||
358 | struct hpfs_sb_info *sbi = hpfs_sb(s); | ||
359 | WARN_ON(!mutex_is_locked(&sbi->hpfs_mutex)); | ||
366 | } | 360 | } |
diff --git a/fs/hpfs/inode.c b/fs/hpfs/inode.c index 87f1f787e767..338cd8368451 100644 --- a/fs/hpfs/inode.c +++ b/fs/hpfs/inode.c | |||
@@ -17,7 +17,6 @@ void hpfs_init_inode(struct inode *i) | |||
17 | i->i_uid = hpfs_sb(sb)->sb_uid; | 17 | i->i_uid = hpfs_sb(sb)->sb_uid; |
18 | i->i_gid = hpfs_sb(sb)->sb_gid; | 18 | i->i_gid = hpfs_sb(sb)->sb_gid; |
19 | i->i_mode = hpfs_sb(sb)->sb_mode; | 19 | i->i_mode = hpfs_sb(sb)->sb_mode; |
20 | hpfs_inode->i_conv = hpfs_sb(sb)->sb_conv; | ||
21 | i->i_size = -1; | 20 | i->i_size = -1; |
22 | i->i_blocks = -1; | 21 | i->i_blocks = -1; |
23 | 22 | ||
@@ -116,8 +115,8 @@ void hpfs_read_inode(struct inode *i) | |||
116 | i->i_mode |= S_IFDIR; | 115 | i->i_mode |= S_IFDIR; |
117 | i->i_op = &hpfs_dir_iops; | 116 | i->i_op = &hpfs_dir_iops; |
118 | i->i_fop = &hpfs_dir_ops; | 117 | i->i_fop = &hpfs_dir_ops; |
119 | hpfs_inode->i_parent_dir = fnode->up; | 118 | hpfs_inode->i_parent_dir = le32_to_cpu(fnode->up); |
120 | hpfs_inode->i_dno = fnode->u.external[0].disk_secno; | 119 | hpfs_inode->i_dno = le32_to_cpu(fnode->u.external[0].disk_secno); |
121 | if (hpfs_sb(sb)->sb_chk >= 2) { | 120 | if (hpfs_sb(sb)->sb_chk >= 2) { |
122 | struct buffer_head *bh0; | 121 | struct buffer_head *bh0; |
123 | if (hpfs_map_fnode(sb, hpfs_inode->i_parent_dir, &bh0)) brelse(bh0); | 122 | if (hpfs_map_fnode(sb, hpfs_inode->i_parent_dir, &bh0)) brelse(bh0); |
@@ -133,7 +132,7 @@ void hpfs_read_inode(struct inode *i) | |||
133 | i->i_op = &hpfs_file_iops; | 132 | i->i_op = &hpfs_file_iops; |
134 | i->i_fop = &hpfs_file_ops; | 133 | i->i_fop = &hpfs_file_ops; |
135 | i->i_nlink = 1; | 134 | i->i_nlink = 1; |
136 | i->i_size = fnode->file_size; | 135 | i->i_size = le32_to_cpu(fnode->file_size); |
137 | i->i_blocks = ((i->i_size + 511) >> 9) + 1; | 136 | i->i_blocks = ((i->i_size + 511) >> 9) + 1; |
138 | i->i_data.a_ops = &hpfs_aops; | 137 | i->i_data.a_ops = &hpfs_aops; |
139 | hpfs_i(i)->mmu_private = i->i_size; | 138 | hpfs_i(i)->mmu_private = i->i_size; |
@@ -144,7 +143,7 @@ void hpfs_read_inode(struct inode *i) | |||
144 | static void hpfs_write_inode_ea(struct inode *i, struct fnode *fnode) | 143 | static void hpfs_write_inode_ea(struct inode *i, struct fnode *fnode) |
145 | { | 144 | { |
146 | struct hpfs_inode_info *hpfs_inode = hpfs_i(i); | 145 | struct hpfs_inode_info *hpfs_inode = hpfs_i(i); |
147 | /*if (fnode->acl_size_l || fnode->acl_size_s) { | 146 | /*if (le32_to_cpu(fnode->acl_size_l) || le16_to_cpu(fnode->acl_size_s)) { |
148 | Some unknown structures like ACL may be in fnode, | 147 | Some unknown structures like ACL may be in fnode, |
149 | we'd better not overwrite them | 148 | we'd better not overwrite them |
150 | hpfs_error(i->i_sb, "fnode %08x has some unknown HPFS386 stuctures", i->i_ino); | 149 | hpfs_error(i->i_sb, "fnode %08x has some unknown HPFS386 stuctures", i->i_ino); |
@@ -187,9 +186,7 @@ void hpfs_write_inode(struct inode *i) | |||
187 | kfree(hpfs_inode->i_rddir_off); | 186 | kfree(hpfs_inode->i_rddir_off); |
188 | hpfs_inode->i_rddir_off = NULL; | 187 | hpfs_inode->i_rddir_off = NULL; |
189 | } | 188 | } |
190 | mutex_lock(&hpfs_inode->i_parent_mutex); | ||
191 | if (!i->i_nlink) { | 189 | if (!i->i_nlink) { |
192 | mutex_unlock(&hpfs_inode->i_parent_mutex); | ||
193 | return; | 190 | return; |
194 | } | 191 | } |
195 | parent = iget_locked(i->i_sb, hpfs_inode->i_parent_dir); | 192 | parent = iget_locked(i->i_sb, hpfs_inode->i_parent_dir); |
@@ -200,14 +197,9 @@ void hpfs_write_inode(struct inode *i) | |||
200 | hpfs_read_inode(parent); | 197 | hpfs_read_inode(parent); |
201 | unlock_new_inode(parent); | 198 | unlock_new_inode(parent); |
202 | } | 199 | } |
203 | mutex_lock(&hpfs_inode->i_mutex); | ||
204 | hpfs_write_inode_nolock(i); | 200 | hpfs_write_inode_nolock(i); |
205 | mutex_unlock(&hpfs_inode->i_mutex); | ||
206 | iput(parent); | 201 | iput(parent); |
207 | } else { | ||
208 | mark_inode_dirty(i); | ||
209 | } | 202 | } |
210 | mutex_unlock(&hpfs_inode->i_parent_mutex); | ||
211 | } | 203 | } |
212 | 204 | ||
213 | void hpfs_write_inode_nolock(struct inode *i) | 205 | void hpfs_write_inode_nolock(struct inode *i) |
@@ -226,30 +218,30 @@ void hpfs_write_inode_nolock(struct inode *i) | |||
226 | } | 218 | } |
227 | } else de = NULL; | 219 | } else de = NULL; |
228 | if (S_ISREG(i->i_mode)) { | 220 | if (S_ISREG(i->i_mode)) { |
229 | fnode->file_size = i->i_size; | 221 | fnode->file_size = cpu_to_le32(i->i_size); |
230 | if (de) de->file_size = i->i_size; | 222 | if (de) de->file_size = cpu_to_le32(i->i_size); |
231 | } else if (S_ISDIR(i->i_mode)) { | 223 | } else if (S_ISDIR(i->i_mode)) { |
232 | fnode->file_size = 0; | 224 | fnode->file_size = cpu_to_le32(0); |
233 | if (de) de->file_size = 0; | 225 | if (de) de->file_size = cpu_to_le32(0); |
234 | } | 226 | } |
235 | hpfs_write_inode_ea(i, fnode); | 227 | hpfs_write_inode_ea(i, fnode); |
236 | if (de) { | 228 | if (de) { |
237 | de->write_date = gmt_to_local(i->i_sb, i->i_mtime.tv_sec); | 229 | de->write_date = cpu_to_le32(gmt_to_local(i->i_sb, i->i_mtime.tv_sec)); |
238 | de->read_date = gmt_to_local(i->i_sb, i->i_atime.tv_sec); | 230 | de->read_date = cpu_to_le32(gmt_to_local(i->i_sb, i->i_atime.tv_sec)); |
239 | de->creation_date = gmt_to_local(i->i_sb, i->i_ctime.tv_sec); | 231 | de->creation_date = cpu_to_le32(gmt_to_local(i->i_sb, i->i_ctime.tv_sec)); |
240 | de->read_only = !(i->i_mode & 0222); | 232 | de->read_only = !(i->i_mode & 0222); |
241 | de->ea_size = hpfs_inode->i_ea_size; | 233 | de->ea_size = cpu_to_le32(hpfs_inode->i_ea_size); |
242 | hpfs_mark_4buffers_dirty(&qbh); | 234 | hpfs_mark_4buffers_dirty(&qbh); |
243 | hpfs_brelse4(&qbh); | 235 | hpfs_brelse4(&qbh); |
244 | } | 236 | } |
245 | if (S_ISDIR(i->i_mode)) { | 237 | if (S_ISDIR(i->i_mode)) { |
246 | if ((de = map_dirent(i, hpfs_inode->i_dno, "\001\001", 2, NULL, &qbh))) { | 238 | if ((de = map_dirent(i, hpfs_inode->i_dno, "\001\001", 2, NULL, &qbh))) { |
247 | de->write_date = gmt_to_local(i->i_sb, i->i_mtime.tv_sec); | 239 | de->write_date = cpu_to_le32(gmt_to_local(i->i_sb, i->i_mtime.tv_sec)); |
248 | de->read_date = gmt_to_local(i->i_sb, i->i_atime.tv_sec); | 240 | de->read_date = cpu_to_le32(gmt_to_local(i->i_sb, i->i_atime.tv_sec)); |
249 | de->creation_date = gmt_to_local(i->i_sb, i->i_ctime.tv_sec); | 241 | de->creation_date = cpu_to_le32(gmt_to_local(i->i_sb, i->i_ctime.tv_sec)); |
250 | de->read_only = !(i->i_mode & 0222); | 242 | de->read_only = !(i->i_mode & 0222); |
251 | de->ea_size = /*hpfs_inode->i_ea_size*/0; | 243 | de->ea_size = cpu_to_le32(/*hpfs_inode->i_ea_size*/0); |
252 | de->file_size = 0; | 244 | de->file_size = cpu_to_le32(0); |
253 | hpfs_mark_4buffers_dirty(&qbh); | 245 | hpfs_mark_4buffers_dirty(&qbh); |
254 | hpfs_brelse4(&qbh); | 246 | hpfs_brelse4(&qbh); |
255 | } else | 247 | } else |
@@ -269,6 +261,10 @@ int hpfs_setattr(struct dentry *dentry, struct iattr *attr) | |||
269 | hpfs_lock(inode->i_sb); | 261 | hpfs_lock(inode->i_sb); |
270 | if (inode->i_ino == hpfs_sb(inode->i_sb)->sb_root) | 262 | if (inode->i_ino == hpfs_sb(inode->i_sb)->sb_root) |
271 | goto out_unlock; | 263 | goto out_unlock; |
264 | if ((attr->ia_valid & ATTR_UID) && attr->ia_uid >= 0x10000) | ||
265 | goto out_unlock; | ||
266 | if ((attr->ia_valid & ATTR_GID) && attr->ia_gid >= 0x10000) | ||
267 | goto out_unlock; | ||
272 | if ((attr->ia_valid & ATTR_SIZE) && attr->ia_size > inode->i_size) | 268 | if ((attr->ia_valid & ATTR_SIZE) && attr->ia_size > inode->i_size) |
273 | goto out_unlock; | 269 | goto out_unlock; |
274 | 270 | ||
@@ -284,7 +280,6 @@ int hpfs_setattr(struct dentry *dentry, struct iattr *attr) | |||
284 | } | 280 | } |
285 | 281 | ||
286 | setattr_copy(inode, attr); | 282 | setattr_copy(inode, attr); |
287 | mark_inode_dirty(inode); | ||
288 | 283 | ||
289 | hpfs_write_inode(inode); | 284 | hpfs_write_inode(inode); |
290 | 285 | ||
diff --git a/fs/hpfs/map.c b/fs/hpfs/map.c index 840d033ecee8..a790821366a7 100644 --- a/fs/hpfs/map.c +++ b/fs/hpfs/map.c | |||
@@ -21,7 +21,7 @@ unsigned int *hpfs_map_bitmap(struct super_block *s, unsigned bmp_block, | |||
21 | hpfs_error(s, "hpfs_map_bitmap called with bad parameter: %08x at %s", bmp_block, id); | 21 | hpfs_error(s, "hpfs_map_bitmap called with bad parameter: %08x at %s", bmp_block, id); |
22 | return NULL; | 22 | return NULL; |
23 | } | 23 | } |
24 | sec = hpfs_sb(s)->sb_bmp_dir[bmp_block]; | 24 | sec = le32_to_cpu(hpfs_sb(s)->sb_bmp_dir[bmp_block]); |
25 | if (!sec || sec > hpfs_sb(s)->sb_fs_size-4) { | 25 | if (!sec || sec > hpfs_sb(s)->sb_fs_size-4) { |
26 | hpfs_error(s, "invalid bitmap block pointer %08x -> %08x at %s", bmp_block, sec, id); | 26 | hpfs_error(s, "invalid bitmap block pointer %08x -> %08x at %s", bmp_block, sec, id); |
27 | return NULL; | 27 | return NULL; |
@@ -46,18 +46,18 @@ unsigned char *hpfs_load_code_page(struct super_block *s, secno cps) | |||
46 | struct code_page_data *cpd; | 46 | struct code_page_data *cpd; |
47 | struct code_page_directory *cp = hpfs_map_sector(s, cps, &bh, 0); | 47 | struct code_page_directory *cp = hpfs_map_sector(s, cps, &bh, 0); |
48 | if (!cp) return NULL; | 48 | if (!cp) return NULL; |
49 | if (cp->magic != CP_DIR_MAGIC) { | 49 | if (le32_to_cpu(cp->magic) != CP_DIR_MAGIC) { |
50 | printk("HPFS: Code page directory magic doesn't match (magic = %08x)\n", cp->magic); | 50 | printk("HPFS: Code page directory magic doesn't match (magic = %08x)\n", le32_to_cpu(cp->magic)); |
51 | brelse(bh); | 51 | brelse(bh); |
52 | return NULL; | 52 | return NULL; |
53 | } | 53 | } |
54 | if (!cp->n_code_pages) { | 54 | if (!le32_to_cpu(cp->n_code_pages)) { |
55 | printk("HPFS: n_code_pages == 0\n"); | 55 | printk("HPFS: n_code_pages == 0\n"); |
56 | brelse(bh); | 56 | brelse(bh); |
57 | return NULL; | 57 | return NULL; |
58 | } | 58 | } |
59 | cpds = cp->array[0].code_page_data; | 59 | cpds = le32_to_cpu(cp->array[0].code_page_data); |
60 | cpi = cp->array[0].index; | 60 | cpi = le16_to_cpu(cp->array[0].index); |
61 | brelse(bh); | 61 | brelse(bh); |
62 | 62 | ||
63 | if (cpi >= 3) { | 63 | if (cpi >= 3) { |
@@ -66,12 +66,12 @@ unsigned char *hpfs_load_code_page(struct super_block *s, secno cps) | |||
66 | } | 66 | } |
67 | 67 | ||
68 | if (!(cpd = hpfs_map_sector(s, cpds, &bh, 0))) return NULL; | 68 | if (!(cpd = hpfs_map_sector(s, cpds, &bh, 0))) return NULL; |
69 | if ((unsigned)cpd->offs[cpi] > 0x178) { | 69 | if (le16_to_cpu(cpd->offs[cpi]) > 0x178) { |
70 | printk("HPFS: Code page index out of sector\n"); | 70 | printk("HPFS: Code page index out of sector\n"); |
71 | brelse(bh); | 71 | brelse(bh); |
72 | return NULL; | 72 | return NULL; |
73 | } | 73 | } |
74 | ptr = (unsigned char *)cpd + cpd->offs[cpi] + 6; | 74 | ptr = (unsigned char *)cpd + le16_to_cpu(cpd->offs[cpi]) + 6; |
75 | if (!(cp_table = kmalloc(256, GFP_KERNEL))) { | 75 | if (!(cp_table = kmalloc(256, GFP_KERNEL))) { |
76 | printk("HPFS: out of memory for code page table\n"); | 76 | printk("HPFS: out of memory for code page table\n"); |
77 | brelse(bh); | 77 | brelse(bh); |
@@ -125,7 +125,7 @@ struct fnode *hpfs_map_fnode(struct super_block *s, ino_t ino, struct buffer_hea | |||
125 | if (hpfs_sb(s)->sb_chk) { | 125 | if (hpfs_sb(s)->sb_chk) { |
126 | struct extended_attribute *ea; | 126 | struct extended_attribute *ea; |
127 | struct extended_attribute *ea_end; | 127 | struct extended_attribute *ea_end; |
128 | if (fnode->magic != FNODE_MAGIC) { | 128 | if (le32_to_cpu(fnode->magic) != FNODE_MAGIC) { |
129 | hpfs_error(s, "bad magic on fnode %08lx", | 129 | hpfs_error(s, "bad magic on fnode %08lx", |
130 | (unsigned long)ino); | 130 | (unsigned long)ino); |
131 | goto bail; | 131 | goto bail; |
@@ -138,7 +138,7 @@ struct fnode *hpfs_map_fnode(struct super_block *s, ino_t ino, struct buffer_hea | |||
138 | (unsigned long)ino); | 138 | (unsigned long)ino); |
139 | goto bail; | 139 | goto bail; |
140 | } | 140 | } |
141 | if (fnode->btree.first_free != | 141 | if (le16_to_cpu(fnode->btree.first_free) != |
142 | 8 + fnode->btree.n_used_nodes * (fnode->btree.internal ? 8 : 12)) { | 142 | 8 + fnode->btree.n_used_nodes * (fnode->btree.internal ? 8 : 12)) { |
143 | hpfs_error(s, | 143 | hpfs_error(s, |
144 | "bad first_free pointer in fnode %08lx", | 144 | "bad first_free pointer in fnode %08lx", |
@@ -146,12 +146,12 @@ struct fnode *hpfs_map_fnode(struct super_block *s, ino_t ino, struct buffer_hea | |||
146 | goto bail; | 146 | goto bail; |
147 | } | 147 | } |
148 | } | 148 | } |
149 | if (fnode->ea_size_s && ((signed int)fnode->ea_offs < 0xc4 || | 149 | if (le16_to_cpu(fnode->ea_size_s) && (le16_to_cpu(fnode->ea_offs) < 0xc4 || |
150 | (signed int)fnode->ea_offs + fnode->acl_size_s + fnode->ea_size_s > 0x200)) { | 150 | le16_to_cpu(fnode->ea_offs) + le16_to_cpu(fnode->acl_size_s) + le16_to_cpu(fnode->ea_size_s) > 0x200)) { |
151 | hpfs_error(s, | 151 | hpfs_error(s, |
152 | "bad EA info in fnode %08lx: ea_offs == %04x ea_size_s == %04x", | 152 | "bad EA info in fnode %08lx: ea_offs == %04x ea_size_s == %04x", |
153 | (unsigned long)ino, | 153 | (unsigned long)ino, |
154 | fnode->ea_offs, fnode->ea_size_s); | 154 | le16_to_cpu(fnode->ea_offs), le16_to_cpu(fnode->ea_size_s)); |
155 | goto bail; | 155 | goto bail; |
156 | } | 156 | } |
157 | ea = fnode_ea(fnode); | 157 | ea = fnode_ea(fnode); |
@@ -178,16 +178,20 @@ struct anode *hpfs_map_anode(struct super_block *s, anode_secno ano, struct buff | |||
178 | if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, ano, 1, "anode")) return NULL; | 178 | if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, ano, 1, "anode")) return NULL; |
179 | if ((anode = hpfs_map_sector(s, ano, bhp, ANODE_RD_AHEAD))) | 179 | if ((anode = hpfs_map_sector(s, ano, bhp, ANODE_RD_AHEAD))) |
180 | if (hpfs_sb(s)->sb_chk) { | 180 | if (hpfs_sb(s)->sb_chk) { |
181 | if (anode->magic != ANODE_MAGIC || anode->self != ano) { | 181 | if (le32_to_cpu(anode->magic) != ANODE_MAGIC) { |
182 | hpfs_error(s, "bad magic on anode %08x", ano); | 182 | hpfs_error(s, "bad magic on anode %08x", ano); |
183 | goto bail; | 183 | goto bail; |
184 | } | 184 | } |
185 | if (le32_to_cpu(anode->self) != ano) { | ||
186 | hpfs_error(s, "self pointer invalid on anode %08x", ano); | ||
187 | goto bail; | ||
188 | } | ||
185 | if ((unsigned)anode->btree.n_used_nodes + (unsigned)anode->btree.n_free_nodes != | 189 | if ((unsigned)anode->btree.n_used_nodes + (unsigned)anode->btree.n_free_nodes != |
186 | (anode->btree.internal ? 60 : 40)) { | 190 | (anode->btree.internal ? 60 : 40)) { |
187 | hpfs_error(s, "bad number of nodes in anode %08x", ano); | 191 | hpfs_error(s, "bad number of nodes in anode %08x", ano); |
188 | goto bail; | 192 | goto bail; |
189 | } | 193 | } |
190 | if (anode->btree.first_free != | 194 | if (le16_to_cpu(anode->btree.first_free) != |
191 | 8 + anode->btree.n_used_nodes * (anode->btree.internal ? 8 : 12)) { | 195 | 8 + anode->btree.n_used_nodes * (anode->btree.internal ? 8 : 12)) { |
192 | hpfs_error(s, "bad first_free pointer in anode %08x", ano); | 196 | hpfs_error(s, "bad first_free pointer in anode %08x", ano); |
193 | goto bail; | 197 | goto bail; |
@@ -219,26 +223,26 @@ struct dnode *hpfs_map_dnode(struct super_block *s, unsigned secno, | |||
219 | unsigned p, pp = 0; | 223 | unsigned p, pp = 0; |
220 | unsigned char *d = (unsigned char *)dnode; | 224 | unsigned char *d = (unsigned char *)dnode; |
221 | int b = 0; | 225 | int b = 0; |
222 | if (dnode->magic != DNODE_MAGIC) { | 226 | if (le32_to_cpu(dnode->magic) != DNODE_MAGIC) { |
223 | hpfs_error(s, "bad magic on dnode %08x", secno); | 227 | hpfs_error(s, "bad magic on dnode %08x", secno); |
224 | goto bail; | 228 | goto bail; |
225 | } | 229 | } |
226 | if (dnode->self != secno) | 230 | if (le32_to_cpu(dnode->self) != secno) |
227 | hpfs_error(s, "bad self pointer on dnode %08x self = %08x", secno, dnode->self); | 231 | hpfs_error(s, "bad self pointer on dnode %08x self = %08x", secno, le32_to_cpu(dnode->self)); |
228 | /* Check dirents - bad dirents would cause infinite | 232 | /* Check dirents - bad dirents would cause infinite |
229 | loops or shooting to memory */ | 233 | loops or shooting to memory */ |
230 | if (dnode->first_free > 2048/* || dnode->first_free < 84*/) { | 234 | if (le32_to_cpu(dnode->first_free) > 2048) { |
231 | hpfs_error(s, "dnode %08x has first_free == %08x", secno, dnode->first_free); | 235 | hpfs_error(s, "dnode %08x has first_free == %08x", secno, le32_to_cpu(dnode->first_free)); |
232 | goto bail; | 236 | goto bail; |
233 | } | 237 | } |
234 | for (p = 20; p < dnode->first_free; p += d[p] + (d[p+1] << 8)) { | 238 | for (p = 20; p < le32_to_cpu(dnode->first_free); p += d[p] + (d[p+1] << 8)) { |
235 | struct hpfs_dirent *de = (struct hpfs_dirent *)((char *)dnode + p); | 239 | struct hpfs_dirent *de = (struct hpfs_dirent *)((char *)dnode + p); |
236 | if (de->length > 292 || (de->length < 32) || (de->length & 3) || p + de->length > 2048) { | 240 | if (le16_to_cpu(de->length) > 292 || (le16_to_cpu(de->length) < 32) || (le16_to_cpu(de->length) & 3) || p + le16_to_cpu(de->length) > 2048) { |
237 | hpfs_error(s, "bad dirent size in dnode %08x, dirent %03x, last %03x", secno, p, pp); | 241 | hpfs_error(s, "bad dirent size in dnode %08x, dirent %03x, last %03x", secno, p, pp); |
238 | goto bail; | 242 | goto bail; |
239 | } | 243 | } |
240 | if (((31 + de->namelen + de->down*4 + 3) & ~3) != de->length) { | 244 | if (((31 + de->namelen + de->down*4 + 3) & ~3) != le16_to_cpu(de->length)) { |
241 | if (((31 + de->namelen + de->down*4 + 3) & ~3) < de->length && s->s_flags & MS_RDONLY) goto ok; | 245 | if (((31 + de->namelen + de->down*4 + 3) & ~3) < le16_to_cpu(de->length) && s->s_flags & MS_RDONLY) goto ok; |
242 | hpfs_error(s, "namelen does not match dirent size in dnode %08x, dirent %03x, last %03x", secno, p, pp); | 246 | hpfs_error(s, "namelen does not match dirent size in dnode %08x, dirent %03x, last %03x", secno, p, pp); |
243 | goto bail; | 247 | goto bail; |
244 | } | 248 | } |
@@ -251,7 +255,7 @@ struct dnode *hpfs_map_dnode(struct super_block *s, unsigned secno, | |||
251 | pp = p; | 255 | pp = p; |
252 | 256 | ||
253 | } | 257 | } |
254 | if (p != dnode->first_free) { | 258 | if (p != le32_to_cpu(dnode->first_free)) { |
255 | hpfs_error(s, "size on last dirent does not match first_free; dnode %08x", secno); | 259 | hpfs_error(s, "size on last dirent does not match first_free; dnode %08x", secno); |
256 | goto bail; | 260 | goto bail; |
257 | } | 261 | } |
@@ -277,7 +281,7 @@ dnode_secno hpfs_fnode_dno(struct super_block *s, ino_t ino) | |||
277 | if (!fnode) | 281 | if (!fnode) |
278 | return 0; | 282 | return 0; |
279 | 283 | ||
280 | dno = fnode->u.external[0].disk_secno; | 284 | dno = le32_to_cpu(fnode->u.external[0].disk_secno); |
281 | brelse(bh); | 285 | brelse(bh); |
282 | return dno; | 286 | return dno; |
283 | } | 287 | } |
diff --git a/fs/hpfs/name.c b/fs/hpfs/name.c index f24736d7a439..9acdf338def0 100644 --- a/fs/hpfs/name.c +++ b/fs/hpfs/name.c | |||
@@ -8,39 +8,6 @@ | |||
8 | 8 | ||
9 | #include "hpfs_fn.h" | 9 | #include "hpfs_fn.h" |
10 | 10 | ||
11 | static const char *text_postfix[]={ | ||
12 | ".ASM", ".BAS", ".BAT", ".C", ".CC", ".CFG", ".CMD", ".CON", ".CPP", ".DEF", | ||
13 | ".DOC", ".DPR", ".ERX", ".H", ".HPP", ".HTM", ".HTML", ".JAVA", ".LOG", ".PAS", | ||
14 | ".RC", ".TEX", ".TXT", ".Y", ""}; | ||
15 | |||
16 | static const char *text_prefix[]={ | ||
17 | "AUTOEXEC.", "CHANGES", "COPYING", "CONFIG.", "CREDITS", "FAQ", "FILE_ID.DIZ", | ||
18 | "MAKEFILE", "READ.ME", "README", "TERMCAP", ""}; | ||
19 | |||
20 | void hpfs_decide_conv(struct inode *inode, const unsigned char *name, unsigned len) | ||
21 | { | ||
22 | struct hpfs_inode_info *hpfs_inode = hpfs_i(inode); | ||
23 | int i; | ||
24 | if (hpfs_inode->i_conv != CONV_AUTO) return; | ||
25 | for (i = 0; *text_postfix[i]; i++) { | ||
26 | int l = strlen(text_postfix[i]); | ||
27 | if (l <= len) | ||
28 | if (!hpfs_compare_names(inode->i_sb, text_postfix[i], l, name + len - l, l, 0)) | ||
29 | goto text; | ||
30 | } | ||
31 | for (i = 0; *text_prefix[i]; i++) { | ||
32 | int l = strlen(text_prefix[i]); | ||
33 | if (l <= len) | ||
34 | if (!hpfs_compare_names(inode->i_sb, text_prefix[i], l, name, l, 0)) | ||
35 | goto text; | ||
36 | } | ||
37 | hpfs_inode->i_conv = CONV_BINARY; | ||
38 | return; | ||
39 | text: | ||
40 | hpfs_inode->i_conv = CONV_TEXT; | ||
41 | return; | ||
42 | } | ||
43 | |||
44 | static inline int not_allowed_char(unsigned char c) | 11 | static inline int not_allowed_char(unsigned char c) |
45 | { | 12 | { |
46 | return c<' ' || c=='"' || c=='*' || c=='/' || c==':' || c=='<' || | 13 | return c<' ' || c=='"' || c=='*' || c=='/' || c==':' || c=='<' || |
diff --git a/fs/hpfs/namei.c b/fs/hpfs/namei.c index d5f8c8a19023..1f05839c27a7 100644 --- a/fs/hpfs/namei.c +++ b/fs/hpfs/namei.c | |||
@@ -29,7 +29,7 @@ static int hpfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
29 | fnode = hpfs_alloc_fnode(dir->i_sb, hpfs_i(dir)->i_dno, &fno, &bh); | 29 | fnode = hpfs_alloc_fnode(dir->i_sb, hpfs_i(dir)->i_dno, &fno, &bh); |
30 | if (!fnode) | 30 | if (!fnode) |
31 | goto bail; | 31 | goto bail; |
32 | dnode = hpfs_alloc_dnode(dir->i_sb, fno, &dno, &qbh0, 1); | 32 | dnode = hpfs_alloc_dnode(dir->i_sb, fno, &dno, &qbh0); |
33 | if (!dnode) | 33 | if (!dnode) |
34 | goto bail1; | 34 | goto bail1; |
35 | memset(&dee, 0, sizeof dee); | 35 | memset(&dee, 0, sizeof dee); |
@@ -37,8 +37,8 @@ static int hpfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
37 | if (!(mode & 0222)) dee.read_only = 1; | 37 | if (!(mode & 0222)) dee.read_only = 1; |
38 | /*dee.archive = 0;*/ | 38 | /*dee.archive = 0;*/ |
39 | dee.hidden = name[0] == '.'; | 39 | dee.hidden = name[0] == '.'; |
40 | dee.fnode = fno; | 40 | dee.fnode = cpu_to_le32(fno); |
41 | dee.creation_date = dee.write_date = dee.read_date = gmt_to_local(dir->i_sb, get_seconds()); | 41 | dee.creation_date = dee.write_date = dee.read_date = cpu_to_le32(gmt_to_local(dir->i_sb, get_seconds())); |
42 | result = new_inode(dir->i_sb); | 42 | result = new_inode(dir->i_sb); |
43 | if (!result) | 43 | if (!result) |
44 | goto bail2; | 44 | goto bail2; |
@@ -46,7 +46,7 @@ static int hpfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
46 | result->i_ino = fno; | 46 | result->i_ino = fno; |
47 | hpfs_i(result)->i_parent_dir = dir->i_ino; | 47 | hpfs_i(result)->i_parent_dir = dir->i_ino; |
48 | hpfs_i(result)->i_dno = dno; | 48 | hpfs_i(result)->i_dno = dno; |
49 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, dee.creation_date); | 49 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, le32_to_cpu(dee.creation_date)); |
50 | result->i_ctime.tv_nsec = 0; | 50 | result->i_ctime.tv_nsec = 0; |
51 | result->i_mtime.tv_nsec = 0; | 51 | result->i_mtime.tv_nsec = 0; |
52 | result->i_atime.tv_nsec = 0; | 52 | result->i_atime.tv_nsec = 0; |
@@ -60,8 +60,7 @@ static int hpfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
60 | if (dee.read_only) | 60 | if (dee.read_only) |
61 | result->i_mode &= ~0222; | 61 | result->i_mode &= ~0222; |
62 | 62 | ||
63 | mutex_lock(&hpfs_i(dir)->i_mutex); | 63 | r = hpfs_add_dirent(dir, name, len, &dee); |
64 | r = hpfs_add_dirent(dir, name, len, &dee, 0); | ||
65 | if (r == 1) | 64 | if (r == 1) |
66 | goto bail3; | 65 | goto bail3; |
67 | if (r == -1) { | 66 | if (r == -1) { |
@@ -70,21 +69,21 @@ static int hpfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
70 | } | 69 | } |
71 | fnode->len = len; | 70 | fnode->len = len; |
72 | memcpy(fnode->name, name, len > 15 ? 15 : len); | 71 | memcpy(fnode->name, name, len > 15 ? 15 : len); |
73 | fnode->up = dir->i_ino; | 72 | fnode->up = cpu_to_le32(dir->i_ino); |
74 | fnode->dirflag = 1; | 73 | fnode->dirflag = 1; |
75 | fnode->btree.n_free_nodes = 7; | 74 | fnode->btree.n_free_nodes = 7; |
76 | fnode->btree.n_used_nodes = 1; | 75 | fnode->btree.n_used_nodes = 1; |
77 | fnode->btree.first_free = 0x14; | 76 | fnode->btree.first_free = cpu_to_le16(0x14); |
78 | fnode->u.external[0].disk_secno = dno; | 77 | fnode->u.external[0].disk_secno = cpu_to_le32(dno); |
79 | fnode->u.external[0].file_secno = -1; | 78 | fnode->u.external[0].file_secno = cpu_to_le32(-1); |
80 | dnode->root_dnode = 1; | 79 | dnode->root_dnode = 1; |
81 | dnode->up = fno; | 80 | dnode->up = cpu_to_le32(fno); |
82 | de = hpfs_add_de(dir->i_sb, dnode, "\001\001", 2, 0); | 81 | de = hpfs_add_de(dir->i_sb, dnode, "\001\001", 2, 0); |
83 | de->creation_date = de->write_date = de->read_date = gmt_to_local(dir->i_sb, get_seconds()); | 82 | de->creation_date = de->write_date = de->read_date = cpu_to_le32(gmt_to_local(dir->i_sb, get_seconds())); |
84 | if (!(mode & 0222)) de->read_only = 1; | 83 | if (!(mode & 0222)) de->read_only = 1; |
85 | de->first = de->directory = 1; | 84 | de->first = de->directory = 1; |
86 | /*de->hidden = de->system = 0;*/ | 85 | /*de->hidden = de->system = 0;*/ |
87 | de->fnode = fno; | 86 | de->fnode = cpu_to_le32(fno); |
88 | mark_buffer_dirty(bh); | 87 | mark_buffer_dirty(bh); |
89 | brelse(bh); | 88 | brelse(bh); |
90 | hpfs_mark_4buffers_dirty(&qbh0); | 89 | hpfs_mark_4buffers_dirty(&qbh0); |
@@ -101,11 +100,9 @@ static int hpfs_mkdir(struct inode *dir, struct dentry *dentry, int mode) | |||
101 | hpfs_write_inode_nolock(result); | 100 | hpfs_write_inode_nolock(result); |
102 | } | 101 | } |
103 | d_instantiate(dentry, result); | 102 | d_instantiate(dentry, result); |
104 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
105 | hpfs_unlock(dir->i_sb); | 103 | hpfs_unlock(dir->i_sb); |
106 | return 0; | 104 | return 0; |
107 | bail3: | 105 | bail3: |
108 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
109 | iput(result); | 106 | iput(result); |
110 | bail2: | 107 | bail2: |
111 | hpfs_brelse4(&qbh0); | 108 | hpfs_brelse4(&qbh0); |
@@ -140,8 +137,8 @@ static int hpfs_create(struct inode *dir, struct dentry *dentry, int mode, struc | |||
140 | if (!(mode & 0222)) dee.read_only = 1; | 137 | if (!(mode & 0222)) dee.read_only = 1; |
141 | dee.archive = 1; | 138 | dee.archive = 1; |
142 | dee.hidden = name[0] == '.'; | 139 | dee.hidden = name[0] == '.'; |
143 | dee.fnode = fno; | 140 | dee.fnode = cpu_to_le32(fno); |
144 | dee.creation_date = dee.write_date = dee.read_date = gmt_to_local(dir->i_sb, get_seconds()); | 141 | dee.creation_date = dee.write_date = dee.read_date = cpu_to_le32(gmt_to_local(dir->i_sb, get_seconds())); |
145 | 142 | ||
146 | result = new_inode(dir->i_sb); | 143 | result = new_inode(dir->i_sb); |
147 | if (!result) | 144 | if (!result) |
@@ -154,9 +151,8 @@ static int hpfs_create(struct inode *dir, struct dentry *dentry, int mode, struc | |||
154 | result->i_op = &hpfs_file_iops; | 151 | result->i_op = &hpfs_file_iops; |
155 | result->i_fop = &hpfs_file_ops; | 152 | result->i_fop = &hpfs_file_ops; |
156 | result->i_nlink = 1; | 153 | result->i_nlink = 1; |
157 | hpfs_decide_conv(result, name, len); | ||
158 | hpfs_i(result)->i_parent_dir = dir->i_ino; | 154 | hpfs_i(result)->i_parent_dir = dir->i_ino; |
159 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, dee.creation_date); | 155 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, le32_to_cpu(dee.creation_date)); |
160 | result->i_ctime.tv_nsec = 0; | 156 | result->i_ctime.tv_nsec = 0; |
161 | result->i_mtime.tv_nsec = 0; | 157 | result->i_mtime.tv_nsec = 0; |
162 | result->i_atime.tv_nsec = 0; | 158 | result->i_atime.tv_nsec = 0; |
@@ -168,8 +164,7 @@ static int hpfs_create(struct inode *dir, struct dentry *dentry, int mode, struc | |||
168 | result->i_data.a_ops = &hpfs_aops; | 164 | result->i_data.a_ops = &hpfs_aops; |
169 | hpfs_i(result)->mmu_private = 0; | 165 | hpfs_i(result)->mmu_private = 0; |
170 | 166 | ||
171 | mutex_lock(&hpfs_i(dir)->i_mutex); | 167 | r = hpfs_add_dirent(dir, name, len, &dee); |
172 | r = hpfs_add_dirent(dir, name, len, &dee, 0); | ||
173 | if (r == 1) | 168 | if (r == 1) |
174 | goto bail2; | 169 | goto bail2; |
175 | if (r == -1) { | 170 | if (r == -1) { |
@@ -178,7 +173,7 @@ static int hpfs_create(struct inode *dir, struct dentry *dentry, int mode, struc | |||
178 | } | 173 | } |
179 | fnode->len = len; | 174 | fnode->len = len; |
180 | memcpy(fnode->name, name, len > 15 ? 15 : len); | 175 | memcpy(fnode->name, name, len > 15 ? 15 : len); |
181 | fnode->up = dir->i_ino; | 176 | fnode->up = cpu_to_le32(dir->i_ino); |
182 | mark_buffer_dirty(bh); | 177 | mark_buffer_dirty(bh); |
183 | brelse(bh); | 178 | brelse(bh); |
184 | 179 | ||
@@ -193,12 +188,10 @@ static int hpfs_create(struct inode *dir, struct dentry *dentry, int mode, struc | |||
193 | hpfs_write_inode_nolock(result); | 188 | hpfs_write_inode_nolock(result); |
194 | } | 189 | } |
195 | d_instantiate(dentry, result); | 190 | d_instantiate(dentry, result); |
196 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
197 | hpfs_unlock(dir->i_sb); | 191 | hpfs_unlock(dir->i_sb); |
198 | return 0; | 192 | return 0; |
199 | 193 | ||
200 | bail2: | 194 | bail2: |
201 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
202 | iput(result); | 195 | iput(result); |
203 | bail1: | 196 | bail1: |
204 | brelse(bh); | 197 | brelse(bh); |
@@ -232,8 +225,8 @@ static int hpfs_mknod(struct inode *dir, struct dentry *dentry, int mode, dev_t | |||
232 | if (!(mode & 0222)) dee.read_only = 1; | 225 | if (!(mode & 0222)) dee.read_only = 1; |
233 | dee.archive = 1; | 226 | dee.archive = 1; |
234 | dee.hidden = name[0] == '.'; | 227 | dee.hidden = name[0] == '.'; |
235 | dee.fnode = fno; | 228 | dee.fnode = cpu_to_le32(fno); |
236 | dee.creation_date = dee.write_date = dee.read_date = gmt_to_local(dir->i_sb, get_seconds()); | 229 | dee.creation_date = dee.write_date = dee.read_date = cpu_to_le32(gmt_to_local(dir->i_sb, get_seconds())); |
237 | 230 | ||
238 | result = new_inode(dir->i_sb); | 231 | result = new_inode(dir->i_sb); |
239 | if (!result) | 232 | if (!result) |
@@ -242,7 +235,7 @@ static int hpfs_mknod(struct inode *dir, struct dentry *dentry, int mode, dev_t | |||
242 | hpfs_init_inode(result); | 235 | hpfs_init_inode(result); |
243 | result->i_ino = fno; | 236 | result->i_ino = fno; |
244 | hpfs_i(result)->i_parent_dir = dir->i_ino; | 237 | hpfs_i(result)->i_parent_dir = dir->i_ino; |
245 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, dee.creation_date); | 238 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, le32_to_cpu(dee.creation_date)); |
246 | result->i_ctime.tv_nsec = 0; | 239 | result->i_ctime.tv_nsec = 0; |
247 | result->i_mtime.tv_nsec = 0; | 240 | result->i_mtime.tv_nsec = 0; |
248 | result->i_atime.tv_nsec = 0; | 241 | result->i_atime.tv_nsec = 0; |
@@ -254,8 +247,7 @@ static int hpfs_mknod(struct inode *dir, struct dentry *dentry, int mode, dev_t | |||
254 | result->i_blocks = 1; | 247 | result->i_blocks = 1; |
255 | init_special_inode(result, mode, rdev); | 248 | init_special_inode(result, mode, rdev); |
256 | 249 | ||
257 | mutex_lock(&hpfs_i(dir)->i_mutex); | 250 | r = hpfs_add_dirent(dir, name, len, &dee); |
258 | r = hpfs_add_dirent(dir, name, len, &dee, 0); | ||
259 | if (r == 1) | 251 | if (r == 1) |
260 | goto bail2; | 252 | goto bail2; |
261 | if (r == -1) { | 253 | if (r == -1) { |
@@ -264,19 +256,17 @@ static int hpfs_mknod(struct inode *dir, struct dentry *dentry, int mode, dev_t | |||
264 | } | 256 | } |
265 | fnode->len = len; | 257 | fnode->len = len; |
266 | memcpy(fnode->name, name, len > 15 ? 15 : len); | 258 | memcpy(fnode->name, name, len > 15 ? 15 : len); |
267 | fnode->up = dir->i_ino; | 259 | fnode->up = cpu_to_le32(dir->i_ino); |
268 | mark_buffer_dirty(bh); | 260 | mark_buffer_dirty(bh); |
269 | 261 | ||
270 | insert_inode_hash(result); | 262 | insert_inode_hash(result); |
271 | 263 | ||
272 | hpfs_write_inode_nolock(result); | 264 | hpfs_write_inode_nolock(result); |
273 | d_instantiate(dentry, result); | 265 | d_instantiate(dentry, result); |
274 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
275 | brelse(bh); | 266 | brelse(bh); |
276 | hpfs_unlock(dir->i_sb); | 267 | hpfs_unlock(dir->i_sb); |
277 | return 0; | 268 | return 0; |
278 | bail2: | 269 | bail2: |
279 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
280 | iput(result); | 270 | iput(result); |
281 | bail1: | 271 | bail1: |
282 | brelse(bh); | 272 | brelse(bh); |
@@ -310,8 +300,8 @@ static int hpfs_symlink(struct inode *dir, struct dentry *dentry, const char *sy | |||
310 | memset(&dee, 0, sizeof dee); | 300 | memset(&dee, 0, sizeof dee); |
311 | dee.archive = 1; | 301 | dee.archive = 1; |
312 | dee.hidden = name[0] == '.'; | 302 | dee.hidden = name[0] == '.'; |
313 | dee.fnode = fno; | 303 | dee.fnode = cpu_to_le32(fno); |
314 | dee.creation_date = dee.write_date = dee.read_date = gmt_to_local(dir->i_sb, get_seconds()); | 304 | dee.creation_date = dee.write_date = dee.read_date = cpu_to_le32(gmt_to_local(dir->i_sb, get_seconds())); |
315 | 305 | ||
316 | result = new_inode(dir->i_sb); | 306 | result = new_inode(dir->i_sb); |
317 | if (!result) | 307 | if (!result) |
@@ -319,7 +309,7 @@ static int hpfs_symlink(struct inode *dir, struct dentry *dentry, const char *sy | |||
319 | result->i_ino = fno; | 309 | result->i_ino = fno; |
320 | hpfs_init_inode(result); | 310 | hpfs_init_inode(result); |
321 | hpfs_i(result)->i_parent_dir = dir->i_ino; | 311 | hpfs_i(result)->i_parent_dir = dir->i_ino; |
322 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, dee.creation_date); | 312 | result->i_ctime.tv_sec = result->i_mtime.tv_sec = result->i_atime.tv_sec = local_to_gmt(dir->i_sb, le32_to_cpu(dee.creation_date)); |
323 | result->i_ctime.tv_nsec = 0; | 313 | result->i_ctime.tv_nsec = 0; |
324 | result->i_mtime.tv_nsec = 0; | 314 | result->i_mtime.tv_nsec = 0; |
325 | result->i_atime.tv_nsec = 0; | 315 | result->i_atime.tv_nsec = 0; |
@@ -333,8 +323,7 @@ static int hpfs_symlink(struct inode *dir, struct dentry *dentry, const char *sy | |||
333 | result->i_op = &page_symlink_inode_operations; | 323 | result->i_op = &page_symlink_inode_operations; |
334 | result->i_data.a_ops = &hpfs_symlink_aops; | 324 | result->i_data.a_ops = &hpfs_symlink_aops; |
335 | 325 | ||
336 | mutex_lock(&hpfs_i(dir)->i_mutex); | 326 | r = hpfs_add_dirent(dir, name, len, &dee); |
337 | r = hpfs_add_dirent(dir, name, len, &dee, 0); | ||
338 | if (r == 1) | 327 | if (r == 1) |
339 | goto bail2; | 328 | goto bail2; |
340 | if (r == -1) { | 329 | if (r == -1) { |
@@ -343,7 +332,7 @@ static int hpfs_symlink(struct inode *dir, struct dentry *dentry, const char *sy | |||
343 | } | 332 | } |
344 | fnode->len = len; | 333 | fnode->len = len; |
345 | memcpy(fnode->name, name, len > 15 ? 15 : len); | 334 | memcpy(fnode->name, name, len > 15 ? 15 : len); |
346 | fnode->up = dir->i_ino; | 335 | fnode->up = cpu_to_le32(dir->i_ino); |
347 | hpfs_set_ea(result, fnode, "SYMLINK", symlink, strlen(symlink)); | 336 | hpfs_set_ea(result, fnode, "SYMLINK", symlink, strlen(symlink)); |
348 | mark_buffer_dirty(bh); | 337 | mark_buffer_dirty(bh); |
349 | brelse(bh); | 338 | brelse(bh); |
@@ -352,11 +341,9 @@ static int hpfs_symlink(struct inode *dir, struct dentry *dentry, const char *sy | |||
352 | 341 | ||
353 | hpfs_write_inode_nolock(result); | 342 | hpfs_write_inode_nolock(result); |
354 | d_instantiate(dentry, result); | 343 | d_instantiate(dentry, result); |
355 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
356 | hpfs_unlock(dir->i_sb); | 344 | hpfs_unlock(dir->i_sb); |
357 | return 0; | 345 | return 0; |
358 | bail2: | 346 | bail2: |
359 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
360 | iput(result); | 347 | iput(result); |
361 | bail1: | 348 | bail1: |
362 | brelse(bh); | 349 | brelse(bh); |
@@ -374,7 +361,6 @@ static int hpfs_unlink(struct inode *dir, struct dentry *dentry) | |||
374 | struct hpfs_dirent *de; | 361 | struct hpfs_dirent *de; |
375 | struct inode *inode = dentry->d_inode; | 362 | struct inode *inode = dentry->d_inode; |
376 | dnode_secno dno; | 363 | dnode_secno dno; |
377 | fnode_secno fno; | ||
378 | int r; | 364 | int r; |
379 | int rep = 0; | 365 | int rep = 0; |
380 | int err; | 366 | int err; |
@@ -382,8 +368,6 @@ static int hpfs_unlink(struct inode *dir, struct dentry *dentry) | |||
382 | hpfs_lock(dir->i_sb); | 368 | hpfs_lock(dir->i_sb); |
383 | hpfs_adjust_length(name, &len); | 369 | hpfs_adjust_length(name, &len); |
384 | again: | 370 | again: |
385 | mutex_lock(&hpfs_i(inode)->i_parent_mutex); | ||
386 | mutex_lock(&hpfs_i(dir)->i_mutex); | ||
387 | err = -ENOENT; | 371 | err = -ENOENT; |
388 | de = map_dirent(dir, hpfs_i(dir)->i_dno, name, len, &dno, &qbh); | 372 | de = map_dirent(dir, hpfs_i(dir)->i_dno, name, len, &dno, &qbh); |
389 | if (!de) | 373 | if (!de) |
@@ -397,7 +381,6 @@ again: | |||
397 | if (de->directory) | 381 | if (de->directory) |
398 | goto out1; | 382 | goto out1; |
399 | 383 | ||
400 | fno = de->fnode; | ||
401 | r = hpfs_remove_dirent(dir, dno, de, &qbh, 1); | 384 | r = hpfs_remove_dirent(dir, dno, de, &qbh, 1); |
402 | switch (r) { | 385 | switch (r) { |
403 | case 1: | 386 | case 1: |
@@ -410,8 +393,6 @@ again: | |||
410 | if (rep++) | 393 | if (rep++) |
411 | break; | 394 | break; |
412 | 395 | ||
413 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
414 | mutex_unlock(&hpfs_i(inode)->i_parent_mutex); | ||
415 | dentry_unhash(dentry); | 396 | dentry_unhash(dentry); |
416 | if (!d_unhashed(dentry)) { | 397 | if (!d_unhashed(dentry)) { |
417 | dput(dentry); | 398 | dput(dentry); |
@@ -445,8 +426,6 @@ again: | |||
445 | out1: | 426 | out1: |
446 | hpfs_brelse4(&qbh); | 427 | hpfs_brelse4(&qbh); |
447 | out: | 428 | out: |
448 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
449 | mutex_unlock(&hpfs_i(inode)->i_parent_mutex); | ||
450 | hpfs_unlock(dir->i_sb); | 429 | hpfs_unlock(dir->i_sb); |
451 | return err; | 430 | return err; |
452 | } | 431 | } |
@@ -459,15 +438,12 @@ static int hpfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
459 | struct hpfs_dirent *de; | 438 | struct hpfs_dirent *de; |
460 | struct inode *inode = dentry->d_inode; | 439 | struct inode *inode = dentry->d_inode; |
461 | dnode_secno dno; | 440 | dnode_secno dno; |
462 | fnode_secno fno; | ||
463 | int n_items = 0; | 441 | int n_items = 0; |
464 | int err; | 442 | int err; |
465 | int r; | 443 | int r; |
466 | 444 | ||
467 | hpfs_adjust_length(name, &len); | 445 | hpfs_adjust_length(name, &len); |
468 | hpfs_lock(dir->i_sb); | 446 | hpfs_lock(dir->i_sb); |
469 | mutex_lock(&hpfs_i(inode)->i_parent_mutex); | ||
470 | mutex_lock(&hpfs_i(dir)->i_mutex); | ||
471 | err = -ENOENT; | 447 | err = -ENOENT; |
472 | de = map_dirent(dir, hpfs_i(dir)->i_dno, name, len, &dno, &qbh); | 448 | de = map_dirent(dir, hpfs_i(dir)->i_dno, name, len, &dno, &qbh); |
473 | if (!de) | 449 | if (!de) |
@@ -486,7 +462,6 @@ static int hpfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
486 | if (n_items) | 462 | if (n_items) |
487 | goto out1; | 463 | goto out1; |
488 | 464 | ||
489 | fno = de->fnode; | ||
490 | r = hpfs_remove_dirent(dir, dno, de, &qbh, 1); | 465 | r = hpfs_remove_dirent(dir, dno, de, &qbh, 1); |
491 | switch (r) { | 466 | switch (r) { |
492 | case 1: | 467 | case 1: |
@@ -505,8 +480,6 @@ static int hpfs_rmdir(struct inode *dir, struct dentry *dentry) | |||
505 | out1: | 480 | out1: |
506 | hpfs_brelse4(&qbh); | 481 | hpfs_brelse4(&qbh); |
507 | out: | 482 | out: |
508 | mutex_unlock(&hpfs_i(dir)->i_mutex); | ||
509 | mutex_unlock(&hpfs_i(inode)->i_parent_mutex); | ||
510 | hpfs_unlock(dir->i_sb); | 483 | hpfs_unlock(dir->i_sb); |
511 | return err; | 484 | return err; |
512 | } | 485 | } |
@@ -568,12 +541,6 @@ static int hpfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
568 | 541 | ||
569 | hpfs_lock(i->i_sb); | 542 | hpfs_lock(i->i_sb); |
570 | /* order doesn't matter, due to VFS exclusion */ | 543 | /* order doesn't matter, due to VFS exclusion */ |
571 | mutex_lock(&hpfs_i(i)->i_parent_mutex); | ||
572 | if (new_inode) | ||
573 | mutex_lock(&hpfs_i(new_inode)->i_parent_mutex); | ||
574 | mutex_lock(&hpfs_i(old_dir)->i_mutex); | ||
575 | if (new_dir != old_dir) | ||
576 | mutex_lock(&hpfs_i(new_dir)->i_mutex); | ||
577 | 544 | ||
578 | /* Erm? Moving over the empty non-busy directory is perfectly legal */ | 545 | /* Erm? Moving over the empty non-busy directory is perfectly legal */ |
579 | if (new_inode && S_ISDIR(new_inode->i_mode)) { | 546 | if (new_inode && S_ISDIR(new_inode->i_mode)) { |
@@ -610,9 +577,7 @@ static int hpfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
610 | 577 | ||
611 | if (new_dir == old_dir) hpfs_brelse4(&qbh); | 578 | if (new_dir == old_dir) hpfs_brelse4(&qbh); |
612 | 579 | ||
613 | hpfs_lock_creation(i->i_sb); | 580 | if ((r = hpfs_add_dirent(new_dir, new_name, new_len, &de))) { |
614 | if ((r = hpfs_add_dirent(new_dir, new_name, new_len, &de, 1))) { | ||
615 | hpfs_unlock_creation(i->i_sb); | ||
616 | if (r == -1) hpfs_error(new_dir->i_sb, "hpfs_rename: dirent already exists!"); | 581 | if (r == -1) hpfs_error(new_dir->i_sb, "hpfs_rename: dirent already exists!"); |
617 | err = r == 1 ? -ENOSPC : -EFSERROR; | 582 | err = r == 1 ? -ENOSPC : -EFSERROR; |
618 | if (new_dir != old_dir) hpfs_brelse4(&qbh); | 583 | if (new_dir != old_dir) hpfs_brelse4(&qbh); |
@@ -621,20 +586,17 @@ static int hpfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
621 | 586 | ||
622 | if (new_dir == old_dir) | 587 | if (new_dir == old_dir) |
623 | if (!(dep = map_dirent(old_dir, hpfs_i(old_dir)->i_dno, old_name, old_len, &dno, &qbh))) { | 588 | if (!(dep = map_dirent(old_dir, hpfs_i(old_dir)->i_dno, old_name, old_len, &dno, &qbh))) { |
624 | hpfs_unlock_creation(i->i_sb); | ||
625 | hpfs_error(i->i_sb, "lookup succeeded but map dirent failed at #2"); | 589 | hpfs_error(i->i_sb, "lookup succeeded but map dirent failed at #2"); |
626 | err = -ENOENT; | 590 | err = -ENOENT; |
627 | goto end1; | 591 | goto end1; |
628 | } | 592 | } |
629 | 593 | ||
630 | if ((r = hpfs_remove_dirent(old_dir, dno, dep, &qbh, 0))) { | 594 | if ((r = hpfs_remove_dirent(old_dir, dno, dep, &qbh, 0))) { |
631 | hpfs_unlock_creation(i->i_sb); | ||
632 | hpfs_error(i->i_sb, "hpfs_rename: could not remove dirent"); | 595 | hpfs_error(i->i_sb, "hpfs_rename: could not remove dirent"); |
633 | err = r == 2 ? -ENOSPC : -EFSERROR; | 596 | err = r == 2 ? -ENOSPC : -EFSERROR; |
634 | goto end1; | 597 | goto end1; |
635 | } | 598 | } |
636 | hpfs_unlock_creation(i->i_sb); | 599 | |
637 | |||
638 | end: | 600 | end: |
639 | hpfs_i(i)->i_parent_dir = new_dir->i_ino; | 601 | hpfs_i(i)->i_parent_dir = new_dir->i_ino; |
640 | if (S_ISDIR(i->i_mode)) { | 602 | if (S_ISDIR(i->i_mode)) { |
@@ -642,22 +604,14 @@ static int hpfs_rename(struct inode *old_dir, struct dentry *old_dentry, | |||
642 | drop_nlink(old_dir); | 604 | drop_nlink(old_dir); |
643 | } | 605 | } |
644 | if ((fnode = hpfs_map_fnode(i->i_sb, i->i_ino, &bh))) { | 606 | if ((fnode = hpfs_map_fnode(i->i_sb, i->i_ino, &bh))) { |
645 | fnode->up = new_dir->i_ino; | 607 | fnode->up = cpu_to_le32(new_dir->i_ino); |
646 | fnode->len = new_len; | 608 | fnode->len = new_len; |
647 | memcpy(fnode->name, new_name, new_len>15?15:new_len); | 609 | memcpy(fnode->name, new_name, new_len>15?15:new_len); |
648 | if (new_len < 15) memset(&fnode->name[new_len], 0, 15 - new_len); | 610 | if (new_len < 15) memset(&fnode->name[new_len], 0, 15 - new_len); |
649 | mark_buffer_dirty(bh); | 611 | mark_buffer_dirty(bh); |
650 | brelse(bh); | 612 | brelse(bh); |
651 | } | 613 | } |
652 | hpfs_i(i)->i_conv = hpfs_sb(i->i_sb)->sb_conv; | ||
653 | hpfs_decide_conv(i, new_name, new_len); | ||
654 | end1: | 614 | end1: |
655 | if (old_dir != new_dir) | ||
656 | mutex_unlock(&hpfs_i(new_dir)->i_mutex); | ||
657 | mutex_unlock(&hpfs_i(old_dir)->i_mutex); | ||
658 | mutex_unlock(&hpfs_i(i)->i_parent_mutex); | ||
659 | if (new_inode) | ||
660 | mutex_unlock(&hpfs_i(new_inode)->i_parent_mutex); | ||
661 | hpfs_unlock(i->i_sb); | 615 | hpfs_unlock(i->i_sb); |
662 | return err; | 616 | return err; |
663 | } | 617 | } |
diff --git a/fs/hpfs/super.c b/fs/hpfs/super.c index c89b40808587..98580a3b5005 100644 --- a/fs/hpfs/super.c +++ b/fs/hpfs/super.c | |||
@@ -18,15 +18,16 @@ | |||
18 | 18 | ||
19 | /* Mark the filesystem dirty, so that chkdsk checks it when os/2 booted */ | 19 | /* Mark the filesystem dirty, so that chkdsk checks it when os/2 booted */ |
20 | 20 | ||
21 | static void mark_dirty(struct super_block *s) | 21 | static void mark_dirty(struct super_block *s, int remount) |
22 | { | 22 | { |
23 | if (hpfs_sb(s)->sb_chkdsk && !(s->s_flags & MS_RDONLY)) { | 23 | if (hpfs_sb(s)->sb_chkdsk && (remount || !(s->s_flags & MS_RDONLY))) { |
24 | struct buffer_head *bh; | 24 | struct buffer_head *bh; |
25 | struct hpfs_spare_block *sb; | 25 | struct hpfs_spare_block *sb; |
26 | if ((sb = hpfs_map_sector(s, 17, &bh, 0))) { | 26 | if ((sb = hpfs_map_sector(s, 17, &bh, 0))) { |
27 | sb->dirty = 1; | 27 | sb->dirty = 1; |
28 | sb->old_wrote = 0; | 28 | sb->old_wrote = 0; |
29 | mark_buffer_dirty(bh); | 29 | mark_buffer_dirty(bh); |
30 | sync_dirty_buffer(bh); | ||
30 | brelse(bh); | 31 | brelse(bh); |
31 | } | 32 | } |
32 | } | 33 | } |
@@ -40,10 +41,12 @@ static void unmark_dirty(struct super_block *s) | |||
40 | struct buffer_head *bh; | 41 | struct buffer_head *bh; |
41 | struct hpfs_spare_block *sb; | 42 | struct hpfs_spare_block *sb; |
42 | if (s->s_flags & MS_RDONLY) return; | 43 | if (s->s_flags & MS_RDONLY) return; |
44 | sync_blockdev(s->s_bdev); | ||
43 | if ((sb = hpfs_map_sector(s, 17, &bh, 0))) { | 45 | if ((sb = hpfs_map_sector(s, 17, &bh, 0))) { |
44 | sb->dirty = hpfs_sb(s)->sb_chkdsk > 1 - hpfs_sb(s)->sb_was_error; | 46 | sb->dirty = hpfs_sb(s)->sb_chkdsk > 1 - hpfs_sb(s)->sb_was_error; |
45 | sb->old_wrote = hpfs_sb(s)->sb_chkdsk >= 2 && !hpfs_sb(s)->sb_was_error; | 47 | sb->old_wrote = hpfs_sb(s)->sb_chkdsk >= 2 && !hpfs_sb(s)->sb_was_error; |
46 | mark_buffer_dirty(bh); | 48 | mark_buffer_dirty(bh); |
49 | sync_dirty_buffer(bh); | ||
47 | brelse(bh); | 50 | brelse(bh); |
48 | } | 51 | } |
49 | } | 52 | } |
@@ -63,13 +66,13 @@ void hpfs_error(struct super_block *s, const char *fmt, ...) | |||
63 | if (!hpfs_sb(s)->sb_was_error) { | 66 | if (!hpfs_sb(s)->sb_was_error) { |
64 | if (hpfs_sb(s)->sb_err == 2) { | 67 | if (hpfs_sb(s)->sb_err == 2) { |
65 | printk("; crashing the system because you wanted it\n"); | 68 | printk("; crashing the system because you wanted it\n"); |
66 | mark_dirty(s); | 69 | mark_dirty(s, 0); |
67 | panic("HPFS panic"); | 70 | panic("HPFS panic"); |
68 | } else if (hpfs_sb(s)->sb_err == 1) { | 71 | } else if (hpfs_sb(s)->sb_err == 1) { |
69 | if (s->s_flags & MS_RDONLY) printk("; already mounted read-only\n"); | 72 | if (s->s_flags & MS_RDONLY) printk("; already mounted read-only\n"); |
70 | else { | 73 | else { |
71 | printk("; remounting read-only\n"); | 74 | printk("; remounting read-only\n"); |
72 | mark_dirty(s); | 75 | mark_dirty(s, 0); |
73 | s->s_flags |= MS_RDONLY; | 76 | s->s_flags |= MS_RDONLY; |
74 | } | 77 | } |
75 | } else if (s->s_flags & MS_RDONLY) printk("; going on - but anything won't be destroyed because it's read-only\n"); | 78 | } else if (s->s_flags & MS_RDONLY) printk("; going on - but anything won't be destroyed because it's read-only\n"); |
@@ -102,9 +105,12 @@ static void hpfs_put_super(struct super_block *s) | |||
102 | { | 105 | { |
103 | struct hpfs_sb_info *sbi = hpfs_sb(s); | 106 | struct hpfs_sb_info *sbi = hpfs_sb(s); |
104 | 107 | ||
108 | hpfs_lock(s); | ||
109 | unmark_dirty(s); | ||
110 | hpfs_unlock(s); | ||
111 | |||
105 | kfree(sbi->sb_cp_table); | 112 | kfree(sbi->sb_cp_table); |
106 | kfree(sbi->sb_bmp_dir); | 113 | kfree(sbi->sb_bmp_dir); |
107 | unmark_dirty(s); | ||
108 | s->s_fs_info = NULL; | 114 | s->s_fs_info = NULL; |
109 | kfree(sbi); | 115 | kfree(sbi); |
110 | } | 116 | } |
@@ -129,7 +135,7 @@ static unsigned count_bitmaps(struct super_block *s) | |||
129 | n_bands = (hpfs_sb(s)->sb_fs_size + 0x3fff) >> 14; | 135 | n_bands = (hpfs_sb(s)->sb_fs_size + 0x3fff) >> 14; |
130 | count = 0; | 136 | count = 0; |
131 | for (n = 0; n < n_bands; n++) | 137 | for (n = 0; n < n_bands; n++) |
132 | count += hpfs_count_one_bitmap(s, hpfs_sb(s)->sb_bmp_dir[n]); | 138 | count += hpfs_count_one_bitmap(s, le32_to_cpu(hpfs_sb(s)->sb_bmp_dir[n])); |
133 | return count; | 139 | return count; |
134 | } | 140 | } |
135 | 141 | ||
@@ -188,8 +194,6 @@ static void init_once(void *foo) | |||
188 | { | 194 | { |
189 | struct hpfs_inode_info *ei = (struct hpfs_inode_info *) foo; | 195 | struct hpfs_inode_info *ei = (struct hpfs_inode_info *) foo; |
190 | 196 | ||
191 | mutex_init(&ei->i_mutex); | ||
192 | mutex_init(&ei->i_parent_mutex); | ||
193 | inode_init_once(&ei->vfs_inode); | 197 | inode_init_once(&ei->vfs_inode); |
194 | } | 198 | } |
195 | 199 | ||
@@ -218,7 +222,6 @@ static void destroy_inodecache(void) | |||
218 | 222 | ||
219 | enum { | 223 | enum { |
220 | Opt_help, Opt_uid, Opt_gid, Opt_umask, Opt_case_lower, Opt_case_asis, | 224 | Opt_help, Opt_uid, Opt_gid, Opt_umask, Opt_case_lower, Opt_case_asis, |
221 | Opt_conv_binary, Opt_conv_text, Opt_conv_auto, | ||
222 | Opt_check_none, Opt_check_normal, Opt_check_strict, | 225 | Opt_check_none, Opt_check_normal, Opt_check_strict, |
223 | Opt_err_cont, Opt_err_ro, Opt_err_panic, | 226 | Opt_err_cont, Opt_err_ro, Opt_err_panic, |
224 | Opt_eas_no, Opt_eas_ro, Opt_eas_rw, | 227 | Opt_eas_no, Opt_eas_ro, Opt_eas_rw, |
@@ -233,9 +236,6 @@ static const match_table_t tokens = { | |||
233 | {Opt_umask, "umask=%o"}, | 236 | {Opt_umask, "umask=%o"}, |
234 | {Opt_case_lower, "case=lower"}, | 237 | {Opt_case_lower, "case=lower"}, |
235 | {Opt_case_asis, "case=asis"}, | 238 | {Opt_case_asis, "case=asis"}, |
236 | {Opt_conv_binary, "conv=binary"}, | ||
237 | {Opt_conv_text, "conv=text"}, | ||
238 | {Opt_conv_auto, "conv=auto"}, | ||
239 | {Opt_check_none, "check=none"}, | 239 | {Opt_check_none, "check=none"}, |
240 | {Opt_check_normal, "check=normal"}, | 240 | {Opt_check_normal, "check=normal"}, |
241 | {Opt_check_strict, "check=strict"}, | 241 | {Opt_check_strict, "check=strict"}, |
@@ -253,7 +253,7 @@ static const match_table_t tokens = { | |||
253 | }; | 253 | }; |
254 | 254 | ||
255 | static int parse_opts(char *opts, uid_t *uid, gid_t *gid, umode_t *umask, | 255 | static int parse_opts(char *opts, uid_t *uid, gid_t *gid, umode_t *umask, |
256 | int *lowercase, int *conv, int *eas, int *chk, int *errs, | 256 | int *lowercase, int *eas, int *chk, int *errs, |
257 | int *chkdsk, int *timeshift) | 257 | int *chkdsk, int *timeshift) |
258 | { | 258 | { |
259 | char *p; | 259 | char *p; |
@@ -295,15 +295,6 @@ static int parse_opts(char *opts, uid_t *uid, gid_t *gid, umode_t *umask, | |||
295 | case Opt_case_asis: | 295 | case Opt_case_asis: |
296 | *lowercase = 0; | 296 | *lowercase = 0; |
297 | break; | 297 | break; |
298 | case Opt_conv_binary: | ||
299 | *conv = CONV_BINARY; | ||
300 | break; | ||
301 | case Opt_conv_text: | ||
302 | *conv = CONV_TEXT; | ||
303 | break; | ||
304 | case Opt_conv_auto: | ||
305 | *conv = CONV_AUTO; | ||
306 | break; | ||
307 | case Opt_check_none: | 298 | case Opt_check_none: |
308 | *chk = 0; | 299 | *chk = 0; |
309 | break; | 300 | break; |
@@ -370,9 +361,6 @@ HPFS filesystem options:\n\ | |||
370 | umask=xxx set mode of files that don't have mode specified in eas\n\ | 361 | umask=xxx set mode of files that don't have mode specified in eas\n\ |
371 | case=lower lowercase all files\n\ | 362 | case=lower lowercase all files\n\ |
372 | case=asis do not lowercase files (default)\n\ | 363 | case=asis do not lowercase files (default)\n\ |
373 | conv=binary do not convert CR/LF -> LF (default)\n\ | ||
374 | conv=auto convert only files with known text extensions\n\ | ||
375 | conv=text convert all files\n\ | ||
376 | check=none no fs checks - kernel may crash on corrupted filesystem\n\ | 364 | check=none no fs checks - kernel may crash on corrupted filesystem\n\ |
377 | check=normal do some checks - it should not crash (default)\n\ | 365 | check=normal do some checks - it should not crash (default)\n\ |
378 | check=strict do extra time-consuming checks, used for debugging\n\ | 366 | check=strict do extra time-consuming checks, used for debugging\n\ |
@@ -394,7 +382,7 @@ static int hpfs_remount_fs(struct super_block *s, int *flags, char *data) | |||
394 | uid_t uid; | 382 | uid_t uid; |
395 | gid_t gid; | 383 | gid_t gid; |
396 | umode_t umask; | 384 | umode_t umask; |
397 | int lowercase, conv, eas, chk, errs, chkdsk, timeshift; | 385 | int lowercase, eas, chk, errs, chkdsk, timeshift; |
398 | int o; | 386 | int o; |
399 | struct hpfs_sb_info *sbi = hpfs_sb(s); | 387 | struct hpfs_sb_info *sbi = hpfs_sb(s); |
400 | char *new_opts = kstrdup(data, GFP_KERNEL); | 388 | char *new_opts = kstrdup(data, GFP_KERNEL); |
@@ -405,11 +393,11 @@ static int hpfs_remount_fs(struct super_block *s, int *flags, char *data) | |||
405 | lock_super(s); | 393 | lock_super(s); |
406 | uid = sbi->sb_uid; gid = sbi->sb_gid; | 394 | uid = sbi->sb_uid; gid = sbi->sb_gid; |
407 | umask = 0777 & ~sbi->sb_mode; | 395 | umask = 0777 & ~sbi->sb_mode; |
408 | lowercase = sbi->sb_lowercase; conv = sbi->sb_conv; | 396 | lowercase = sbi->sb_lowercase; |
409 | eas = sbi->sb_eas; chk = sbi->sb_chk; chkdsk = sbi->sb_chkdsk; | 397 | eas = sbi->sb_eas; chk = sbi->sb_chk; chkdsk = sbi->sb_chkdsk; |
410 | errs = sbi->sb_err; timeshift = sbi->sb_timeshift; | 398 | errs = sbi->sb_err; timeshift = sbi->sb_timeshift; |
411 | 399 | ||
412 | if (!(o = parse_opts(data, &uid, &gid, &umask, &lowercase, &conv, | 400 | if (!(o = parse_opts(data, &uid, &gid, &umask, &lowercase, |
413 | &eas, &chk, &errs, &chkdsk, ×hift))) { | 401 | &eas, &chk, &errs, &chkdsk, ×hift))) { |
414 | printk("HPFS: bad mount options.\n"); | 402 | printk("HPFS: bad mount options.\n"); |
415 | goto out_err; | 403 | goto out_err; |
@@ -427,11 +415,11 @@ static int hpfs_remount_fs(struct super_block *s, int *flags, char *data) | |||
427 | 415 | ||
428 | sbi->sb_uid = uid; sbi->sb_gid = gid; | 416 | sbi->sb_uid = uid; sbi->sb_gid = gid; |
429 | sbi->sb_mode = 0777 & ~umask; | 417 | sbi->sb_mode = 0777 & ~umask; |
430 | sbi->sb_lowercase = lowercase; sbi->sb_conv = conv; | 418 | sbi->sb_lowercase = lowercase; |
431 | sbi->sb_eas = eas; sbi->sb_chk = chk; sbi->sb_chkdsk = chkdsk; | 419 | sbi->sb_eas = eas; sbi->sb_chk = chk; sbi->sb_chkdsk = chkdsk; |
432 | sbi->sb_err = errs; sbi->sb_timeshift = timeshift; | 420 | sbi->sb_err = errs; sbi->sb_timeshift = timeshift; |
433 | 421 | ||
434 | if (!(*flags & MS_RDONLY)) mark_dirty(s); | 422 | if (!(*flags & MS_RDONLY)) mark_dirty(s, 1); |
435 | 423 | ||
436 | replace_mount_options(s, new_opts); | 424 | replace_mount_options(s, new_opts); |
437 | 425 | ||
@@ -471,7 +459,7 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
471 | uid_t uid; | 459 | uid_t uid; |
472 | gid_t gid; | 460 | gid_t gid; |
473 | umode_t umask; | 461 | umode_t umask; |
474 | int lowercase, conv, eas, chk, errs, chkdsk, timeshift; | 462 | int lowercase, eas, chk, errs, chkdsk, timeshift; |
475 | 463 | ||
476 | dnode_secno root_dno; | 464 | dnode_secno root_dno; |
477 | struct hpfs_dirent *de = NULL; | 465 | struct hpfs_dirent *de = NULL; |
@@ -479,11 +467,6 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
479 | 467 | ||
480 | int o; | 468 | int o; |
481 | 469 | ||
482 | if (num_possible_cpus() > 1) { | ||
483 | printk(KERN_ERR "HPFS is not SMP safe\n"); | ||
484 | return -EINVAL; | ||
485 | } | ||
486 | |||
487 | save_mount_options(s, options); | 470 | save_mount_options(s, options); |
488 | 471 | ||
489 | sbi = kzalloc(sizeof(*sbi), GFP_KERNEL); | 472 | sbi = kzalloc(sizeof(*sbi), GFP_KERNEL); |
@@ -495,20 +478,20 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
495 | sbi->sb_bmp_dir = NULL; | 478 | sbi->sb_bmp_dir = NULL; |
496 | sbi->sb_cp_table = NULL; | 479 | sbi->sb_cp_table = NULL; |
497 | 480 | ||
498 | mutex_init(&sbi->hpfs_creation_de); | 481 | mutex_init(&sbi->hpfs_mutex); |
482 | hpfs_lock(s); | ||
499 | 483 | ||
500 | uid = current_uid(); | 484 | uid = current_uid(); |
501 | gid = current_gid(); | 485 | gid = current_gid(); |
502 | umask = current_umask(); | 486 | umask = current_umask(); |
503 | lowercase = 0; | 487 | lowercase = 0; |
504 | conv = CONV_BINARY; | ||
505 | eas = 2; | 488 | eas = 2; |
506 | chk = 1; | 489 | chk = 1; |
507 | errs = 1; | 490 | errs = 1; |
508 | chkdsk = 1; | 491 | chkdsk = 1; |
509 | timeshift = 0; | 492 | timeshift = 0; |
510 | 493 | ||
511 | if (!(o = parse_opts(options, &uid, &gid, &umask, &lowercase, &conv, | 494 | if (!(o = parse_opts(options, &uid, &gid, &umask, &lowercase, |
512 | &eas, &chk, &errs, &chkdsk, ×hift))) { | 495 | &eas, &chk, &errs, &chkdsk, ×hift))) { |
513 | printk("HPFS: bad mount options.\n"); | 496 | printk("HPFS: bad mount options.\n"); |
514 | goto bail0; | 497 | goto bail0; |
@@ -526,9 +509,9 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
526 | if (!(spareblock = hpfs_map_sector(s, 17, &bh2, 0))) goto bail3; | 509 | if (!(spareblock = hpfs_map_sector(s, 17, &bh2, 0))) goto bail3; |
527 | 510 | ||
528 | /* Check magics */ | 511 | /* Check magics */ |
529 | if (/*bootblock->magic != BB_MAGIC | 512 | if (/*le16_to_cpu(bootblock->magic) != BB_MAGIC |
530 | ||*/ superblock->magic != SB_MAGIC | 513 | ||*/ le32_to_cpu(superblock->magic) != SB_MAGIC |
531 | || spareblock->magic != SP_MAGIC) { | 514 | || le32_to_cpu(spareblock->magic) != SP_MAGIC) { |
532 | if (!silent) printk("HPFS: Bad magic ... probably not HPFS\n"); | 515 | if (!silent) printk("HPFS: Bad magic ... probably not HPFS\n"); |
533 | goto bail4; | 516 | goto bail4; |
534 | } | 517 | } |
@@ -549,19 +532,18 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
549 | s->s_op = &hpfs_sops; | 532 | s->s_op = &hpfs_sops; |
550 | s->s_d_op = &hpfs_dentry_operations; | 533 | s->s_d_op = &hpfs_dentry_operations; |
551 | 534 | ||
552 | sbi->sb_root = superblock->root; | 535 | sbi->sb_root = le32_to_cpu(superblock->root); |
553 | sbi->sb_fs_size = superblock->n_sectors; | 536 | sbi->sb_fs_size = le32_to_cpu(superblock->n_sectors); |
554 | sbi->sb_bitmaps = superblock->bitmaps; | 537 | sbi->sb_bitmaps = le32_to_cpu(superblock->bitmaps); |
555 | sbi->sb_dirband_start = superblock->dir_band_start; | 538 | sbi->sb_dirband_start = le32_to_cpu(superblock->dir_band_start); |
556 | sbi->sb_dirband_size = superblock->n_dir_band; | 539 | sbi->sb_dirband_size = le32_to_cpu(superblock->n_dir_band); |
557 | sbi->sb_dmap = superblock->dir_band_bitmap; | 540 | sbi->sb_dmap = le32_to_cpu(superblock->dir_band_bitmap); |
558 | sbi->sb_uid = uid; | 541 | sbi->sb_uid = uid; |
559 | sbi->sb_gid = gid; | 542 | sbi->sb_gid = gid; |
560 | sbi->sb_mode = 0777 & ~umask; | 543 | sbi->sb_mode = 0777 & ~umask; |
561 | sbi->sb_n_free = -1; | 544 | sbi->sb_n_free = -1; |
562 | sbi->sb_n_free_dnodes = -1; | 545 | sbi->sb_n_free_dnodes = -1; |
563 | sbi->sb_lowercase = lowercase; | 546 | sbi->sb_lowercase = lowercase; |
564 | sbi->sb_conv = conv; | ||
565 | sbi->sb_eas = eas; | 547 | sbi->sb_eas = eas; |
566 | sbi->sb_chk = chk; | 548 | sbi->sb_chk = chk; |
567 | sbi->sb_chkdsk = chkdsk; | 549 | sbi->sb_chkdsk = chkdsk; |
@@ -573,7 +555,7 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
573 | sbi->sb_max_fwd_alloc = 0xffffff; | 555 | sbi->sb_max_fwd_alloc = 0xffffff; |
574 | 556 | ||
575 | /* Load bitmap directory */ | 557 | /* Load bitmap directory */ |
576 | if (!(sbi->sb_bmp_dir = hpfs_load_bitmap_directory(s, superblock->bitmaps))) | 558 | if (!(sbi->sb_bmp_dir = hpfs_load_bitmap_directory(s, le32_to_cpu(superblock->bitmaps)))) |
577 | goto bail4; | 559 | goto bail4; |
578 | 560 | ||
579 | /* Check for general fs errors*/ | 561 | /* Check for general fs errors*/ |
@@ -591,20 +573,20 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
591 | mark_buffer_dirty(bh2); | 573 | mark_buffer_dirty(bh2); |
592 | } | 574 | } |
593 | 575 | ||
594 | if (spareblock->hotfixes_used || spareblock->n_spares_used) { | 576 | if (le32_to_cpu(spareblock->hotfixes_used) || le32_to_cpu(spareblock->n_spares_used)) { |
595 | if (errs >= 2) { | 577 | if (errs >= 2) { |
596 | printk("HPFS: Hotfixes not supported here, try chkdsk\n"); | 578 | printk("HPFS: Hotfixes not supported here, try chkdsk\n"); |
597 | mark_dirty(s); | 579 | mark_dirty(s, 0); |
598 | goto bail4; | 580 | goto bail4; |
599 | } | 581 | } |
600 | hpfs_error(s, "hotfixes not supported here, try chkdsk"); | 582 | hpfs_error(s, "hotfixes not supported here, try chkdsk"); |
601 | if (errs == 0) printk("HPFS: Proceeding, but your filesystem will be probably corrupted by this driver...\n"); | 583 | if (errs == 0) printk("HPFS: Proceeding, but your filesystem will be probably corrupted by this driver...\n"); |
602 | else printk("HPFS: This driver may read bad files or crash when operating on disk with hotfixes.\n"); | 584 | else printk("HPFS: This driver may read bad files or crash when operating on disk with hotfixes.\n"); |
603 | } | 585 | } |
604 | if (spareblock->n_dnode_spares != spareblock->n_dnode_spares_free) { | 586 | if (le32_to_cpu(spareblock->n_dnode_spares) != le32_to_cpu(spareblock->n_dnode_spares_free)) { |
605 | if (errs >= 2) { | 587 | if (errs >= 2) { |
606 | printk("HPFS: Spare dnodes used, try chkdsk\n"); | 588 | printk("HPFS: Spare dnodes used, try chkdsk\n"); |
607 | mark_dirty(s); | 589 | mark_dirty(s, 0); |
608 | goto bail4; | 590 | goto bail4; |
609 | } | 591 | } |
610 | hpfs_error(s, "warning: spare dnodes used, try chkdsk"); | 592 | hpfs_error(s, "warning: spare dnodes used, try chkdsk"); |
@@ -612,26 +594,26 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
612 | } | 594 | } |
613 | if (chk) { | 595 | if (chk) { |
614 | unsigned a; | 596 | unsigned a; |
615 | if (superblock->dir_band_end - superblock->dir_band_start + 1 != superblock->n_dir_band || | 597 | if (le32_to_cpu(superblock->dir_band_end) - le32_to_cpu(superblock->dir_band_start) + 1 != le32_to_cpu(superblock->n_dir_band) || |
616 | superblock->dir_band_end < superblock->dir_band_start || superblock->n_dir_band > 0x4000) { | 598 | le32_to_cpu(superblock->dir_band_end) < le32_to_cpu(superblock->dir_band_start) || le32_to_cpu(superblock->n_dir_band) > 0x4000) { |
617 | hpfs_error(s, "dir band size mismatch: dir_band_start==%08x, dir_band_end==%08x, n_dir_band==%08x", | 599 | hpfs_error(s, "dir band size mismatch: dir_band_start==%08x, dir_band_end==%08x, n_dir_band==%08x", |
618 | superblock->dir_band_start, superblock->dir_band_end, superblock->n_dir_band); | 600 | le32_to_cpu(superblock->dir_band_start), le32_to_cpu(superblock->dir_band_end), le32_to_cpu(superblock->n_dir_band)); |
619 | goto bail4; | 601 | goto bail4; |
620 | } | 602 | } |
621 | a = sbi->sb_dirband_size; | 603 | a = sbi->sb_dirband_size; |
622 | sbi->sb_dirband_size = 0; | 604 | sbi->sb_dirband_size = 0; |
623 | if (hpfs_chk_sectors(s, superblock->dir_band_start, superblock->n_dir_band, "dir_band") || | 605 | if (hpfs_chk_sectors(s, le32_to_cpu(superblock->dir_band_start), le32_to_cpu(superblock->n_dir_band), "dir_band") || |
624 | hpfs_chk_sectors(s, superblock->dir_band_bitmap, 4, "dir_band_bitmap") || | 606 | hpfs_chk_sectors(s, le32_to_cpu(superblock->dir_band_bitmap), 4, "dir_band_bitmap") || |
625 | hpfs_chk_sectors(s, superblock->bitmaps, 4, "bitmaps")) { | 607 | hpfs_chk_sectors(s, le32_to_cpu(superblock->bitmaps), 4, "bitmaps")) { |
626 | mark_dirty(s); | 608 | mark_dirty(s, 0); |
627 | goto bail4; | 609 | goto bail4; |
628 | } | 610 | } |
629 | sbi->sb_dirband_size = a; | 611 | sbi->sb_dirband_size = a; |
630 | } else printk("HPFS: You really don't want any checks? You are crazy...\n"); | 612 | } else printk("HPFS: You really don't want any checks? You are crazy...\n"); |
631 | 613 | ||
632 | /* Load code page table */ | 614 | /* Load code page table */ |
633 | if (spareblock->n_code_pages) | 615 | if (le32_to_cpu(spareblock->n_code_pages)) |
634 | if (!(sbi->sb_cp_table = hpfs_load_code_page(s, spareblock->code_page_dir))) | 616 | if (!(sbi->sb_cp_table = hpfs_load_code_page(s, le32_to_cpu(spareblock->code_page_dir)))) |
635 | printk("HPFS: Warning: code page support is disabled\n"); | 617 | printk("HPFS: Warning: code page support is disabled\n"); |
636 | 618 | ||
637 | brelse(bh2); | 619 | brelse(bh2); |
@@ -660,13 +642,13 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
660 | if (!de) | 642 | if (!de) |
661 | hpfs_error(s, "unable to find root dir"); | 643 | hpfs_error(s, "unable to find root dir"); |
662 | else { | 644 | else { |
663 | root->i_atime.tv_sec = local_to_gmt(s, de->read_date); | 645 | root->i_atime.tv_sec = local_to_gmt(s, le32_to_cpu(de->read_date)); |
664 | root->i_atime.tv_nsec = 0; | 646 | root->i_atime.tv_nsec = 0; |
665 | root->i_mtime.tv_sec = local_to_gmt(s, de->write_date); | 647 | root->i_mtime.tv_sec = local_to_gmt(s, le32_to_cpu(de->write_date)); |
666 | root->i_mtime.tv_nsec = 0; | 648 | root->i_mtime.tv_nsec = 0; |
667 | root->i_ctime.tv_sec = local_to_gmt(s, de->creation_date); | 649 | root->i_ctime.tv_sec = local_to_gmt(s, le32_to_cpu(de->creation_date)); |
668 | root->i_ctime.tv_nsec = 0; | 650 | root->i_ctime.tv_nsec = 0; |
669 | hpfs_i(root)->i_ea_size = de->ea_size; | 651 | hpfs_i(root)->i_ea_size = le16_to_cpu(de->ea_size); |
670 | hpfs_i(root)->i_parent_dir = root->i_ino; | 652 | hpfs_i(root)->i_parent_dir = root->i_ino; |
671 | if (root->i_size == -1) | 653 | if (root->i_size == -1) |
672 | root->i_size = 2048; | 654 | root->i_size = 2048; |
@@ -674,6 +656,7 @@ static int hpfs_fill_super(struct super_block *s, void *options, int silent) | |||
674 | root->i_blocks = 5; | 656 | root->i_blocks = 5; |
675 | hpfs_brelse4(&qbh); | 657 | hpfs_brelse4(&qbh); |
676 | } | 658 | } |
659 | hpfs_unlock(s); | ||
677 | return 0; | 660 | return 0; |
678 | 661 | ||
679 | bail4: brelse(bh2); | 662 | bail4: brelse(bh2); |
@@ -681,6 +664,7 @@ bail3: brelse(bh1); | |||
681 | bail2: brelse(bh0); | 664 | bail2: brelse(bh0); |
682 | bail1: | 665 | bail1: |
683 | bail0: | 666 | bail0: |
667 | hpfs_unlock(s); | ||
684 | kfree(sbi->sb_bmp_dir); | 668 | kfree(sbi->sb_bmp_dir); |
685 | kfree(sbi->sb_cp_table); | 669 | kfree(sbi->sb_cp_table); |
686 | s->s_fs_info = NULL; | 670 | s->s_fs_info = NULL; |
diff --git a/fs/logfs/super.c b/fs/logfs/super.c index 33435e4b14d2..ce03a182c771 100644 --- a/fs/logfs/super.c +++ b/fs/logfs/super.c | |||
@@ -480,10 +480,6 @@ static int logfs_read_sb(struct super_block *sb, int read_only) | |||
480 | !read_only) | 480 | !read_only) |
481 | return -EIO; | 481 | return -EIO; |
482 | 482 | ||
483 | mutex_init(&super->s_dirop_mutex); | ||
484 | mutex_init(&super->s_object_alias_mutex); | ||
485 | INIT_LIST_HEAD(&super->s_freeing_list); | ||
486 | |||
487 | ret = logfs_init_rw(sb); | 483 | ret = logfs_init_rw(sb); |
488 | if (ret) | 484 | if (ret) |
489 | return ret; | 485 | return ret; |
@@ -601,6 +597,10 @@ static struct dentry *logfs_mount(struct file_system_type *type, int flags, | |||
601 | if (!super) | 597 | if (!super) |
602 | return ERR_PTR(-ENOMEM); | 598 | return ERR_PTR(-ENOMEM); |
603 | 599 | ||
600 | mutex_init(&super->s_dirop_mutex); | ||
601 | mutex_init(&super->s_object_alias_mutex); | ||
602 | INIT_LIST_HEAD(&super->s_freeing_list); | ||
603 | |||
604 | if (!devname) | 604 | if (!devname) |
605 | err = logfs_get_sb_bdev(super, type, devname); | 605 | err = logfs_get_sb_bdev(super, type, devname); |
606 | else if (strncmp(devname, "mtd", 3)) | 606 | else if (strncmp(devname, "mtd", 3)) |
diff --git a/fs/namei.c b/fs/namei.c index 54fc993e3027..e3c4f112ebf7 100644 --- a/fs/namei.c +++ b/fs/namei.c | |||
@@ -179,7 +179,7 @@ EXPORT_SYMBOL(putname); | |||
179 | static int acl_permission_check(struct inode *inode, int mask, unsigned int flags, | 179 | static int acl_permission_check(struct inode *inode, int mask, unsigned int flags, |
180 | int (*check_acl)(struct inode *inode, int mask, unsigned int flags)) | 180 | int (*check_acl)(struct inode *inode, int mask, unsigned int flags)) |
181 | { | 181 | { |
182 | umode_t mode = inode->i_mode; | 182 | unsigned int mode = inode->i_mode; |
183 | 183 | ||
184 | mask &= MAY_READ | MAY_WRITE | MAY_EXEC; | 184 | mask &= MAY_READ | MAY_WRITE | MAY_EXEC; |
185 | 185 | ||
diff --git a/fs/nfs/nfs4filelayout.c b/fs/nfs/nfs4filelayout.c index 6f8192f4cfc7..be79dc9f386d 100644 --- a/fs/nfs/nfs4filelayout.c +++ b/fs/nfs/nfs4filelayout.c | |||
@@ -117,6 +117,8 @@ static int filelayout_async_handle_error(struct rpc_task *task, | |||
117 | case -EKEYEXPIRED: | 117 | case -EKEYEXPIRED: |
118 | rpc_delay(task, FILELAYOUT_POLL_RETRY_MAX); | 118 | rpc_delay(task, FILELAYOUT_POLL_RETRY_MAX); |
119 | break; | 119 | break; |
120 | case -NFS4ERR_RETRY_UNCACHED_REP: | ||
121 | break; | ||
120 | default: | 122 | default: |
121 | dprintk("%s DS error. Retry through MDS %d\n", __func__, | 123 | dprintk("%s DS error. Retry through MDS %d\n", __func__, |
122 | task->tk_status); | 124 | task->tk_status); |
@@ -416,7 +418,8 @@ static int | |||
416 | filelayout_check_layout(struct pnfs_layout_hdr *lo, | 418 | filelayout_check_layout(struct pnfs_layout_hdr *lo, |
417 | struct nfs4_filelayout_segment *fl, | 419 | struct nfs4_filelayout_segment *fl, |
418 | struct nfs4_layoutget_res *lgr, | 420 | struct nfs4_layoutget_res *lgr, |
419 | struct nfs4_deviceid *id) | 421 | struct nfs4_deviceid *id, |
422 | gfp_t gfp_flags) | ||
420 | { | 423 | { |
421 | struct nfs4_file_layout_dsaddr *dsaddr; | 424 | struct nfs4_file_layout_dsaddr *dsaddr; |
422 | int status = -EINVAL; | 425 | int status = -EINVAL; |
@@ -439,7 +442,7 @@ filelayout_check_layout(struct pnfs_layout_hdr *lo, | |||
439 | /* find and reference the deviceid */ | 442 | /* find and reference the deviceid */ |
440 | dsaddr = nfs4_fl_find_get_deviceid(id); | 443 | dsaddr = nfs4_fl_find_get_deviceid(id); |
441 | if (dsaddr == NULL) { | 444 | if (dsaddr == NULL) { |
442 | dsaddr = get_device_info(lo->plh_inode, id); | 445 | dsaddr = get_device_info(lo->plh_inode, id, gfp_flags); |
443 | if (dsaddr == NULL) | 446 | if (dsaddr == NULL) |
444 | goto out; | 447 | goto out; |
445 | } | 448 | } |
@@ -500,7 +503,8 @@ static int | |||
500 | filelayout_decode_layout(struct pnfs_layout_hdr *flo, | 503 | filelayout_decode_layout(struct pnfs_layout_hdr *flo, |
501 | struct nfs4_filelayout_segment *fl, | 504 | struct nfs4_filelayout_segment *fl, |
502 | struct nfs4_layoutget_res *lgr, | 505 | struct nfs4_layoutget_res *lgr, |
503 | struct nfs4_deviceid *id) | 506 | struct nfs4_deviceid *id, |
507 | gfp_t gfp_flags) | ||
504 | { | 508 | { |
505 | struct xdr_stream stream; | 509 | struct xdr_stream stream; |
506 | struct xdr_buf buf = { | 510 | struct xdr_buf buf = { |
@@ -516,7 +520,7 @@ filelayout_decode_layout(struct pnfs_layout_hdr *flo, | |||
516 | 520 | ||
517 | dprintk("%s: set_layout_map Begin\n", __func__); | 521 | dprintk("%s: set_layout_map Begin\n", __func__); |
518 | 522 | ||
519 | scratch = alloc_page(GFP_KERNEL); | 523 | scratch = alloc_page(gfp_flags); |
520 | if (!scratch) | 524 | if (!scratch) |
521 | return -ENOMEM; | 525 | return -ENOMEM; |
522 | 526 | ||
@@ -554,13 +558,13 @@ filelayout_decode_layout(struct pnfs_layout_hdr *flo, | |||
554 | goto out_err; | 558 | goto out_err; |
555 | 559 | ||
556 | fl->fh_array = kzalloc(fl->num_fh * sizeof(struct nfs_fh *), | 560 | fl->fh_array = kzalloc(fl->num_fh * sizeof(struct nfs_fh *), |
557 | GFP_KERNEL); | 561 | gfp_flags); |
558 | if (!fl->fh_array) | 562 | if (!fl->fh_array) |
559 | goto out_err; | 563 | goto out_err; |
560 | 564 | ||
561 | for (i = 0; i < fl->num_fh; i++) { | 565 | for (i = 0; i < fl->num_fh; i++) { |
562 | /* Do we want to use a mempool here? */ | 566 | /* Do we want to use a mempool here? */ |
563 | fl->fh_array[i] = kmalloc(sizeof(struct nfs_fh), GFP_KERNEL); | 567 | fl->fh_array[i] = kmalloc(sizeof(struct nfs_fh), gfp_flags); |
564 | if (!fl->fh_array[i]) | 568 | if (!fl->fh_array[i]) |
565 | goto out_err_free; | 569 | goto out_err_free; |
566 | 570 | ||
@@ -605,19 +609,20 @@ filelayout_free_lseg(struct pnfs_layout_segment *lseg) | |||
605 | 609 | ||
606 | static struct pnfs_layout_segment * | 610 | static struct pnfs_layout_segment * |
607 | filelayout_alloc_lseg(struct pnfs_layout_hdr *layoutid, | 611 | filelayout_alloc_lseg(struct pnfs_layout_hdr *layoutid, |
608 | struct nfs4_layoutget_res *lgr) | 612 | struct nfs4_layoutget_res *lgr, |
613 | gfp_t gfp_flags) | ||
609 | { | 614 | { |
610 | struct nfs4_filelayout_segment *fl; | 615 | struct nfs4_filelayout_segment *fl; |
611 | int rc; | 616 | int rc; |
612 | struct nfs4_deviceid id; | 617 | struct nfs4_deviceid id; |
613 | 618 | ||
614 | dprintk("--> %s\n", __func__); | 619 | dprintk("--> %s\n", __func__); |
615 | fl = kzalloc(sizeof(*fl), GFP_KERNEL); | 620 | fl = kzalloc(sizeof(*fl), gfp_flags); |
616 | if (!fl) | 621 | if (!fl) |
617 | return NULL; | 622 | return NULL; |
618 | 623 | ||
619 | rc = filelayout_decode_layout(layoutid, fl, lgr, &id); | 624 | rc = filelayout_decode_layout(layoutid, fl, lgr, &id, gfp_flags); |
620 | if (rc != 0 || filelayout_check_layout(layoutid, fl, lgr, &id)) { | 625 | if (rc != 0 || filelayout_check_layout(layoutid, fl, lgr, &id, gfp_flags)) { |
621 | _filelayout_free_lseg(fl); | 626 | _filelayout_free_lseg(fl); |
622 | return NULL; | 627 | return NULL; |
623 | } | 628 | } |
@@ -633,7 +638,7 @@ filelayout_alloc_lseg(struct pnfs_layout_hdr *layoutid, | |||
633 | int size = (fl->stripe_type == STRIPE_SPARSE) ? | 638 | int size = (fl->stripe_type == STRIPE_SPARSE) ? |
634 | fl->dsaddr->ds_num : fl->dsaddr->stripe_count; | 639 | fl->dsaddr->ds_num : fl->dsaddr->stripe_count; |
635 | 640 | ||
636 | fl->commit_buckets = kcalloc(size, sizeof(struct list_head), GFP_KERNEL); | 641 | fl->commit_buckets = kcalloc(size, sizeof(struct list_head), gfp_flags); |
637 | if (!fl->commit_buckets) { | 642 | if (!fl->commit_buckets) { |
638 | filelayout_free_lseg(&fl->generic_hdr); | 643 | filelayout_free_lseg(&fl->generic_hdr); |
639 | return NULL; | 644 | return NULL; |
diff --git a/fs/nfs/nfs4filelayout.h b/fs/nfs/nfs4filelayout.h index 7c44579f5832..2b461d77b43a 100644 --- a/fs/nfs/nfs4filelayout.h +++ b/fs/nfs/nfs4filelayout.h | |||
@@ -104,6 +104,6 @@ extern struct nfs4_file_layout_dsaddr * | |||
104 | nfs4_fl_find_get_deviceid(struct nfs4_deviceid *dev_id); | 104 | nfs4_fl_find_get_deviceid(struct nfs4_deviceid *dev_id); |
105 | extern void nfs4_fl_put_deviceid(struct nfs4_file_layout_dsaddr *dsaddr); | 105 | extern void nfs4_fl_put_deviceid(struct nfs4_file_layout_dsaddr *dsaddr); |
106 | struct nfs4_file_layout_dsaddr * | 106 | struct nfs4_file_layout_dsaddr * |
107 | get_device_info(struct inode *inode, struct nfs4_deviceid *dev_id); | 107 | get_device_info(struct inode *inode, struct nfs4_deviceid *dev_id, gfp_t gfp_flags); |
108 | 108 | ||
109 | #endif /* FS_NFS_NFS4FILELAYOUT_H */ | 109 | #endif /* FS_NFS_NFS4FILELAYOUT_H */ |
diff --git a/fs/nfs/nfs4filelayoutdev.c b/fs/nfs/nfs4filelayoutdev.c index de5350f2b249..db07c7af1395 100644 --- a/fs/nfs/nfs4filelayoutdev.c +++ b/fs/nfs/nfs4filelayoutdev.c | |||
@@ -225,11 +225,11 @@ nfs4_fl_free_deviceid(struct nfs4_file_layout_dsaddr *dsaddr) | |||
225 | } | 225 | } |
226 | 226 | ||
227 | static struct nfs4_pnfs_ds * | 227 | static struct nfs4_pnfs_ds * |
228 | nfs4_pnfs_ds_add(struct inode *inode, u32 ip_addr, u32 port) | 228 | nfs4_pnfs_ds_add(struct inode *inode, u32 ip_addr, u32 port, gfp_t gfp_flags) |
229 | { | 229 | { |
230 | struct nfs4_pnfs_ds *tmp_ds, *ds; | 230 | struct nfs4_pnfs_ds *tmp_ds, *ds; |
231 | 231 | ||
232 | ds = kzalloc(sizeof(*tmp_ds), GFP_KERNEL); | 232 | ds = kzalloc(sizeof(*tmp_ds), gfp_flags); |
233 | if (!ds) | 233 | if (!ds) |
234 | goto out; | 234 | goto out; |
235 | 235 | ||
@@ -261,7 +261,7 @@ out: | |||
261 | * Currently only support ipv4, and one multi-path address. | 261 | * Currently only support ipv4, and one multi-path address. |
262 | */ | 262 | */ |
263 | static struct nfs4_pnfs_ds * | 263 | static struct nfs4_pnfs_ds * |
264 | decode_and_add_ds(struct xdr_stream *streamp, struct inode *inode) | 264 | decode_and_add_ds(struct xdr_stream *streamp, struct inode *inode, gfp_t gfp_flags) |
265 | { | 265 | { |
266 | struct nfs4_pnfs_ds *ds = NULL; | 266 | struct nfs4_pnfs_ds *ds = NULL; |
267 | char *buf; | 267 | char *buf; |
@@ -303,7 +303,7 @@ decode_and_add_ds(struct xdr_stream *streamp, struct inode *inode) | |||
303 | rlen); | 303 | rlen); |
304 | goto out_err; | 304 | goto out_err; |
305 | } | 305 | } |
306 | buf = kmalloc(rlen + 1, GFP_KERNEL); | 306 | buf = kmalloc(rlen + 1, gfp_flags); |
307 | if (!buf) { | 307 | if (!buf) { |
308 | dprintk("%s: Not enough memory\n", __func__); | 308 | dprintk("%s: Not enough memory\n", __func__); |
309 | goto out_err; | 309 | goto out_err; |
@@ -333,7 +333,7 @@ decode_and_add_ds(struct xdr_stream *streamp, struct inode *inode) | |||
333 | sscanf(pstr, "-%d-%d", &tmp[0], &tmp[1]); | 333 | sscanf(pstr, "-%d-%d", &tmp[0], &tmp[1]); |
334 | port = htons((tmp[0] << 8) | (tmp[1])); | 334 | port = htons((tmp[0] << 8) | (tmp[1])); |
335 | 335 | ||
336 | ds = nfs4_pnfs_ds_add(inode, ip_addr, port); | 336 | ds = nfs4_pnfs_ds_add(inode, ip_addr, port, gfp_flags); |
337 | dprintk("%s: Decoded address and port %s\n", __func__, buf); | 337 | dprintk("%s: Decoded address and port %s\n", __func__, buf); |
338 | out_free: | 338 | out_free: |
339 | kfree(buf); | 339 | kfree(buf); |
@@ -343,7 +343,7 @@ out_err: | |||
343 | 343 | ||
344 | /* Decode opaque device data and return the result */ | 344 | /* Decode opaque device data and return the result */ |
345 | static struct nfs4_file_layout_dsaddr* | 345 | static struct nfs4_file_layout_dsaddr* |
346 | decode_device(struct inode *ino, struct pnfs_device *pdev) | 346 | decode_device(struct inode *ino, struct pnfs_device *pdev, gfp_t gfp_flags) |
347 | { | 347 | { |
348 | int i; | 348 | int i; |
349 | u32 cnt, num; | 349 | u32 cnt, num; |
@@ -362,7 +362,7 @@ decode_device(struct inode *ino, struct pnfs_device *pdev) | |||
362 | struct page *scratch; | 362 | struct page *scratch; |
363 | 363 | ||
364 | /* set up xdr stream */ | 364 | /* set up xdr stream */ |
365 | scratch = alloc_page(GFP_KERNEL); | 365 | scratch = alloc_page(gfp_flags); |
366 | if (!scratch) | 366 | if (!scratch) |
367 | goto out_err; | 367 | goto out_err; |
368 | 368 | ||
@@ -384,7 +384,7 @@ decode_device(struct inode *ino, struct pnfs_device *pdev) | |||
384 | } | 384 | } |
385 | 385 | ||
386 | /* read stripe indices */ | 386 | /* read stripe indices */ |
387 | stripe_indices = kcalloc(cnt, sizeof(u8), GFP_KERNEL); | 387 | stripe_indices = kcalloc(cnt, sizeof(u8), gfp_flags); |
388 | if (!stripe_indices) | 388 | if (!stripe_indices) |
389 | goto out_err_free_scratch; | 389 | goto out_err_free_scratch; |
390 | 390 | ||
@@ -423,7 +423,7 @@ decode_device(struct inode *ino, struct pnfs_device *pdev) | |||
423 | 423 | ||
424 | dsaddr = kzalloc(sizeof(*dsaddr) + | 424 | dsaddr = kzalloc(sizeof(*dsaddr) + |
425 | (sizeof(struct nfs4_pnfs_ds *) * (num - 1)), | 425 | (sizeof(struct nfs4_pnfs_ds *) * (num - 1)), |
426 | GFP_KERNEL); | 426 | gfp_flags); |
427 | if (!dsaddr) | 427 | if (!dsaddr) |
428 | goto out_err_free_stripe_indices; | 428 | goto out_err_free_stripe_indices; |
429 | 429 | ||
@@ -452,7 +452,7 @@ decode_device(struct inode *ino, struct pnfs_device *pdev) | |||
452 | for (j = 0; j < mp_count; j++) { | 452 | for (j = 0; j < mp_count; j++) { |
453 | if (j == 0) { | 453 | if (j == 0) { |
454 | dsaddr->ds_list[i] = decode_and_add_ds(&stream, | 454 | dsaddr->ds_list[i] = decode_and_add_ds(&stream, |
455 | ino); | 455 | ino, gfp_flags); |
456 | if (dsaddr->ds_list[i] == NULL) | 456 | if (dsaddr->ds_list[i] == NULL) |
457 | goto out_err_free_deviceid; | 457 | goto out_err_free_deviceid; |
458 | } else { | 458 | } else { |
@@ -503,12 +503,12 @@ out_err: | |||
503 | * available devices. | 503 | * available devices. |
504 | */ | 504 | */ |
505 | static struct nfs4_file_layout_dsaddr * | 505 | static struct nfs4_file_layout_dsaddr * |
506 | decode_and_add_device(struct inode *inode, struct pnfs_device *dev) | 506 | decode_and_add_device(struct inode *inode, struct pnfs_device *dev, gfp_t gfp_flags) |
507 | { | 507 | { |
508 | struct nfs4_file_layout_dsaddr *d, *new; | 508 | struct nfs4_file_layout_dsaddr *d, *new; |
509 | long hash; | 509 | long hash; |
510 | 510 | ||
511 | new = decode_device(inode, dev); | 511 | new = decode_device(inode, dev, gfp_flags); |
512 | if (!new) { | 512 | if (!new) { |
513 | printk(KERN_WARNING "%s: Could not decode or add device\n", | 513 | printk(KERN_WARNING "%s: Could not decode or add device\n", |
514 | __func__); | 514 | __func__); |
@@ -537,7 +537,7 @@ decode_and_add_device(struct inode *inode, struct pnfs_device *dev) | |||
537 | * of available devices, and return it. | 537 | * of available devices, and return it. |
538 | */ | 538 | */ |
539 | struct nfs4_file_layout_dsaddr * | 539 | struct nfs4_file_layout_dsaddr * |
540 | get_device_info(struct inode *inode, struct nfs4_deviceid *dev_id) | 540 | get_device_info(struct inode *inode, struct nfs4_deviceid *dev_id, gfp_t gfp_flags) |
541 | { | 541 | { |
542 | struct pnfs_device *pdev = NULL; | 542 | struct pnfs_device *pdev = NULL; |
543 | u32 max_resp_sz; | 543 | u32 max_resp_sz; |
@@ -556,17 +556,17 @@ get_device_info(struct inode *inode, struct nfs4_deviceid *dev_id) | |||
556 | dprintk("%s inode %p max_resp_sz %u max_pages %d\n", | 556 | dprintk("%s inode %p max_resp_sz %u max_pages %d\n", |
557 | __func__, inode, max_resp_sz, max_pages); | 557 | __func__, inode, max_resp_sz, max_pages); |
558 | 558 | ||
559 | pdev = kzalloc(sizeof(struct pnfs_device), GFP_KERNEL); | 559 | pdev = kzalloc(sizeof(struct pnfs_device), gfp_flags); |
560 | if (pdev == NULL) | 560 | if (pdev == NULL) |
561 | return NULL; | 561 | return NULL; |
562 | 562 | ||
563 | pages = kzalloc(max_pages * sizeof(struct page *), GFP_KERNEL); | 563 | pages = kzalloc(max_pages * sizeof(struct page *), gfp_flags); |
564 | if (pages == NULL) { | 564 | if (pages == NULL) { |
565 | kfree(pdev); | 565 | kfree(pdev); |
566 | return NULL; | 566 | return NULL; |
567 | } | 567 | } |
568 | for (i = 0; i < max_pages; i++) { | 568 | for (i = 0; i < max_pages; i++) { |
569 | pages[i] = alloc_page(GFP_KERNEL); | 569 | pages[i] = alloc_page(gfp_flags); |
570 | if (!pages[i]) | 570 | if (!pages[i]) |
571 | goto out_free; | 571 | goto out_free; |
572 | } | 572 | } |
@@ -587,7 +587,7 @@ get_device_info(struct inode *inode, struct nfs4_deviceid *dev_id) | |||
587 | * Found new device, need to decode it and then add it to the | 587 | * Found new device, need to decode it and then add it to the |
588 | * list of known devices for this mountpoint. | 588 | * list of known devices for this mountpoint. |
589 | */ | 589 | */ |
590 | dsaddr = decode_and_add_device(inode, pdev); | 590 | dsaddr = decode_and_add_device(inode, pdev, gfp_flags); |
591 | out_free: | 591 | out_free: |
592 | for (i = 0; i < max_pages; i++) | 592 | for (i = 0; i < max_pages; i++) |
593 | __free_page(pages[i]); | 593 | __free_page(pages[i]); |
diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c index 69c0f3c5ee7a..cf1b339c3937 100644 --- a/fs/nfs/nfs4proc.c +++ b/fs/nfs/nfs4proc.c | |||
@@ -300,6 +300,7 @@ static int nfs4_handle_exception(struct nfs_server *server, int errorcode, struc | |||
300 | ret = nfs4_delay(server->client, &exception->timeout); | 300 | ret = nfs4_delay(server->client, &exception->timeout); |
301 | if (ret != 0) | 301 | if (ret != 0) |
302 | break; | 302 | break; |
303 | case -NFS4ERR_RETRY_UNCACHED_REP: | ||
303 | case -NFS4ERR_OLD_STATEID: | 304 | case -NFS4ERR_OLD_STATEID: |
304 | exception->retry = 1; | 305 | exception->retry = 1; |
305 | break; | 306 | break; |
@@ -3695,6 +3696,7 @@ nfs4_async_handle_error(struct rpc_task *task, const struct nfs_server *server, | |||
3695 | rpc_delay(task, NFS4_POLL_RETRY_MAX); | 3696 | rpc_delay(task, NFS4_POLL_RETRY_MAX); |
3696 | task->tk_status = 0; | 3697 | task->tk_status = 0; |
3697 | return -EAGAIN; | 3698 | return -EAGAIN; |
3699 | case -NFS4ERR_RETRY_UNCACHED_REP: | ||
3698 | case -NFS4ERR_OLD_STATEID: | 3700 | case -NFS4ERR_OLD_STATEID: |
3699 | task->tk_status = 0; | 3701 | task->tk_status = 0; |
3700 | return -EAGAIN; | 3702 | return -EAGAIN; |
@@ -4844,6 +4846,8 @@ static void nfs4_get_lease_time_done(struct rpc_task *task, void *calldata) | |||
4844 | dprintk("%s Retry: tk_status %d\n", __func__, task->tk_status); | 4846 | dprintk("%s Retry: tk_status %d\n", __func__, task->tk_status); |
4845 | rpc_delay(task, NFS4_POLL_RETRY_MIN); | 4847 | rpc_delay(task, NFS4_POLL_RETRY_MIN); |
4846 | task->tk_status = 0; | 4848 | task->tk_status = 0; |
4849 | /* fall through */ | ||
4850 | case -NFS4ERR_RETRY_UNCACHED_REP: | ||
4847 | nfs_restart_rpc(task, data->clp); | 4851 | nfs_restart_rpc(task, data->clp); |
4848 | return; | 4852 | return; |
4849 | } | 4853 | } |
@@ -5479,6 +5483,8 @@ static int nfs41_reclaim_complete_handle_errors(struct rpc_task *task, struct nf | |||
5479 | break; | 5483 | break; |
5480 | case -NFS4ERR_DELAY: | 5484 | case -NFS4ERR_DELAY: |
5481 | rpc_delay(task, NFS4_POLL_RETRY_MAX); | 5485 | rpc_delay(task, NFS4_POLL_RETRY_MAX); |
5486 | /* fall through */ | ||
5487 | case -NFS4ERR_RETRY_UNCACHED_REP: | ||
5482 | return -EAGAIN; | 5488 | return -EAGAIN; |
5483 | default: | 5489 | default: |
5484 | nfs4_schedule_lease_recovery(clp); | 5490 | nfs4_schedule_lease_recovery(clp); |
diff --git a/fs/nfs/pnfs.c b/fs/nfs/pnfs.c index ff681ab65d31..f57f5281a520 100644 --- a/fs/nfs/pnfs.c +++ b/fs/nfs/pnfs.c | |||
@@ -383,6 +383,7 @@ pnfs_destroy_all_layouts(struct nfs_client *clp) | |||
383 | plh_layouts); | 383 | plh_layouts); |
384 | dprintk("%s freeing layout for inode %lu\n", __func__, | 384 | dprintk("%s freeing layout for inode %lu\n", __func__, |
385 | lo->plh_inode->i_ino); | 385 | lo->plh_inode->i_ino); |
386 | list_del_init(&lo->plh_layouts); | ||
386 | pnfs_destroy_layout(NFS_I(lo->plh_inode)); | 387 | pnfs_destroy_layout(NFS_I(lo->plh_inode)); |
387 | } | 388 | } |
388 | } | 389 | } |
@@ -466,7 +467,8 @@ pnfs_choose_layoutget_stateid(nfs4_stateid *dst, struct pnfs_layout_hdr *lo, | |||
466 | static struct pnfs_layout_segment * | 467 | static struct pnfs_layout_segment * |
467 | send_layoutget(struct pnfs_layout_hdr *lo, | 468 | send_layoutget(struct pnfs_layout_hdr *lo, |
468 | struct nfs_open_context *ctx, | 469 | struct nfs_open_context *ctx, |
469 | u32 iomode) | 470 | u32 iomode, |
471 | gfp_t gfp_flags) | ||
470 | { | 472 | { |
471 | struct inode *ino = lo->plh_inode; | 473 | struct inode *ino = lo->plh_inode; |
472 | struct nfs_server *server = NFS_SERVER(ino); | 474 | struct nfs_server *server = NFS_SERVER(ino); |
@@ -479,7 +481,7 @@ send_layoutget(struct pnfs_layout_hdr *lo, | |||
479 | dprintk("--> %s\n", __func__); | 481 | dprintk("--> %s\n", __func__); |
480 | 482 | ||
481 | BUG_ON(ctx == NULL); | 483 | BUG_ON(ctx == NULL); |
482 | lgp = kzalloc(sizeof(*lgp), GFP_KERNEL); | 484 | lgp = kzalloc(sizeof(*lgp), gfp_flags); |
483 | if (lgp == NULL) | 485 | if (lgp == NULL) |
484 | return NULL; | 486 | return NULL; |
485 | 487 | ||
@@ -487,12 +489,12 @@ send_layoutget(struct pnfs_layout_hdr *lo, | |||
487 | max_resp_sz = server->nfs_client->cl_session->fc_attrs.max_resp_sz; | 489 | max_resp_sz = server->nfs_client->cl_session->fc_attrs.max_resp_sz; |
488 | max_pages = max_resp_sz >> PAGE_SHIFT; | 490 | max_pages = max_resp_sz >> PAGE_SHIFT; |
489 | 491 | ||
490 | pages = kzalloc(max_pages * sizeof(struct page *), GFP_KERNEL); | 492 | pages = kzalloc(max_pages * sizeof(struct page *), gfp_flags); |
491 | if (!pages) | 493 | if (!pages) |
492 | goto out_err_free; | 494 | goto out_err_free; |
493 | 495 | ||
494 | for (i = 0; i < max_pages; i++) { | 496 | for (i = 0; i < max_pages; i++) { |
495 | pages[i] = alloc_page(GFP_KERNEL); | 497 | pages[i] = alloc_page(gfp_flags); |
496 | if (!pages[i]) | 498 | if (!pages[i]) |
497 | goto out_err_free; | 499 | goto out_err_free; |
498 | } | 500 | } |
@@ -508,6 +510,7 @@ send_layoutget(struct pnfs_layout_hdr *lo, | |||
508 | lgp->args.layout.pages = pages; | 510 | lgp->args.layout.pages = pages; |
509 | lgp->args.layout.pglen = max_pages * PAGE_SIZE; | 511 | lgp->args.layout.pglen = max_pages * PAGE_SIZE; |
510 | lgp->lsegpp = &lseg; | 512 | lgp->lsegpp = &lseg; |
513 | lgp->gfp_flags = gfp_flags; | ||
511 | 514 | ||
512 | /* Synchronously retrieve layout information from server and | 515 | /* Synchronously retrieve layout information from server and |
513 | * store in lseg. | 516 | * store in lseg. |
@@ -665,11 +668,11 @@ pnfs_insert_layout(struct pnfs_layout_hdr *lo, | |||
665 | } | 668 | } |
666 | 669 | ||
667 | static struct pnfs_layout_hdr * | 670 | static struct pnfs_layout_hdr * |
668 | alloc_init_layout_hdr(struct inode *ino) | 671 | alloc_init_layout_hdr(struct inode *ino, gfp_t gfp_flags) |
669 | { | 672 | { |
670 | struct pnfs_layout_hdr *lo; | 673 | struct pnfs_layout_hdr *lo; |
671 | 674 | ||
672 | lo = kzalloc(sizeof(struct pnfs_layout_hdr), GFP_KERNEL); | 675 | lo = kzalloc(sizeof(struct pnfs_layout_hdr), gfp_flags); |
673 | if (!lo) | 676 | if (!lo) |
674 | return NULL; | 677 | return NULL; |
675 | atomic_set(&lo->plh_refcount, 1); | 678 | atomic_set(&lo->plh_refcount, 1); |
@@ -681,7 +684,7 @@ alloc_init_layout_hdr(struct inode *ino) | |||
681 | } | 684 | } |
682 | 685 | ||
683 | static struct pnfs_layout_hdr * | 686 | static struct pnfs_layout_hdr * |
684 | pnfs_find_alloc_layout(struct inode *ino) | 687 | pnfs_find_alloc_layout(struct inode *ino, gfp_t gfp_flags) |
685 | { | 688 | { |
686 | struct nfs_inode *nfsi = NFS_I(ino); | 689 | struct nfs_inode *nfsi = NFS_I(ino); |
687 | struct pnfs_layout_hdr *new = NULL; | 690 | struct pnfs_layout_hdr *new = NULL; |
@@ -696,7 +699,7 @@ pnfs_find_alloc_layout(struct inode *ino) | |||
696 | return nfsi->layout; | 699 | return nfsi->layout; |
697 | } | 700 | } |
698 | spin_unlock(&ino->i_lock); | 701 | spin_unlock(&ino->i_lock); |
699 | new = alloc_init_layout_hdr(ino); | 702 | new = alloc_init_layout_hdr(ino, gfp_flags); |
700 | spin_lock(&ino->i_lock); | 703 | spin_lock(&ino->i_lock); |
701 | 704 | ||
702 | if (likely(nfsi->layout == NULL)) /* Won the race? */ | 705 | if (likely(nfsi->layout == NULL)) /* Won the race? */ |
@@ -756,7 +759,8 @@ pnfs_find_lseg(struct pnfs_layout_hdr *lo, u32 iomode) | |||
756 | struct pnfs_layout_segment * | 759 | struct pnfs_layout_segment * |
757 | pnfs_update_layout(struct inode *ino, | 760 | pnfs_update_layout(struct inode *ino, |
758 | struct nfs_open_context *ctx, | 761 | struct nfs_open_context *ctx, |
759 | enum pnfs_iomode iomode) | 762 | enum pnfs_iomode iomode, |
763 | gfp_t gfp_flags) | ||
760 | { | 764 | { |
761 | struct nfs_inode *nfsi = NFS_I(ino); | 765 | struct nfs_inode *nfsi = NFS_I(ino); |
762 | struct nfs_client *clp = NFS_SERVER(ino)->nfs_client; | 766 | struct nfs_client *clp = NFS_SERVER(ino)->nfs_client; |
@@ -767,7 +771,7 @@ pnfs_update_layout(struct inode *ino, | |||
767 | if (!pnfs_enabled_sb(NFS_SERVER(ino))) | 771 | if (!pnfs_enabled_sb(NFS_SERVER(ino))) |
768 | return NULL; | 772 | return NULL; |
769 | spin_lock(&ino->i_lock); | 773 | spin_lock(&ino->i_lock); |
770 | lo = pnfs_find_alloc_layout(ino); | 774 | lo = pnfs_find_alloc_layout(ino, gfp_flags); |
771 | if (lo == NULL) { | 775 | if (lo == NULL) { |
772 | dprintk("%s ERROR: can't get pnfs_layout_hdr\n", __func__); | 776 | dprintk("%s ERROR: can't get pnfs_layout_hdr\n", __func__); |
773 | goto out_unlock; | 777 | goto out_unlock; |
@@ -807,7 +811,7 @@ pnfs_update_layout(struct inode *ino, | |||
807 | spin_unlock(&clp->cl_lock); | 811 | spin_unlock(&clp->cl_lock); |
808 | } | 812 | } |
809 | 813 | ||
810 | lseg = send_layoutget(lo, ctx, iomode); | 814 | lseg = send_layoutget(lo, ctx, iomode, gfp_flags); |
811 | if (!lseg && first) { | 815 | if (!lseg && first) { |
812 | spin_lock(&clp->cl_lock); | 816 | spin_lock(&clp->cl_lock); |
813 | list_del_init(&lo->plh_layouts); | 817 | list_del_init(&lo->plh_layouts); |
@@ -846,7 +850,7 @@ pnfs_layout_process(struct nfs4_layoutget *lgp) | |||
846 | goto out; | 850 | goto out; |
847 | } | 851 | } |
848 | /* Inject layout blob into I/O device driver */ | 852 | /* Inject layout blob into I/O device driver */ |
849 | lseg = NFS_SERVER(ino)->pnfs_curr_ld->alloc_lseg(lo, res); | 853 | lseg = NFS_SERVER(ino)->pnfs_curr_ld->alloc_lseg(lo, res, lgp->gfp_flags); |
850 | if (!lseg || IS_ERR(lseg)) { | 854 | if (!lseg || IS_ERR(lseg)) { |
851 | if (!lseg) | 855 | if (!lseg) |
852 | status = -ENOMEM; | 856 | status = -ENOMEM; |
@@ -899,7 +903,8 @@ static int pnfs_read_pg_test(struct nfs_pageio_descriptor *pgio, | |||
899 | /* This is first coelesce call for a series of nfs_pages */ | 903 | /* This is first coelesce call for a series of nfs_pages */ |
900 | pgio->pg_lseg = pnfs_update_layout(pgio->pg_inode, | 904 | pgio->pg_lseg = pnfs_update_layout(pgio->pg_inode, |
901 | prev->wb_context, | 905 | prev->wb_context, |
902 | IOMODE_READ); | 906 | IOMODE_READ, |
907 | GFP_KERNEL); | ||
903 | } | 908 | } |
904 | return NFS_SERVER(pgio->pg_inode)->pnfs_curr_ld->pg_test(pgio, prev, req); | 909 | return NFS_SERVER(pgio->pg_inode)->pnfs_curr_ld->pg_test(pgio, prev, req); |
905 | } | 910 | } |
@@ -921,7 +926,8 @@ static int pnfs_write_pg_test(struct nfs_pageio_descriptor *pgio, | |||
921 | /* This is first coelesce call for a series of nfs_pages */ | 926 | /* This is first coelesce call for a series of nfs_pages */ |
922 | pgio->pg_lseg = pnfs_update_layout(pgio->pg_inode, | 927 | pgio->pg_lseg = pnfs_update_layout(pgio->pg_inode, |
923 | prev->wb_context, | 928 | prev->wb_context, |
924 | IOMODE_RW); | 929 | IOMODE_RW, |
930 | GFP_NOFS); | ||
925 | } | 931 | } |
926 | return NFS_SERVER(pgio->pg_inode)->pnfs_curr_ld->pg_test(pgio, prev, req); | 932 | return NFS_SERVER(pgio->pg_inode)->pnfs_curr_ld->pg_test(pgio, prev, req); |
927 | } | 933 | } |
diff --git a/fs/nfs/pnfs.h b/fs/nfs/pnfs.h index bc4827202e7a..0c015bad9e7a 100644 --- a/fs/nfs/pnfs.h +++ b/fs/nfs/pnfs.h | |||
@@ -70,7 +70,7 @@ struct pnfs_layoutdriver_type { | |||
70 | const u32 id; | 70 | const u32 id; |
71 | const char *name; | 71 | const char *name; |
72 | struct module *owner; | 72 | struct module *owner; |
73 | struct pnfs_layout_segment * (*alloc_lseg) (struct pnfs_layout_hdr *layoutid, struct nfs4_layoutget_res *lgr); | 73 | struct pnfs_layout_segment * (*alloc_lseg) (struct pnfs_layout_hdr *layoutid, struct nfs4_layoutget_res *lgr, gfp_t gfp_flags); |
74 | void (*free_lseg) (struct pnfs_layout_segment *lseg); | 74 | void (*free_lseg) (struct pnfs_layout_segment *lseg); |
75 | 75 | ||
76 | /* test for nfs page cache coalescing */ | 76 | /* test for nfs page cache coalescing */ |
@@ -126,7 +126,7 @@ void get_layout_hdr(struct pnfs_layout_hdr *lo); | |||
126 | void put_lseg(struct pnfs_layout_segment *lseg); | 126 | void put_lseg(struct pnfs_layout_segment *lseg); |
127 | struct pnfs_layout_segment * | 127 | struct pnfs_layout_segment * |
128 | pnfs_update_layout(struct inode *ino, struct nfs_open_context *ctx, | 128 | pnfs_update_layout(struct inode *ino, struct nfs_open_context *ctx, |
129 | enum pnfs_iomode access_type); | 129 | enum pnfs_iomode access_type, gfp_t gfp_flags); |
130 | void set_pnfs_layoutdriver(struct nfs_server *, u32 id); | 130 | void set_pnfs_layoutdriver(struct nfs_server *, u32 id); |
131 | void unset_pnfs_layoutdriver(struct nfs_server *); | 131 | void unset_pnfs_layoutdriver(struct nfs_server *); |
132 | enum pnfs_try_status pnfs_try_to_write_data(struct nfs_write_data *, | 132 | enum pnfs_try_status pnfs_try_to_write_data(struct nfs_write_data *, |
@@ -245,7 +245,7 @@ static inline void put_lseg(struct pnfs_layout_segment *lseg) | |||
245 | 245 | ||
246 | static inline struct pnfs_layout_segment * | 246 | static inline struct pnfs_layout_segment * |
247 | pnfs_update_layout(struct inode *ino, struct nfs_open_context *ctx, | 247 | pnfs_update_layout(struct inode *ino, struct nfs_open_context *ctx, |
248 | enum pnfs_iomode access_type) | 248 | enum pnfs_iomode access_type, gfp_t gfp_flags) |
249 | { | 249 | { |
250 | return NULL; | 250 | return NULL; |
251 | } | 251 | } |
diff --git a/fs/nfs/read.c b/fs/nfs/read.c index 7cded2b12a05..2bcf0dc306a1 100644 --- a/fs/nfs/read.c +++ b/fs/nfs/read.c | |||
@@ -288,7 +288,7 @@ static int nfs_pagein_multi(struct nfs_pageio_descriptor *desc) | |||
288 | atomic_set(&req->wb_complete, requests); | 288 | atomic_set(&req->wb_complete, requests); |
289 | 289 | ||
290 | BUG_ON(desc->pg_lseg != NULL); | 290 | BUG_ON(desc->pg_lseg != NULL); |
291 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_READ); | 291 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_READ, GFP_KERNEL); |
292 | ClearPageError(page); | 292 | ClearPageError(page); |
293 | offset = 0; | 293 | offset = 0; |
294 | nbytes = desc->pg_count; | 294 | nbytes = desc->pg_count; |
@@ -351,7 +351,7 @@ static int nfs_pagein_one(struct nfs_pageio_descriptor *desc) | |||
351 | } | 351 | } |
352 | req = nfs_list_entry(data->pages.next); | 352 | req = nfs_list_entry(data->pages.next); |
353 | if ((!lseg) && list_is_singular(&data->pages)) | 353 | if ((!lseg) && list_is_singular(&data->pages)) |
354 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_READ); | 354 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_READ, GFP_KERNEL); |
355 | 355 | ||
356 | ret = nfs_read_rpcsetup(req, data, &nfs_read_full_ops, desc->pg_count, | 356 | ret = nfs_read_rpcsetup(req, data, &nfs_read_full_ops, desc->pg_count, |
357 | 0, lseg); | 357 | 0, lseg); |
diff --git a/fs/nfs/write.c b/fs/nfs/write.c index 3bd5d7e80f6c..49c715b4ac92 100644 --- a/fs/nfs/write.c +++ b/fs/nfs/write.c | |||
@@ -939,7 +939,7 @@ static int nfs_flush_multi(struct nfs_pageio_descriptor *desc) | |||
939 | atomic_set(&req->wb_complete, requests); | 939 | atomic_set(&req->wb_complete, requests); |
940 | 940 | ||
941 | BUG_ON(desc->pg_lseg); | 941 | BUG_ON(desc->pg_lseg); |
942 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_RW); | 942 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_RW, GFP_NOFS); |
943 | ClearPageError(page); | 943 | ClearPageError(page); |
944 | offset = 0; | 944 | offset = 0; |
945 | nbytes = desc->pg_count; | 945 | nbytes = desc->pg_count; |
@@ -1013,7 +1013,7 @@ static int nfs_flush_one(struct nfs_pageio_descriptor *desc) | |||
1013 | } | 1013 | } |
1014 | req = nfs_list_entry(data->pages.next); | 1014 | req = nfs_list_entry(data->pages.next); |
1015 | if ((!lseg) && list_is_singular(&data->pages)) | 1015 | if ((!lseg) && list_is_singular(&data->pages)) |
1016 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_RW); | 1016 | lseg = pnfs_update_layout(desc->pg_inode, req->wb_context, IOMODE_RW, GFP_NOFS); |
1017 | 1017 | ||
1018 | if ((desc->pg_ioflags & FLUSH_COND_STABLE) && | 1018 | if ((desc->pg_ioflags & FLUSH_COND_STABLE) && |
1019 | (desc->pg_moreio || NFS_I(desc->pg_inode)->ncommit)) | 1019 | (desc->pg_moreio || NFS_I(desc->pg_inode)->ncommit)) |
diff --git a/fs/nilfs2/alloc.c b/fs/nilfs2/alloc.c index 0a0a66d98cce..f7684483785e 100644 --- a/fs/nilfs2/alloc.c +++ b/fs/nilfs2/alloc.c | |||
@@ -646,7 +646,7 @@ int nilfs_palloc_freev(struct inode *inode, __u64 *entry_nrs, size_t nitems) | |||
646 | unsigned long group, group_offset; | 646 | unsigned long group, group_offset; |
647 | int i, j, n, ret; | 647 | int i, j, n, ret; |
648 | 648 | ||
649 | for (i = 0; i < nitems; i += n) { | 649 | for (i = 0; i < nitems; i = j) { |
650 | group = nilfs_palloc_group(inode, entry_nrs[i], &group_offset); | 650 | group = nilfs_palloc_group(inode, entry_nrs[i], &group_offset); |
651 | ret = nilfs_palloc_get_desc_block(inode, group, 0, &desc_bh); | 651 | ret = nilfs_palloc_get_desc_block(inode, group, 0, &desc_bh); |
652 | if (ret < 0) | 652 | if (ret < 0) |
diff --git a/fs/ocfs2/cluster/heartbeat.c b/fs/ocfs2/cluster/heartbeat.c index 643720209a98..9a3e6bbff27b 100644 --- a/fs/ocfs2/cluster/heartbeat.c +++ b/fs/ocfs2/cluster/heartbeat.c | |||
@@ -539,25 +539,41 @@ static int o2hb_verify_crc(struct o2hb_region *reg, | |||
539 | 539 | ||
540 | /* We want to make sure that nobody is heartbeating on top of us -- | 540 | /* We want to make sure that nobody is heartbeating on top of us -- |
541 | * this will help detect an invalid configuration. */ | 541 | * this will help detect an invalid configuration. */ |
542 | static int o2hb_check_last_timestamp(struct o2hb_region *reg) | 542 | static void o2hb_check_last_timestamp(struct o2hb_region *reg) |
543 | { | 543 | { |
544 | int node_num, ret; | ||
545 | struct o2hb_disk_slot *slot; | 544 | struct o2hb_disk_slot *slot; |
546 | struct o2hb_disk_heartbeat_block *hb_block; | 545 | struct o2hb_disk_heartbeat_block *hb_block; |
546 | char *errstr; | ||
547 | 547 | ||
548 | node_num = o2nm_this_node(); | 548 | slot = ®->hr_slots[o2nm_this_node()]; |
549 | |||
550 | ret = 1; | ||
551 | slot = ®->hr_slots[node_num]; | ||
552 | /* Don't check on our 1st timestamp */ | 549 | /* Don't check on our 1st timestamp */ |
553 | if (slot->ds_last_time) { | 550 | if (!slot->ds_last_time) |
554 | hb_block = slot->ds_raw_block; | 551 | return; |
555 | 552 | ||
556 | if (le64_to_cpu(hb_block->hb_seq) != slot->ds_last_time) | 553 | hb_block = slot->ds_raw_block; |
557 | ret = 0; | 554 | if (le64_to_cpu(hb_block->hb_seq) == slot->ds_last_time && |
558 | } | 555 | le64_to_cpu(hb_block->hb_generation) == slot->ds_last_generation && |
556 | hb_block->hb_node == slot->ds_node_num) | ||
557 | return; | ||
559 | 558 | ||
560 | return ret; | 559 | #define ERRSTR1 "Another node is heartbeating on device" |
560 | #define ERRSTR2 "Heartbeat generation mismatch on device" | ||
561 | #define ERRSTR3 "Heartbeat sequence mismatch on device" | ||
562 | |||
563 | if (hb_block->hb_node != slot->ds_node_num) | ||
564 | errstr = ERRSTR1; | ||
565 | else if (le64_to_cpu(hb_block->hb_generation) != | ||
566 | slot->ds_last_generation) | ||
567 | errstr = ERRSTR2; | ||
568 | else | ||
569 | errstr = ERRSTR3; | ||
570 | |||
571 | mlog(ML_ERROR, "%s (%s): expected(%u:0x%llx, 0x%llx), " | ||
572 | "ondisk(%u:0x%llx, 0x%llx)\n", errstr, reg->hr_dev_name, | ||
573 | slot->ds_node_num, (unsigned long long)slot->ds_last_generation, | ||
574 | (unsigned long long)slot->ds_last_time, hb_block->hb_node, | ||
575 | (unsigned long long)le64_to_cpu(hb_block->hb_generation), | ||
576 | (unsigned long long)le64_to_cpu(hb_block->hb_seq)); | ||
561 | } | 577 | } |
562 | 578 | ||
563 | static inline void o2hb_prepare_block(struct o2hb_region *reg, | 579 | static inline void o2hb_prepare_block(struct o2hb_region *reg, |
@@ -983,9 +999,7 @@ static int o2hb_do_disk_heartbeat(struct o2hb_region *reg) | |||
983 | /* With an up to date view of the slots, we can check that no | 999 | /* With an up to date view of the slots, we can check that no |
984 | * other node has been improperly configured to heartbeat in | 1000 | * other node has been improperly configured to heartbeat in |
985 | * our slot. */ | 1001 | * our slot. */ |
986 | if (!o2hb_check_last_timestamp(reg)) | 1002 | o2hb_check_last_timestamp(reg); |
987 | mlog(ML_ERROR, "Device \"%s\": another node is heartbeating " | ||
988 | "in our slot!\n", reg->hr_dev_name); | ||
989 | 1003 | ||
990 | /* fill in the proper info for our next heartbeat */ | 1004 | /* fill in the proper info for our next heartbeat */ |
991 | o2hb_prepare_block(reg, reg->hr_generation); | 1005 | o2hb_prepare_block(reg, reg->hr_generation); |
@@ -999,8 +1013,8 @@ static int o2hb_do_disk_heartbeat(struct o2hb_region *reg) | |||
999 | } | 1013 | } |
1000 | 1014 | ||
1001 | i = -1; | 1015 | i = -1; |
1002 | while((i = find_next_bit(configured_nodes, O2NM_MAX_NODES, i + 1)) < O2NM_MAX_NODES) { | 1016 | while((i = find_next_bit(configured_nodes, |
1003 | 1017 | O2NM_MAX_NODES, i + 1)) < O2NM_MAX_NODES) { | |
1004 | change |= o2hb_check_slot(reg, ®->hr_slots[i]); | 1018 | change |= o2hb_check_slot(reg, ®->hr_slots[i]); |
1005 | } | 1019 | } |
1006 | 1020 | ||
@@ -1690,6 +1704,7 @@ static ssize_t o2hb_region_dev_write(struct o2hb_region *reg, | |||
1690 | struct file *filp = NULL; | 1704 | struct file *filp = NULL; |
1691 | struct inode *inode = NULL; | 1705 | struct inode *inode = NULL; |
1692 | ssize_t ret = -EINVAL; | 1706 | ssize_t ret = -EINVAL; |
1707 | int live_threshold; | ||
1693 | 1708 | ||
1694 | if (reg->hr_bdev) | 1709 | if (reg->hr_bdev) |
1695 | goto out; | 1710 | goto out; |
@@ -1766,8 +1781,18 @@ static ssize_t o2hb_region_dev_write(struct o2hb_region *reg, | |||
1766 | * A node is considered live after it has beat LIVE_THRESHOLD | 1781 | * A node is considered live after it has beat LIVE_THRESHOLD |
1767 | * times. We're not steady until we've given them a chance | 1782 | * times. We're not steady until we've given them a chance |
1768 | * _after_ our first read. | 1783 | * _after_ our first read. |
1784 | * The default threshold is bare minimum so as to limit the delay | ||
1785 | * during mounts. For global heartbeat, the threshold doubled for the | ||
1786 | * first region. | ||
1769 | */ | 1787 | */ |
1770 | atomic_set(®->hr_steady_iterations, O2HB_LIVE_THRESHOLD + 1); | 1788 | live_threshold = O2HB_LIVE_THRESHOLD; |
1789 | if (o2hb_global_heartbeat_active()) { | ||
1790 | spin_lock(&o2hb_live_lock); | ||
1791 | if (o2hb_pop_count(&o2hb_region_bitmap, O2NM_MAX_REGIONS) == 1) | ||
1792 | live_threshold <<= 1; | ||
1793 | spin_unlock(&o2hb_live_lock); | ||
1794 | } | ||
1795 | atomic_set(®->hr_steady_iterations, live_threshold + 1); | ||
1771 | 1796 | ||
1772 | hb_task = kthread_run(o2hb_thread, reg, "o2hb-%s", | 1797 | hb_task = kthread_run(o2hb_thread, reg, "o2hb-%s", |
1773 | reg->hr_item.ci_name); | 1798 | reg->hr_item.ci_name); |
diff --git a/fs/ocfs2/dir.c b/fs/ocfs2/dir.c index 9fe5b8fd658f..8582e3f4f120 100644 --- a/fs/ocfs2/dir.c +++ b/fs/ocfs2/dir.c | |||
@@ -2868,7 +2868,7 @@ static int ocfs2_expand_inline_dir(struct inode *dir, struct buffer_head *di_bh, | |||
2868 | bytes = blocks_wanted << sb->s_blocksize_bits; | 2868 | bytes = blocks_wanted << sb->s_blocksize_bits; |
2869 | struct ocfs2_super *osb = OCFS2_SB(dir->i_sb); | 2869 | struct ocfs2_super *osb = OCFS2_SB(dir->i_sb); |
2870 | struct ocfs2_inode_info *oi = OCFS2_I(dir); | 2870 | struct ocfs2_inode_info *oi = OCFS2_I(dir); |
2871 | struct ocfs2_alloc_context *data_ac; | 2871 | struct ocfs2_alloc_context *data_ac = NULL; |
2872 | struct ocfs2_alloc_context *meta_ac = NULL; | 2872 | struct ocfs2_alloc_context *meta_ac = NULL; |
2873 | struct buffer_head *dirdata_bh = NULL; | 2873 | struct buffer_head *dirdata_bh = NULL; |
2874 | struct buffer_head *dx_root_bh = NULL; | 2874 | struct buffer_head *dx_root_bh = NULL; |
diff --git a/fs/ocfs2/dlm/dlmdomain.c b/fs/ocfs2/dlm/dlmdomain.c index 7540a492eaba..3b179d6cbde0 100644 --- a/fs/ocfs2/dlm/dlmdomain.c +++ b/fs/ocfs2/dlm/dlmdomain.c | |||
@@ -1614,7 +1614,8 @@ static int dlm_try_to_join_domain(struct dlm_ctxt *dlm) | |||
1614 | spin_unlock(&dlm->spinlock); | 1614 | spin_unlock(&dlm->spinlock); |
1615 | 1615 | ||
1616 | /* Support for global heartbeat and node info was added in 1.1 */ | 1616 | /* Support for global heartbeat and node info was added in 1.1 */ |
1617 | if (dlm_protocol.pv_major > 1 || dlm_protocol.pv_minor > 0) { | 1617 | if (dlm->dlm_locking_proto.pv_major > 1 || |
1618 | dlm->dlm_locking_proto.pv_minor > 0) { | ||
1618 | status = dlm_send_nodeinfo(dlm, ctxt->yes_resp_map); | 1619 | status = dlm_send_nodeinfo(dlm, ctxt->yes_resp_map); |
1619 | if (status) { | 1620 | if (status) { |
1620 | mlog_errno(status); | 1621 | mlog_errno(status); |
diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c index fede57ed005f..84d166328cf7 100644 --- a/fs/ocfs2/dlm/dlmmaster.c +++ b/fs/ocfs2/dlm/dlmmaster.c | |||
@@ -2574,6 +2574,9 @@ fail: | |||
2574 | res->state &= ~DLM_LOCK_RES_MIGRATING; | 2574 | res->state &= ~DLM_LOCK_RES_MIGRATING; |
2575 | wake = 1; | 2575 | wake = 1; |
2576 | spin_unlock(&res->spinlock); | 2576 | spin_unlock(&res->spinlock); |
2577 | if (dlm_is_host_down(ret)) | ||
2578 | dlm_wait_for_node_death(dlm, target, | ||
2579 | DLM_NODE_DEATH_WAIT_MAX); | ||
2577 | goto leave; | 2580 | goto leave; |
2578 | } | 2581 | } |
2579 | 2582 | ||
diff --git a/fs/ocfs2/file.c b/fs/ocfs2/file.c index 41565ae52856..89659d6dc206 100644 --- a/fs/ocfs2/file.c +++ b/fs/ocfs2/file.c | |||
@@ -1607,6 +1607,9 @@ static void ocfs2_calc_trunc_pos(struct inode *inode, | |||
1607 | range = le32_to_cpu(rec->e_cpos) + ocfs2_rec_clusters(el, rec); | 1607 | range = le32_to_cpu(rec->e_cpos) + ocfs2_rec_clusters(el, rec); |
1608 | 1608 | ||
1609 | if (le32_to_cpu(rec->e_cpos) >= trunc_start) { | 1609 | if (le32_to_cpu(rec->e_cpos) >= trunc_start) { |
1610 | /* | ||
1611 | * remove an entire extent record. | ||
1612 | */ | ||
1610 | *trunc_cpos = le32_to_cpu(rec->e_cpos); | 1613 | *trunc_cpos = le32_to_cpu(rec->e_cpos); |
1611 | /* | 1614 | /* |
1612 | * Skip holes if any. | 1615 | * Skip holes if any. |
@@ -1617,7 +1620,16 @@ static void ocfs2_calc_trunc_pos(struct inode *inode, | |||
1617 | *blkno = le64_to_cpu(rec->e_blkno); | 1620 | *blkno = le64_to_cpu(rec->e_blkno); |
1618 | *trunc_end = le32_to_cpu(rec->e_cpos); | 1621 | *trunc_end = le32_to_cpu(rec->e_cpos); |
1619 | } else if (range > trunc_start) { | 1622 | } else if (range > trunc_start) { |
1623 | /* | ||
1624 | * remove a partial extent record, which means we're | ||
1625 | * removing the last extent record. | ||
1626 | */ | ||
1620 | *trunc_cpos = trunc_start; | 1627 | *trunc_cpos = trunc_start; |
1628 | /* | ||
1629 | * skip hole if any. | ||
1630 | */ | ||
1631 | if (range < *trunc_end) | ||
1632 | *trunc_end = range; | ||
1621 | *trunc_len = *trunc_end - trunc_start; | 1633 | *trunc_len = *trunc_end - trunc_start; |
1622 | coff = trunc_start - le32_to_cpu(rec->e_cpos); | 1634 | coff = trunc_start - le32_to_cpu(rec->e_cpos); |
1623 | *blkno = le64_to_cpu(rec->e_blkno) + | 1635 | *blkno = le64_to_cpu(rec->e_blkno) + |
diff --git a/fs/ocfs2/journal.c b/fs/ocfs2/journal.c index b141a44605ca..295d56454e8b 100644 --- a/fs/ocfs2/journal.c +++ b/fs/ocfs2/journal.c | |||
@@ -1260,6 +1260,9 @@ void ocfs2_complete_mount_recovery(struct ocfs2_super *osb) | |||
1260 | { | 1260 | { |
1261 | struct ocfs2_journal *journal = osb->journal; | 1261 | struct ocfs2_journal *journal = osb->journal; |
1262 | 1262 | ||
1263 | if (ocfs2_is_hard_readonly(osb)) | ||
1264 | return; | ||
1265 | |||
1263 | /* No need to queue up our truncate_log as regular cleanup will catch | 1266 | /* No need to queue up our truncate_log as regular cleanup will catch |
1264 | * that */ | 1267 | * that */ |
1265 | ocfs2_queue_recovery_completion(journal, osb->slot_num, | 1268 | ocfs2_queue_recovery_completion(journal, osb->slot_num, |
diff --git a/fs/partitions/efi.c b/fs/partitions/efi.c index ac0ccb5026a2..19d6750d1d6c 100644 --- a/fs/partitions/efi.c +++ b/fs/partitions/efi.c | |||
@@ -348,6 +348,12 @@ static int is_gpt_valid(struct parsed_partitions *state, u64 lba, | |||
348 | goto fail; | 348 | goto fail; |
349 | } | 349 | } |
350 | 350 | ||
351 | /* Check that sizeof_partition_entry has the correct value */ | ||
352 | if (le32_to_cpu((*gpt)->sizeof_partition_entry) != sizeof(gpt_entry)) { | ||
353 | pr_debug("GUID Partitition Entry Size check failed.\n"); | ||
354 | goto fail; | ||
355 | } | ||
356 | |||
351 | if (!(*ptes = alloc_read_gpt_entries(state, *gpt))) | 357 | if (!(*ptes = alloc_read_gpt_entries(state, *gpt))) |
352 | goto fail; | 358 | goto fail; |
353 | 359 | ||
diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c index 2e7addfd9803..318d8654989b 100644 --- a/fs/proc/task_mmu.c +++ b/fs/proc/task_mmu.c | |||
@@ -214,7 +214,7 @@ static void show_map_vma(struct seq_file *m, struct vm_area_struct *vma) | |||
214 | int flags = vma->vm_flags; | 214 | int flags = vma->vm_flags; |
215 | unsigned long ino = 0; | 215 | unsigned long ino = 0; |
216 | unsigned long long pgoff = 0; | 216 | unsigned long long pgoff = 0; |
217 | unsigned long start; | 217 | unsigned long start, end; |
218 | dev_t dev = 0; | 218 | dev_t dev = 0; |
219 | int len; | 219 | int len; |
220 | 220 | ||
@@ -227,13 +227,15 @@ static void show_map_vma(struct seq_file *m, struct vm_area_struct *vma) | |||
227 | 227 | ||
228 | /* We don't show the stack guard page in /proc/maps */ | 228 | /* We don't show the stack guard page in /proc/maps */ |
229 | start = vma->vm_start; | 229 | start = vma->vm_start; |
230 | if (vma->vm_flags & VM_GROWSDOWN) | 230 | if (stack_guard_page_start(vma, start)) |
231 | if (!vma_stack_continue(vma->vm_prev, vma->vm_start)) | 231 | start += PAGE_SIZE; |
232 | start += PAGE_SIZE; | 232 | end = vma->vm_end; |
233 | if (stack_guard_page_end(vma, end)) | ||
234 | end -= PAGE_SIZE; | ||
233 | 235 | ||
234 | seq_printf(m, "%08lx-%08lx %c%c%c%c %08llx %02x:%02x %lu %n", | 236 | seq_printf(m, "%08lx-%08lx %c%c%c%c %08llx %02x:%02x %lu %n", |
235 | start, | 237 | start, |
236 | vma->vm_end, | 238 | end, |
237 | flags & VM_READ ? 'r' : '-', | 239 | flags & VM_READ ? 'r' : '-', |
238 | flags & VM_WRITE ? 'w' : '-', | 240 | flags & VM_WRITE ? 'w' : '-', |
239 | flags & VM_EXEC ? 'x' : '-', | 241 | flags & VM_EXEC ? 'x' : '-', |
diff --git a/fs/ubifs/log.c b/fs/ubifs/log.c index 4d0cb1241460..40fa780ebea7 100644 --- a/fs/ubifs/log.c +++ b/fs/ubifs/log.c | |||
@@ -175,26 +175,6 @@ void ubifs_add_bud(struct ubifs_info *c, struct ubifs_bud *bud) | |||
175 | } | 175 | } |
176 | 176 | ||
177 | /** | 177 | /** |
178 | * ubifs_create_buds_lists - create journal head buds lists for remount rw. | ||
179 | * @c: UBIFS file-system description object | ||
180 | */ | ||
181 | void ubifs_create_buds_lists(struct ubifs_info *c) | ||
182 | { | ||
183 | struct rb_node *p; | ||
184 | |||
185 | spin_lock(&c->buds_lock); | ||
186 | p = rb_first(&c->buds); | ||
187 | while (p) { | ||
188 | struct ubifs_bud *bud = rb_entry(p, struct ubifs_bud, rb); | ||
189 | struct ubifs_jhead *jhead = &c->jheads[bud->jhead]; | ||
190 | |||
191 | list_add_tail(&bud->list, &jhead->buds_list); | ||
192 | p = rb_next(p); | ||
193 | } | ||
194 | spin_unlock(&c->buds_lock); | ||
195 | } | ||
196 | |||
197 | /** | ||
198 | * ubifs_add_bud_to_log - add a new bud to the log. | 178 | * ubifs_add_bud_to_log - add a new bud to the log. |
199 | * @c: UBIFS file-system description object | 179 | * @c: UBIFS file-system description object |
200 | * @jhead: journal head the bud belongs to | 180 | * @jhead: journal head the bud belongs to |
diff --git a/fs/ubifs/replay.c b/fs/ubifs/replay.c index eed0fcff8d73..d3d6d365bfc1 100644 --- a/fs/ubifs/replay.c +++ b/fs/ubifs/replay.c | |||
@@ -59,6 +59,7 @@ enum { | |||
59 | * @new_size: truncation new size | 59 | * @new_size: truncation new size |
60 | * @free: amount of free space in a bud | 60 | * @free: amount of free space in a bud |
61 | * @dirty: amount of dirty space in a bud from padding and deletion nodes | 61 | * @dirty: amount of dirty space in a bud from padding and deletion nodes |
62 | * @jhead: journal head number of the bud | ||
62 | * | 63 | * |
63 | * UBIFS journal replay must compare node sequence numbers, which means it must | 64 | * UBIFS journal replay must compare node sequence numbers, which means it must |
64 | * build a tree of node information to insert into the TNC. | 65 | * build a tree of node information to insert into the TNC. |
@@ -80,6 +81,7 @@ struct replay_entry { | |||
80 | struct { | 81 | struct { |
81 | int free; | 82 | int free; |
82 | int dirty; | 83 | int dirty; |
84 | int jhead; | ||
83 | }; | 85 | }; |
84 | }; | 86 | }; |
85 | }; | 87 | }; |
@@ -159,6 +161,11 @@ static int set_bud_lprops(struct ubifs_info *c, struct replay_entry *r) | |||
159 | err = PTR_ERR(lp); | 161 | err = PTR_ERR(lp); |
160 | goto out; | 162 | goto out; |
161 | } | 163 | } |
164 | |||
165 | /* Make sure the journal head points to the latest bud */ | ||
166 | err = ubifs_wbuf_seek_nolock(&c->jheads[r->jhead].wbuf, r->lnum, | ||
167 | c->leb_size - r->free, UBI_SHORTTERM); | ||
168 | |||
162 | out: | 169 | out: |
163 | ubifs_release_lprops(c); | 170 | ubifs_release_lprops(c); |
164 | return err; | 171 | return err; |
@@ -627,10 +634,6 @@ static int replay_bud(struct ubifs_info *c, int lnum, int offs, int jhead, | |||
627 | ubifs_assert(sleb->endpt - offs >= used); | 634 | ubifs_assert(sleb->endpt - offs >= used); |
628 | ubifs_assert(sleb->endpt % c->min_io_size == 0); | 635 | ubifs_assert(sleb->endpt % c->min_io_size == 0); |
629 | 636 | ||
630 | if (sleb->endpt + c->min_io_size <= c->leb_size && !c->ro_mount) | ||
631 | err = ubifs_wbuf_seek_nolock(&c->jheads[jhead].wbuf, lnum, | ||
632 | sleb->endpt, UBI_SHORTTERM); | ||
633 | |||
634 | *dirty = sleb->endpt - offs - used; | 637 | *dirty = sleb->endpt - offs - used; |
635 | *free = c->leb_size - sleb->endpt; | 638 | *free = c->leb_size - sleb->endpt; |
636 | 639 | ||
@@ -653,12 +656,14 @@ out_dump: | |||
653 | * @sqnum: sequence number | 656 | * @sqnum: sequence number |
654 | * @free: amount of free space in bud | 657 | * @free: amount of free space in bud |
655 | * @dirty: amount of dirty space from padding and deletion nodes | 658 | * @dirty: amount of dirty space from padding and deletion nodes |
659 | * @jhead: journal head number for the bud | ||
656 | * | 660 | * |
657 | * This function inserts a reference node to the replay tree and returns zero | 661 | * This function inserts a reference node to the replay tree and returns zero |
658 | * in case of success or a negative error code in case of failure. | 662 | * in case of success or a negative error code in case of failure. |
659 | */ | 663 | */ |
660 | static int insert_ref_node(struct ubifs_info *c, int lnum, int offs, | 664 | static int insert_ref_node(struct ubifs_info *c, int lnum, int offs, |
661 | unsigned long long sqnum, int free, int dirty) | 665 | unsigned long long sqnum, int free, int dirty, |
666 | int jhead) | ||
662 | { | 667 | { |
663 | struct rb_node **p = &c->replay_tree.rb_node, *parent = NULL; | 668 | struct rb_node **p = &c->replay_tree.rb_node, *parent = NULL; |
664 | struct replay_entry *r; | 669 | struct replay_entry *r; |
@@ -688,6 +693,7 @@ static int insert_ref_node(struct ubifs_info *c, int lnum, int offs, | |||
688 | r->flags = REPLAY_REF; | 693 | r->flags = REPLAY_REF; |
689 | r->free = free; | 694 | r->free = free; |
690 | r->dirty = dirty; | 695 | r->dirty = dirty; |
696 | r->jhead = jhead; | ||
691 | 697 | ||
692 | rb_link_node(&r->rb, parent, p); | 698 | rb_link_node(&r->rb, parent, p); |
693 | rb_insert_color(&r->rb, &c->replay_tree); | 699 | rb_insert_color(&r->rb, &c->replay_tree); |
@@ -712,7 +718,7 @@ static int replay_buds(struct ubifs_info *c) | |||
712 | if (err) | 718 | if (err) |
713 | return err; | 719 | return err; |
714 | err = insert_ref_node(c, b->bud->lnum, b->bud->start, b->sqnum, | 720 | err = insert_ref_node(c, b->bud->lnum, b->bud->start, b->sqnum, |
715 | free, dirty); | 721 | free, dirty, b->bud->jhead); |
716 | if (err) | 722 | if (err) |
717 | return err; | 723 | return err; |
718 | } | 724 | } |
diff --git a/fs/ubifs/super.c b/fs/ubifs/super.c index be6c7b008f38..04ad07f4fcc3 100644 --- a/fs/ubifs/super.c +++ b/fs/ubifs/super.c | |||
@@ -1257,12 +1257,12 @@ static int mount_ubifs(struct ubifs_info *c) | |||
1257 | goto out_free; | 1257 | goto out_free; |
1258 | } | 1258 | } |
1259 | 1259 | ||
1260 | err = alloc_wbufs(c); | ||
1261 | if (err) | ||
1262 | goto out_cbuf; | ||
1263 | |||
1260 | sprintf(c->bgt_name, BGT_NAME_PATTERN, c->vi.ubi_num, c->vi.vol_id); | 1264 | sprintf(c->bgt_name, BGT_NAME_PATTERN, c->vi.ubi_num, c->vi.vol_id); |
1261 | if (!c->ro_mount) { | 1265 | if (!c->ro_mount) { |
1262 | err = alloc_wbufs(c); | ||
1263 | if (err) | ||
1264 | goto out_cbuf; | ||
1265 | |||
1266 | /* Create background thread */ | 1266 | /* Create background thread */ |
1267 | c->bgt = kthread_create(ubifs_bg_thread, c, "%s", c->bgt_name); | 1267 | c->bgt = kthread_create(ubifs_bg_thread, c, "%s", c->bgt_name); |
1268 | if (IS_ERR(c->bgt)) { | 1268 | if (IS_ERR(c->bgt)) { |
@@ -1631,12 +1631,6 @@ static int ubifs_remount_rw(struct ubifs_info *c) | |||
1631 | if (err) | 1631 | if (err) |
1632 | goto out; | 1632 | goto out; |
1633 | 1633 | ||
1634 | err = alloc_wbufs(c); | ||
1635 | if (err) | ||
1636 | goto out; | ||
1637 | |||
1638 | ubifs_create_buds_lists(c); | ||
1639 | |||
1640 | /* Create background thread */ | 1634 | /* Create background thread */ |
1641 | c->bgt = kthread_create(ubifs_bg_thread, c, "%s", c->bgt_name); | 1635 | c->bgt = kthread_create(ubifs_bg_thread, c, "%s", c->bgt_name); |
1642 | if (IS_ERR(c->bgt)) { | 1636 | if (IS_ERR(c->bgt)) { |
@@ -1744,7 +1738,6 @@ static void ubifs_remount_ro(struct ubifs_info *c) | |||
1744 | if (err) | 1738 | if (err) |
1745 | ubifs_ro_mode(c, err); | 1739 | ubifs_ro_mode(c, err); |
1746 | 1740 | ||
1747 | free_wbufs(c); | ||
1748 | vfree(c->orph_buf); | 1741 | vfree(c->orph_buf); |
1749 | c->orph_buf = NULL; | 1742 | c->orph_buf = NULL; |
1750 | kfree(c->write_reserve_buf); | 1743 | kfree(c->write_reserve_buf); |
diff --git a/fs/xfs/linux-2.6/xfs_sync.c b/fs/xfs/linux-2.6/xfs_sync.c index e4f9c1b0836c..3e898a48122d 100644 --- a/fs/xfs/linux-2.6/xfs_sync.c +++ b/fs/xfs/linux-2.6/xfs_sync.c | |||
@@ -926,6 +926,7 @@ restart: | |||
926 | XFS_LOOKUP_BATCH, | 926 | XFS_LOOKUP_BATCH, |
927 | XFS_ICI_RECLAIM_TAG); | 927 | XFS_ICI_RECLAIM_TAG); |
928 | if (!nr_found) { | 928 | if (!nr_found) { |
929 | done = 1; | ||
929 | rcu_read_unlock(); | 930 | rcu_read_unlock(); |
930 | break; | 931 | break; |
931 | } | 932 | } |
diff --git a/fs/xfs/xfs_trans_ail.c b/fs/xfs/xfs_trans_ail.c index acdb92f14d51..5fc2380092c8 100644 --- a/fs/xfs/xfs_trans_ail.c +++ b/fs/xfs/xfs_trans_ail.c | |||
@@ -346,20 +346,23 @@ xfs_ail_delete( | |||
346 | */ | 346 | */ |
347 | STATIC void | 347 | STATIC void |
348 | xfs_ail_worker( | 348 | xfs_ail_worker( |
349 | struct work_struct *work) | 349 | struct work_struct *work) |
350 | { | 350 | { |
351 | struct xfs_ail *ailp = container_of(to_delayed_work(work), | 351 | struct xfs_ail *ailp = container_of(to_delayed_work(work), |
352 | struct xfs_ail, xa_work); | 352 | struct xfs_ail, xa_work); |
353 | long tout; | 353 | xfs_mount_t *mp = ailp->xa_mount; |
354 | xfs_lsn_t target = ailp->xa_target; | ||
355 | xfs_lsn_t lsn; | ||
356 | xfs_log_item_t *lip; | ||
357 | int flush_log, count, stuck; | ||
358 | xfs_mount_t *mp = ailp->xa_mount; | ||
359 | struct xfs_ail_cursor *cur = &ailp->xa_cursors; | 354 | struct xfs_ail_cursor *cur = &ailp->xa_cursors; |
360 | int push_xfsbufd = 0; | 355 | xfs_log_item_t *lip; |
356 | xfs_lsn_t lsn; | ||
357 | xfs_lsn_t target; | ||
358 | long tout = 10; | ||
359 | int flush_log = 0; | ||
360 | int stuck = 0; | ||
361 | int count = 0; | ||
362 | int push_xfsbufd = 0; | ||
361 | 363 | ||
362 | spin_lock(&ailp->xa_lock); | 364 | spin_lock(&ailp->xa_lock); |
365 | target = ailp->xa_target; | ||
363 | xfs_trans_ail_cursor_init(ailp, cur); | 366 | xfs_trans_ail_cursor_init(ailp, cur); |
364 | lip = xfs_trans_ail_cursor_first(ailp, cur, ailp->xa_last_pushed_lsn); | 367 | lip = xfs_trans_ail_cursor_first(ailp, cur, ailp->xa_last_pushed_lsn); |
365 | if (!lip || XFS_FORCED_SHUTDOWN(mp)) { | 368 | if (!lip || XFS_FORCED_SHUTDOWN(mp)) { |
@@ -368,8 +371,7 @@ xfs_ail_worker( | |||
368 | */ | 371 | */ |
369 | xfs_trans_ail_cursor_done(ailp, cur); | 372 | xfs_trans_ail_cursor_done(ailp, cur); |
370 | spin_unlock(&ailp->xa_lock); | 373 | spin_unlock(&ailp->xa_lock); |
371 | ailp->xa_last_pushed_lsn = 0; | 374 | goto out_done; |
372 | return; | ||
373 | } | 375 | } |
374 | 376 | ||
375 | XFS_STATS_INC(xs_push_ail); | 377 | XFS_STATS_INC(xs_push_ail); |
@@ -386,8 +388,7 @@ xfs_ail_worker( | |||
386 | * lots of contention on the AIL lists. | 388 | * lots of contention on the AIL lists. |
387 | */ | 389 | */ |
388 | lsn = lip->li_lsn; | 390 | lsn = lip->li_lsn; |
389 | flush_log = stuck = count = 0; | 391 | while ((XFS_LSN_CMP(lip->li_lsn, target) <= 0)) { |
390 | while ((XFS_LSN_CMP(lip->li_lsn, target) < 0)) { | ||
391 | int lock_result; | 392 | int lock_result; |
392 | /* | 393 | /* |
393 | * If we can lock the item without sleeping, unlock the AIL | 394 | * If we can lock the item without sleeping, unlock the AIL |
@@ -480,21 +481,25 @@ xfs_ail_worker( | |||
480 | } | 481 | } |
481 | 482 | ||
482 | /* assume we have more work to do in a short while */ | 483 | /* assume we have more work to do in a short while */ |
483 | tout = 10; | 484 | out_done: |
484 | if (!count) { | 485 | if (!count) { |
485 | /* We're past our target or empty, so idle */ | 486 | /* We're past our target or empty, so idle */ |
486 | ailp->xa_last_pushed_lsn = 0; | 487 | ailp->xa_last_pushed_lsn = 0; |
487 | 488 | ||
488 | /* | 489 | /* |
489 | * Check for an updated push target before clearing the | 490 | * We clear the XFS_AIL_PUSHING_BIT first before checking |
490 | * XFS_AIL_PUSHING_BIT. If the target changed, we've got more | 491 | * whether the target has changed. If the target has changed, |
491 | * work to do. Wait a bit longer before starting that work. | 492 | * this pushes the requeue race directly onto the result of the |
493 | * atomic test/set bit, so we are guaranteed that either the | ||
494 | * the pusher that changed the target or ourselves will requeue | ||
495 | * the work (but not both). | ||
492 | */ | 496 | */ |
497 | clear_bit(XFS_AIL_PUSHING_BIT, &ailp->xa_flags); | ||
493 | smp_rmb(); | 498 | smp_rmb(); |
494 | if (ailp->xa_target == target) { | 499 | if (XFS_LSN_CMP(ailp->xa_target, target) == 0 || |
495 | clear_bit(XFS_AIL_PUSHING_BIT, &ailp->xa_flags); | 500 | test_and_set_bit(XFS_AIL_PUSHING_BIT, &ailp->xa_flags)) |
496 | return; | 501 | return; |
497 | } | 502 | |
498 | tout = 50; | 503 | tout = 50; |
499 | } else if (XFS_LSN_CMP(lsn, target) >= 0) { | 504 | } else if (XFS_LSN_CMP(lsn, target) >= 0) { |
500 | /* | 505 | /* |
@@ -553,7 +558,7 @@ xfs_ail_push( | |||
553 | * the XFS_AIL_PUSHING_BIT. | 558 | * the XFS_AIL_PUSHING_BIT. |
554 | */ | 559 | */ |
555 | smp_wmb(); | 560 | smp_wmb(); |
556 | ailp->xa_target = threshold_lsn; | 561 | xfs_trans_ail_copy_lsn(ailp, &ailp->xa_target, &threshold_lsn); |
557 | if (!test_and_set_bit(XFS_AIL_PUSHING_BIT, &ailp->xa_flags)) | 562 | if (!test_and_set_bit(XFS_AIL_PUSHING_BIT, &ailp->xa_flags)) |
558 | queue_delayed_work(xfs_syncd_wq, &ailp->xa_work, 0); | 563 | queue_delayed_work(xfs_syncd_wq, &ailp->xa_work, 0); |
559 | } | 564 | } |