diff options
Diffstat (limited to 'fs/reiserfs/inode.c')
-rw-r--r-- | fs/reiserfs/inode.c | 4915 |
1 files changed, 2521 insertions, 2394 deletions
diff --git a/fs/reiserfs/inode.c b/fs/reiserfs/inode.c index 289d864fe731..1aaf2c7d44e6 100644 --- a/fs/reiserfs/inode.c +++ b/fs/reiserfs/inode.c | |||
@@ -18,107 +18,109 @@ | |||
18 | #include <linux/writeback.h> | 18 | #include <linux/writeback.h> |
19 | #include <linux/quotaops.h> | 19 | #include <linux/quotaops.h> |
20 | 20 | ||
21 | extern int reiserfs_default_io_size; /* default io size devuned in super.c */ | 21 | extern int reiserfs_default_io_size; /* default io size devuned in super.c */ |
22 | 22 | ||
23 | static int reiserfs_commit_write(struct file *f, struct page *page, | 23 | static int reiserfs_commit_write(struct file *f, struct page *page, |
24 | unsigned from, unsigned to); | 24 | unsigned from, unsigned to); |
25 | static int reiserfs_prepare_write(struct file *f, struct page *page, | 25 | static int reiserfs_prepare_write(struct file *f, struct page *page, |
26 | unsigned from, unsigned to); | 26 | unsigned from, unsigned to); |
27 | 27 | ||
28 | void reiserfs_delete_inode (struct inode * inode) | 28 | void reiserfs_delete_inode(struct inode *inode) |
29 | { | 29 | { |
30 | /* We need blocks for transaction + (user+group) quota update (possibly delete) */ | 30 | /* We need blocks for transaction + (user+group) quota update (possibly delete) */ |
31 | int jbegin_count = JOURNAL_PER_BALANCE_CNT * 2 + 2 * REISERFS_QUOTA_INIT_BLOCKS(inode->i_sb); | 31 | int jbegin_count = |
32 | struct reiserfs_transaction_handle th ; | 32 | JOURNAL_PER_BALANCE_CNT * 2 + |
33 | 33 | 2 * REISERFS_QUOTA_INIT_BLOCKS(inode->i_sb); | |
34 | reiserfs_write_lock(inode->i_sb); | 34 | struct reiserfs_transaction_handle th; |
35 | 35 | ||
36 | /* The = 0 happens when we abort creating a new inode for some reason like lack of space.. */ | 36 | reiserfs_write_lock(inode->i_sb); |
37 | if (!(inode->i_state & I_NEW) && INODE_PKEY(inode)->k_objectid != 0) { /* also handles bad_inode case */ | ||
38 | down (&inode->i_sem); | ||
39 | 37 | ||
40 | reiserfs_delete_xattrs (inode); | 38 | /* The = 0 happens when we abort creating a new inode for some reason like lack of space.. */ |
39 | if (!(inode->i_state & I_NEW) && INODE_PKEY(inode)->k_objectid != 0) { /* also handles bad_inode case */ | ||
40 | down(&inode->i_sem); | ||
41 | 41 | ||
42 | if (journal_begin(&th, inode->i_sb, jbegin_count)) { | 42 | reiserfs_delete_xattrs(inode); |
43 | up (&inode->i_sem); | ||
44 | goto out; | ||
45 | } | ||
46 | reiserfs_update_inode_transaction(inode) ; | ||
47 | 43 | ||
48 | if (reiserfs_delete_object (&th, inode)) { | 44 | if (journal_begin(&th, inode->i_sb, jbegin_count)) { |
49 | up (&inode->i_sem); | 45 | up(&inode->i_sem); |
50 | goto out; | 46 | goto out; |
51 | } | 47 | } |
48 | reiserfs_update_inode_transaction(inode); | ||
52 | 49 | ||
53 | /* Do quota update inside a transaction for journaled quotas. We must do that | 50 | if (reiserfs_delete_object(&th, inode)) { |
54 | * after delete_object so that quota updates go into the same transaction as | 51 | up(&inode->i_sem); |
55 | * stat data deletion */ | 52 | goto out; |
56 | DQUOT_FREE_INODE(inode); | 53 | } |
57 | 54 | ||
58 | if (journal_end(&th, inode->i_sb, jbegin_count)) { | 55 | /* Do quota update inside a transaction for journaled quotas. We must do that |
59 | up (&inode->i_sem); | 56 | * after delete_object so that quota updates go into the same transaction as |
60 | goto out; | 57 | * stat data deletion */ |
61 | } | 58 | DQUOT_FREE_INODE(inode); |
59 | |||
60 | if (journal_end(&th, inode->i_sb, jbegin_count)) { | ||
61 | up(&inode->i_sem); | ||
62 | goto out; | ||
63 | } | ||
62 | 64 | ||
63 | up (&inode->i_sem); | 65 | up(&inode->i_sem); |
64 | 66 | ||
65 | /* all items of file are deleted, so we can remove "save" link */ | 67 | /* all items of file are deleted, so we can remove "save" link */ |
66 | remove_save_link (inode, 0/* not truncate */); /* we can't do anything | 68 | remove_save_link(inode, 0 /* not truncate */ ); /* we can't do anything |
67 | * about an error here */ | 69 | * about an error here */ |
68 | } else { | 70 | } else { |
69 | /* no object items are in the tree */ | 71 | /* no object items are in the tree */ |
70 | ; | 72 | ; |
71 | } | 73 | } |
72 | out: | 74 | out: |
73 | clear_inode (inode); /* note this must go after the journal_end to prevent deadlock */ | 75 | clear_inode(inode); /* note this must go after the journal_end to prevent deadlock */ |
74 | inode->i_blocks = 0; | 76 | inode->i_blocks = 0; |
75 | reiserfs_write_unlock(inode->i_sb); | 77 | reiserfs_write_unlock(inode->i_sb); |
76 | } | 78 | } |
77 | 79 | ||
78 | static void _make_cpu_key (struct cpu_key * key, int version, __u32 dirid, __u32 objectid, | 80 | static void _make_cpu_key(struct cpu_key *key, int version, __u32 dirid, |
79 | loff_t offset, int type, int length ) | 81 | __u32 objectid, loff_t offset, int type, int length) |
80 | { | 82 | { |
81 | key->version = version; | 83 | key->version = version; |
82 | 84 | ||
83 | key->on_disk_key.k_dir_id = dirid; | 85 | key->on_disk_key.k_dir_id = dirid; |
84 | key->on_disk_key.k_objectid = objectid; | 86 | key->on_disk_key.k_objectid = objectid; |
85 | set_cpu_key_k_offset (key, offset); | 87 | set_cpu_key_k_offset(key, offset); |
86 | set_cpu_key_k_type (key, type); | 88 | set_cpu_key_k_type(key, type); |
87 | key->key_length = length; | 89 | key->key_length = length; |
88 | } | 90 | } |
89 | 91 | ||
90 | |||
91 | /* take base of inode_key (it comes from inode always) (dirid, objectid) and version from an inode, set | 92 | /* take base of inode_key (it comes from inode always) (dirid, objectid) and version from an inode, set |
92 | offset and type of key */ | 93 | offset and type of key */ |
93 | void make_cpu_key (struct cpu_key * key, struct inode * inode, loff_t offset, | 94 | void make_cpu_key(struct cpu_key *key, struct inode *inode, loff_t offset, |
94 | int type, int length ) | 95 | int type, int length) |
95 | { | 96 | { |
96 | _make_cpu_key (key, get_inode_item_key_version (inode), le32_to_cpu (INODE_PKEY (inode)->k_dir_id), | 97 | _make_cpu_key(key, get_inode_item_key_version(inode), |
97 | le32_to_cpu (INODE_PKEY (inode)->k_objectid), | 98 | le32_to_cpu(INODE_PKEY(inode)->k_dir_id), |
98 | offset, type, length); | 99 | le32_to_cpu(INODE_PKEY(inode)->k_objectid), offset, type, |
100 | length); | ||
99 | } | 101 | } |
100 | 102 | ||
101 | |||
102 | // | 103 | // |
103 | // when key is 0, do not set version and short key | 104 | // when key is 0, do not set version and short key |
104 | // | 105 | // |
105 | inline void make_le_item_head (struct item_head * ih, const struct cpu_key * key, | 106 | inline void make_le_item_head(struct item_head *ih, const struct cpu_key *key, |
106 | int version, | 107 | int version, |
107 | loff_t offset, int type, int length, | 108 | loff_t offset, int type, int length, |
108 | int entry_count/*or ih_free_space*/) | 109 | int entry_count /*or ih_free_space */ ) |
109 | { | 110 | { |
110 | if (key) { | 111 | if (key) { |
111 | ih->ih_key.k_dir_id = cpu_to_le32 (key->on_disk_key.k_dir_id); | 112 | ih->ih_key.k_dir_id = cpu_to_le32(key->on_disk_key.k_dir_id); |
112 | ih->ih_key.k_objectid = cpu_to_le32 (key->on_disk_key.k_objectid); | 113 | ih->ih_key.k_objectid = |
113 | } | 114 | cpu_to_le32(key->on_disk_key.k_objectid); |
114 | put_ih_version( ih, version ); | 115 | } |
115 | set_le_ih_k_offset (ih, offset); | 116 | put_ih_version(ih, version); |
116 | set_le_ih_k_type (ih, type); | 117 | set_le_ih_k_offset(ih, offset); |
117 | put_ih_item_len( ih, length ); | 118 | set_le_ih_k_type(ih, type); |
118 | /* set_ih_free_space (ih, 0);*/ | 119 | put_ih_item_len(ih, length); |
119 | // for directory items it is entry count, for directs and stat | 120 | /* set_ih_free_space (ih, 0); */ |
120 | // datas - 0xffff, for indirects - 0 | 121 | // for directory items it is entry count, for directs and stat |
121 | put_ih_entry_count( ih, entry_count ); | 122 | // datas - 0xffff, for indirects - 0 |
123 | put_ih_entry_count(ih, entry_count); | ||
122 | } | 124 | } |
123 | 125 | ||
124 | // | 126 | // |
@@ -153,84 +155,84 @@ inline void make_le_item_head (struct item_head * ih, const struct cpu_key * key | |||
153 | ** to be unmapped, so that block_prepare_write will correctly call | 155 | ** to be unmapped, so that block_prepare_write will correctly call |
154 | ** reiserfs_get_block to convert the tail into an unformatted node | 156 | ** reiserfs_get_block to convert the tail into an unformatted node |
155 | */ | 157 | */ |
156 | static inline void fix_tail_page_for_writing(struct page *page) { | 158 | static inline void fix_tail_page_for_writing(struct page *page) |
157 | struct buffer_head *head, *next, *bh ; | 159 | { |
158 | 160 | struct buffer_head *head, *next, *bh; | |
159 | if (page && page_has_buffers(page)) { | 161 | |
160 | head = page_buffers(page) ; | 162 | if (page && page_has_buffers(page)) { |
161 | bh = head ; | 163 | head = page_buffers(page); |
162 | do { | 164 | bh = head; |
163 | next = bh->b_this_page ; | 165 | do { |
164 | if (buffer_mapped(bh) && bh->b_blocknr == 0) { | 166 | next = bh->b_this_page; |
165 | reiserfs_unmap_buffer(bh) ; | 167 | if (buffer_mapped(bh) && bh->b_blocknr == 0) { |
166 | } | 168 | reiserfs_unmap_buffer(bh); |
167 | bh = next ; | 169 | } |
168 | } while (bh != head) ; | 170 | bh = next; |
169 | } | 171 | } while (bh != head); |
172 | } | ||
170 | } | 173 | } |
171 | 174 | ||
172 | /* reiserfs_get_block does not need to allocate a block only if it has been | 175 | /* reiserfs_get_block does not need to allocate a block only if it has been |
173 | done already or non-hole position has been found in the indirect item */ | 176 | done already or non-hole position has been found in the indirect item */ |
174 | static inline int allocation_needed (int retval, b_blocknr_t allocated, | 177 | static inline int allocation_needed(int retval, b_blocknr_t allocated, |
175 | struct item_head * ih, | 178 | struct item_head *ih, |
176 | __le32 * item, int pos_in_item) | 179 | __le32 * item, int pos_in_item) |
177 | { | 180 | { |
178 | if (allocated) | 181 | if (allocated) |
179 | return 0; | 182 | return 0; |
180 | if (retval == POSITION_FOUND && is_indirect_le_ih (ih) && | 183 | if (retval == POSITION_FOUND && is_indirect_le_ih(ih) && |
181 | get_block_num(item, pos_in_item)) | 184 | get_block_num(item, pos_in_item)) |
182 | return 0; | 185 | return 0; |
183 | return 1; | 186 | return 1; |
184 | } | 187 | } |
185 | 188 | ||
186 | static inline int indirect_item_found (int retval, struct item_head * ih) | 189 | static inline int indirect_item_found(int retval, struct item_head *ih) |
187 | { | 190 | { |
188 | return (retval == POSITION_FOUND) && is_indirect_le_ih (ih); | 191 | return (retval == POSITION_FOUND) && is_indirect_le_ih(ih); |
189 | } | 192 | } |
190 | 193 | ||
191 | 194 | static inline void set_block_dev_mapped(struct buffer_head *bh, | |
192 | static inline void set_block_dev_mapped (struct buffer_head * bh, | 195 | b_blocknr_t block, struct inode *inode) |
193 | b_blocknr_t block, struct inode * inode) | ||
194 | { | 196 | { |
195 | map_bh(bh, inode->i_sb, block); | 197 | map_bh(bh, inode->i_sb, block); |
196 | } | 198 | } |
197 | 199 | ||
198 | |||
199 | // | 200 | // |
200 | // files which were created in the earlier version can not be longer, | 201 | // files which were created in the earlier version can not be longer, |
201 | // than 2 gb | 202 | // than 2 gb |
202 | // | 203 | // |
203 | static int file_capable (struct inode * inode, long block) | 204 | static int file_capable(struct inode *inode, long block) |
204 | { | 205 | { |
205 | if (get_inode_item_key_version (inode) != KEY_FORMAT_3_5 || // it is new file. | 206 | if (get_inode_item_key_version(inode) != KEY_FORMAT_3_5 || // it is new file. |
206 | block < (1 << (31 - inode->i_sb->s_blocksize_bits))) // old file, but 'block' is inside of 2gb | 207 | block < (1 << (31 - inode->i_sb->s_blocksize_bits))) // old file, but 'block' is inside of 2gb |
207 | return 1; | 208 | return 1; |
208 | 209 | ||
209 | return 0; | 210 | return 0; |
210 | } | 211 | } |
211 | 212 | ||
212 | /*static*/ int restart_transaction(struct reiserfs_transaction_handle *th, | 213 | /*static*/ int restart_transaction(struct reiserfs_transaction_handle *th, |
213 | struct inode *inode, struct path *path) { | 214 | struct inode *inode, struct path *path) |
214 | struct super_block *s = th->t_super ; | 215 | { |
215 | int len = th->t_blocks_allocated ; | 216 | struct super_block *s = th->t_super; |
216 | int err; | 217 | int len = th->t_blocks_allocated; |
217 | 218 | int err; | |
218 | BUG_ON (!th->t_trans_id); | 219 | |
219 | BUG_ON (!th->t_refcount); | 220 | BUG_ON(!th->t_trans_id); |
220 | 221 | BUG_ON(!th->t_refcount); | |
221 | /* we cannot restart while nested */ | 222 | |
222 | if (th->t_refcount > 1) { | 223 | /* we cannot restart while nested */ |
223 | return 0 ; | 224 | if (th->t_refcount > 1) { |
224 | } | 225 | return 0; |
225 | pathrelse(path) ; | 226 | } |
226 | reiserfs_update_sd(th, inode) ; | 227 | pathrelse(path); |
227 | err = journal_end(th, s, len) ; | 228 | reiserfs_update_sd(th, inode); |
228 | if (!err) { | 229 | err = journal_end(th, s, len); |
229 | err = journal_begin(th, s, JOURNAL_PER_BALANCE_CNT * 6) ; | 230 | if (!err) { |
230 | if (!err) | 231 | err = journal_begin(th, s, JOURNAL_PER_BALANCE_CNT * 6); |
231 | reiserfs_update_inode_transaction(inode) ; | 232 | if (!err) |
232 | } | 233 | reiserfs_update_inode_transaction(inode); |
233 | return err; | 234 | } |
235 | return err; | ||
234 | } | 236 | } |
235 | 237 | ||
236 | // it is called by get_block when create == 0. Returns block number | 238 | // it is called by get_block when create == 0. Returns block number |
@@ -241,190 +243,192 @@ static int file_capable (struct inode * inode, long block) | |||
241 | // Please improve the english/clarity in the comment above, as it is | 243 | // Please improve the english/clarity in the comment above, as it is |
242 | // hard to understand. | 244 | // hard to understand. |
243 | 245 | ||
244 | static int _get_block_create_0 (struct inode * inode, long block, | 246 | static int _get_block_create_0(struct inode *inode, long block, |
245 | struct buffer_head * bh_result, | 247 | struct buffer_head *bh_result, int args) |
246 | int args) | ||
247 | { | 248 | { |
248 | INITIALIZE_PATH (path); | 249 | INITIALIZE_PATH(path); |
249 | struct cpu_key key; | 250 | struct cpu_key key; |
250 | struct buffer_head * bh; | 251 | struct buffer_head *bh; |
251 | struct item_head * ih, tmp_ih; | 252 | struct item_head *ih, tmp_ih; |
252 | int fs_gen ; | 253 | int fs_gen; |
253 | int blocknr; | 254 | int blocknr; |
254 | char * p = NULL; | 255 | char *p = NULL; |
255 | int chars; | 256 | int chars; |
256 | int ret ; | 257 | int ret; |
257 | int result ; | 258 | int result; |
258 | int done = 0 ; | 259 | int done = 0; |
259 | unsigned long offset ; | 260 | unsigned long offset; |
260 | 261 | ||
261 | // prepare the key to look for the 'block'-th block of file | 262 | // prepare the key to look for the 'block'-th block of file |
262 | make_cpu_key (&key, inode, | 263 | make_cpu_key(&key, inode, |
263 | (loff_t)block * inode->i_sb->s_blocksize + 1, TYPE_ANY, 3); | 264 | (loff_t) block * inode->i_sb->s_blocksize + 1, TYPE_ANY, |
264 | 265 | 3); | |
265 | research: | 266 | |
266 | result = search_for_position_by_key (inode->i_sb, &key, &path) ; | 267 | research: |
267 | if (result != POSITION_FOUND) { | 268 | result = search_for_position_by_key(inode->i_sb, &key, &path); |
268 | pathrelse (&path); | 269 | if (result != POSITION_FOUND) { |
269 | if (p) | 270 | pathrelse(&path); |
270 | kunmap(bh_result->b_page) ; | 271 | if (p) |
271 | if (result == IO_ERROR) | 272 | kunmap(bh_result->b_page); |
272 | return -EIO; | 273 | if (result == IO_ERROR) |
273 | // We do not return -ENOENT if there is a hole but page is uptodate, because it means | 274 | return -EIO; |
274 | // That there is some MMAPED data associated with it that is yet to be written to disk. | 275 | // We do not return -ENOENT if there is a hole but page is uptodate, because it means |
275 | if ((args & GET_BLOCK_NO_HOLE) && !PageUptodate(bh_result->b_page) ) { | 276 | // That there is some MMAPED data associated with it that is yet to be written to disk. |
276 | return -ENOENT ; | 277 | if ((args & GET_BLOCK_NO_HOLE) |
277 | } | 278 | && !PageUptodate(bh_result->b_page)) { |
278 | return 0 ; | 279 | return -ENOENT; |
279 | } | 280 | } |
280 | 281 | return 0; | |
281 | // | 282 | } |
282 | bh = get_last_bh (&path); | 283 | // |
283 | ih = get_ih (&path); | 284 | bh = get_last_bh(&path); |
284 | if (is_indirect_le_ih (ih)) { | 285 | ih = get_ih(&path); |
285 | __le32 * ind_item = (__le32 *)B_I_PITEM (bh, ih); | 286 | if (is_indirect_le_ih(ih)) { |
286 | 287 | __le32 *ind_item = (__le32 *) B_I_PITEM(bh, ih); | |
287 | /* FIXME: here we could cache indirect item or part of it in | 288 | |
288 | the inode to avoid search_by_key in case of subsequent | 289 | /* FIXME: here we could cache indirect item or part of it in |
289 | access to file */ | 290 | the inode to avoid search_by_key in case of subsequent |
290 | blocknr = get_block_num(ind_item, path.pos_in_item) ; | 291 | access to file */ |
291 | ret = 0 ; | 292 | blocknr = get_block_num(ind_item, path.pos_in_item); |
292 | if (blocknr) { | 293 | ret = 0; |
293 | map_bh(bh_result, inode->i_sb, blocknr); | 294 | if (blocknr) { |
294 | if (path.pos_in_item == ((ih_item_len(ih) / UNFM_P_SIZE) - 1)) { | 295 | map_bh(bh_result, inode->i_sb, blocknr); |
295 | set_buffer_boundary(bh_result); | 296 | if (path.pos_in_item == |
296 | } | 297 | ((ih_item_len(ih) / UNFM_P_SIZE) - 1)) { |
297 | } else | 298 | set_buffer_boundary(bh_result); |
298 | // We do not return -ENOENT if there is a hole but page is uptodate, because it means | 299 | } |
299 | // That there is some MMAPED data associated with it that is yet to be written to disk. | 300 | } else |
300 | if ((args & GET_BLOCK_NO_HOLE) && !PageUptodate(bh_result->b_page) ) { | 301 | // We do not return -ENOENT if there is a hole but page is uptodate, because it means |
301 | ret = -ENOENT ; | 302 | // That there is some MMAPED data associated with it that is yet to be written to disk. |
302 | } | 303 | if ((args & GET_BLOCK_NO_HOLE) |
303 | 304 | && !PageUptodate(bh_result->b_page)) { | |
304 | pathrelse (&path); | 305 | ret = -ENOENT; |
305 | if (p) | 306 | } |
306 | kunmap(bh_result->b_page) ; | 307 | |
307 | return ret ; | 308 | pathrelse(&path); |
308 | } | 309 | if (p) |
309 | 310 | kunmap(bh_result->b_page); | |
310 | // requested data are in direct item(s) | 311 | return ret; |
311 | if (!(args & GET_BLOCK_READ_DIRECT)) { | 312 | } |
312 | // we are called by bmap. FIXME: we can not map block of file | 313 | // requested data are in direct item(s) |
313 | // when it is stored in direct item(s) | 314 | if (!(args & GET_BLOCK_READ_DIRECT)) { |
314 | pathrelse (&path); | 315 | // we are called by bmap. FIXME: we can not map block of file |
315 | if (p) | 316 | // when it is stored in direct item(s) |
316 | kunmap(bh_result->b_page) ; | 317 | pathrelse(&path); |
317 | return -ENOENT; | 318 | if (p) |
318 | } | 319 | kunmap(bh_result->b_page); |
319 | 320 | return -ENOENT; | |
320 | /* if we've got a direct item, and the buffer or page was uptodate, | 321 | } |
321 | ** we don't want to pull data off disk again. skip to the | 322 | |
322 | ** end, where we map the buffer and return | 323 | /* if we've got a direct item, and the buffer or page was uptodate, |
323 | */ | 324 | ** we don't want to pull data off disk again. skip to the |
324 | if (buffer_uptodate(bh_result)) { | 325 | ** end, where we map the buffer and return |
325 | goto finished ; | 326 | */ |
326 | } else | 327 | if (buffer_uptodate(bh_result)) { |
327 | /* | 328 | goto finished; |
328 | ** grab_tail_page can trigger calls to reiserfs_get_block on up to date | 329 | } else |
329 | ** pages without any buffers. If the page is up to date, we don't want | 330 | /* |
330 | ** read old data off disk. Set the up to date bit on the buffer instead | 331 | ** grab_tail_page can trigger calls to reiserfs_get_block on up to date |
331 | ** and jump to the end | 332 | ** pages without any buffers. If the page is up to date, we don't want |
332 | */ | 333 | ** read old data off disk. Set the up to date bit on the buffer instead |
333 | if (!bh_result->b_page || PageUptodate(bh_result->b_page)) { | 334 | ** and jump to the end |
335 | */ | ||
336 | if (!bh_result->b_page || PageUptodate(bh_result->b_page)) { | ||
334 | set_buffer_uptodate(bh_result); | 337 | set_buffer_uptodate(bh_result); |
335 | goto finished ; | 338 | goto finished; |
336 | } | 339 | } |
337 | 340 | // read file tail into part of page | |
338 | // read file tail into part of page | 341 | offset = (cpu_key_k_offset(&key) - 1) & (PAGE_CACHE_SIZE - 1); |
339 | offset = (cpu_key_k_offset(&key) - 1) & (PAGE_CACHE_SIZE - 1) ; | 342 | fs_gen = get_generation(inode->i_sb); |
340 | fs_gen = get_generation(inode->i_sb) ; | 343 | copy_item_head(&tmp_ih, ih); |
341 | copy_item_head (&tmp_ih, ih); | 344 | |
342 | 345 | /* we only want to kmap if we are reading the tail into the page. | |
343 | /* we only want to kmap if we are reading the tail into the page. | 346 | ** this is not the common case, so we don't kmap until we are |
344 | ** this is not the common case, so we don't kmap until we are | 347 | ** sure we need to. But, this means the item might move if |
345 | ** sure we need to. But, this means the item might move if | 348 | ** kmap schedules |
346 | ** kmap schedules | 349 | */ |
347 | */ | 350 | if (!p) { |
348 | if (!p) { | 351 | p = (char *)kmap(bh_result->b_page); |
349 | p = (char *)kmap(bh_result->b_page) ; | 352 | if (fs_changed(fs_gen, inode->i_sb) |
350 | if (fs_changed (fs_gen, inode->i_sb) && item_moved (&tmp_ih, &path)) { | 353 | && item_moved(&tmp_ih, &path)) { |
351 | goto research; | 354 | goto research; |
352 | } | 355 | } |
353 | } | 356 | } |
354 | p += offset ; | 357 | p += offset; |
355 | memset (p, 0, inode->i_sb->s_blocksize); | 358 | memset(p, 0, inode->i_sb->s_blocksize); |
356 | do { | 359 | do { |
357 | if (!is_direct_le_ih (ih)) { | 360 | if (!is_direct_le_ih(ih)) { |
358 | BUG (); | 361 | BUG(); |
359 | } | 362 | } |
360 | /* make sure we don't read more bytes than actually exist in | 363 | /* make sure we don't read more bytes than actually exist in |
361 | ** the file. This can happen in odd cases where i_size isn't | 364 | ** the file. This can happen in odd cases where i_size isn't |
362 | ** correct, and when direct item padding results in a few | 365 | ** correct, and when direct item padding results in a few |
363 | ** extra bytes at the end of the direct item | 366 | ** extra bytes at the end of the direct item |
364 | */ | 367 | */ |
365 | if ((le_ih_k_offset(ih) + path.pos_in_item) > inode->i_size) | 368 | if ((le_ih_k_offset(ih) + path.pos_in_item) > inode->i_size) |
366 | break ; | 369 | break; |
367 | if ((le_ih_k_offset(ih) - 1 + ih_item_len(ih)) > inode->i_size) { | 370 | if ((le_ih_k_offset(ih) - 1 + ih_item_len(ih)) > inode->i_size) { |
368 | chars = inode->i_size - (le_ih_k_offset(ih) - 1) - path.pos_in_item; | 371 | chars = |
369 | done = 1 ; | 372 | inode->i_size - (le_ih_k_offset(ih) - 1) - |
370 | } else { | 373 | path.pos_in_item; |
371 | chars = ih_item_len(ih) - path.pos_in_item; | 374 | done = 1; |
372 | } | 375 | } else { |
373 | memcpy (p, B_I_PITEM (bh, ih) + path.pos_in_item, chars); | 376 | chars = ih_item_len(ih) - path.pos_in_item; |
374 | 377 | } | |
375 | if (done) | 378 | memcpy(p, B_I_PITEM(bh, ih) + path.pos_in_item, chars); |
376 | break ; | 379 | |
377 | 380 | if (done) | |
378 | p += chars; | 381 | break; |
379 | 382 | ||
380 | if (PATH_LAST_POSITION (&path) != (B_NR_ITEMS (bh) - 1)) | 383 | p += chars; |
381 | // we done, if read direct item is not the last item of | 384 | |
382 | // node FIXME: we could try to check right delimiting key | 385 | if (PATH_LAST_POSITION(&path) != (B_NR_ITEMS(bh) - 1)) |
383 | // to see whether direct item continues in the right | 386 | // we done, if read direct item is not the last item of |
384 | // neighbor or rely on i_size | 387 | // node FIXME: we could try to check right delimiting key |
385 | break; | 388 | // to see whether direct item continues in the right |
386 | 389 | // neighbor or rely on i_size | |
387 | // update key to look for the next piece | 390 | break; |
388 | set_cpu_key_k_offset (&key, cpu_key_k_offset (&key) + chars); | 391 | |
389 | result = search_for_position_by_key (inode->i_sb, &key, &path); | 392 | // update key to look for the next piece |
390 | if (result != POSITION_FOUND) | 393 | set_cpu_key_k_offset(&key, cpu_key_k_offset(&key) + chars); |
391 | // i/o error most likely | 394 | result = search_for_position_by_key(inode->i_sb, &key, &path); |
392 | break; | 395 | if (result != POSITION_FOUND) |
393 | bh = get_last_bh (&path); | 396 | // i/o error most likely |
394 | ih = get_ih (&path); | 397 | break; |
395 | } while (1); | 398 | bh = get_last_bh(&path); |
396 | 399 | ih = get_ih(&path); | |
397 | flush_dcache_page(bh_result->b_page) ; | 400 | } while (1); |
398 | kunmap(bh_result->b_page) ; | 401 | |
399 | 402 | flush_dcache_page(bh_result->b_page); | |
400 | finished: | 403 | kunmap(bh_result->b_page); |
401 | pathrelse (&path); | 404 | |
402 | 405 | finished: | |
403 | if (result == IO_ERROR) | 406 | pathrelse(&path); |
404 | return -EIO; | 407 | |
405 | 408 | if (result == IO_ERROR) | |
406 | /* this buffer has valid data, but isn't valid for io. mapping it to | 409 | return -EIO; |
407 | * block #0 tells the rest of reiserfs it just has a tail in it | ||
408 | */ | ||
409 | map_bh(bh_result, inode->i_sb, 0); | ||
410 | set_buffer_uptodate (bh_result); | ||
411 | return 0; | ||
412 | } | ||
413 | 410 | ||
411 | /* this buffer has valid data, but isn't valid for io. mapping it to | ||
412 | * block #0 tells the rest of reiserfs it just has a tail in it | ||
413 | */ | ||
414 | map_bh(bh_result, inode->i_sb, 0); | ||
415 | set_buffer_uptodate(bh_result); | ||
416 | return 0; | ||
417 | } | ||
414 | 418 | ||
415 | // this is called to create file map. So, _get_block_create_0 will not | 419 | // this is called to create file map. So, _get_block_create_0 will not |
416 | // read direct item | 420 | // read direct item |
417 | static int reiserfs_bmap (struct inode * inode, sector_t block, | 421 | static int reiserfs_bmap(struct inode *inode, sector_t block, |
418 | struct buffer_head * bh_result, int create) | 422 | struct buffer_head *bh_result, int create) |
419 | { | 423 | { |
420 | if (!file_capable (inode, block)) | 424 | if (!file_capable(inode, block)) |
421 | return -EFBIG; | 425 | return -EFBIG; |
422 | 426 | ||
423 | reiserfs_write_lock(inode->i_sb); | 427 | reiserfs_write_lock(inode->i_sb); |
424 | /* do not read the direct item */ | 428 | /* do not read the direct item */ |
425 | _get_block_create_0 (inode, block, bh_result, 0) ; | 429 | _get_block_create_0(inode, block, bh_result, 0); |
426 | reiserfs_write_unlock(inode->i_sb); | 430 | reiserfs_write_unlock(inode->i_sb); |
427 | return 0; | 431 | return 0; |
428 | } | 432 | } |
429 | 433 | ||
430 | /* special version of get_block that is only used by grab_tail_page right | 434 | /* special version of get_block that is only used by grab_tail_page right |
@@ -444,9 +448,11 @@ static int reiserfs_bmap (struct inode * inode, sector_t block, | |||
444 | ** don't want to send create == GET_BLOCK_NO_HOLE to reiserfs_get_block, | 448 | ** don't want to send create == GET_BLOCK_NO_HOLE to reiserfs_get_block, |
445 | ** don't use this function. | 449 | ** don't use this function. |
446 | */ | 450 | */ |
447 | static int reiserfs_get_block_create_0 (struct inode * inode, sector_t block, | 451 | static int reiserfs_get_block_create_0(struct inode *inode, sector_t block, |
448 | struct buffer_head * bh_result, int create) { | 452 | struct buffer_head *bh_result, |
449 | return reiserfs_get_block(inode, block, bh_result, GET_BLOCK_NO_HOLE) ; | 453 | int create) |
454 | { | ||
455 | return reiserfs_get_block(inode, block, bh_result, GET_BLOCK_NO_HOLE); | ||
450 | } | 456 | } |
451 | 457 | ||
452 | /* This is special helper for reiserfs_get_block in case we are executing | 458 | /* This is special helper for reiserfs_get_block in case we are executing |
@@ -457,43 +463,42 @@ static int reiserfs_get_blocks_direct_io(struct inode *inode, | |||
457 | struct buffer_head *bh_result, | 463 | struct buffer_head *bh_result, |
458 | int create) | 464 | int create) |
459 | { | 465 | { |
460 | int ret ; | 466 | int ret; |
461 | 467 | ||
462 | bh_result->b_page = NULL; | 468 | bh_result->b_page = NULL; |
463 | |||
464 | /* We set the b_size before reiserfs_get_block call since it is | ||
465 | referenced in convert_tail_for_hole() that may be called from | ||
466 | reiserfs_get_block() */ | ||
467 | bh_result->b_size = (1 << inode->i_blkbits); | ||
468 | |||
469 | ret = reiserfs_get_block(inode, iblock, bh_result, | ||
470 | create | GET_BLOCK_NO_DANGLE) ; | ||
471 | if (ret) | ||
472 | goto out; | ||
473 | |||
474 | /* don't allow direct io onto tail pages */ | ||
475 | if (buffer_mapped(bh_result) && bh_result->b_blocknr == 0) { | ||
476 | /* make sure future calls to the direct io funcs for this offset | ||
477 | ** in the file fail by unmapping the buffer | ||
478 | */ | ||
479 | clear_buffer_mapped(bh_result); | ||
480 | ret = -EINVAL ; | ||
481 | } | ||
482 | /* Possible unpacked tail. Flush the data before pages have | ||
483 | disappeared */ | ||
484 | if (REISERFS_I(inode)->i_flags & i_pack_on_close_mask) { | ||
485 | int err; | ||
486 | lock_kernel(); | ||
487 | err = reiserfs_commit_for_inode(inode); | ||
488 | REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask; | ||
489 | unlock_kernel(); | ||
490 | if (err < 0) | ||
491 | ret = err; | ||
492 | } | ||
493 | out: | ||
494 | return ret ; | ||
495 | } | ||
496 | 469 | ||
470 | /* We set the b_size before reiserfs_get_block call since it is | ||
471 | referenced in convert_tail_for_hole() that may be called from | ||
472 | reiserfs_get_block() */ | ||
473 | bh_result->b_size = (1 << inode->i_blkbits); | ||
474 | |||
475 | ret = reiserfs_get_block(inode, iblock, bh_result, | ||
476 | create | GET_BLOCK_NO_DANGLE); | ||
477 | if (ret) | ||
478 | goto out; | ||
479 | |||
480 | /* don't allow direct io onto tail pages */ | ||
481 | if (buffer_mapped(bh_result) && bh_result->b_blocknr == 0) { | ||
482 | /* make sure future calls to the direct io funcs for this offset | ||
483 | ** in the file fail by unmapping the buffer | ||
484 | */ | ||
485 | clear_buffer_mapped(bh_result); | ||
486 | ret = -EINVAL; | ||
487 | } | ||
488 | /* Possible unpacked tail. Flush the data before pages have | ||
489 | disappeared */ | ||
490 | if (REISERFS_I(inode)->i_flags & i_pack_on_close_mask) { | ||
491 | int err; | ||
492 | lock_kernel(); | ||
493 | err = reiserfs_commit_for_inode(inode); | ||
494 | REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask; | ||
495 | unlock_kernel(); | ||
496 | if (err < 0) | ||
497 | ret = err; | ||
498 | } | ||
499 | out: | ||
500 | return ret; | ||
501 | } | ||
497 | 502 | ||
498 | /* | 503 | /* |
499 | ** helper function for when reiserfs_get_block is called for a hole | 504 | ** helper function for when reiserfs_get_block is called for a hole |
@@ -505,490 +510,547 @@ out: | |||
505 | ** you should not be in a transaction, or have any paths held when you | 510 | ** you should not be in a transaction, or have any paths held when you |
506 | ** call this. | 511 | ** call this. |
507 | */ | 512 | */ |
508 | static int convert_tail_for_hole(struct inode *inode, | 513 | static int convert_tail_for_hole(struct inode *inode, |
509 | struct buffer_head *bh_result, | 514 | struct buffer_head *bh_result, |
510 | loff_t tail_offset) { | 515 | loff_t tail_offset) |
511 | unsigned long index ; | 516 | { |
512 | unsigned long tail_end ; | 517 | unsigned long index; |
513 | unsigned long tail_start ; | 518 | unsigned long tail_end; |
514 | struct page * tail_page ; | 519 | unsigned long tail_start; |
515 | struct page * hole_page = bh_result->b_page ; | 520 | struct page *tail_page; |
516 | int retval = 0 ; | 521 | struct page *hole_page = bh_result->b_page; |
517 | 522 | int retval = 0; | |
518 | if ((tail_offset & (bh_result->b_size - 1)) != 1) | 523 | |
519 | return -EIO ; | 524 | if ((tail_offset & (bh_result->b_size - 1)) != 1) |
520 | 525 | return -EIO; | |
521 | /* always try to read until the end of the block */ | 526 | |
522 | tail_start = tail_offset & (PAGE_CACHE_SIZE - 1) ; | 527 | /* always try to read until the end of the block */ |
523 | tail_end = (tail_start | (bh_result->b_size - 1)) + 1 ; | 528 | tail_start = tail_offset & (PAGE_CACHE_SIZE - 1); |
524 | 529 | tail_end = (tail_start | (bh_result->b_size - 1)) + 1; | |
525 | index = tail_offset >> PAGE_CACHE_SHIFT ; | 530 | |
526 | /* hole_page can be zero in case of direct_io, we are sure | 531 | index = tail_offset >> PAGE_CACHE_SHIFT; |
527 | that we cannot get here if we write with O_DIRECT into | 532 | /* hole_page can be zero in case of direct_io, we are sure |
528 | tail page */ | 533 | that we cannot get here if we write with O_DIRECT into |
529 | if (!hole_page || index != hole_page->index) { | 534 | tail page */ |
530 | tail_page = grab_cache_page(inode->i_mapping, index) ; | 535 | if (!hole_page || index != hole_page->index) { |
531 | retval = -ENOMEM; | 536 | tail_page = grab_cache_page(inode->i_mapping, index); |
532 | if (!tail_page) { | 537 | retval = -ENOMEM; |
533 | goto out ; | 538 | if (!tail_page) { |
534 | } | 539 | goto out; |
535 | } else { | 540 | } |
536 | tail_page = hole_page ; | 541 | } else { |
537 | } | 542 | tail_page = hole_page; |
538 | 543 | } | |
539 | /* we don't have to make sure the conversion did not happen while | 544 | |
540 | ** we were locking the page because anyone that could convert | 545 | /* we don't have to make sure the conversion did not happen while |
541 | ** must first take i_sem. | 546 | ** we were locking the page because anyone that could convert |
542 | ** | 547 | ** must first take i_sem. |
543 | ** We must fix the tail page for writing because it might have buffers | 548 | ** |
544 | ** that are mapped, but have a block number of 0. This indicates tail | 549 | ** We must fix the tail page for writing because it might have buffers |
545 | ** data that has been read directly into the page, and block_prepare_write | 550 | ** that are mapped, but have a block number of 0. This indicates tail |
546 | ** won't trigger a get_block in this case. | 551 | ** data that has been read directly into the page, and block_prepare_write |
547 | */ | 552 | ** won't trigger a get_block in this case. |
548 | fix_tail_page_for_writing(tail_page) ; | 553 | */ |
549 | retval = reiserfs_prepare_write(NULL, tail_page, tail_start, tail_end); | 554 | fix_tail_page_for_writing(tail_page); |
550 | if (retval) | 555 | retval = reiserfs_prepare_write(NULL, tail_page, tail_start, tail_end); |
551 | goto unlock ; | 556 | if (retval) |
552 | 557 | goto unlock; | |
553 | /* tail conversion might change the data in the page */ | 558 | |
554 | flush_dcache_page(tail_page) ; | 559 | /* tail conversion might change the data in the page */ |
555 | 560 | flush_dcache_page(tail_page); | |
556 | retval = reiserfs_commit_write(NULL, tail_page, tail_start, tail_end) ; | 561 | |
557 | 562 | retval = reiserfs_commit_write(NULL, tail_page, tail_start, tail_end); | |
558 | unlock: | 563 | |
559 | if (tail_page != hole_page) { | 564 | unlock: |
560 | unlock_page(tail_page) ; | 565 | if (tail_page != hole_page) { |
561 | page_cache_release(tail_page) ; | 566 | unlock_page(tail_page); |
562 | } | 567 | page_cache_release(tail_page); |
563 | out: | 568 | } |
564 | return retval ; | 569 | out: |
570 | return retval; | ||
565 | } | 571 | } |
566 | 572 | ||
567 | static inline int _allocate_block(struct reiserfs_transaction_handle *th, | 573 | static inline int _allocate_block(struct reiserfs_transaction_handle *th, |
568 | long block, | 574 | long block, |
569 | struct inode *inode, | 575 | struct inode *inode, |
570 | b_blocknr_t *allocated_block_nr, | 576 | b_blocknr_t * allocated_block_nr, |
571 | struct path * path, | 577 | struct path *path, int flags) |
572 | int flags) { | 578 | { |
573 | BUG_ON (!th->t_trans_id); | 579 | BUG_ON(!th->t_trans_id); |
574 | 580 | ||
575 | #ifdef REISERFS_PREALLOCATE | 581 | #ifdef REISERFS_PREALLOCATE |
576 | if (!(flags & GET_BLOCK_NO_ISEM)) { | 582 | if (!(flags & GET_BLOCK_NO_ISEM)) { |
577 | return reiserfs_new_unf_blocknrs2(th, inode, allocated_block_nr, path, block); | 583 | return reiserfs_new_unf_blocknrs2(th, inode, allocated_block_nr, |
578 | } | 584 | path, block); |
585 | } | ||
579 | #endif | 586 | #endif |
580 | return reiserfs_new_unf_blocknrs (th, inode, allocated_block_nr, path, block); | 587 | return reiserfs_new_unf_blocknrs(th, inode, allocated_block_nr, path, |
588 | block); | ||
581 | } | 589 | } |
582 | 590 | ||
583 | int reiserfs_get_block (struct inode * inode, sector_t block, | 591 | int reiserfs_get_block(struct inode *inode, sector_t block, |
584 | struct buffer_head * bh_result, int create) | 592 | struct buffer_head *bh_result, int create) |
585 | { | 593 | { |
586 | int repeat, retval = 0; | 594 | int repeat, retval = 0; |
587 | b_blocknr_t allocated_block_nr = 0;// b_blocknr_t is (unsigned) 32 bit int | 595 | b_blocknr_t allocated_block_nr = 0; // b_blocknr_t is (unsigned) 32 bit int |
588 | INITIALIZE_PATH(path); | 596 | INITIALIZE_PATH(path); |
589 | int pos_in_item; | 597 | int pos_in_item; |
590 | struct cpu_key key; | 598 | struct cpu_key key; |
591 | struct buffer_head * bh, * unbh = NULL; | 599 | struct buffer_head *bh, *unbh = NULL; |
592 | struct item_head * ih, tmp_ih; | 600 | struct item_head *ih, tmp_ih; |
593 | __le32 * item; | 601 | __le32 *item; |
594 | int done; | 602 | int done; |
595 | int fs_gen; | 603 | int fs_gen; |
596 | struct reiserfs_transaction_handle *th = NULL; | 604 | struct reiserfs_transaction_handle *th = NULL; |
597 | /* space reserved in transaction batch: | 605 | /* space reserved in transaction batch: |
598 | . 3 balancings in direct->indirect conversion | 606 | . 3 balancings in direct->indirect conversion |
599 | . 1 block involved into reiserfs_update_sd() | 607 | . 1 block involved into reiserfs_update_sd() |
600 | XXX in practically impossible worst case direct2indirect() | 608 | XXX in practically impossible worst case direct2indirect() |
601 | can incur (much) more than 3 balancings. | 609 | can incur (much) more than 3 balancings. |
602 | quota update for user, group */ | 610 | quota update for user, group */ |
603 | int jbegin_count = JOURNAL_PER_BALANCE_CNT * 3 + 1 + 2 * REISERFS_QUOTA_TRANS_BLOCKS(inode->i_sb); | 611 | int jbegin_count = |
604 | int version; | 612 | JOURNAL_PER_BALANCE_CNT * 3 + 1 + |
605 | int dangle = 1; | 613 | 2 * REISERFS_QUOTA_TRANS_BLOCKS(inode->i_sb); |
606 | loff_t new_offset = (((loff_t)block) << inode->i_sb->s_blocksize_bits) + 1 ; | 614 | int version; |
607 | 615 | int dangle = 1; | |
608 | /* bad.... */ | 616 | loff_t new_offset = |
609 | reiserfs_write_lock(inode->i_sb); | 617 | (((loff_t) block) << inode->i_sb->s_blocksize_bits) + 1; |
610 | version = get_inode_item_key_version (inode); | 618 | |
611 | 619 | /* bad.... */ | |
612 | if (block < 0) { | 620 | reiserfs_write_lock(inode->i_sb); |
613 | reiserfs_write_unlock(inode->i_sb); | 621 | version = get_inode_item_key_version(inode); |
614 | return -EIO; | ||
615 | } | ||
616 | 622 | ||
617 | if (!file_capable (inode, block)) { | 623 | if (block < 0) { |
618 | reiserfs_write_unlock(inode->i_sb); | 624 | reiserfs_write_unlock(inode->i_sb); |
619 | return -EFBIG; | 625 | return -EIO; |
620 | } | 626 | } |
621 | |||
622 | /* if !create, we aren't changing the FS, so we don't need to | ||
623 | ** log anything, so we don't need to start a transaction | ||
624 | */ | ||
625 | if (!(create & GET_BLOCK_CREATE)) { | ||
626 | int ret ; | ||
627 | /* find number of block-th logical block of the file */ | ||
628 | ret = _get_block_create_0 (inode, block, bh_result, | ||
629 | create | GET_BLOCK_READ_DIRECT) ; | ||
630 | reiserfs_write_unlock(inode->i_sb); | ||
631 | return ret; | ||
632 | } | ||
633 | /* | ||
634 | * if we're already in a transaction, make sure to close | ||
635 | * any new transactions we start in this func | ||
636 | */ | ||
637 | if ((create & GET_BLOCK_NO_DANGLE) || | ||
638 | reiserfs_transaction_running(inode->i_sb)) | ||
639 | dangle = 0; | ||
640 | |||
641 | /* If file is of such a size, that it might have a tail and tails are enabled | ||
642 | ** we should mark it as possibly needing tail packing on close | ||
643 | */ | ||
644 | if ( (have_large_tails (inode->i_sb) && inode->i_size < i_block_size (inode)*4) || | ||
645 | (have_small_tails (inode->i_sb) && inode->i_size < i_block_size(inode)) ) | ||
646 | REISERFS_I(inode)->i_flags |= i_pack_on_close_mask ; | ||
647 | |||
648 | /* set the key of the first byte in the 'block'-th block of file */ | ||
649 | make_cpu_key (&key, inode, new_offset, | ||
650 | TYPE_ANY, 3/*key length*/); | ||
651 | if ((new_offset + inode->i_sb->s_blocksize - 1) > inode->i_size) { | ||
652 | start_trans: | ||
653 | th = reiserfs_persistent_transaction(inode->i_sb, jbegin_count); | ||
654 | if (!th) { | ||
655 | retval = -ENOMEM; | ||
656 | goto failure; | ||
657 | } | ||
658 | reiserfs_update_inode_transaction(inode) ; | ||
659 | } | ||
660 | research: | ||
661 | |||
662 | retval = search_for_position_by_key (inode->i_sb, &key, &path); | ||
663 | if (retval == IO_ERROR) { | ||
664 | retval = -EIO; | ||
665 | goto failure; | ||
666 | } | ||
667 | |||
668 | bh = get_last_bh (&path); | ||
669 | ih = get_ih (&path); | ||
670 | item = get_item (&path); | ||
671 | pos_in_item = path.pos_in_item; | ||
672 | |||
673 | fs_gen = get_generation (inode->i_sb); | ||
674 | copy_item_head (&tmp_ih, ih); | ||
675 | |||
676 | if (allocation_needed (retval, allocated_block_nr, ih, item, pos_in_item)) { | ||
677 | /* we have to allocate block for the unformatted node */ | ||
678 | if (!th) { | ||
679 | pathrelse(&path) ; | ||
680 | goto start_trans; | ||
681 | } | ||
682 | |||
683 | repeat = _allocate_block(th, block, inode, &allocated_block_nr, &path, create); | ||
684 | |||
685 | if (repeat == NO_DISK_SPACE || repeat == QUOTA_EXCEEDED) { | ||
686 | /* restart the transaction to give the journal a chance to free | ||
687 | ** some blocks. releases the path, so we have to go back to | ||
688 | ** research if we succeed on the second try | ||
689 | */ | ||
690 | SB_JOURNAL(inode->i_sb)->j_next_async_flush = 1; | ||
691 | retval = restart_transaction(th, inode, &path) ; | ||
692 | if (retval) | ||
693 | goto failure; | ||
694 | repeat = _allocate_block(th, block, inode, &allocated_block_nr, NULL, create); | ||
695 | |||
696 | if (repeat != NO_DISK_SPACE && repeat != QUOTA_EXCEEDED) { | ||
697 | goto research ; | ||
698 | } | ||
699 | if (repeat == QUOTA_EXCEEDED) | ||
700 | retval = -EDQUOT; | ||
701 | else | ||
702 | retval = -ENOSPC; | ||
703 | goto failure; | ||
704 | } | ||
705 | |||
706 | if (fs_changed (fs_gen, inode->i_sb) && item_moved (&tmp_ih, &path)) { | ||
707 | goto research; | ||
708 | } | ||
709 | } | ||
710 | |||
711 | if (indirect_item_found (retval, ih)) { | ||
712 | b_blocknr_t unfm_ptr; | ||
713 | /* 'block'-th block is in the file already (there is | ||
714 | corresponding cell in some indirect item). But it may be | ||
715 | zero unformatted node pointer (hole) */ | ||
716 | unfm_ptr = get_block_num (item, pos_in_item); | ||
717 | if (unfm_ptr == 0) { | ||
718 | /* use allocated block to plug the hole */ | ||
719 | reiserfs_prepare_for_journal(inode->i_sb, bh, 1) ; | ||
720 | if (fs_changed (fs_gen, inode->i_sb) && item_moved (&tmp_ih, &path)) { | ||
721 | reiserfs_restore_prepared_buffer(inode->i_sb, bh) ; | ||
722 | goto research; | ||
723 | } | ||
724 | set_buffer_new(bh_result); | ||
725 | if (buffer_dirty(bh_result) && reiserfs_data_ordered(inode->i_sb)) | ||
726 | reiserfs_add_ordered_list(inode, bh_result); | ||
727 | put_block_num(item, pos_in_item, allocated_block_nr) ; | ||
728 | unfm_ptr = allocated_block_nr; | ||
729 | journal_mark_dirty (th, inode->i_sb, bh); | ||
730 | reiserfs_update_sd(th, inode) ; | ||
731 | } | ||
732 | set_block_dev_mapped(bh_result, unfm_ptr, inode); | ||
733 | pathrelse (&path); | ||
734 | retval = 0; | ||
735 | if (!dangle && th) | ||
736 | retval = reiserfs_end_persistent_transaction(th); | ||
737 | 627 | ||
738 | reiserfs_write_unlock(inode->i_sb); | 628 | if (!file_capable(inode, block)) { |
739 | 629 | reiserfs_write_unlock(inode->i_sb); | |
740 | /* the item was found, so new blocks were not added to the file | 630 | return -EFBIG; |
741 | ** there is no need to make sure the inode is updated with this | 631 | } |
742 | ** transaction | 632 | |
743 | */ | 633 | /* if !create, we aren't changing the FS, so we don't need to |
744 | return retval; | 634 | ** log anything, so we don't need to start a transaction |
745 | } | 635 | */ |
746 | 636 | if (!(create & GET_BLOCK_CREATE)) { | |
747 | if (!th) { | 637 | int ret; |
748 | pathrelse(&path) ; | 638 | /* find number of block-th logical block of the file */ |
749 | goto start_trans; | 639 | ret = _get_block_create_0(inode, block, bh_result, |
750 | } | 640 | create | GET_BLOCK_READ_DIRECT); |
751 | 641 | reiserfs_write_unlock(inode->i_sb); | |
752 | /* desired position is not found or is in the direct item. We have | 642 | return ret; |
753 | to append file with holes up to 'block'-th block converting | 643 | } |
754 | direct items to indirect one if necessary */ | 644 | /* |
755 | done = 0; | 645 | * if we're already in a transaction, make sure to close |
756 | do { | 646 | * any new transactions we start in this func |
757 | if (is_statdata_le_ih (ih)) { | 647 | */ |
758 | __le32 unp = 0; | 648 | if ((create & GET_BLOCK_NO_DANGLE) || |
759 | struct cpu_key tmp_key; | 649 | reiserfs_transaction_running(inode->i_sb)) |
760 | 650 | dangle = 0; | |
761 | /* indirect item has to be inserted */ | 651 | |
762 | make_le_item_head (&tmp_ih, &key, version, 1, TYPE_INDIRECT, | 652 | /* If file is of such a size, that it might have a tail and tails are enabled |
763 | UNFM_P_SIZE, 0/* free_space */); | 653 | ** we should mark it as possibly needing tail packing on close |
764 | 654 | */ | |
765 | if (cpu_key_k_offset (&key) == 1) { | 655 | if ((have_large_tails(inode->i_sb) |
766 | /* we are going to add 'block'-th block to the file. Use | 656 | && inode->i_size < i_block_size(inode) * 4) |
767 | allocated block for that */ | 657 | || (have_small_tails(inode->i_sb) |
768 | unp = cpu_to_le32 (allocated_block_nr); | 658 | && inode->i_size < i_block_size(inode))) |
769 | set_block_dev_mapped (bh_result, allocated_block_nr, inode); | 659 | REISERFS_I(inode)->i_flags |= i_pack_on_close_mask; |
770 | set_buffer_new(bh_result); | 660 | |
771 | done = 1; | 661 | /* set the key of the first byte in the 'block'-th block of file */ |
772 | } | 662 | make_cpu_key(&key, inode, new_offset, TYPE_ANY, 3 /*key length */ ); |
773 | tmp_key = key; // ;) | 663 | if ((new_offset + inode->i_sb->s_blocksize - 1) > inode->i_size) { |
774 | set_cpu_key_k_offset (&tmp_key, 1); | 664 | start_trans: |
775 | PATH_LAST_POSITION(&path) ++; | 665 | th = reiserfs_persistent_transaction(inode->i_sb, jbegin_count); |
776 | 666 | if (!th) { | |
777 | retval = reiserfs_insert_item (th, &path, &tmp_key, &tmp_ih, inode, (char *)&unp); | 667 | retval = -ENOMEM; |
778 | if (retval) { | ||
779 | reiserfs_free_block (th, inode, allocated_block_nr, 1); | ||
780 | goto failure; // retval == -ENOSPC, -EDQUOT or -EIO or -EEXIST | ||
781 | } | ||
782 | //mark_tail_converted (inode); | ||
783 | } else if (is_direct_le_ih (ih)) { | ||
784 | /* direct item has to be converted */ | ||
785 | loff_t tail_offset; | ||
786 | |||
787 | tail_offset = ((le_ih_k_offset (ih) - 1) & ~(inode->i_sb->s_blocksize - 1)) + 1; | ||
788 | if (tail_offset == cpu_key_k_offset (&key)) { | ||
789 | /* direct item we just found fits into block we have | ||
790 | to map. Convert it into unformatted node: use | ||
791 | bh_result for the conversion */ | ||
792 | set_block_dev_mapped (bh_result, allocated_block_nr, inode); | ||
793 | unbh = bh_result; | ||
794 | done = 1; | ||
795 | } else { | ||
796 | /* we have to padd file tail stored in direct item(s) | ||
797 | up to block size and convert it to unformatted | ||
798 | node. FIXME: this should also get into page cache */ | ||
799 | |||
800 | pathrelse(&path) ; | ||
801 | /* | ||
802 | * ugly, but we can only end the transaction if | ||
803 | * we aren't nested | ||
804 | */ | ||
805 | BUG_ON (!th->t_refcount); | ||
806 | if (th->t_refcount == 1) { | ||
807 | retval = reiserfs_end_persistent_transaction(th); | ||
808 | th = NULL; | ||
809 | if (retval) | ||
810 | goto failure; | 668 | goto failure; |
811 | } | 669 | } |
670 | reiserfs_update_inode_transaction(inode); | ||
671 | } | ||
672 | research: | ||
812 | 673 | ||
813 | retval = convert_tail_for_hole(inode, bh_result, tail_offset) ; | 674 | retval = search_for_position_by_key(inode->i_sb, &key, &path); |
814 | if (retval) { | ||
815 | if ( retval != -ENOSPC ) | ||
816 | reiserfs_warning (inode->i_sb, "clm-6004: convert tail failed inode %lu, error %d", inode->i_ino, retval) ; | ||
817 | if (allocated_block_nr) { | ||
818 | /* the bitmap, the super, and the stat data == 3 */ | ||
819 | if (!th) | ||
820 | th = reiserfs_persistent_transaction(inode->i_sb,3); | ||
821 | if (th) | ||
822 | reiserfs_free_block (th,inode,allocated_block_nr,1); | ||
823 | } | ||
824 | goto failure ; | ||
825 | } | ||
826 | goto research ; | ||
827 | } | ||
828 | retval = direct2indirect (th, inode, &path, unbh, tail_offset); | ||
829 | if (retval) { | ||
830 | reiserfs_unmap_buffer(unbh); | ||
831 | reiserfs_free_block (th, inode, allocated_block_nr, 1); | ||
832 | goto failure; | ||
833 | } | ||
834 | /* it is important the set_buffer_uptodate is done after | ||
835 | ** the direct2indirect. The buffer might contain valid | ||
836 | ** data newer than the data on disk (read by readpage, changed, | ||
837 | ** and then sent here by writepage). direct2indirect needs | ||
838 | ** to know if unbh was already up to date, so it can decide | ||
839 | ** if the data in unbh needs to be replaced with data from | ||
840 | ** the disk | ||
841 | */ | ||
842 | set_buffer_uptodate (unbh); | ||
843 | |||
844 | /* unbh->b_page == NULL in case of DIRECT_IO request, this means | ||
845 | buffer will disappear shortly, so it should not be added to | ||
846 | */ | ||
847 | if ( unbh->b_page ) { | ||
848 | /* we've converted the tail, so we must | ||
849 | ** flush unbh before the transaction commits | ||
850 | */ | ||
851 | reiserfs_add_tail_list(inode, unbh) ; | ||
852 | |||
853 | /* mark it dirty now to prevent commit_write from adding | ||
854 | ** this buffer to the inode's dirty buffer list | ||
855 | */ | ||
856 | /* | ||
857 | * AKPM: changed __mark_buffer_dirty to mark_buffer_dirty(). | ||
858 | * It's still atomic, but it sets the page dirty too, | ||
859 | * which makes it eligible for writeback at any time by the | ||
860 | * VM (which was also the case with __mark_buffer_dirty()) | ||
861 | */ | ||
862 | mark_buffer_dirty(unbh) ; | ||
863 | } | ||
864 | } else { | ||
865 | /* append indirect item with holes if needed, when appending | ||
866 | pointer to 'block'-th block use block, which is already | ||
867 | allocated */ | ||
868 | struct cpu_key tmp_key; | ||
869 | unp_t unf_single=0; // We use this in case we need to allocate only | ||
870 | // one block which is a fastpath | ||
871 | unp_t *un; | ||
872 | __u64 max_to_insert=MAX_ITEM_LEN(inode->i_sb->s_blocksize)/UNFM_P_SIZE; | ||
873 | __u64 blocks_needed; | ||
874 | |||
875 | RFALSE( pos_in_item != ih_item_len(ih) / UNFM_P_SIZE, | ||
876 | "vs-804: invalid position for append"); | ||
877 | /* indirect item has to be appended, set up key of that position */ | ||
878 | make_cpu_key (&tmp_key, inode, | ||
879 | le_key_k_offset (version, &(ih->ih_key)) + op_bytes_number (ih, inode->i_sb->s_blocksize), | ||
880 | //pos_in_item * inode->i_sb->s_blocksize, | ||
881 | TYPE_INDIRECT, 3);// key type is unimportant | ||
882 | |||
883 | blocks_needed = 1 + ((cpu_key_k_offset (&key) - cpu_key_k_offset (&tmp_key)) >> inode->i_sb->s_blocksize_bits); | ||
884 | RFALSE( blocks_needed < 0, "green-805: invalid offset"); | ||
885 | |||
886 | if ( blocks_needed == 1 ) { | ||
887 | un = &unf_single; | ||
888 | } else { | ||
889 | un=kmalloc( min(blocks_needed,max_to_insert)*UNFM_P_SIZE, | ||
890 | GFP_ATOMIC); // We need to avoid scheduling. | ||
891 | if ( !un) { | ||
892 | un = &unf_single; | ||
893 | blocks_needed = 1; | ||
894 | max_to_insert = 0; | ||
895 | } else | ||
896 | memset(un, 0, UNFM_P_SIZE * min(blocks_needed,max_to_insert)); | ||
897 | } | ||
898 | if ( blocks_needed <= max_to_insert) { | ||
899 | /* we are going to add target block to the file. Use allocated | ||
900 | block for that */ | ||
901 | un[blocks_needed-1] = cpu_to_le32 (allocated_block_nr); | ||
902 | set_block_dev_mapped (bh_result, allocated_block_nr, inode); | ||
903 | set_buffer_new(bh_result); | ||
904 | done = 1; | ||
905 | } else { | ||
906 | /* paste hole to the indirect item */ | ||
907 | /* If kmalloc failed, max_to_insert becomes zero and it means we | ||
908 | only have space for one block */ | ||
909 | blocks_needed=max_to_insert?max_to_insert:1; | ||
910 | } | ||
911 | retval = reiserfs_paste_into_item (th, &path, &tmp_key, inode, (char *)un, UNFM_P_SIZE * blocks_needed); | ||
912 | |||
913 | if (blocks_needed != 1) | ||
914 | kfree(un); | ||
915 | |||
916 | if (retval) { | ||
917 | reiserfs_free_block (th, inode, allocated_block_nr, 1); | ||
918 | goto failure; | ||
919 | } | ||
920 | if (!done) { | ||
921 | /* We need to mark new file size in case this function will be | ||
922 | interrupted/aborted later on. And we may do this only for | ||
923 | holes. */ | ||
924 | inode->i_size += inode->i_sb->s_blocksize * blocks_needed; | ||
925 | } | ||
926 | } | ||
927 | |||
928 | if (done == 1) | ||
929 | break; | ||
930 | |||
931 | /* this loop could log more blocks than we had originally asked | ||
932 | ** for. So, we have to allow the transaction to end if it is | ||
933 | ** too big or too full. Update the inode so things are | ||
934 | ** consistent if we crash before the function returns | ||
935 | ** | ||
936 | ** release the path so that anybody waiting on the path before | ||
937 | ** ending their transaction will be able to continue. | ||
938 | */ | ||
939 | if (journal_transaction_should_end(th, th->t_blocks_allocated)) { | ||
940 | retval = restart_transaction(th, inode, &path) ; | ||
941 | if (retval) | ||
942 | goto failure; | ||
943 | } | ||
944 | /* inserting indirect pointers for a hole can take a | ||
945 | ** long time. reschedule if needed | ||
946 | */ | ||
947 | cond_resched(); | ||
948 | |||
949 | retval = search_for_position_by_key (inode->i_sb, &key, &path); | ||
950 | if (retval == IO_ERROR) { | 675 | if (retval == IO_ERROR) { |
951 | retval = -EIO; | 676 | retval = -EIO; |
952 | goto failure; | 677 | goto failure; |
953 | } | 678 | } |
954 | if (retval == POSITION_FOUND) { | 679 | |
955 | reiserfs_warning (inode->i_sb, "vs-825: reiserfs_get_block: " | 680 | bh = get_last_bh(&path); |
956 | "%K should not be found", &key); | 681 | ih = get_ih(&path); |
957 | retval = -EEXIST; | 682 | item = get_item(&path); |
958 | if (allocated_block_nr) | ||
959 | reiserfs_free_block (th, inode, allocated_block_nr, 1); | ||
960 | pathrelse(&path) ; | ||
961 | goto failure; | ||
962 | } | ||
963 | bh = get_last_bh (&path); | ||
964 | ih = get_ih (&path); | ||
965 | item = get_item (&path); | ||
966 | pos_in_item = path.pos_in_item; | 683 | pos_in_item = path.pos_in_item; |
967 | } while (1); | ||
968 | 684 | ||
685 | fs_gen = get_generation(inode->i_sb); | ||
686 | copy_item_head(&tmp_ih, ih); | ||
687 | |||
688 | if (allocation_needed | ||
689 | (retval, allocated_block_nr, ih, item, pos_in_item)) { | ||
690 | /* we have to allocate block for the unformatted node */ | ||
691 | if (!th) { | ||
692 | pathrelse(&path); | ||
693 | goto start_trans; | ||
694 | } | ||
695 | |||
696 | repeat = | ||
697 | _allocate_block(th, block, inode, &allocated_block_nr, | ||
698 | &path, create); | ||
699 | |||
700 | if (repeat == NO_DISK_SPACE || repeat == QUOTA_EXCEEDED) { | ||
701 | /* restart the transaction to give the journal a chance to free | ||
702 | ** some blocks. releases the path, so we have to go back to | ||
703 | ** research if we succeed on the second try | ||
704 | */ | ||
705 | SB_JOURNAL(inode->i_sb)->j_next_async_flush = 1; | ||
706 | retval = restart_transaction(th, inode, &path); | ||
707 | if (retval) | ||
708 | goto failure; | ||
709 | repeat = | ||
710 | _allocate_block(th, block, inode, | ||
711 | &allocated_block_nr, NULL, create); | ||
712 | |||
713 | if (repeat != NO_DISK_SPACE && repeat != QUOTA_EXCEEDED) { | ||
714 | goto research; | ||
715 | } | ||
716 | if (repeat == QUOTA_EXCEEDED) | ||
717 | retval = -EDQUOT; | ||
718 | else | ||
719 | retval = -ENOSPC; | ||
720 | goto failure; | ||
721 | } | ||
722 | |||
723 | if (fs_changed(fs_gen, inode->i_sb) | ||
724 | && item_moved(&tmp_ih, &path)) { | ||
725 | goto research; | ||
726 | } | ||
727 | } | ||
728 | |||
729 | if (indirect_item_found(retval, ih)) { | ||
730 | b_blocknr_t unfm_ptr; | ||
731 | /* 'block'-th block is in the file already (there is | ||
732 | corresponding cell in some indirect item). But it may be | ||
733 | zero unformatted node pointer (hole) */ | ||
734 | unfm_ptr = get_block_num(item, pos_in_item); | ||
735 | if (unfm_ptr == 0) { | ||
736 | /* use allocated block to plug the hole */ | ||
737 | reiserfs_prepare_for_journal(inode->i_sb, bh, 1); | ||
738 | if (fs_changed(fs_gen, inode->i_sb) | ||
739 | && item_moved(&tmp_ih, &path)) { | ||
740 | reiserfs_restore_prepared_buffer(inode->i_sb, | ||
741 | bh); | ||
742 | goto research; | ||
743 | } | ||
744 | set_buffer_new(bh_result); | ||
745 | if (buffer_dirty(bh_result) | ||
746 | && reiserfs_data_ordered(inode->i_sb)) | ||
747 | reiserfs_add_ordered_list(inode, bh_result); | ||
748 | put_block_num(item, pos_in_item, allocated_block_nr); | ||
749 | unfm_ptr = allocated_block_nr; | ||
750 | journal_mark_dirty(th, inode->i_sb, bh); | ||
751 | reiserfs_update_sd(th, inode); | ||
752 | } | ||
753 | set_block_dev_mapped(bh_result, unfm_ptr, inode); | ||
754 | pathrelse(&path); | ||
755 | retval = 0; | ||
756 | if (!dangle && th) | ||
757 | retval = reiserfs_end_persistent_transaction(th); | ||
758 | |||
759 | reiserfs_write_unlock(inode->i_sb); | ||
760 | |||
761 | /* the item was found, so new blocks were not added to the file | ||
762 | ** there is no need to make sure the inode is updated with this | ||
763 | ** transaction | ||
764 | */ | ||
765 | return retval; | ||
766 | } | ||
767 | |||
768 | if (!th) { | ||
769 | pathrelse(&path); | ||
770 | goto start_trans; | ||
771 | } | ||
772 | |||
773 | /* desired position is not found or is in the direct item. We have | ||
774 | to append file with holes up to 'block'-th block converting | ||
775 | direct items to indirect one if necessary */ | ||
776 | done = 0; | ||
777 | do { | ||
778 | if (is_statdata_le_ih(ih)) { | ||
779 | __le32 unp = 0; | ||
780 | struct cpu_key tmp_key; | ||
781 | |||
782 | /* indirect item has to be inserted */ | ||
783 | make_le_item_head(&tmp_ih, &key, version, 1, | ||
784 | TYPE_INDIRECT, UNFM_P_SIZE, | ||
785 | 0 /* free_space */ ); | ||
786 | |||
787 | if (cpu_key_k_offset(&key) == 1) { | ||
788 | /* we are going to add 'block'-th block to the file. Use | ||
789 | allocated block for that */ | ||
790 | unp = cpu_to_le32(allocated_block_nr); | ||
791 | set_block_dev_mapped(bh_result, | ||
792 | allocated_block_nr, inode); | ||
793 | set_buffer_new(bh_result); | ||
794 | done = 1; | ||
795 | } | ||
796 | tmp_key = key; // ;) | ||
797 | set_cpu_key_k_offset(&tmp_key, 1); | ||
798 | PATH_LAST_POSITION(&path)++; | ||
799 | |||
800 | retval = | ||
801 | reiserfs_insert_item(th, &path, &tmp_key, &tmp_ih, | ||
802 | inode, (char *)&unp); | ||
803 | if (retval) { | ||
804 | reiserfs_free_block(th, inode, | ||
805 | allocated_block_nr, 1); | ||
806 | goto failure; // retval == -ENOSPC, -EDQUOT or -EIO or -EEXIST | ||
807 | } | ||
808 | //mark_tail_converted (inode); | ||
809 | } else if (is_direct_le_ih(ih)) { | ||
810 | /* direct item has to be converted */ | ||
811 | loff_t tail_offset; | ||
812 | |||
813 | tail_offset = | ||
814 | ((le_ih_k_offset(ih) - | ||
815 | 1) & ~(inode->i_sb->s_blocksize - 1)) + 1; | ||
816 | if (tail_offset == cpu_key_k_offset(&key)) { | ||
817 | /* direct item we just found fits into block we have | ||
818 | to map. Convert it into unformatted node: use | ||
819 | bh_result for the conversion */ | ||
820 | set_block_dev_mapped(bh_result, | ||
821 | allocated_block_nr, inode); | ||
822 | unbh = bh_result; | ||
823 | done = 1; | ||
824 | } else { | ||
825 | /* we have to padd file tail stored in direct item(s) | ||
826 | up to block size and convert it to unformatted | ||
827 | node. FIXME: this should also get into page cache */ | ||
828 | |||
829 | pathrelse(&path); | ||
830 | /* | ||
831 | * ugly, but we can only end the transaction if | ||
832 | * we aren't nested | ||
833 | */ | ||
834 | BUG_ON(!th->t_refcount); | ||
835 | if (th->t_refcount == 1) { | ||
836 | retval = | ||
837 | reiserfs_end_persistent_transaction | ||
838 | (th); | ||
839 | th = NULL; | ||
840 | if (retval) | ||
841 | goto failure; | ||
842 | } | ||
843 | |||
844 | retval = | ||
845 | convert_tail_for_hole(inode, bh_result, | ||
846 | tail_offset); | ||
847 | if (retval) { | ||
848 | if (retval != -ENOSPC) | ||
849 | reiserfs_warning(inode->i_sb, | ||
850 | "clm-6004: convert tail failed inode %lu, error %d", | ||
851 | inode->i_ino, | ||
852 | retval); | ||
853 | if (allocated_block_nr) { | ||
854 | /* the bitmap, the super, and the stat data == 3 */ | ||
855 | if (!th) | ||
856 | th = reiserfs_persistent_transaction(inode->i_sb, 3); | ||
857 | if (th) | ||
858 | reiserfs_free_block(th, | ||
859 | inode, | ||
860 | allocated_block_nr, | ||
861 | 1); | ||
862 | } | ||
863 | goto failure; | ||
864 | } | ||
865 | goto research; | ||
866 | } | ||
867 | retval = | ||
868 | direct2indirect(th, inode, &path, unbh, | ||
869 | tail_offset); | ||
870 | if (retval) { | ||
871 | reiserfs_unmap_buffer(unbh); | ||
872 | reiserfs_free_block(th, inode, | ||
873 | allocated_block_nr, 1); | ||
874 | goto failure; | ||
875 | } | ||
876 | /* it is important the set_buffer_uptodate is done after | ||
877 | ** the direct2indirect. The buffer might contain valid | ||
878 | ** data newer than the data on disk (read by readpage, changed, | ||
879 | ** and then sent here by writepage). direct2indirect needs | ||
880 | ** to know if unbh was already up to date, so it can decide | ||
881 | ** if the data in unbh needs to be replaced with data from | ||
882 | ** the disk | ||
883 | */ | ||
884 | set_buffer_uptodate(unbh); | ||
885 | |||
886 | /* unbh->b_page == NULL in case of DIRECT_IO request, this means | ||
887 | buffer will disappear shortly, so it should not be added to | ||
888 | */ | ||
889 | if (unbh->b_page) { | ||
890 | /* we've converted the tail, so we must | ||
891 | ** flush unbh before the transaction commits | ||
892 | */ | ||
893 | reiserfs_add_tail_list(inode, unbh); | ||
894 | |||
895 | /* mark it dirty now to prevent commit_write from adding | ||
896 | ** this buffer to the inode's dirty buffer list | ||
897 | */ | ||
898 | /* | ||
899 | * AKPM: changed __mark_buffer_dirty to mark_buffer_dirty(). | ||
900 | * It's still atomic, but it sets the page dirty too, | ||
901 | * which makes it eligible for writeback at any time by the | ||
902 | * VM (which was also the case with __mark_buffer_dirty()) | ||
903 | */ | ||
904 | mark_buffer_dirty(unbh); | ||
905 | } | ||
906 | } else { | ||
907 | /* append indirect item with holes if needed, when appending | ||
908 | pointer to 'block'-th block use block, which is already | ||
909 | allocated */ | ||
910 | struct cpu_key tmp_key; | ||
911 | unp_t unf_single = 0; // We use this in case we need to allocate only | ||
912 | // one block which is a fastpath | ||
913 | unp_t *un; | ||
914 | __u64 max_to_insert = | ||
915 | MAX_ITEM_LEN(inode->i_sb->s_blocksize) / | ||
916 | UNFM_P_SIZE; | ||
917 | __u64 blocks_needed; | ||
918 | |||
919 | RFALSE(pos_in_item != ih_item_len(ih) / UNFM_P_SIZE, | ||
920 | "vs-804: invalid position for append"); | ||
921 | /* indirect item has to be appended, set up key of that position */ | ||
922 | make_cpu_key(&tmp_key, inode, | ||
923 | le_key_k_offset(version, | ||
924 | &(ih->ih_key)) + | ||
925 | op_bytes_number(ih, | ||
926 | inode->i_sb->s_blocksize), | ||
927 | //pos_in_item * inode->i_sb->s_blocksize, | ||
928 | TYPE_INDIRECT, 3); // key type is unimportant | ||
929 | |||
930 | blocks_needed = | ||
931 | 1 + | ||
932 | ((cpu_key_k_offset(&key) - | ||
933 | cpu_key_k_offset(&tmp_key)) >> inode->i_sb-> | ||
934 | s_blocksize_bits); | ||
935 | RFALSE(blocks_needed < 0, "green-805: invalid offset"); | ||
936 | |||
937 | if (blocks_needed == 1) { | ||
938 | un = &unf_single; | ||
939 | } else { | ||
940 | un = kmalloc(min(blocks_needed, max_to_insert) * UNFM_P_SIZE, GFP_ATOMIC); // We need to avoid scheduling. | ||
941 | if (!un) { | ||
942 | un = &unf_single; | ||
943 | blocks_needed = 1; | ||
944 | max_to_insert = 0; | ||
945 | } else | ||
946 | memset(un, 0, | ||
947 | UNFM_P_SIZE * min(blocks_needed, | ||
948 | max_to_insert)); | ||
949 | } | ||
950 | if (blocks_needed <= max_to_insert) { | ||
951 | /* we are going to add target block to the file. Use allocated | ||
952 | block for that */ | ||
953 | un[blocks_needed - 1] = | ||
954 | cpu_to_le32(allocated_block_nr); | ||
955 | set_block_dev_mapped(bh_result, | ||
956 | allocated_block_nr, inode); | ||
957 | set_buffer_new(bh_result); | ||
958 | done = 1; | ||
959 | } else { | ||
960 | /* paste hole to the indirect item */ | ||
961 | /* If kmalloc failed, max_to_insert becomes zero and it means we | ||
962 | only have space for one block */ | ||
963 | blocks_needed = | ||
964 | max_to_insert ? max_to_insert : 1; | ||
965 | } | ||
966 | retval = | ||
967 | reiserfs_paste_into_item(th, &path, &tmp_key, inode, | ||
968 | (char *)un, | ||
969 | UNFM_P_SIZE * | ||
970 | blocks_needed); | ||
971 | |||
972 | if (blocks_needed != 1) | ||
973 | kfree(un); | ||
974 | |||
975 | if (retval) { | ||
976 | reiserfs_free_block(th, inode, | ||
977 | allocated_block_nr, 1); | ||
978 | goto failure; | ||
979 | } | ||
980 | if (!done) { | ||
981 | /* We need to mark new file size in case this function will be | ||
982 | interrupted/aborted later on. And we may do this only for | ||
983 | holes. */ | ||
984 | inode->i_size += | ||
985 | inode->i_sb->s_blocksize * blocks_needed; | ||
986 | } | ||
987 | } | ||
969 | 988 | ||
970 | retval = 0; | 989 | if (done == 1) |
990 | break; | ||
971 | 991 | ||
972 | failure: | 992 | /* this loop could log more blocks than we had originally asked |
973 | if (th && (!dangle || (retval && !th->t_trans_id))) { | 993 | ** for. So, we have to allow the transaction to end if it is |
974 | int err; | 994 | ** too big or too full. Update the inode so things are |
975 | if (th->t_trans_id) | 995 | ** consistent if we crash before the function returns |
976 | reiserfs_update_sd(th, inode); | 996 | ** |
977 | err = reiserfs_end_persistent_transaction(th); | 997 | ** release the path so that anybody waiting on the path before |
978 | if (err) | 998 | ** ending their transaction will be able to continue. |
979 | retval = err; | 999 | */ |
980 | } | 1000 | if (journal_transaction_should_end(th, th->t_blocks_allocated)) { |
1001 | retval = restart_transaction(th, inode, &path); | ||
1002 | if (retval) | ||
1003 | goto failure; | ||
1004 | } | ||
1005 | /* inserting indirect pointers for a hole can take a | ||
1006 | ** long time. reschedule if needed | ||
1007 | */ | ||
1008 | cond_resched(); | ||
981 | 1009 | ||
982 | reiserfs_write_unlock(inode->i_sb); | 1010 | retval = search_for_position_by_key(inode->i_sb, &key, &path); |
983 | reiserfs_check_path(&path) ; | 1011 | if (retval == IO_ERROR) { |
984 | return retval; | 1012 | retval = -EIO; |
1013 | goto failure; | ||
1014 | } | ||
1015 | if (retval == POSITION_FOUND) { | ||
1016 | reiserfs_warning(inode->i_sb, | ||
1017 | "vs-825: reiserfs_get_block: " | ||
1018 | "%K should not be found", &key); | ||
1019 | retval = -EEXIST; | ||
1020 | if (allocated_block_nr) | ||
1021 | reiserfs_free_block(th, inode, | ||
1022 | allocated_block_nr, 1); | ||
1023 | pathrelse(&path); | ||
1024 | goto failure; | ||
1025 | } | ||
1026 | bh = get_last_bh(&path); | ||
1027 | ih = get_ih(&path); | ||
1028 | item = get_item(&path); | ||
1029 | pos_in_item = path.pos_in_item; | ||
1030 | } while (1); | ||
1031 | |||
1032 | retval = 0; | ||
1033 | |||
1034 | failure: | ||
1035 | if (th && (!dangle || (retval && !th->t_trans_id))) { | ||
1036 | int err; | ||
1037 | if (th->t_trans_id) | ||
1038 | reiserfs_update_sd(th, inode); | ||
1039 | err = reiserfs_end_persistent_transaction(th); | ||
1040 | if (err) | ||
1041 | retval = err; | ||
1042 | } | ||
1043 | |||
1044 | reiserfs_write_unlock(inode->i_sb); | ||
1045 | reiserfs_check_path(&path); | ||
1046 | return retval; | ||
985 | } | 1047 | } |
986 | 1048 | ||
987 | static int | 1049 | static int |
988 | reiserfs_readpages(struct file *file, struct address_space *mapping, | 1050 | reiserfs_readpages(struct file *file, struct address_space *mapping, |
989 | struct list_head *pages, unsigned nr_pages) | 1051 | struct list_head *pages, unsigned nr_pages) |
990 | { | 1052 | { |
991 | return mpage_readpages(mapping, pages, nr_pages, reiserfs_get_block); | 1053 | return mpage_readpages(mapping, pages, nr_pages, reiserfs_get_block); |
992 | } | 1054 | } |
993 | 1055 | ||
994 | /* Compute real number of used bytes by file | 1056 | /* Compute real number of used bytes by file |
@@ -996,51 +1058,56 @@ reiserfs_readpages(struct file *file, struct address_space *mapping, | |||
996 | */ | 1058 | */ |
997 | static int real_space_diff(struct inode *inode, int sd_size) | 1059 | static int real_space_diff(struct inode *inode, int sd_size) |
998 | { | 1060 | { |
999 | int bytes; | 1061 | int bytes; |
1000 | loff_t blocksize = inode->i_sb->s_blocksize ; | 1062 | loff_t blocksize = inode->i_sb->s_blocksize; |
1001 | 1063 | ||
1002 | if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) | 1064 | if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) |
1003 | return sd_size ; | 1065 | return sd_size; |
1004 | 1066 | ||
1005 | /* End of file is also in full block with indirect reference, so round | 1067 | /* End of file is also in full block with indirect reference, so round |
1006 | ** up to the next block. | 1068 | ** up to the next block. |
1007 | ** | 1069 | ** |
1008 | ** there is just no way to know if the tail is actually packed | 1070 | ** there is just no way to know if the tail is actually packed |
1009 | ** on the file, so we have to assume it isn't. When we pack the | 1071 | ** on the file, so we have to assume it isn't. When we pack the |
1010 | ** tail, we add 4 bytes to pretend there really is an unformatted | 1072 | ** tail, we add 4 bytes to pretend there really is an unformatted |
1011 | ** node pointer | 1073 | ** node pointer |
1012 | */ | 1074 | */ |
1013 | bytes = ((inode->i_size + (blocksize-1)) >> inode->i_sb->s_blocksize_bits) * UNFM_P_SIZE + sd_size; | 1075 | bytes = |
1014 | return bytes ; | 1076 | ((inode->i_size + |
1077 | (blocksize - 1)) >> inode->i_sb->s_blocksize_bits) * UNFM_P_SIZE + | ||
1078 | sd_size; | ||
1079 | return bytes; | ||
1015 | } | 1080 | } |
1016 | 1081 | ||
1017 | static inline loff_t to_real_used_space(struct inode *inode, ulong blocks, | 1082 | static inline loff_t to_real_used_space(struct inode *inode, ulong blocks, |
1018 | int sd_size) | 1083 | int sd_size) |
1019 | { | 1084 | { |
1020 | if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) { | 1085 | if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) { |
1021 | return inode->i_size + (loff_t)(real_space_diff(inode, sd_size)) ; | 1086 | return inode->i_size + |
1022 | } | 1087 | (loff_t) (real_space_diff(inode, sd_size)); |
1023 | return ((loff_t)real_space_diff(inode, sd_size)) + (((loff_t)blocks) << 9); | 1088 | } |
1089 | return ((loff_t) real_space_diff(inode, sd_size)) + | ||
1090 | (((loff_t) blocks) << 9); | ||
1024 | } | 1091 | } |
1025 | 1092 | ||
1026 | /* Compute number of blocks used by file in ReiserFS counting */ | 1093 | /* Compute number of blocks used by file in ReiserFS counting */ |
1027 | static inline ulong to_fake_used_blocks(struct inode *inode, int sd_size) | 1094 | static inline ulong to_fake_used_blocks(struct inode *inode, int sd_size) |
1028 | { | 1095 | { |
1029 | loff_t bytes = inode_get_bytes(inode) ; | 1096 | loff_t bytes = inode_get_bytes(inode); |
1030 | loff_t real_space = real_space_diff(inode, sd_size) ; | 1097 | loff_t real_space = real_space_diff(inode, sd_size); |
1031 | 1098 | ||
1032 | /* keeps fsck and non-quota versions of reiserfs happy */ | 1099 | /* keeps fsck and non-quota versions of reiserfs happy */ |
1033 | if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) { | 1100 | if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) { |
1034 | bytes += (loff_t)511 ; | 1101 | bytes += (loff_t) 511; |
1035 | } | 1102 | } |
1036 | 1103 | ||
1037 | /* files from before the quota patch might i_blocks such that | 1104 | /* files from before the quota patch might i_blocks such that |
1038 | ** bytes < real_space. Deal with that here to prevent it from | 1105 | ** bytes < real_space. Deal with that here to prevent it from |
1039 | ** going negative. | 1106 | ** going negative. |
1040 | */ | 1107 | */ |
1041 | if (bytes < real_space) | 1108 | if (bytes < real_space) |
1042 | return 0 ; | 1109 | return 0; |
1043 | return (bytes - real_space) >> 9; | 1110 | return (bytes - real_space) >> 9; |
1044 | } | 1111 | } |
1045 | 1112 | ||
1046 | // | 1113 | // |
@@ -1051,263 +1118,269 @@ static inline ulong to_fake_used_blocks(struct inode *inode, int sd_size) | |||
1051 | // | 1118 | // |
1052 | 1119 | ||
1053 | // called by read_locked_inode | 1120 | // called by read_locked_inode |
1054 | static void init_inode (struct inode * inode, struct path * path) | 1121 | static void init_inode(struct inode *inode, struct path *path) |
1055 | { | 1122 | { |
1056 | struct buffer_head * bh; | 1123 | struct buffer_head *bh; |
1057 | struct item_head * ih; | 1124 | struct item_head *ih; |
1058 | __u32 rdev; | 1125 | __u32 rdev; |
1059 | //int version = ITEM_VERSION_1; | 1126 | //int version = ITEM_VERSION_1; |
1060 | 1127 | ||
1061 | bh = PATH_PLAST_BUFFER (path); | 1128 | bh = PATH_PLAST_BUFFER(path); |
1062 | ih = PATH_PITEM_HEAD (path); | 1129 | ih = PATH_PITEM_HEAD(path); |
1063 | 1130 | ||
1064 | 1131 | copy_key(INODE_PKEY(inode), &(ih->ih_key)); | |
1065 | copy_key (INODE_PKEY (inode), &(ih->ih_key)); | 1132 | inode->i_blksize = reiserfs_default_io_size; |
1066 | inode->i_blksize = reiserfs_default_io_size; | 1133 | |
1067 | 1134 | INIT_LIST_HEAD(&(REISERFS_I(inode)->i_prealloc_list)); | |
1068 | INIT_LIST_HEAD(&(REISERFS_I(inode)->i_prealloc_list )); | 1135 | REISERFS_I(inode)->i_flags = 0; |
1069 | REISERFS_I(inode)->i_flags = 0; | 1136 | REISERFS_I(inode)->i_prealloc_block = 0; |
1070 | REISERFS_I(inode)->i_prealloc_block = 0; | 1137 | REISERFS_I(inode)->i_prealloc_count = 0; |
1071 | REISERFS_I(inode)->i_prealloc_count = 0; | 1138 | REISERFS_I(inode)->i_trans_id = 0; |
1072 | REISERFS_I(inode)->i_trans_id = 0; | 1139 | REISERFS_I(inode)->i_jl = NULL; |
1073 | REISERFS_I(inode)->i_jl = NULL; | 1140 | REISERFS_I(inode)->i_acl_access = NULL; |
1074 | REISERFS_I(inode)->i_acl_access = NULL; | 1141 | REISERFS_I(inode)->i_acl_default = NULL; |
1075 | REISERFS_I(inode)->i_acl_default = NULL; | 1142 | init_rwsem(&REISERFS_I(inode)->xattr_sem); |
1076 | init_rwsem (&REISERFS_I(inode)->xattr_sem); | 1143 | |
1077 | 1144 | if (stat_data_v1(ih)) { | |
1078 | if (stat_data_v1 (ih)) { | 1145 | struct stat_data_v1 *sd = |
1079 | struct stat_data_v1 * sd = (struct stat_data_v1 *)B_I_PITEM (bh, ih); | 1146 | (struct stat_data_v1 *)B_I_PITEM(bh, ih); |
1080 | unsigned long blocks; | 1147 | unsigned long blocks; |
1081 | 1148 | ||
1082 | set_inode_item_key_version (inode, KEY_FORMAT_3_5); | 1149 | set_inode_item_key_version(inode, KEY_FORMAT_3_5); |
1083 | set_inode_sd_version (inode, STAT_DATA_V1); | 1150 | set_inode_sd_version(inode, STAT_DATA_V1); |
1084 | inode->i_mode = sd_v1_mode(sd); | 1151 | inode->i_mode = sd_v1_mode(sd); |
1085 | inode->i_nlink = sd_v1_nlink(sd); | 1152 | inode->i_nlink = sd_v1_nlink(sd); |
1086 | inode->i_uid = sd_v1_uid(sd); | 1153 | inode->i_uid = sd_v1_uid(sd); |
1087 | inode->i_gid = sd_v1_gid(sd); | 1154 | inode->i_gid = sd_v1_gid(sd); |
1088 | inode->i_size = sd_v1_size(sd); | 1155 | inode->i_size = sd_v1_size(sd); |
1089 | inode->i_atime.tv_sec = sd_v1_atime(sd); | 1156 | inode->i_atime.tv_sec = sd_v1_atime(sd); |
1090 | inode->i_mtime.tv_sec = sd_v1_mtime(sd); | 1157 | inode->i_mtime.tv_sec = sd_v1_mtime(sd); |
1091 | inode->i_ctime.tv_sec = sd_v1_ctime(sd); | 1158 | inode->i_ctime.tv_sec = sd_v1_ctime(sd); |
1092 | inode->i_atime.tv_nsec = 0; | 1159 | inode->i_atime.tv_nsec = 0; |
1093 | inode->i_ctime.tv_nsec = 0; | 1160 | inode->i_ctime.tv_nsec = 0; |
1094 | inode->i_mtime.tv_nsec = 0; | 1161 | inode->i_mtime.tv_nsec = 0; |
1095 | 1162 | ||
1096 | inode->i_blocks = sd_v1_blocks(sd); | 1163 | inode->i_blocks = sd_v1_blocks(sd); |
1097 | inode->i_generation = le32_to_cpu (INODE_PKEY (inode)->k_dir_id); | 1164 | inode->i_generation = le32_to_cpu(INODE_PKEY(inode)->k_dir_id); |
1098 | blocks = (inode->i_size + 511) >> 9; | 1165 | blocks = (inode->i_size + 511) >> 9; |
1099 | blocks = _ROUND_UP (blocks, inode->i_sb->s_blocksize >> 9); | 1166 | blocks = _ROUND_UP(blocks, inode->i_sb->s_blocksize >> 9); |
1100 | if (inode->i_blocks > blocks) { | 1167 | if (inode->i_blocks > blocks) { |
1101 | // there was a bug in <=3.5.23 when i_blocks could take negative | 1168 | // there was a bug in <=3.5.23 when i_blocks could take negative |
1102 | // values. Starting from 3.5.17 this value could even be stored in | 1169 | // values. Starting from 3.5.17 this value could even be stored in |
1103 | // stat data. For such files we set i_blocks based on file | 1170 | // stat data. For such files we set i_blocks based on file |
1104 | // size. Just 2 notes: this can be wrong for sparce files. On-disk value will be | 1171 | // size. Just 2 notes: this can be wrong for sparce files. On-disk value will be |
1105 | // only updated if file's inode will ever change | 1172 | // only updated if file's inode will ever change |
1106 | inode->i_blocks = blocks; | 1173 | inode->i_blocks = blocks; |
1107 | } | 1174 | } |
1108 | |||
1109 | rdev = sd_v1_rdev(sd); | ||
1110 | REISERFS_I(inode)->i_first_direct_byte = sd_v1_first_direct_byte(sd); | ||
1111 | /* an early bug in the quota code can give us an odd number for the | ||
1112 | ** block count. This is incorrect, fix it here. | ||
1113 | */ | ||
1114 | if (inode->i_blocks & 1) { | ||
1115 | inode->i_blocks++ ; | ||
1116 | } | ||
1117 | inode_set_bytes(inode, to_real_used_space(inode, inode->i_blocks, | ||
1118 | SD_V1_SIZE)); | ||
1119 | /* nopack is initially zero for v1 objects. For v2 objects, | ||
1120 | nopack is initialised from sd_attrs */ | ||
1121 | REISERFS_I(inode)->i_flags &= ~i_nopack_mask; | ||
1122 | } else { | ||
1123 | // new stat data found, but object may have old items | ||
1124 | // (directories and symlinks) | ||
1125 | struct stat_data * sd = (struct stat_data *)B_I_PITEM (bh, ih); | ||
1126 | |||
1127 | inode->i_mode = sd_v2_mode(sd); | ||
1128 | inode->i_nlink = sd_v2_nlink(sd); | ||
1129 | inode->i_uid = sd_v2_uid(sd); | ||
1130 | inode->i_size = sd_v2_size(sd); | ||
1131 | inode->i_gid = sd_v2_gid(sd); | ||
1132 | inode->i_mtime.tv_sec = sd_v2_mtime(sd); | ||
1133 | inode->i_atime.tv_sec = sd_v2_atime(sd); | ||
1134 | inode->i_ctime.tv_sec = sd_v2_ctime(sd); | ||
1135 | inode->i_ctime.tv_nsec = 0; | ||
1136 | inode->i_mtime.tv_nsec = 0; | ||
1137 | inode->i_atime.tv_nsec = 0; | ||
1138 | inode->i_blocks = sd_v2_blocks(sd); | ||
1139 | rdev = sd_v2_rdev(sd); | ||
1140 | if( S_ISCHR( inode -> i_mode ) || S_ISBLK( inode -> i_mode ) ) | ||
1141 | inode->i_generation = le32_to_cpu (INODE_PKEY (inode)->k_dir_id); | ||
1142 | else | ||
1143 | inode->i_generation = sd_v2_generation(sd); | ||
1144 | 1175 | ||
1145 | if (S_ISDIR (inode->i_mode) || S_ISLNK (inode->i_mode)) | 1176 | rdev = sd_v1_rdev(sd); |
1146 | set_inode_item_key_version (inode, KEY_FORMAT_3_5); | 1177 | REISERFS_I(inode)->i_first_direct_byte = |
1147 | else | 1178 | sd_v1_first_direct_byte(sd); |
1148 | set_inode_item_key_version (inode, KEY_FORMAT_3_6); | 1179 | /* an early bug in the quota code can give us an odd number for the |
1149 | REISERFS_I(inode)->i_first_direct_byte = 0; | 1180 | ** block count. This is incorrect, fix it here. |
1150 | set_inode_sd_version (inode, STAT_DATA_V2); | 1181 | */ |
1151 | inode_set_bytes(inode, to_real_used_space(inode, inode->i_blocks, | 1182 | if (inode->i_blocks & 1) { |
1152 | SD_V2_SIZE)); | 1183 | inode->i_blocks++; |
1153 | /* read persistent inode attributes from sd and initalise | 1184 | } |
1154 | generic inode flags from them */ | 1185 | inode_set_bytes(inode, |
1155 | REISERFS_I(inode)->i_attrs = sd_v2_attrs( sd ); | 1186 | to_real_used_space(inode, inode->i_blocks, |
1156 | sd_attrs_to_i_attrs( sd_v2_attrs( sd ), inode ); | 1187 | SD_V1_SIZE)); |
1157 | } | 1188 | /* nopack is initially zero for v1 objects. For v2 objects, |
1158 | 1189 | nopack is initialised from sd_attrs */ | |
1159 | pathrelse (path); | 1190 | REISERFS_I(inode)->i_flags &= ~i_nopack_mask; |
1160 | if (S_ISREG (inode->i_mode)) { | 1191 | } else { |
1161 | inode->i_op = &reiserfs_file_inode_operations; | 1192 | // new stat data found, but object may have old items |
1162 | inode->i_fop = &reiserfs_file_operations; | 1193 | // (directories and symlinks) |
1163 | inode->i_mapping->a_ops = &reiserfs_address_space_operations ; | 1194 | struct stat_data *sd = (struct stat_data *)B_I_PITEM(bh, ih); |
1164 | } else if (S_ISDIR (inode->i_mode)) { | 1195 | |
1165 | inode->i_op = &reiserfs_dir_inode_operations; | 1196 | inode->i_mode = sd_v2_mode(sd); |
1166 | inode->i_fop = &reiserfs_dir_operations; | 1197 | inode->i_nlink = sd_v2_nlink(sd); |
1167 | } else if (S_ISLNK (inode->i_mode)) { | 1198 | inode->i_uid = sd_v2_uid(sd); |
1168 | inode->i_op = &reiserfs_symlink_inode_operations; | 1199 | inode->i_size = sd_v2_size(sd); |
1169 | inode->i_mapping->a_ops = &reiserfs_address_space_operations; | 1200 | inode->i_gid = sd_v2_gid(sd); |
1170 | } else { | 1201 | inode->i_mtime.tv_sec = sd_v2_mtime(sd); |
1171 | inode->i_blocks = 0; | 1202 | inode->i_atime.tv_sec = sd_v2_atime(sd); |
1172 | inode->i_op = &reiserfs_special_inode_operations; | 1203 | inode->i_ctime.tv_sec = sd_v2_ctime(sd); |
1173 | init_special_inode(inode, inode->i_mode, new_decode_dev(rdev)); | 1204 | inode->i_ctime.tv_nsec = 0; |
1174 | } | 1205 | inode->i_mtime.tv_nsec = 0; |
1175 | } | 1206 | inode->i_atime.tv_nsec = 0; |
1207 | inode->i_blocks = sd_v2_blocks(sd); | ||
1208 | rdev = sd_v2_rdev(sd); | ||
1209 | if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode)) | ||
1210 | inode->i_generation = | ||
1211 | le32_to_cpu(INODE_PKEY(inode)->k_dir_id); | ||
1212 | else | ||
1213 | inode->i_generation = sd_v2_generation(sd); | ||
1176 | 1214 | ||
1215 | if (S_ISDIR(inode->i_mode) || S_ISLNK(inode->i_mode)) | ||
1216 | set_inode_item_key_version(inode, KEY_FORMAT_3_5); | ||
1217 | else | ||
1218 | set_inode_item_key_version(inode, KEY_FORMAT_3_6); | ||
1219 | REISERFS_I(inode)->i_first_direct_byte = 0; | ||
1220 | set_inode_sd_version(inode, STAT_DATA_V2); | ||
1221 | inode_set_bytes(inode, | ||
1222 | to_real_used_space(inode, inode->i_blocks, | ||
1223 | SD_V2_SIZE)); | ||
1224 | /* read persistent inode attributes from sd and initalise | ||
1225 | generic inode flags from them */ | ||
1226 | REISERFS_I(inode)->i_attrs = sd_v2_attrs(sd); | ||
1227 | sd_attrs_to_i_attrs(sd_v2_attrs(sd), inode); | ||
1228 | } | ||
1229 | |||
1230 | pathrelse(path); | ||
1231 | if (S_ISREG(inode->i_mode)) { | ||
1232 | inode->i_op = &reiserfs_file_inode_operations; | ||
1233 | inode->i_fop = &reiserfs_file_operations; | ||
1234 | inode->i_mapping->a_ops = &reiserfs_address_space_operations; | ||
1235 | } else if (S_ISDIR(inode->i_mode)) { | ||
1236 | inode->i_op = &reiserfs_dir_inode_operations; | ||
1237 | inode->i_fop = &reiserfs_dir_operations; | ||
1238 | } else if (S_ISLNK(inode->i_mode)) { | ||
1239 | inode->i_op = &reiserfs_symlink_inode_operations; | ||
1240 | inode->i_mapping->a_ops = &reiserfs_address_space_operations; | ||
1241 | } else { | ||
1242 | inode->i_blocks = 0; | ||
1243 | inode->i_op = &reiserfs_special_inode_operations; | ||
1244 | init_special_inode(inode, inode->i_mode, new_decode_dev(rdev)); | ||
1245 | } | ||
1246 | } | ||
1177 | 1247 | ||
1178 | // update new stat data with inode fields | 1248 | // update new stat data with inode fields |
1179 | static void inode2sd (void * sd, struct inode * inode, loff_t size) | 1249 | static void inode2sd(void *sd, struct inode *inode, loff_t size) |
1180 | { | 1250 | { |
1181 | struct stat_data * sd_v2 = (struct stat_data *)sd; | 1251 | struct stat_data *sd_v2 = (struct stat_data *)sd; |
1182 | __u16 flags; | 1252 | __u16 flags; |
1183 | 1253 | ||
1184 | set_sd_v2_mode(sd_v2, inode->i_mode ); | 1254 | set_sd_v2_mode(sd_v2, inode->i_mode); |
1185 | set_sd_v2_nlink(sd_v2, inode->i_nlink ); | 1255 | set_sd_v2_nlink(sd_v2, inode->i_nlink); |
1186 | set_sd_v2_uid(sd_v2, inode->i_uid ); | 1256 | set_sd_v2_uid(sd_v2, inode->i_uid); |
1187 | set_sd_v2_size(sd_v2, size ); | 1257 | set_sd_v2_size(sd_v2, size); |
1188 | set_sd_v2_gid(sd_v2, inode->i_gid ); | 1258 | set_sd_v2_gid(sd_v2, inode->i_gid); |
1189 | set_sd_v2_mtime(sd_v2, inode->i_mtime.tv_sec ); | 1259 | set_sd_v2_mtime(sd_v2, inode->i_mtime.tv_sec); |
1190 | set_sd_v2_atime(sd_v2, inode->i_atime.tv_sec ); | 1260 | set_sd_v2_atime(sd_v2, inode->i_atime.tv_sec); |
1191 | set_sd_v2_ctime(sd_v2, inode->i_ctime.tv_sec ); | 1261 | set_sd_v2_ctime(sd_v2, inode->i_ctime.tv_sec); |
1192 | set_sd_v2_blocks(sd_v2, to_fake_used_blocks(inode, SD_V2_SIZE)); | 1262 | set_sd_v2_blocks(sd_v2, to_fake_used_blocks(inode, SD_V2_SIZE)); |
1193 | if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode)) | 1263 | if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode)) |
1194 | set_sd_v2_rdev(sd_v2, new_encode_dev(inode->i_rdev)); | 1264 | set_sd_v2_rdev(sd_v2, new_encode_dev(inode->i_rdev)); |
1195 | else | 1265 | else |
1196 | set_sd_v2_generation(sd_v2, inode->i_generation); | 1266 | set_sd_v2_generation(sd_v2, inode->i_generation); |
1197 | flags = REISERFS_I(inode)->i_attrs; | 1267 | flags = REISERFS_I(inode)->i_attrs; |
1198 | i_attrs_to_sd_attrs( inode, &flags ); | 1268 | i_attrs_to_sd_attrs(inode, &flags); |
1199 | set_sd_v2_attrs( sd_v2, flags ); | 1269 | set_sd_v2_attrs(sd_v2, flags); |
1200 | } | 1270 | } |
1201 | 1271 | ||
1202 | |||
1203 | // used to copy inode's fields to old stat data | 1272 | // used to copy inode's fields to old stat data |
1204 | static void inode2sd_v1 (void * sd, struct inode * inode, loff_t size) | 1273 | static void inode2sd_v1(void *sd, struct inode *inode, loff_t size) |
1205 | { | 1274 | { |
1206 | struct stat_data_v1 * sd_v1 = (struct stat_data_v1 *)sd; | 1275 | struct stat_data_v1 *sd_v1 = (struct stat_data_v1 *)sd; |
1207 | 1276 | ||
1208 | set_sd_v1_mode(sd_v1, inode->i_mode ); | 1277 | set_sd_v1_mode(sd_v1, inode->i_mode); |
1209 | set_sd_v1_uid(sd_v1, inode->i_uid ); | 1278 | set_sd_v1_uid(sd_v1, inode->i_uid); |
1210 | set_sd_v1_gid(sd_v1, inode->i_gid ); | 1279 | set_sd_v1_gid(sd_v1, inode->i_gid); |
1211 | set_sd_v1_nlink(sd_v1, inode->i_nlink ); | 1280 | set_sd_v1_nlink(sd_v1, inode->i_nlink); |
1212 | set_sd_v1_size(sd_v1, size ); | 1281 | set_sd_v1_size(sd_v1, size); |
1213 | set_sd_v1_atime(sd_v1, inode->i_atime.tv_sec ); | 1282 | set_sd_v1_atime(sd_v1, inode->i_atime.tv_sec); |
1214 | set_sd_v1_ctime(sd_v1, inode->i_ctime.tv_sec ); | 1283 | set_sd_v1_ctime(sd_v1, inode->i_ctime.tv_sec); |
1215 | set_sd_v1_mtime(sd_v1, inode->i_mtime.tv_sec ); | 1284 | set_sd_v1_mtime(sd_v1, inode->i_mtime.tv_sec); |
1216 | 1285 | ||
1217 | if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode)) | 1286 | if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode)) |
1218 | set_sd_v1_rdev(sd_v1, new_encode_dev(inode->i_rdev)); | 1287 | set_sd_v1_rdev(sd_v1, new_encode_dev(inode->i_rdev)); |
1219 | else | 1288 | else |
1220 | set_sd_v1_blocks(sd_v1, to_fake_used_blocks(inode, SD_V1_SIZE)); | 1289 | set_sd_v1_blocks(sd_v1, to_fake_used_blocks(inode, SD_V1_SIZE)); |
1221 | |||
1222 | // Sigh. i_first_direct_byte is back | ||
1223 | set_sd_v1_first_direct_byte(sd_v1, REISERFS_I(inode)->i_first_direct_byte); | ||
1224 | } | ||
1225 | 1290 | ||
1291 | // Sigh. i_first_direct_byte is back | ||
1292 | set_sd_v1_first_direct_byte(sd_v1, | ||
1293 | REISERFS_I(inode)->i_first_direct_byte); | ||
1294 | } | ||
1226 | 1295 | ||
1227 | /* NOTE, you must prepare the buffer head before sending it here, | 1296 | /* NOTE, you must prepare the buffer head before sending it here, |
1228 | ** and then log it after the call | 1297 | ** and then log it after the call |
1229 | */ | 1298 | */ |
1230 | static void update_stat_data (struct path * path, struct inode * inode, | 1299 | static void update_stat_data(struct path *path, struct inode *inode, |
1231 | loff_t size) | 1300 | loff_t size) |
1232 | { | 1301 | { |
1233 | struct buffer_head * bh; | 1302 | struct buffer_head *bh; |
1234 | struct item_head * ih; | 1303 | struct item_head *ih; |
1235 | 1304 | ||
1236 | bh = PATH_PLAST_BUFFER (path); | 1305 | bh = PATH_PLAST_BUFFER(path); |
1237 | ih = PATH_PITEM_HEAD (path); | 1306 | ih = PATH_PITEM_HEAD(path); |
1238 | 1307 | ||
1239 | if (!is_statdata_le_ih (ih)) | 1308 | if (!is_statdata_le_ih(ih)) |
1240 | reiserfs_panic (inode->i_sb, "vs-13065: update_stat_data: key %k, found item %h", | 1309 | reiserfs_panic(inode->i_sb, |
1241 | INODE_PKEY (inode), ih); | 1310 | "vs-13065: update_stat_data: key %k, found item %h", |
1242 | 1311 | INODE_PKEY(inode), ih); | |
1243 | if (stat_data_v1 (ih)) { | 1312 | |
1244 | // path points to old stat data | 1313 | if (stat_data_v1(ih)) { |
1245 | inode2sd_v1 (B_I_PITEM (bh, ih), inode, size); | 1314 | // path points to old stat data |
1246 | } else { | 1315 | inode2sd_v1(B_I_PITEM(bh, ih), inode, size); |
1247 | inode2sd (B_I_PITEM (bh, ih), inode, size); | 1316 | } else { |
1248 | } | 1317 | inode2sd(B_I_PITEM(bh, ih), inode, size); |
1249 | 1318 | } | |
1250 | return; | ||
1251 | } | ||
1252 | 1319 | ||
1320 | return; | ||
1321 | } | ||
1253 | 1322 | ||
1254 | void reiserfs_update_sd_size (struct reiserfs_transaction_handle *th, | 1323 | void reiserfs_update_sd_size(struct reiserfs_transaction_handle *th, |
1255 | struct inode * inode, loff_t size) | 1324 | struct inode *inode, loff_t size) |
1256 | { | 1325 | { |
1257 | struct cpu_key key; | 1326 | struct cpu_key key; |
1258 | INITIALIZE_PATH(path); | 1327 | INITIALIZE_PATH(path); |
1259 | struct buffer_head *bh ; | 1328 | struct buffer_head *bh; |
1260 | int fs_gen ; | 1329 | int fs_gen; |
1261 | struct item_head *ih, tmp_ih ; | 1330 | struct item_head *ih, tmp_ih; |
1262 | int retval; | 1331 | int retval; |
1263 | 1332 | ||
1264 | BUG_ON (!th->t_trans_id); | 1333 | BUG_ON(!th->t_trans_id); |
1265 | 1334 | ||
1266 | make_cpu_key (&key, inode, SD_OFFSET, TYPE_STAT_DATA, 3);//key type is unimportant | 1335 | make_cpu_key(&key, inode, SD_OFFSET, TYPE_STAT_DATA, 3); //key type is unimportant |
1267 | 1336 | ||
1268 | for(;;) { | 1337 | for (;;) { |
1269 | int pos; | 1338 | int pos; |
1270 | /* look for the object's stat data */ | 1339 | /* look for the object's stat data */ |
1271 | retval = search_item (inode->i_sb, &key, &path); | 1340 | retval = search_item(inode->i_sb, &key, &path); |
1272 | if (retval == IO_ERROR) { | 1341 | if (retval == IO_ERROR) { |
1273 | reiserfs_warning (inode->i_sb, "vs-13050: reiserfs_update_sd: " | 1342 | reiserfs_warning(inode->i_sb, |
1274 | "i/o failure occurred trying to update %K stat data", | 1343 | "vs-13050: reiserfs_update_sd: " |
1275 | &key); | 1344 | "i/o failure occurred trying to update %K stat data", |
1276 | return; | 1345 | &key); |
1277 | } | 1346 | return; |
1278 | if (retval == ITEM_NOT_FOUND) { | 1347 | } |
1279 | pos = PATH_LAST_POSITION (&path); | 1348 | if (retval == ITEM_NOT_FOUND) { |
1280 | pathrelse(&path) ; | 1349 | pos = PATH_LAST_POSITION(&path); |
1281 | if (inode->i_nlink == 0) { | 1350 | pathrelse(&path); |
1282 | /*reiserfs_warning (inode->i_sb, "vs-13050: reiserfs_update_sd: i_nlink == 0, stat data not found");*/ | 1351 | if (inode->i_nlink == 0) { |
1283 | return; | 1352 | /*reiserfs_warning (inode->i_sb, "vs-13050: reiserfs_update_sd: i_nlink == 0, stat data not found"); */ |
1284 | } | 1353 | return; |
1285 | reiserfs_warning (inode->i_sb, "vs-13060: reiserfs_update_sd: " | 1354 | } |
1286 | "stat data of object %k (nlink == %d) not found (pos %d)", | 1355 | reiserfs_warning(inode->i_sb, |
1287 | INODE_PKEY (inode), inode->i_nlink, pos); | 1356 | "vs-13060: reiserfs_update_sd: " |
1288 | reiserfs_check_path(&path) ; | 1357 | "stat data of object %k (nlink == %d) not found (pos %d)", |
1289 | return; | 1358 | INODE_PKEY(inode), inode->i_nlink, |
1290 | } | 1359 | pos); |
1291 | 1360 | reiserfs_check_path(&path); | |
1292 | /* sigh, prepare_for_journal might schedule. When it schedules the | 1361 | return; |
1293 | ** FS might change. We have to detect that, and loop back to the | 1362 | } |
1294 | ** search if the stat data item has moved | 1363 | |
1295 | */ | 1364 | /* sigh, prepare_for_journal might schedule. When it schedules the |
1296 | bh = get_last_bh(&path) ; | 1365 | ** FS might change. We have to detect that, and loop back to the |
1297 | ih = get_ih(&path) ; | 1366 | ** search if the stat data item has moved |
1298 | copy_item_head (&tmp_ih, ih); | 1367 | */ |
1299 | fs_gen = get_generation (inode->i_sb); | 1368 | bh = get_last_bh(&path); |
1300 | reiserfs_prepare_for_journal(inode->i_sb, bh, 1) ; | 1369 | ih = get_ih(&path); |
1301 | if (fs_changed (fs_gen, inode->i_sb) && item_moved(&tmp_ih, &path)) { | 1370 | copy_item_head(&tmp_ih, ih); |
1302 | reiserfs_restore_prepared_buffer(inode->i_sb, bh) ; | 1371 | fs_gen = get_generation(inode->i_sb); |
1303 | continue ; /* Stat_data item has been moved after scheduling. */ | 1372 | reiserfs_prepare_for_journal(inode->i_sb, bh, 1); |
1304 | } | 1373 | if (fs_changed(fs_gen, inode->i_sb) |
1305 | break; | 1374 | && item_moved(&tmp_ih, &path)) { |
1306 | } | 1375 | reiserfs_restore_prepared_buffer(inode->i_sb, bh); |
1307 | update_stat_data (&path, inode, size); | 1376 | continue; /* Stat_data item has been moved after scheduling. */ |
1308 | journal_mark_dirty(th, th->t_super, bh) ; | 1377 | } |
1309 | pathrelse (&path); | 1378 | break; |
1310 | return; | 1379 | } |
1380 | update_stat_data(&path, inode, size); | ||
1381 | journal_mark_dirty(th, th->t_super, bh); | ||
1382 | pathrelse(&path); | ||
1383 | return; | ||
1311 | } | 1384 | } |
1312 | 1385 | ||
1313 | /* reiserfs_read_locked_inode is called to read the inode off disk, and it | 1386 | /* reiserfs_read_locked_inode is called to read the inode off disk, and it |
@@ -1316,9 +1389,10 @@ void reiserfs_update_sd_size (struct reiserfs_transaction_handle *th, | |||
1316 | ** corresponding iput might try to delete whatever object the inode last | 1389 | ** corresponding iput might try to delete whatever object the inode last |
1317 | ** represented. | 1390 | ** represented. |
1318 | */ | 1391 | */ |
1319 | static void reiserfs_make_bad_inode(struct inode *inode) { | 1392 | static void reiserfs_make_bad_inode(struct inode *inode) |
1320 | memset(INODE_PKEY(inode), 0, KEY_SIZE); | 1393 | { |
1321 | make_bad_inode(inode); | 1394 | memset(INODE_PKEY(inode), 0, KEY_SIZE); |
1395 | make_bad_inode(inode); | ||
1322 | } | 1396 | } |
1323 | 1397 | ||
1324 | // | 1398 | // |
@@ -1326,77 +1400,79 @@ static void reiserfs_make_bad_inode(struct inode *inode) { | |||
1326 | // evolved as the prototype did | 1400 | // evolved as the prototype did |
1327 | // | 1401 | // |
1328 | 1402 | ||
1329 | int reiserfs_init_locked_inode (struct inode * inode, void *p) | 1403 | int reiserfs_init_locked_inode(struct inode *inode, void *p) |
1330 | { | 1404 | { |
1331 | struct reiserfs_iget_args *args = (struct reiserfs_iget_args *)p ; | 1405 | struct reiserfs_iget_args *args = (struct reiserfs_iget_args *)p; |
1332 | inode->i_ino = args->objectid; | 1406 | inode->i_ino = args->objectid; |
1333 | INODE_PKEY(inode)->k_dir_id = cpu_to_le32(args->dirid); | 1407 | INODE_PKEY(inode)->k_dir_id = cpu_to_le32(args->dirid); |
1334 | return 0; | 1408 | return 0; |
1335 | } | 1409 | } |
1336 | 1410 | ||
1337 | /* looks for stat data in the tree, and fills up the fields of in-core | 1411 | /* looks for stat data in the tree, and fills up the fields of in-core |
1338 | inode stat data fields */ | 1412 | inode stat data fields */ |
1339 | void reiserfs_read_locked_inode (struct inode * inode, struct reiserfs_iget_args *args) | 1413 | void reiserfs_read_locked_inode(struct inode *inode, |
1414 | struct reiserfs_iget_args *args) | ||
1340 | { | 1415 | { |
1341 | INITIALIZE_PATH (path_to_sd); | 1416 | INITIALIZE_PATH(path_to_sd); |
1342 | struct cpu_key key; | 1417 | struct cpu_key key; |
1343 | unsigned long dirino; | 1418 | unsigned long dirino; |
1344 | int retval; | 1419 | int retval; |
1345 | 1420 | ||
1346 | dirino = args->dirid ; | 1421 | dirino = args->dirid; |
1347 | 1422 | ||
1348 | /* set version 1, version 2 could be used too, because stat data | 1423 | /* set version 1, version 2 could be used too, because stat data |
1349 | key is the same in both versions */ | 1424 | key is the same in both versions */ |
1350 | key.version = KEY_FORMAT_3_5; | 1425 | key.version = KEY_FORMAT_3_5; |
1351 | key.on_disk_key.k_dir_id = dirino; | 1426 | key.on_disk_key.k_dir_id = dirino; |
1352 | key.on_disk_key.k_objectid = inode->i_ino; | 1427 | key.on_disk_key.k_objectid = inode->i_ino; |
1353 | key.on_disk_key.k_offset = 0; | 1428 | key.on_disk_key.k_offset = 0; |
1354 | key.on_disk_key.k_type = 0; | 1429 | key.on_disk_key.k_type = 0; |
1355 | 1430 | ||
1356 | /* look for the object's stat data */ | 1431 | /* look for the object's stat data */ |
1357 | retval = search_item (inode->i_sb, &key, &path_to_sd); | 1432 | retval = search_item(inode->i_sb, &key, &path_to_sd); |
1358 | if (retval == IO_ERROR) { | 1433 | if (retval == IO_ERROR) { |
1359 | reiserfs_warning (inode->i_sb, "vs-13070: reiserfs_read_locked_inode: " | 1434 | reiserfs_warning(inode->i_sb, |
1360 | "i/o failure occurred trying to find stat data of %K", | 1435 | "vs-13070: reiserfs_read_locked_inode: " |
1361 | &key); | 1436 | "i/o failure occurred trying to find stat data of %K", |
1362 | reiserfs_make_bad_inode(inode) ; | 1437 | &key); |
1363 | return; | 1438 | reiserfs_make_bad_inode(inode); |
1364 | } | 1439 | return; |
1365 | if (retval != ITEM_FOUND) { | 1440 | } |
1366 | /* a stale NFS handle can trigger this without it being an error */ | 1441 | if (retval != ITEM_FOUND) { |
1367 | pathrelse (&path_to_sd); | 1442 | /* a stale NFS handle can trigger this without it being an error */ |
1368 | reiserfs_make_bad_inode(inode) ; | 1443 | pathrelse(&path_to_sd); |
1369 | inode->i_nlink = 0; | 1444 | reiserfs_make_bad_inode(inode); |
1370 | return; | 1445 | inode->i_nlink = 0; |
1371 | } | 1446 | return; |
1372 | 1447 | } | |
1373 | init_inode (inode, &path_to_sd); | 1448 | |
1374 | 1449 | init_inode(inode, &path_to_sd); | |
1375 | /* It is possible that knfsd is trying to access inode of a file | 1450 | |
1376 | that is being removed from the disk by some other thread. As we | 1451 | /* It is possible that knfsd is trying to access inode of a file |
1377 | update sd on unlink all that is required is to check for nlink | 1452 | that is being removed from the disk by some other thread. As we |
1378 | here. This bug was first found by Sizif when debugging | 1453 | update sd on unlink all that is required is to check for nlink |
1379 | SquidNG/Butterfly, forgotten, and found again after Philippe | 1454 | here. This bug was first found by Sizif when debugging |
1380 | Gramoulle <philippe.gramoulle@mmania.com> reproduced it. | 1455 | SquidNG/Butterfly, forgotten, and found again after Philippe |
1381 | 1456 | Gramoulle <philippe.gramoulle@mmania.com> reproduced it. | |
1382 | More logical fix would require changes in fs/inode.c:iput() to | 1457 | |
1383 | remove inode from hash-table _after_ fs cleaned disk stuff up and | 1458 | More logical fix would require changes in fs/inode.c:iput() to |
1384 | in iget() to return NULL if I_FREEING inode is found in | 1459 | remove inode from hash-table _after_ fs cleaned disk stuff up and |
1385 | hash-table. */ | 1460 | in iget() to return NULL if I_FREEING inode is found in |
1386 | /* Currently there is one place where it's ok to meet inode with | 1461 | hash-table. */ |
1387 | nlink==0: processing of open-unlinked and half-truncated files | 1462 | /* Currently there is one place where it's ok to meet inode with |
1388 | during mount (fs/reiserfs/super.c:finish_unfinished()). */ | 1463 | nlink==0: processing of open-unlinked and half-truncated files |
1389 | if( ( inode -> i_nlink == 0 ) && | 1464 | during mount (fs/reiserfs/super.c:finish_unfinished()). */ |
1390 | ! REISERFS_SB(inode -> i_sb) -> s_is_unlinked_ok ) { | 1465 | if ((inode->i_nlink == 0) && |
1391 | reiserfs_warning (inode->i_sb, | 1466 | !REISERFS_SB(inode->i_sb)->s_is_unlinked_ok) { |
1392 | "vs-13075: reiserfs_read_locked_inode: " | 1467 | reiserfs_warning(inode->i_sb, |
1393 | "dead inode read from disk %K. " | 1468 | "vs-13075: reiserfs_read_locked_inode: " |
1394 | "This is likely to be race with knfsd. Ignore", | 1469 | "dead inode read from disk %K. " |
1395 | &key ); | 1470 | "This is likely to be race with knfsd. Ignore", |
1396 | reiserfs_make_bad_inode( inode ); | 1471 | &key); |
1397 | } | 1472 | reiserfs_make_bad_inode(inode); |
1398 | 1473 | } | |
1399 | reiserfs_check_path(&path_to_sd) ; /* init inode should be relsing */ | 1474 | |
1475 | reiserfs_check_path(&path_to_sd); /* init inode should be relsing */ | ||
1400 | 1476 | ||
1401 | } | 1477 | } |
1402 | 1478 | ||
@@ -1412,140 +1488,148 @@ void reiserfs_read_locked_inode (struct inode * inode, struct reiserfs_iget_args | |||
1412 | * inode numbers (objectids) are distinguished by parent directory ids. | 1488 | * inode numbers (objectids) are distinguished by parent directory ids. |
1413 | * | 1489 | * |
1414 | */ | 1490 | */ |
1415 | int reiserfs_find_actor( struct inode *inode, void *opaque ) | 1491 | int reiserfs_find_actor(struct inode *inode, void *opaque) |
1416 | { | 1492 | { |
1417 | struct reiserfs_iget_args *args; | 1493 | struct reiserfs_iget_args *args; |
1418 | 1494 | ||
1419 | args = opaque; | 1495 | args = opaque; |
1420 | /* args is already in CPU order */ | 1496 | /* args is already in CPU order */ |
1421 | return (inode->i_ino == args->objectid) && | 1497 | return (inode->i_ino == args->objectid) && |
1422 | (le32_to_cpu(INODE_PKEY(inode)->k_dir_id) == args->dirid); | 1498 | (le32_to_cpu(INODE_PKEY(inode)->k_dir_id) == args->dirid); |
1423 | } | 1499 | } |
1424 | 1500 | ||
1425 | struct inode * reiserfs_iget (struct super_block * s, const struct cpu_key * key) | 1501 | struct inode *reiserfs_iget(struct super_block *s, const struct cpu_key *key) |
1426 | { | 1502 | { |
1427 | struct inode * inode; | 1503 | struct inode *inode; |
1428 | struct reiserfs_iget_args args ; | 1504 | struct reiserfs_iget_args args; |
1429 | 1505 | ||
1430 | args.objectid = key->on_disk_key.k_objectid ; | 1506 | args.objectid = key->on_disk_key.k_objectid; |
1431 | args.dirid = key->on_disk_key.k_dir_id ; | 1507 | args.dirid = key->on_disk_key.k_dir_id; |
1432 | inode = iget5_locked (s, key->on_disk_key.k_objectid, | 1508 | inode = iget5_locked(s, key->on_disk_key.k_objectid, |
1433 | reiserfs_find_actor, reiserfs_init_locked_inode, (void *)(&args)); | 1509 | reiserfs_find_actor, reiserfs_init_locked_inode, |
1434 | if (!inode) | 1510 | (void *)(&args)); |
1435 | return ERR_PTR(-ENOMEM) ; | 1511 | if (!inode) |
1436 | 1512 | return ERR_PTR(-ENOMEM); | |
1437 | if (inode->i_state & I_NEW) { | 1513 | |
1438 | reiserfs_read_locked_inode(inode, &args); | 1514 | if (inode->i_state & I_NEW) { |
1439 | unlock_new_inode(inode); | 1515 | reiserfs_read_locked_inode(inode, &args); |
1440 | } | 1516 | unlock_new_inode(inode); |
1441 | 1517 | } | |
1442 | if (comp_short_keys (INODE_PKEY (inode), key) || is_bad_inode (inode)) { | 1518 | |
1443 | /* either due to i/o error or a stale NFS handle */ | 1519 | if (comp_short_keys(INODE_PKEY(inode), key) || is_bad_inode(inode)) { |
1444 | iput (inode); | 1520 | /* either due to i/o error or a stale NFS handle */ |
1445 | inode = NULL; | 1521 | iput(inode); |
1446 | } | 1522 | inode = NULL; |
1447 | return inode; | 1523 | } |
1524 | return inode; | ||
1448 | } | 1525 | } |
1449 | 1526 | ||
1450 | struct dentry *reiserfs_get_dentry(struct super_block *sb, void *vobjp) | 1527 | struct dentry *reiserfs_get_dentry(struct super_block *sb, void *vobjp) |
1451 | { | 1528 | { |
1452 | __u32 *data = vobjp; | 1529 | __u32 *data = vobjp; |
1453 | struct cpu_key key ; | 1530 | struct cpu_key key; |
1454 | struct dentry *result; | 1531 | struct dentry *result; |
1455 | struct inode *inode; | 1532 | struct inode *inode; |
1456 | 1533 | ||
1457 | key.on_disk_key.k_objectid = data[0] ; | 1534 | key.on_disk_key.k_objectid = data[0]; |
1458 | key.on_disk_key.k_dir_id = data[1] ; | 1535 | key.on_disk_key.k_dir_id = data[1]; |
1459 | reiserfs_write_lock(sb); | 1536 | reiserfs_write_lock(sb); |
1460 | inode = reiserfs_iget(sb, &key) ; | 1537 | inode = reiserfs_iget(sb, &key); |
1461 | if (inode && !IS_ERR(inode) && data[2] != 0 && | 1538 | if (inode && !IS_ERR(inode) && data[2] != 0 && |
1462 | data[2] != inode->i_generation) { | 1539 | data[2] != inode->i_generation) { |
1463 | iput(inode) ; | 1540 | iput(inode); |
1464 | inode = NULL ; | 1541 | inode = NULL; |
1465 | } | 1542 | } |
1466 | reiserfs_write_unlock(sb); | 1543 | reiserfs_write_unlock(sb); |
1467 | if (!inode) | 1544 | if (!inode) |
1468 | inode = ERR_PTR(-ESTALE); | 1545 | inode = ERR_PTR(-ESTALE); |
1469 | if (IS_ERR(inode)) | 1546 | if (IS_ERR(inode)) |
1470 | return ERR_PTR(PTR_ERR(inode)); | 1547 | return ERR_PTR(PTR_ERR(inode)); |
1471 | result = d_alloc_anon(inode); | 1548 | result = d_alloc_anon(inode); |
1472 | if (!result) { | 1549 | if (!result) { |
1473 | iput(inode); | 1550 | iput(inode); |
1474 | return ERR_PTR(-ENOMEM); | 1551 | return ERR_PTR(-ENOMEM); |
1475 | } | 1552 | } |
1476 | return result; | 1553 | return result; |
1477 | } | 1554 | } |
1478 | 1555 | ||
1479 | struct dentry *reiserfs_decode_fh(struct super_block *sb, __u32 *data, | 1556 | struct dentry *reiserfs_decode_fh(struct super_block *sb, __u32 * data, |
1480 | int len, int fhtype, | 1557 | int len, int fhtype, |
1481 | int (*acceptable)(void *contect, struct dentry *de), | 1558 | int (*acceptable) (void *contect, |
1482 | void *context) { | 1559 | struct dentry * de), |
1483 | __u32 obj[3], parent[3]; | 1560 | void *context) |
1484 | 1561 | { | |
1485 | /* fhtype happens to reflect the number of u32s encoded. | 1562 | __u32 obj[3], parent[3]; |
1486 | * due to a bug in earlier code, fhtype might indicate there | 1563 | |
1487 | * are more u32s then actually fitted. | 1564 | /* fhtype happens to reflect the number of u32s encoded. |
1488 | * so if fhtype seems to be more than len, reduce fhtype. | 1565 | * due to a bug in earlier code, fhtype might indicate there |
1489 | * Valid types are: | 1566 | * are more u32s then actually fitted. |
1490 | * 2 - objectid + dir_id - legacy support | 1567 | * so if fhtype seems to be more than len, reduce fhtype. |
1491 | * 3 - objectid + dir_id + generation | 1568 | * Valid types are: |
1492 | * 4 - objectid + dir_id + objectid and dirid of parent - legacy | 1569 | * 2 - objectid + dir_id - legacy support |
1493 | * 5 - objectid + dir_id + generation + objectid and dirid of parent | 1570 | * 3 - objectid + dir_id + generation |
1494 | * 6 - as above plus generation of directory | 1571 | * 4 - objectid + dir_id + objectid and dirid of parent - legacy |
1495 | * 6 does not fit in NFSv2 handles | 1572 | * 5 - objectid + dir_id + generation + objectid and dirid of parent |
1496 | */ | 1573 | * 6 - as above plus generation of directory |
1497 | if (fhtype > len) { | 1574 | * 6 does not fit in NFSv2 handles |
1498 | if (fhtype != 6 || len != 5) | 1575 | */ |
1499 | reiserfs_warning (sb, "nfsd/reiserfs, fhtype=%d, len=%d - odd", | 1576 | if (fhtype > len) { |
1500 | fhtype, len); | 1577 | if (fhtype != 6 || len != 5) |
1501 | fhtype = 5; | 1578 | reiserfs_warning(sb, |
1502 | } | 1579 | "nfsd/reiserfs, fhtype=%d, len=%d - odd", |
1503 | 1580 | fhtype, len); | |
1504 | obj[0] = data[0]; | 1581 | fhtype = 5; |
1505 | obj[1] = data[1]; | 1582 | } |
1506 | if (fhtype == 3 || fhtype >= 5) | 1583 | |
1507 | obj[2] = data[2]; | 1584 | obj[0] = data[0]; |
1508 | else obj[2] = 0; /* generation number */ | 1585 | obj[1] = data[1]; |
1509 | 1586 | if (fhtype == 3 || fhtype >= 5) | |
1510 | if (fhtype >= 4) { | 1587 | obj[2] = data[2]; |
1511 | parent[0] = data[fhtype>=5?3:2] ; | 1588 | else |
1512 | parent[1] = data[fhtype>=5?4:3] ; | 1589 | obj[2] = 0; /* generation number */ |
1513 | if (fhtype == 6) | ||
1514 | parent[2] = data[5]; | ||
1515 | else parent[2] = 0; | ||
1516 | } | ||
1517 | return sb->s_export_op->find_exported_dentry(sb, obj, fhtype < 4 ? NULL : parent, | ||
1518 | acceptable, context); | ||
1519 | } | ||
1520 | 1590 | ||
1521 | int reiserfs_encode_fh(struct dentry *dentry, __u32 *data, int *lenp, int need_parent) { | 1591 | if (fhtype >= 4) { |
1522 | struct inode *inode = dentry->d_inode ; | 1592 | parent[0] = data[fhtype >= 5 ? 3 : 2]; |
1523 | int maxlen = *lenp; | 1593 | parent[1] = data[fhtype >= 5 ? 4 : 3]; |
1524 | 1594 | if (fhtype == 6) | |
1525 | if (maxlen < 3) | 1595 | parent[2] = data[5]; |
1526 | return 255 ; | 1596 | else |
1527 | 1597 | parent[2] = 0; | |
1528 | data[0] = inode->i_ino ; | 1598 | } |
1529 | data[1] = le32_to_cpu(INODE_PKEY (inode)->k_dir_id) ; | 1599 | return sb->s_export_op->find_exported_dentry(sb, obj, |
1530 | data[2] = inode->i_generation ; | 1600 | fhtype < 4 ? NULL : parent, |
1531 | *lenp = 3 ; | 1601 | acceptable, context); |
1532 | /* no room for directory info? return what we've stored so far */ | ||
1533 | if (maxlen < 5 || ! need_parent) | ||
1534 | return 3 ; | ||
1535 | |||
1536 | spin_lock(&dentry->d_lock); | ||
1537 | inode = dentry->d_parent->d_inode ; | ||
1538 | data[3] = inode->i_ino ; | ||
1539 | data[4] = le32_to_cpu(INODE_PKEY (inode)->k_dir_id) ; | ||
1540 | *lenp = 5 ; | ||
1541 | if (maxlen >= 6) { | ||
1542 | data[5] = inode->i_generation ; | ||
1543 | *lenp = 6 ; | ||
1544 | } | ||
1545 | spin_unlock(&dentry->d_lock); | ||
1546 | return *lenp ; | ||
1547 | } | 1602 | } |
1548 | 1603 | ||
1604 | int reiserfs_encode_fh(struct dentry *dentry, __u32 * data, int *lenp, | ||
1605 | int need_parent) | ||
1606 | { | ||
1607 | struct inode *inode = dentry->d_inode; | ||
1608 | int maxlen = *lenp; | ||
1609 | |||
1610 | if (maxlen < 3) | ||
1611 | return 255; | ||
1612 | |||
1613 | data[0] = inode->i_ino; | ||
1614 | data[1] = le32_to_cpu(INODE_PKEY(inode)->k_dir_id); | ||
1615 | data[2] = inode->i_generation; | ||
1616 | *lenp = 3; | ||
1617 | /* no room for directory info? return what we've stored so far */ | ||
1618 | if (maxlen < 5 || !need_parent) | ||
1619 | return 3; | ||
1620 | |||
1621 | spin_lock(&dentry->d_lock); | ||
1622 | inode = dentry->d_parent->d_inode; | ||
1623 | data[3] = inode->i_ino; | ||
1624 | data[4] = le32_to_cpu(INODE_PKEY(inode)->k_dir_id); | ||
1625 | *lenp = 5; | ||
1626 | if (maxlen >= 6) { | ||
1627 | data[5] = inode->i_generation; | ||
1628 | *lenp = 6; | ||
1629 | } | ||
1630 | spin_unlock(&dentry->d_lock); | ||
1631 | return *lenp; | ||
1632 | } | ||
1549 | 1633 | ||
1550 | /* looks for stat data, then copies fields to it, marks the buffer | 1634 | /* looks for stat data, then copies fields to it, marks the buffer |
1551 | containing stat data as dirty */ | 1635 | containing stat data as dirty */ |
@@ -1554,120 +1638,127 @@ int reiserfs_encode_fh(struct dentry *dentry, __u32 *data, int *lenp, int need_p | |||
1554 | ** to properly mark inodes for datasync and such, but only actually | 1638 | ** to properly mark inodes for datasync and such, but only actually |
1555 | ** does something when called for a synchronous update. | 1639 | ** does something when called for a synchronous update. |
1556 | */ | 1640 | */ |
1557 | int reiserfs_write_inode (struct inode * inode, int do_sync) { | 1641 | int reiserfs_write_inode(struct inode *inode, int do_sync) |
1558 | struct reiserfs_transaction_handle th ; | 1642 | { |
1559 | int jbegin_count = 1 ; | 1643 | struct reiserfs_transaction_handle th; |
1560 | 1644 | int jbegin_count = 1; | |
1561 | if (inode->i_sb->s_flags & MS_RDONLY) | 1645 | |
1562 | return -EROFS; | 1646 | if (inode->i_sb->s_flags & MS_RDONLY) |
1563 | /* memory pressure can sometimes initiate write_inode calls with sync == 1, | 1647 | return -EROFS; |
1564 | ** these cases are just when the system needs ram, not when the | 1648 | /* memory pressure can sometimes initiate write_inode calls with sync == 1, |
1565 | ** inode needs to reach disk for safety, and they can safely be | 1649 | ** these cases are just when the system needs ram, not when the |
1566 | ** ignored because the altered inode has already been logged. | 1650 | ** inode needs to reach disk for safety, and they can safely be |
1567 | */ | 1651 | ** ignored because the altered inode has already been logged. |
1568 | if (do_sync && !(current->flags & PF_MEMALLOC)) { | 1652 | */ |
1569 | reiserfs_write_lock(inode->i_sb); | 1653 | if (do_sync && !(current->flags & PF_MEMALLOC)) { |
1570 | if (!journal_begin(&th, inode->i_sb, jbegin_count)) { | 1654 | reiserfs_write_lock(inode->i_sb); |
1571 | reiserfs_update_sd (&th, inode); | 1655 | if (!journal_begin(&th, inode->i_sb, jbegin_count)) { |
1572 | journal_end_sync(&th, inode->i_sb, jbegin_count) ; | 1656 | reiserfs_update_sd(&th, inode); |
1573 | } | 1657 | journal_end_sync(&th, inode->i_sb, jbegin_count); |
1574 | reiserfs_write_unlock(inode->i_sb); | 1658 | } |
1575 | } | 1659 | reiserfs_write_unlock(inode->i_sb); |
1576 | return 0; | 1660 | } |
1661 | return 0; | ||
1577 | } | 1662 | } |
1578 | 1663 | ||
1579 | /* stat data of new object is inserted already, this inserts the item | 1664 | /* stat data of new object is inserted already, this inserts the item |
1580 | containing "." and ".." entries */ | 1665 | containing "." and ".." entries */ |
1581 | static int reiserfs_new_directory (struct reiserfs_transaction_handle *th, | 1666 | static int reiserfs_new_directory(struct reiserfs_transaction_handle *th, |
1582 | struct inode *inode, | 1667 | struct inode *inode, |
1583 | struct item_head * ih, struct path * path, | 1668 | struct item_head *ih, struct path *path, |
1584 | struct inode * dir) | 1669 | struct inode *dir) |
1585 | { | 1670 | { |
1586 | struct super_block * sb = th->t_super; | 1671 | struct super_block *sb = th->t_super; |
1587 | char empty_dir [EMPTY_DIR_SIZE]; | 1672 | char empty_dir[EMPTY_DIR_SIZE]; |
1588 | char * body = empty_dir; | 1673 | char *body = empty_dir; |
1589 | struct cpu_key key; | 1674 | struct cpu_key key; |
1590 | int retval; | 1675 | int retval; |
1591 | 1676 | ||
1592 | BUG_ON (!th->t_trans_id); | 1677 | BUG_ON(!th->t_trans_id); |
1593 | 1678 | ||
1594 | _make_cpu_key (&key, KEY_FORMAT_3_5, le32_to_cpu (ih->ih_key.k_dir_id), | 1679 | _make_cpu_key(&key, KEY_FORMAT_3_5, le32_to_cpu(ih->ih_key.k_dir_id), |
1595 | le32_to_cpu (ih->ih_key.k_objectid), DOT_OFFSET, TYPE_DIRENTRY, 3/*key length*/); | 1680 | le32_to_cpu(ih->ih_key.k_objectid), DOT_OFFSET, |
1596 | 1681 | TYPE_DIRENTRY, 3 /*key length */ ); | |
1597 | /* compose item head for new item. Directories consist of items of | 1682 | |
1598 | old type (ITEM_VERSION_1). Do not set key (second arg is 0), it | 1683 | /* compose item head for new item. Directories consist of items of |
1599 | is done by reiserfs_new_inode */ | 1684 | old type (ITEM_VERSION_1). Do not set key (second arg is 0), it |
1600 | if (old_format_only (sb)) { | 1685 | is done by reiserfs_new_inode */ |
1601 | make_le_item_head (ih, NULL, KEY_FORMAT_3_5, DOT_OFFSET, TYPE_DIRENTRY, EMPTY_DIR_SIZE_V1, 2); | 1686 | if (old_format_only(sb)) { |
1602 | 1687 | make_le_item_head(ih, NULL, KEY_FORMAT_3_5, DOT_OFFSET, | |
1603 | make_empty_dir_item_v1 (body, ih->ih_key.k_dir_id, ih->ih_key.k_objectid, | 1688 | TYPE_DIRENTRY, EMPTY_DIR_SIZE_V1, 2); |
1604 | INODE_PKEY (dir)->k_dir_id, | 1689 | |
1605 | INODE_PKEY (dir)->k_objectid ); | 1690 | make_empty_dir_item_v1(body, ih->ih_key.k_dir_id, |
1606 | } else { | 1691 | ih->ih_key.k_objectid, |
1607 | make_le_item_head (ih, NULL, KEY_FORMAT_3_5, DOT_OFFSET, TYPE_DIRENTRY, EMPTY_DIR_SIZE, 2); | 1692 | INODE_PKEY(dir)->k_dir_id, |
1608 | 1693 | INODE_PKEY(dir)->k_objectid); | |
1609 | make_empty_dir_item (body, ih->ih_key.k_dir_id, ih->ih_key.k_objectid, | 1694 | } else { |
1610 | INODE_PKEY (dir)->k_dir_id, | 1695 | make_le_item_head(ih, NULL, KEY_FORMAT_3_5, DOT_OFFSET, |
1611 | INODE_PKEY (dir)->k_objectid ); | 1696 | TYPE_DIRENTRY, EMPTY_DIR_SIZE, 2); |
1612 | } | 1697 | |
1613 | 1698 | make_empty_dir_item(body, ih->ih_key.k_dir_id, | |
1614 | /* look for place in the tree for new item */ | 1699 | ih->ih_key.k_objectid, |
1615 | retval = search_item (sb, &key, path); | 1700 | INODE_PKEY(dir)->k_dir_id, |
1616 | if (retval == IO_ERROR) { | 1701 | INODE_PKEY(dir)->k_objectid); |
1617 | reiserfs_warning (sb, "vs-13080: reiserfs_new_directory: " | 1702 | } |
1618 | "i/o failure occurred creating new directory"); | 1703 | |
1619 | return -EIO; | 1704 | /* look for place in the tree for new item */ |
1620 | } | 1705 | retval = search_item(sb, &key, path); |
1621 | if (retval == ITEM_FOUND) { | 1706 | if (retval == IO_ERROR) { |
1622 | pathrelse (path); | 1707 | reiserfs_warning(sb, "vs-13080: reiserfs_new_directory: " |
1623 | reiserfs_warning (sb, "vs-13070: reiserfs_new_directory: " | 1708 | "i/o failure occurred creating new directory"); |
1624 | "object with this key exists (%k)", &(ih->ih_key)); | 1709 | return -EIO; |
1625 | return -EEXIST; | 1710 | } |
1626 | } | 1711 | if (retval == ITEM_FOUND) { |
1627 | 1712 | pathrelse(path); | |
1628 | /* insert item, that is empty directory item */ | 1713 | reiserfs_warning(sb, "vs-13070: reiserfs_new_directory: " |
1629 | return reiserfs_insert_item (th, path, &key, ih, inode, body); | 1714 | "object with this key exists (%k)", |
1630 | } | 1715 | &(ih->ih_key)); |
1716 | return -EEXIST; | ||
1717 | } | ||
1631 | 1718 | ||
1719 | /* insert item, that is empty directory item */ | ||
1720 | return reiserfs_insert_item(th, path, &key, ih, inode, body); | ||
1721 | } | ||
1632 | 1722 | ||
1633 | /* stat data of object has been inserted, this inserts the item | 1723 | /* stat data of object has been inserted, this inserts the item |
1634 | containing the body of symlink */ | 1724 | containing the body of symlink */ |
1635 | static int reiserfs_new_symlink (struct reiserfs_transaction_handle *th, | 1725 | static int reiserfs_new_symlink(struct reiserfs_transaction_handle *th, struct inode *inode, /* Inode of symlink */ |
1636 | struct inode *inode, /* Inode of symlink */ | 1726 | struct item_head *ih, |
1637 | struct item_head * ih, | 1727 | struct path *path, const char *symname, |
1638 | struct path * path, const char * symname, int item_len) | 1728 | int item_len) |
1639 | { | 1729 | { |
1640 | struct super_block * sb = th->t_super; | 1730 | struct super_block *sb = th->t_super; |
1641 | struct cpu_key key; | 1731 | struct cpu_key key; |
1642 | int retval; | 1732 | int retval; |
1643 | 1733 | ||
1644 | BUG_ON (!th->t_trans_id); | 1734 | BUG_ON(!th->t_trans_id); |
1645 | 1735 | ||
1646 | _make_cpu_key (&key, KEY_FORMAT_3_5, | 1736 | _make_cpu_key(&key, KEY_FORMAT_3_5, |
1647 | le32_to_cpu (ih->ih_key.k_dir_id), | 1737 | le32_to_cpu(ih->ih_key.k_dir_id), |
1648 | le32_to_cpu (ih->ih_key.k_objectid), | 1738 | le32_to_cpu(ih->ih_key.k_objectid), |
1649 | 1, TYPE_DIRECT, 3/*key length*/); | 1739 | 1, TYPE_DIRECT, 3 /*key length */ ); |
1650 | 1740 | ||
1651 | make_le_item_head (ih, NULL, KEY_FORMAT_3_5, 1, TYPE_DIRECT, item_len, 0/*free_space*/); | 1741 | make_le_item_head(ih, NULL, KEY_FORMAT_3_5, 1, TYPE_DIRECT, item_len, |
1652 | 1742 | 0 /*free_space */ ); | |
1653 | /* look for place in the tree for new item */ | 1743 | |
1654 | retval = search_item (sb, &key, path); | 1744 | /* look for place in the tree for new item */ |
1655 | if (retval == IO_ERROR) { | 1745 | retval = search_item(sb, &key, path); |
1656 | reiserfs_warning (sb, "vs-13080: reiserfs_new_symlinik: " | 1746 | if (retval == IO_ERROR) { |
1657 | "i/o failure occurred creating new symlink"); | 1747 | reiserfs_warning(sb, "vs-13080: reiserfs_new_symlinik: " |
1658 | return -EIO; | 1748 | "i/o failure occurred creating new symlink"); |
1659 | } | 1749 | return -EIO; |
1660 | if (retval == ITEM_FOUND) { | 1750 | } |
1661 | pathrelse (path); | 1751 | if (retval == ITEM_FOUND) { |
1662 | reiserfs_warning (sb, "vs-13080: reiserfs_new_symlink: " | 1752 | pathrelse(path); |
1663 | "object with this key exists (%k)", &(ih->ih_key)); | 1753 | reiserfs_warning(sb, "vs-13080: reiserfs_new_symlink: " |
1664 | return -EEXIST; | 1754 | "object with this key exists (%k)", |
1665 | } | 1755 | &(ih->ih_key)); |
1666 | 1756 | return -EEXIST; | |
1667 | /* insert item, that is body of symlink */ | 1757 | } |
1668 | return reiserfs_insert_item (th, path, &key, ih, inode, symname); | ||
1669 | } | ||
1670 | 1758 | ||
1759 | /* insert item, that is body of symlink */ | ||
1760 | return reiserfs_insert_item(th, path, &key, ih, inode, symname); | ||
1761 | } | ||
1671 | 1762 | ||
1672 | /* inserts the stat data into the tree, and then calls | 1763 | /* inserts the stat data into the tree, and then calls |
1673 | reiserfs_new_directory (to insert ".", ".." item if new object is | 1764 | reiserfs_new_directory (to insert ".", ".." item if new object is |
@@ -1678,213 +1769,219 @@ static int reiserfs_new_symlink (struct reiserfs_transaction_handle *th, | |||
1678 | non-zero due to an error, we have to drop the quota previously allocated | 1769 | non-zero due to an error, we have to drop the quota previously allocated |
1679 | for the fresh inode. This can only be done outside a transaction, so | 1770 | for the fresh inode. This can only be done outside a transaction, so |
1680 | if we return non-zero, we also end the transaction. */ | 1771 | if we return non-zero, we also end the transaction. */ |
1681 | int reiserfs_new_inode (struct reiserfs_transaction_handle *th, | 1772 | int reiserfs_new_inode(struct reiserfs_transaction_handle *th, |
1682 | struct inode * dir, int mode, | 1773 | struct inode *dir, int mode, const char *symname, |
1683 | const char * symname, | 1774 | /* 0 for regular, EMTRY_DIR_SIZE for dirs, |
1684 | /* 0 for regular, EMTRY_DIR_SIZE for dirs, | 1775 | strlen (symname) for symlinks) */ |
1685 | strlen (symname) for symlinks)*/ | 1776 | loff_t i_size, struct dentry *dentry, |
1686 | loff_t i_size, struct dentry *dentry, | 1777 | struct inode *inode) |
1687 | struct inode *inode) | ||
1688 | { | 1778 | { |
1689 | struct super_block * sb; | 1779 | struct super_block *sb; |
1690 | INITIALIZE_PATH (path_to_key); | 1780 | INITIALIZE_PATH(path_to_key); |
1691 | struct cpu_key key; | 1781 | struct cpu_key key; |
1692 | struct item_head ih; | 1782 | struct item_head ih; |
1693 | struct stat_data sd; | 1783 | struct stat_data sd; |
1694 | int retval; | 1784 | int retval; |
1695 | int err; | 1785 | int err; |
1696 | 1786 | ||
1697 | BUG_ON (!th->t_trans_id); | 1787 | BUG_ON(!th->t_trans_id); |
1698 | 1788 | ||
1699 | if (DQUOT_ALLOC_INODE(inode)) { | 1789 | if (DQUOT_ALLOC_INODE(inode)) { |
1700 | err = -EDQUOT; | 1790 | err = -EDQUOT; |
1701 | goto out_end_trans; | 1791 | goto out_end_trans; |
1702 | } | 1792 | } |
1703 | if (!dir || !dir->i_nlink) { | 1793 | if (!dir || !dir->i_nlink) { |
1704 | err = -EPERM; | 1794 | err = -EPERM; |
1705 | goto out_bad_inode; | 1795 | goto out_bad_inode; |
1706 | } | 1796 | } |
1707 | 1797 | ||
1708 | sb = dir->i_sb; | 1798 | sb = dir->i_sb; |
1709 | 1799 | ||
1710 | /* item head of new item */ | 1800 | /* item head of new item */ |
1711 | ih.ih_key.k_dir_id = reiserfs_choose_packing(dir); | 1801 | ih.ih_key.k_dir_id = reiserfs_choose_packing(dir); |
1712 | ih.ih_key.k_objectid = cpu_to_le32 (reiserfs_get_unused_objectid (th)); | 1802 | ih.ih_key.k_objectid = cpu_to_le32(reiserfs_get_unused_objectid(th)); |
1713 | if (!ih.ih_key.k_objectid) { | 1803 | if (!ih.ih_key.k_objectid) { |
1714 | err = -ENOMEM; | 1804 | err = -ENOMEM; |
1715 | goto out_bad_inode ; | 1805 | goto out_bad_inode; |
1716 | } | 1806 | } |
1717 | if (old_format_only (sb)) | 1807 | if (old_format_only(sb)) |
1718 | /* not a perfect generation count, as object ids can be reused, but | 1808 | /* not a perfect generation count, as object ids can be reused, but |
1719 | ** this is as good as reiserfs can do right now. | 1809 | ** this is as good as reiserfs can do right now. |
1720 | ** note that the private part of inode isn't filled in yet, we have | 1810 | ** note that the private part of inode isn't filled in yet, we have |
1721 | ** to use the directory. | 1811 | ** to use the directory. |
1722 | */ | 1812 | */ |
1723 | inode->i_generation = le32_to_cpu (INODE_PKEY (dir)->k_objectid); | 1813 | inode->i_generation = le32_to_cpu(INODE_PKEY(dir)->k_objectid); |
1724 | else | 1814 | else |
1725 | #if defined( USE_INODE_GENERATION_COUNTER ) | 1815 | #if defined( USE_INODE_GENERATION_COUNTER ) |
1726 | inode->i_generation = le32_to_cpu(REISERFS_SB(sb)->s_rs->s_inode_generation); | 1816 | inode->i_generation = |
1817 | le32_to_cpu(REISERFS_SB(sb)->s_rs->s_inode_generation); | ||
1727 | #else | 1818 | #else |
1728 | inode->i_generation = ++event; | 1819 | inode->i_generation = ++event; |
1729 | #endif | 1820 | #endif |
1730 | 1821 | ||
1731 | /* fill stat data */ | 1822 | /* fill stat data */ |
1732 | inode->i_nlink = (S_ISDIR (mode) ? 2 : 1); | 1823 | inode->i_nlink = (S_ISDIR(mode) ? 2 : 1); |
1733 | 1824 | ||
1734 | /* uid and gid must already be set by the caller for quota init */ | 1825 | /* uid and gid must already be set by the caller for quota init */ |
1735 | 1826 | ||
1736 | /* symlink cannot be immutable or append only, right? */ | 1827 | /* symlink cannot be immutable or append only, right? */ |
1737 | if( S_ISLNK( inode -> i_mode ) ) | 1828 | if (S_ISLNK(inode->i_mode)) |
1738 | inode -> i_flags &= ~ ( S_IMMUTABLE | S_APPEND ); | 1829 | inode->i_flags &= ~(S_IMMUTABLE | S_APPEND); |
1739 | 1830 | ||
1740 | inode->i_mtime = inode->i_atime = inode->i_ctime = | 1831 | inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME_SEC; |
1741 | CURRENT_TIME_SEC; | 1832 | inode->i_size = i_size; |
1742 | inode->i_size = i_size; | 1833 | inode->i_blocks = 0; |
1743 | inode->i_blocks = 0; | 1834 | inode->i_bytes = 0; |
1744 | inode->i_bytes = 0; | 1835 | REISERFS_I(inode)->i_first_direct_byte = S_ISLNK(mode) ? 1 : |
1745 | REISERFS_I(inode)->i_first_direct_byte = S_ISLNK(mode) ? 1 : | 1836 | U32_MAX /*NO_BYTES_IN_DIRECT_ITEM */ ; |
1746 | U32_MAX/*NO_BYTES_IN_DIRECT_ITEM*/; | 1837 | |
1747 | 1838 | INIT_LIST_HEAD(&(REISERFS_I(inode)->i_prealloc_list)); | |
1748 | INIT_LIST_HEAD(&(REISERFS_I(inode)->i_prealloc_list )); | 1839 | REISERFS_I(inode)->i_flags = 0; |
1749 | REISERFS_I(inode)->i_flags = 0; | 1840 | REISERFS_I(inode)->i_prealloc_block = 0; |
1750 | REISERFS_I(inode)->i_prealloc_block = 0; | 1841 | REISERFS_I(inode)->i_prealloc_count = 0; |
1751 | REISERFS_I(inode)->i_prealloc_count = 0; | 1842 | REISERFS_I(inode)->i_trans_id = 0; |
1752 | REISERFS_I(inode)->i_trans_id = 0; | 1843 | REISERFS_I(inode)->i_jl = NULL; |
1753 | REISERFS_I(inode)->i_jl = NULL; | 1844 | REISERFS_I(inode)->i_attrs = |
1754 | REISERFS_I(inode)->i_attrs = | 1845 | REISERFS_I(dir)->i_attrs & REISERFS_INHERIT_MASK; |
1755 | REISERFS_I(dir)->i_attrs & REISERFS_INHERIT_MASK; | 1846 | sd_attrs_to_i_attrs(REISERFS_I(inode)->i_attrs, inode); |
1756 | sd_attrs_to_i_attrs( REISERFS_I(inode) -> i_attrs, inode ); | 1847 | REISERFS_I(inode)->i_acl_access = NULL; |
1757 | REISERFS_I(inode)->i_acl_access = NULL; | 1848 | REISERFS_I(inode)->i_acl_default = NULL; |
1758 | REISERFS_I(inode)->i_acl_default = NULL; | 1849 | init_rwsem(&REISERFS_I(inode)->xattr_sem); |
1759 | init_rwsem (&REISERFS_I(inode)->xattr_sem); | 1850 | |
1760 | 1851 | if (old_format_only(sb)) | |
1761 | if (old_format_only (sb)) | 1852 | make_le_item_head(&ih, NULL, KEY_FORMAT_3_5, SD_OFFSET, |
1762 | make_le_item_head (&ih, NULL, KEY_FORMAT_3_5, SD_OFFSET, TYPE_STAT_DATA, SD_V1_SIZE, MAX_US_INT); | 1853 | TYPE_STAT_DATA, SD_V1_SIZE, MAX_US_INT); |
1763 | else | 1854 | else |
1764 | make_le_item_head (&ih, NULL, KEY_FORMAT_3_6, SD_OFFSET, TYPE_STAT_DATA, SD_SIZE, MAX_US_INT); | 1855 | make_le_item_head(&ih, NULL, KEY_FORMAT_3_6, SD_OFFSET, |
1765 | 1856 | TYPE_STAT_DATA, SD_SIZE, MAX_US_INT); | |
1766 | /* key to search for correct place for new stat data */ | 1857 | |
1767 | _make_cpu_key (&key, KEY_FORMAT_3_6, le32_to_cpu (ih.ih_key.k_dir_id), | 1858 | /* key to search for correct place for new stat data */ |
1768 | le32_to_cpu (ih.ih_key.k_objectid), SD_OFFSET, TYPE_STAT_DATA, 3/*key length*/); | 1859 | _make_cpu_key(&key, KEY_FORMAT_3_6, le32_to_cpu(ih.ih_key.k_dir_id), |
1769 | 1860 | le32_to_cpu(ih.ih_key.k_objectid), SD_OFFSET, | |
1770 | /* find proper place for inserting of stat data */ | 1861 | TYPE_STAT_DATA, 3 /*key length */ ); |
1771 | retval = search_item (sb, &key, &path_to_key); | 1862 | |
1772 | if (retval == IO_ERROR) { | 1863 | /* find proper place for inserting of stat data */ |
1773 | err = -EIO; | 1864 | retval = search_item(sb, &key, &path_to_key); |
1774 | goto out_bad_inode; | 1865 | if (retval == IO_ERROR) { |
1775 | } | 1866 | err = -EIO; |
1776 | if (retval == ITEM_FOUND) { | 1867 | goto out_bad_inode; |
1777 | pathrelse (&path_to_key); | 1868 | } |
1778 | err = -EEXIST; | 1869 | if (retval == ITEM_FOUND) { |
1779 | goto out_bad_inode; | 1870 | pathrelse(&path_to_key); |
1780 | } | 1871 | err = -EEXIST; |
1781 | if (old_format_only (sb)) { | 1872 | goto out_bad_inode; |
1782 | if (inode->i_uid & ~0xffff || inode->i_gid & ~0xffff) { | 1873 | } |
1783 | pathrelse (&path_to_key); | 1874 | if (old_format_only(sb)) { |
1784 | /* i_uid or i_gid is too big to be stored in stat data v3.5 */ | 1875 | if (inode->i_uid & ~0xffff || inode->i_gid & ~0xffff) { |
1785 | err = -EINVAL; | 1876 | pathrelse(&path_to_key); |
1786 | goto out_bad_inode; | 1877 | /* i_uid or i_gid is too big to be stored in stat data v3.5 */ |
1787 | } | 1878 | err = -EINVAL; |
1788 | inode2sd_v1 (&sd, inode, inode->i_size); | 1879 | goto out_bad_inode; |
1789 | } else { | 1880 | } |
1790 | inode2sd (&sd, inode, inode->i_size); | 1881 | inode2sd_v1(&sd, inode, inode->i_size); |
1791 | } | 1882 | } else { |
1792 | // these do not go to on-disk stat data | 1883 | inode2sd(&sd, inode, inode->i_size); |
1793 | inode->i_ino = le32_to_cpu (ih.ih_key.k_objectid); | 1884 | } |
1794 | inode->i_blksize = reiserfs_default_io_size; | 1885 | // these do not go to on-disk stat data |
1795 | 1886 | inode->i_ino = le32_to_cpu(ih.ih_key.k_objectid); | |
1796 | // store in in-core inode the key of stat data and version all | 1887 | inode->i_blksize = reiserfs_default_io_size; |
1797 | // object items will have (directory items will have old offset | 1888 | |
1798 | // format, other new objects will consist of new items) | 1889 | // store in in-core inode the key of stat data and version all |
1799 | memcpy (INODE_PKEY (inode), &(ih.ih_key), KEY_SIZE); | 1890 | // object items will have (directory items will have old offset |
1800 | if (old_format_only (sb) || S_ISDIR(mode) || S_ISLNK(mode)) | 1891 | // format, other new objects will consist of new items) |
1801 | set_inode_item_key_version (inode, KEY_FORMAT_3_5); | 1892 | memcpy(INODE_PKEY(inode), &(ih.ih_key), KEY_SIZE); |
1802 | else | 1893 | if (old_format_only(sb) || S_ISDIR(mode) || S_ISLNK(mode)) |
1803 | set_inode_item_key_version (inode, KEY_FORMAT_3_6); | 1894 | set_inode_item_key_version(inode, KEY_FORMAT_3_5); |
1804 | if (old_format_only (sb)) | 1895 | else |
1805 | set_inode_sd_version (inode, STAT_DATA_V1); | 1896 | set_inode_item_key_version(inode, KEY_FORMAT_3_6); |
1806 | else | 1897 | if (old_format_only(sb)) |
1807 | set_inode_sd_version (inode, STAT_DATA_V2); | 1898 | set_inode_sd_version(inode, STAT_DATA_V1); |
1808 | 1899 | else | |
1809 | /* insert the stat data into the tree */ | 1900 | set_inode_sd_version(inode, STAT_DATA_V2); |
1901 | |||
1902 | /* insert the stat data into the tree */ | ||
1810 | #ifdef DISPLACE_NEW_PACKING_LOCALITIES | 1903 | #ifdef DISPLACE_NEW_PACKING_LOCALITIES |
1811 | if (REISERFS_I(dir)->new_packing_locality) | 1904 | if (REISERFS_I(dir)->new_packing_locality) |
1812 | th->displace_new_blocks = 1; | 1905 | th->displace_new_blocks = 1; |
1813 | #endif | 1906 | #endif |
1814 | retval = reiserfs_insert_item (th, &path_to_key, &key, &ih, inode, (char *)(&sd)); | 1907 | retval = |
1815 | if (retval) { | 1908 | reiserfs_insert_item(th, &path_to_key, &key, &ih, inode, |
1816 | err = retval; | 1909 | (char *)(&sd)); |
1817 | reiserfs_check_path(&path_to_key) ; | 1910 | if (retval) { |
1818 | goto out_bad_inode; | 1911 | err = retval; |
1819 | } | 1912 | reiserfs_check_path(&path_to_key); |
1820 | 1913 | goto out_bad_inode; | |
1914 | } | ||
1821 | #ifdef DISPLACE_NEW_PACKING_LOCALITIES | 1915 | #ifdef DISPLACE_NEW_PACKING_LOCALITIES |
1822 | if (!th->displace_new_blocks) | 1916 | if (!th->displace_new_blocks) |
1823 | REISERFS_I(dir)->new_packing_locality = 0; | 1917 | REISERFS_I(dir)->new_packing_locality = 0; |
1824 | #endif | 1918 | #endif |
1825 | if (S_ISDIR(mode)) { | 1919 | if (S_ISDIR(mode)) { |
1826 | /* insert item with "." and ".." */ | 1920 | /* insert item with "." and ".." */ |
1827 | retval = reiserfs_new_directory (th, inode, &ih, &path_to_key, dir); | 1921 | retval = |
1828 | } | 1922 | reiserfs_new_directory(th, inode, &ih, &path_to_key, dir); |
1829 | 1923 | } | |
1830 | if (S_ISLNK(mode)) { | 1924 | |
1831 | /* insert body of symlink */ | 1925 | if (S_ISLNK(mode)) { |
1832 | if (!old_format_only (sb)) | 1926 | /* insert body of symlink */ |
1833 | i_size = ROUND_UP(i_size); | 1927 | if (!old_format_only(sb)) |
1834 | retval = reiserfs_new_symlink (th, inode, &ih, &path_to_key, symname, i_size); | 1928 | i_size = ROUND_UP(i_size); |
1835 | } | 1929 | retval = |
1836 | if (retval) { | 1930 | reiserfs_new_symlink(th, inode, &ih, &path_to_key, symname, |
1837 | err = retval; | 1931 | i_size); |
1838 | reiserfs_check_path(&path_to_key) ; | 1932 | } |
1839 | journal_end(th, th->t_super, th->t_blocks_allocated); | 1933 | if (retval) { |
1840 | goto out_inserted_sd; | 1934 | err = retval; |
1841 | } | 1935 | reiserfs_check_path(&path_to_key); |
1842 | 1936 | journal_end(th, th->t_super, th->t_blocks_allocated); | |
1843 | /* XXX CHECK THIS */ | 1937 | goto out_inserted_sd; |
1844 | if (reiserfs_posixacl (inode->i_sb)) { | 1938 | } |
1845 | retval = reiserfs_inherit_default_acl (dir, dentry, inode); | 1939 | |
1846 | if (retval) { | 1940 | /* XXX CHECK THIS */ |
1847 | err = retval; | 1941 | if (reiserfs_posixacl(inode->i_sb)) { |
1848 | reiserfs_check_path(&path_to_key) ; | 1942 | retval = reiserfs_inherit_default_acl(dir, dentry, inode); |
1849 | journal_end(th, th->t_super, th->t_blocks_allocated); | 1943 | if (retval) { |
1850 | goto out_inserted_sd; | 1944 | err = retval; |
1851 | } | 1945 | reiserfs_check_path(&path_to_key); |
1852 | } else if (inode->i_sb->s_flags & MS_POSIXACL) { | 1946 | journal_end(th, th->t_super, th->t_blocks_allocated); |
1853 | reiserfs_warning (inode->i_sb, "ACLs aren't enabled in the fs, " | 1947 | goto out_inserted_sd; |
1854 | "but vfs thinks they are!"); | 1948 | } |
1855 | } else if (is_reiserfs_priv_object (dir)) { | 1949 | } else if (inode->i_sb->s_flags & MS_POSIXACL) { |
1856 | reiserfs_mark_inode_private (inode); | 1950 | reiserfs_warning(inode->i_sb, "ACLs aren't enabled in the fs, " |
1857 | } | 1951 | "but vfs thinks they are!"); |
1858 | 1952 | } else if (is_reiserfs_priv_object(dir)) { | |
1859 | insert_inode_hash (inode); | 1953 | reiserfs_mark_inode_private(inode); |
1860 | reiserfs_update_sd(th, inode); | 1954 | } |
1861 | reiserfs_check_path(&path_to_key) ; | 1955 | |
1862 | 1956 | insert_inode_hash(inode); | |
1863 | return 0; | 1957 | reiserfs_update_sd(th, inode); |
1958 | reiserfs_check_path(&path_to_key); | ||
1959 | |||
1960 | return 0; | ||
1864 | 1961 | ||
1865 | /* it looks like you can easily compress these two goto targets into | 1962 | /* it looks like you can easily compress these two goto targets into |
1866 | * one. Keeping it like this doesn't actually hurt anything, and they | 1963 | * one. Keeping it like this doesn't actually hurt anything, and they |
1867 | * are place holders for what the quota code actually needs. | 1964 | * are place holders for what the quota code actually needs. |
1868 | */ | 1965 | */ |
1869 | out_bad_inode: | 1966 | out_bad_inode: |
1870 | /* Invalidate the object, nothing was inserted yet */ | 1967 | /* Invalidate the object, nothing was inserted yet */ |
1871 | INODE_PKEY(inode)->k_objectid = 0; | 1968 | INODE_PKEY(inode)->k_objectid = 0; |
1872 | 1969 | ||
1873 | /* Quota change must be inside a transaction for journaling */ | 1970 | /* Quota change must be inside a transaction for journaling */ |
1874 | DQUOT_FREE_INODE(inode); | 1971 | DQUOT_FREE_INODE(inode); |
1875 | 1972 | ||
1876 | out_end_trans: | 1973 | out_end_trans: |
1877 | journal_end(th, th->t_super, th->t_blocks_allocated) ; | 1974 | journal_end(th, th->t_super, th->t_blocks_allocated); |
1878 | /* Drop can be outside and it needs more credits so it's better to have it outside */ | 1975 | /* Drop can be outside and it needs more credits so it's better to have it outside */ |
1879 | DQUOT_DROP(inode); | 1976 | DQUOT_DROP(inode); |
1880 | inode->i_flags |= S_NOQUOTA; | 1977 | inode->i_flags |= S_NOQUOTA; |
1881 | make_bad_inode(inode); | 1978 | make_bad_inode(inode); |
1882 | 1979 | ||
1883 | out_inserted_sd: | 1980 | out_inserted_sd: |
1884 | inode->i_nlink = 0; | 1981 | inode->i_nlink = 0; |
1885 | th->t_trans_id = 0; /* so the caller can't use this handle later */ | 1982 | th->t_trans_id = 0; /* so the caller can't use this handle later */ |
1886 | iput(inode); | 1983 | iput(inode); |
1887 | return err; | 1984 | return err; |
1888 | } | 1985 | } |
1889 | 1986 | ||
1890 | /* | 1987 | /* |
@@ -1900,77 +1997,78 @@ out_inserted_sd: | |||
1900 | ** | 1997 | ** |
1901 | ** on failure, nonzero is returned, page_result and bh_result are untouched. | 1998 | ** on failure, nonzero is returned, page_result and bh_result are untouched. |
1902 | */ | 1999 | */ |
1903 | static int grab_tail_page(struct inode *p_s_inode, | 2000 | static int grab_tail_page(struct inode *p_s_inode, |
1904 | struct page **page_result, | 2001 | struct page **page_result, |
1905 | struct buffer_head **bh_result) { | 2002 | struct buffer_head **bh_result) |
1906 | 2003 | { | |
1907 | /* we want the page with the last byte in the file, | 2004 | |
1908 | ** not the page that will hold the next byte for appending | 2005 | /* we want the page with the last byte in the file, |
1909 | */ | 2006 | ** not the page that will hold the next byte for appending |
1910 | unsigned long index = (p_s_inode->i_size-1) >> PAGE_CACHE_SHIFT ; | 2007 | */ |
1911 | unsigned long pos = 0 ; | 2008 | unsigned long index = (p_s_inode->i_size - 1) >> PAGE_CACHE_SHIFT; |
1912 | unsigned long start = 0 ; | 2009 | unsigned long pos = 0; |
1913 | unsigned long blocksize = p_s_inode->i_sb->s_blocksize ; | 2010 | unsigned long start = 0; |
1914 | unsigned long offset = (p_s_inode->i_size) & (PAGE_CACHE_SIZE - 1) ; | 2011 | unsigned long blocksize = p_s_inode->i_sb->s_blocksize; |
1915 | struct buffer_head *bh ; | 2012 | unsigned long offset = (p_s_inode->i_size) & (PAGE_CACHE_SIZE - 1); |
1916 | struct buffer_head *head ; | 2013 | struct buffer_head *bh; |
1917 | struct page * page ; | 2014 | struct buffer_head *head; |
1918 | int error ; | 2015 | struct page *page; |
1919 | 2016 | int error; | |
1920 | /* we know that we are only called with inode->i_size > 0. | 2017 | |
1921 | ** we also know that a file tail can never be as big as a block | 2018 | /* we know that we are only called with inode->i_size > 0. |
1922 | ** If i_size % blocksize == 0, our file is currently block aligned | 2019 | ** we also know that a file tail can never be as big as a block |
1923 | ** and it won't need converting or zeroing after a truncate. | 2020 | ** If i_size % blocksize == 0, our file is currently block aligned |
1924 | */ | 2021 | ** and it won't need converting or zeroing after a truncate. |
1925 | if ((offset & (blocksize - 1)) == 0) { | 2022 | */ |
1926 | return -ENOENT ; | 2023 | if ((offset & (blocksize - 1)) == 0) { |
1927 | } | 2024 | return -ENOENT; |
1928 | page = grab_cache_page(p_s_inode->i_mapping, index) ; | 2025 | } |
1929 | error = -ENOMEM ; | 2026 | page = grab_cache_page(p_s_inode->i_mapping, index); |
1930 | if (!page) { | 2027 | error = -ENOMEM; |
1931 | goto out ; | 2028 | if (!page) { |
1932 | } | 2029 | goto out; |
1933 | /* start within the page of the last block in the file */ | 2030 | } |
1934 | start = (offset / blocksize) * blocksize ; | 2031 | /* start within the page of the last block in the file */ |
1935 | 2032 | start = (offset / blocksize) * blocksize; | |
1936 | error = block_prepare_write(page, start, offset, | 2033 | |
1937 | reiserfs_get_block_create_0) ; | 2034 | error = block_prepare_write(page, start, offset, |
1938 | if (error) | 2035 | reiserfs_get_block_create_0); |
1939 | goto unlock ; | 2036 | if (error) |
1940 | 2037 | goto unlock; | |
1941 | head = page_buffers(page) ; | 2038 | |
1942 | bh = head; | 2039 | head = page_buffers(page); |
1943 | do { | 2040 | bh = head; |
1944 | if (pos >= start) { | 2041 | do { |
1945 | break ; | 2042 | if (pos >= start) { |
1946 | } | 2043 | break; |
1947 | bh = bh->b_this_page ; | 2044 | } |
1948 | pos += blocksize ; | 2045 | bh = bh->b_this_page; |
1949 | } while(bh != head) ; | 2046 | pos += blocksize; |
1950 | 2047 | } while (bh != head); | |
1951 | if (!buffer_uptodate(bh)) { | 2048 | |
1952 | /* note, this should never happen, prepare_write should | 2049 | if (!buffer_uptodate(bh)) { |
1953 | ** be taking care of this for us. If the buffer isn't up to date, | 2050 | /* note, this should never happen, prepare_write should |
1954 | ** I've screwed up the code to find the buffer, or the code to | 2051 | ** be taking care of this for us. If the buffer isn't up to date, |
1955 | ** call prepare_write | 2052 | ** I've screwed up the code to find the buffer, or the code to |
1956 | */ | 2053 | ** call prepare_write |
1957 | reiserfs_warning (p_s_inode->i_sb, | 2054 | */ |
1958 | "clm-6000: error reading block %lu on dev %s", | 2055 | reiserfs_warning(p_s_inode->i_sb, |
1959 | bh->b_blocknr, | 2056 | "clm-6000: error reading block %lu on dev %s", |
1960 | reiserfs_bdevname (p_s_inode->i_sb)) ; | 2057 | bh->b_blocknr, |
1961 | error = -EIO ; | 2058 | reiserfs_bdevname(p_s_inode->i_sb)); |
1962 | goto unlock ; | 2059 | error = -EIO; |
1963 | } | 2060 | goto unlock; |
1964 | *bh_result = bh ; | 2061 | } |
1965 | *page_result = page ; | 2062 | *bh_result = bh; |
1966 | 2063 | *page_result = page; | |
1967 | out: | 2064 | |
1968 | return error ; | 2065 | out: |
1969 | 2066 | return error; | |
1970 | unlock: | 2067 | |
1971 | unlock_page(page) ; | 2068 | unlock: |
1972 | page_cache_release(page) ; | 2069 | unlock_page(page); |
1973 | return error ; | 2070 | page_cache_release(page); |
2071 | return error; | ||
1974 | } | 2072 | } |
1975 | 2073 | ||
1976 | /* | 2074 | /* |
@@ -1979,235 +2077,247 @@ unlock: | |||
1979 | ** | 2077 | ** |
1980 | ** some code taken from block_truncate_page | 2078 | ** some code taken from block_truncate_page |
1981 | */ | 2079 | */ |
1982 | int reiserfs_truncate_file(struct inode *p_s_inode, int update_timestamps) { | 2080 | int reiserfs_truncate_file(struct inode *p_s_inode, int update_timestamps) |
1983 | struct reiserfs_transaction_handle th ; | 2081 | { |
1984 | /* we want the offset for the first byte after the end of the file */ | 2082 | struct reiserfs_transaction_handle th; |
1985 | unsigned long offset = p_s_inode->i_size & (PAGE_CACHE_SIZE - 1) ; | 2083 | /* we want the offset for the first byte after the end of the file */ |
1986 | unsigned blocksize = p_s_inode->i_sb->s_blocksize ; | 2084 | unsigned long offset = p_s_inode->i_size & (PAGE_CACHE_SIZE - 1); |
1987 | unsigned length ; | 2085 | unsigned blocksize = p_s_inode->i_sb->s_blocksize; |
1988 | struct page *page = NULL ; | 2086 | unsigned length; |
1989 | int error ; | 2087 | struct page *page = NULL; |
1990 | struct buffer_head *bh = NULL ; | 2088 | int error; |
1991 | 2089 | struct buffer_head *bh = NULL; | |
1992 | reiserfs_write_lock(p_s_inode->i_sb); | 2090 | |
1993 | 2091 | reiserfs_write_lock(p_s_inode->i_sb); | |
1994 | if (p_s_inode->i_size > 0) { | 2092 | |
1995 | if ((error = grab_tail_page(p_s_inode, &page, &bh))) { | 2093 | if (p_s_inode->i_size > 0) { |
1996 | // -ENOENT means we truncated past the end of the file, | 2094 | if ((error = grab_tail_page(p_s_inode, &page, &bh))) { |
1997 | // and get_block_create_0 could not find a block to read in, | 2095 | // -ENOENT means we truncated past the end of the file, |
1998 | // which is ok. | 2096 | // and get_block_create_0 could not find a block to read in, |
1999 | if (error != -ENOENT) | 2097 | // which is ok. |
2000 | reiserfs_warning (p_s_inode->i_sb, | 2098 | if (error != -ENOENT) |
2001 | "clm-6001: grab_tail_page failed %d", | 2099 | reiserfs_warning(p_s_inode->i_sb, |
2002 | error); | 2100 | "clm-6001: grab_tail_page failed %d", |
2003 | page = NULL ; | 2101 | error); |
2004 | bh = NULL ; | 2102 | page = NULL; |
2005 | } | 2103 | bh = NULL; |
2006 | } | 2104 | } |
2007 | 2105 | } | |
2008 | /* so, if page != NULL, we have a buffer head for the offset at | ||
2009 | ** the end of the file. if the bh is mapped, and bh->b_blocknr != 0, | ||
2010 | ** then we have an unformatted node. Otherwise, we have a direct item, | ||
2011 | ** and no zeroing is required on disk. We zero after the truncate, | ||
2012 | ** because the truncate might pack the item anyway | ||
2013 | ** (it will unmap bh if it packs). | ||
2014 | */ | ||
2015 | /* it is enough to reserve space in transaction for 2 balancings: | ||
2016 | one for "save" link adding and another for the first | ||
2017 | cut_from_item. 1 is for update_sd */ | ||
2018 | error = journal_begin (&th, p_s_inode->i_sb, | ||
2019 | JOURNAL_PER_BALANCE_CNT * 2 + 1); | ||
2020 | if (error) | ||
2021 | goto out; | ||
2022 | reiserfs_update_inode_transaction(p_s_inode) ; | ||
2023 | if (update_timestamps) | ||
2024 | /* we are doing real truncate: if the system crashes before the last | ||
2025 | transaction of truncating gets committed - on reboot the file | ||
2026 | either appears truncated properly or not truncated at all */ | ||
2027 | add_save_link (&th, p_s_inode, 1); | ||
2028 | error = reiserfs_do_truncate (&th, p_s_inode, page, update_timestamps) ; | ||
2029 | if (error) | ||
2030 | goto out; | ||
2031 | error = journal_end (&th, p_s_inode->i_sb, JOURNAL_PER_BALANCE_CNT * 2 + 1); | ||
2032 | if (error) | ||
2033 | goto out; | ||
2034 | |||
2035 | if (update_timestamps) { | ||
2036 | error = remove_save_link (p_s_inode, 1/* truncate */); | ||
2037 | if (error) | ||
2038 | goto out; | ||
2039 | } | ||
2040 | |||
2041 | if (page) { | ||
2042 | length = offset & (blocksize - 1) ; | ||
2043 | /* if we are not on a block boundary */ | ||
2044 | if (length) { | ||
2045 | char *kaddr; | ||
2046 | |||
2047 | length = blocksize - length ; | ||
2048 | kaddr = kmap_atomic(page, KM_USER0) ; | ||
2049 | memset(kaddr + offset, 0, length) ; | ||
2050 | flush_dcache_page(page) ; | ||
2051 | kunmap_atomic(kaddr, KM_USER0) ; | ||
2052 | if (buffer_mapped(bh) && bh->b_blocknr != 0) { | ||
2053 | mark_buffer_dirty(bh) ; | ||
2054 | } | ||
2055 | } | ||
2056 | unlock_page(page) ; | ||
2057 | page_cache_release(page) ; | ||
2058 | } | ||
2059 | |||
2060 | reiserfs_write_unlock(p_s_inode->i_sb); | ||
2061 | return 0; | ||
2062 | out: | ||
2063 | if (page) { | ||
2064 | unlock_page (page); | ||
2065 | page_cache_release (page); | ||
2066 | } | ||
2067 | reiserfs_write_unlock(p_s_inode->i_sb); | ||
2068 | return error; | ||
2069 | } | ||
2070 | 2106 | ||
2071 | static int map_block_for_writepage(struct inode *inode, | 2107 | /* so, if page != NULL, we have a buffer head for the offset at |
2072 | struct buffer_head *bh_result, | 2108 | ** the end of the file. if the bh is mapped, and bh->b_blocknr != 0, |
2073 | unsigned long block) { | 2109 | ** then we have an unformatted node. Otherwise, we have a direct item, |
2074 | struct reiserfs_transaction_handle th ; | 2110 | ** and no zeroing is required on disk. We zero after the truncate, |
2075 | int fs_gen ; | 2111 | ** because the truncate might pack the item anyway |
2076 | struct item_head tmp_ih ; | 2112 | ** (it will unmap bh if it packs). |
2077 | struct item_head *ih ; | ||
2078 | struct buffer_head *bh ; | ||
2079 | __le32 *item ; | ||
2080 | struct cpu_key key ; | ||
2081 | INITIALIZE_PATH(path) ; | ||
2082 | int pos_in_item ; | ||
2083 | int jbegin_count = JOURNAL_PER_BALANCE_CNT ; | ||
2084 | loff_t byte_offset = (block << inode->i_sb->s_blocksize_bits) + 1 ; | ||
2085 | int retval ; | ||
2086 | int use_get_block = 0 ; | ||
2087 | int bytes_copied = 0 ; | ||
2088 | int copy_size ; | ||
2089 | int trans_running = 0; | ||
2090 | |||
2091 | /* catch places below that try to log something without starting a trans */ | ||
2092 | th.t_trans_id = 0; | ||
2093 | |||
2094 | if (!buffer_uptodate(bh_result)) { | ||
2095 | return -EIO; | ||
2096 | } | ||
2097 | |||
2098 | kmap(bh_result->b_page) ; | ||
2099 | start_over: | ||
2100 | reiserfs_write_lock(inode->i_sb); | ||
2101 | make_cpu_key(&key, inode, byte_offset, TYPE_ANY, 3) ; | ||
2102 | |||
2103 | research: | ||
2104 | retval = search_for_position_by_key(inode->i_sb, &key, &path) ; | ||
2105 | if (retval != POSITION_FOUND) { | ||
2106 | use_get_block = 1; | ||
2107 | goto out ; | ||
2108 | } | ||
2109 | |||
2110 | bh = get_last_bh(&path) ; | ||
2111 | ih = get_ih(&path) ; | ||
2112 | item = get_item(&path) ; | ||
2113 | pos_in_item = path.pos_in_item ; | ||
2114 | |||
2115 | /* we've found an unformatted node */ | ||
2116 | if (indirect_item_found(retval, ih)) { | ||
2117 | if (bytes_copied > 0) { | ||
2118 | reiserfs_warning (inode->i_sb, "clm-6002: bytes_copied %d", | ||
2119 | bytes_copied) ; | ||
2120 | } | ||
2121 | if (!get_block_num(item, pos_in_item)) { | ||
2122 | /* crap, we are writing to a hole */ | ||
2123 | use_get_block = 1; | ||
2124 | goto out ; | ||
2125 | } | ||
2126 | set_block_dev_mapped(bh_result, get_block_num(item,pos_in_item),inode); | ||
2127 | } else if (is_direct_le_ih(ih)) { | ||
2128 | char *p ; | ||
2129 | p = page_address(bh_result->b_page) ; | ||
2130 | p += (byte_offset -1) & (PAGE_CACHE_SIZE - 1) ; | ||
2131 | copy_size = ih_item_len(ih) - pos_in_item; | ||
2132 | |||
2133 | fs_gen = get_generation(inode->i_sb) ; | ||
2134 | copy_item_head(&tmp_ih, ih) ; | ||
2135 | |||
2136 | if (!trans_running) { | ||
2137 | /* vs-3050 is gone, no need to drop the path */ | ||
2138 | retval = journal_begin(&th, inode->i_sb, jbegin_count) ; | ||
2139 | if (retval) | ||
2140 | goto out; | ||
2141 | reiserfs_update_inode_transaction(inode) ; | ||
2142 | trans_running = 1; | ||
2143 | if (fs_changed(fs_gen, inode->i_sb) && item_moved(&tmp_ih, &path)) { | ||
2144 | reiserfs_restore_prepared_buffer(inode->i_sb, bh) ; | ||
2145 | goto research; | ||
2146 | } | ||
2147 | } | ||
2148 | |||
2149 | reiserfs_prepare_for_journal(inode->i_sb, bh, 1) ; | ||
2150 | |||
2151 | if (fs_changed (fs_gen, inode->i_sb) && item_moved (&tmp_ih, &path)) { | ||
2152 | reiserfs_restore_prepared_buffer(inode->i_sb, bh) ; | ||
2153 | goto research; | ||
2154 | } | ||
2155 | |||
2156 | memcpy( B_I_PITEM(bh, ih) + pos_in_item, p + bytes_copied, copy_size) ; | ||
2157 | |||
2158 | journal_mark_dirty(&th, inode->i_sb, bh) ; | ||
2159 | bytes_copied += copy_size ; | ||
2160 | set_block_dev_mapped(bh_result, 0, inode); | ||
2161 | |||
2162 | /* are there still bytes left? */ | ||
2163 | if (bytes_copied < bh_result->b_size && | ||
2164 | (byte_offset + bytes_copied) < inode->i_size) { | ||
2165 | set_cpu_key_k_offset(&key, cpu_key_k_offset(&key) + copy_size) ; | ||
2166 | goto research ; | ||
2167 | } | ||
2168 | } else { | ||
2169 | reiserfs_warning (inode->i_sb, | ||
2170 | "clm-6003: bad item inode %lu, device %s", | ||
2171 | inode->i_ino, reiserfs_bdevname (inode->i_sb)) ; | ||
2172 | retval = -EIO ; | ||
2173 | goto out ; | ||
2174 | } | ||
2175 | retval = 0 ; | ||
2176 | |||
2177 | out: | ||
2178 | pathrelse(&path) ; | ||
2179 | if (trans_running) { | ||
2180 | int err = journal_end(&th, inode->i_sb, jbegin_count) ; | ||
2181 | if (err) | ||
2182 | retval = err; | ||
2183 | trans_running = 0; | ||
2184 | } | ||
2185 | reiserfs_write_unlock(inode->i_sb); | ||
2186 | |||
2187 | /* this is where we fill in holes in the file. */ | ||
2188 | if (use_get_block) { | ||
2189 | retval = reiserfs_get_block(inode, block, bh_result, | ||
2190 | GET_BLOCK_CREATE | GET_BLOCK_NO_ISEM | | ||
2191 | GET_BLOCK_NO_DANGLE); | ||
2192 | if (!retval) { | ||
2193 | if (!buffer_mapped(bh_result) || bh_result->b_blocknr == 0) { | ||
2194 | /* get_block failed to find a mapped unformatted node. */ | ||
2195 | use_get_block = 0 ; | ||
2196 | goto start_over ; | ||
2197 | } | ||
2198 | } | ||
2199 | } | ||
2200 | kunmap(bh_result->b_page) ; | ||
2201 | |||
2202 | if (!retval && buffer_mapped(bh_result) && bh_result->b_blocknr == 0) { | ||
2203 | /* we've copied data from the page into the direct item, so the | ||
2204 | * buffer in the page is now clean, mark it to reflect that. | ||
2205 | */ | 2113 | */ |
2206 | lock_buffer(bh_result); | 2114 | /* it is enough to reserve space in transaction for 2 balancings: |
2207 | clear_buffer_dirty(bh_result); | 2115 | one for "save" link adding and another for the first |
2208 | unlock_buffer(bh_result); | 2116 | cut_from_item. 1 is for update_sd */ |
2209 | } | 2117 | error = journal_begin(&th, p_s_inode->i_sb, |
2210 | return retval ; | 2118 | JOURNAL_PER_BALANCE_CNT * 2 + 1); |
2119 | if (error) | ||
2120 | goto out; | ||
2121 | reiserfs_update_inode_transaction(p_s_inode); | ||
2122 | if (update_timestamps) | ||
2123 | /* we are doing real truncate: if the system crashes before the last | ||
2124 | transaction of truncating gets committed - on reboot the file | ||
2125 | either appears truncated properly or not truncated at all */ | ||
2126 | add_save_link(&th, p_s_inode, 1); | ||
2127 | error = reiserfs_do_truncate(&th, p_s_inode, page, update_timestamps); | ||
2128 | if (error) | ||
2129 | goto out; | ||
2130 | error = | ||
2131 | journal_end(&th, p_s_inode->i_sb, JOURNAL_PER_BALANCE_CNT * 2 + 1); | ||
2132 | if (error) | ||
2133 | goto out; | ||
2134 | |||
2135 | if (update_timestamps) { | ||
2136 | error = remove_save_link(p_s_inode, 1 /* truncate */ ); | ||
2137 | if (error) | ||
2138 | goto out; | ||
2139 | } | ||
2140 | |||
2141 | if (page) { | ||
2142 | length = offset & (blocksize - 1); | ||
2143 | /* if we are not on a block boundary */ | ||
2144 | if (length) { | ||
2145 | char *kaddr; | ||
2146 | |||
2147 | length = blocksize - length; | ||
2148 | kaddr = kmap_atomic(page, KM_USER0); | ||
2149 | memset(kaddr + offset, 0, length); | ||
2150 | flush_dcache_page(page); | ||
2151 | kunmap_atomic(kaddr, KM_USER0); | ||
2152 | if (buffer_mapped(bh) && bh->b_blocknr != 0) { | ||
2153 | mark_buffer_dirty(bh); | ||
2154 | } | ||
2155 | } | ||
2156 | unlock_page(page); | ||
2157 | page_cache_release(page); | ||
2158 | } | ||
2159 | |||
2160 | reiserfs_write_unlock(p_s_inode->i_sb); | ||
2161 | return 0; | ||
2162 | out: | ||
2163 | if (page) { | ||
2164 | unlock_page(page); | ||
2165 | page_cache_release(page); | ||
2166 | } | ||
2167 | reiserfs_write_unlock(p_s_inode->i_sb); | ||
2168 | return error; | ||
2169 | } | ||
2170 | |||
2171 | static int map_block_for_writepage(struct inode *inode, | ||
2172 | struct buffer_head *bh_result, | ||
2173 | unsigned long block) | ||
2174 | { | ||
2175 | struct reiserfs_transaction_handle th; | ||
2176 | int fs_gen; | ||
2177 | struct item_head tmp_ih; | ||
2178 | struct item_head *ih; | ||
2179 | struct buffer_head *bh; | ||
2180 | __le32 *item; | ||
2181 | struct cpu_key key; | ||
2182 | INITIALIZE_PATH(path); | ||
2183 | int pos_in_item; | ||
2184 | int jbegin_count = JOURNAL_PER_BALANCE_CNT; | ||
2185 | loff_t byte_offset = (block << inode->i_sb->s_blocksize_bits) + 1; | ||
2186 | int retval; | ||
2187 | int use_get_block = 0; | ||
2188 | int bytes_copied = 0; | ||
2189 | int copy_size; | ||
2190 | int trans_running = 0; | ||
2191 | |||
2192 | /* catch places below that try to log something without starting a trans */ | ||
2193 | th.t_trans_id = 0; | ||
2194 | |||
2195 | if (!buffer_uptodate(bh_result)) { | ||
2196 | return -EIO; | ||
2197 | } | ||
2198 | |||
2199 | kmap(bh_result->b_page); | ||
2200 | start_over: | ||
2201 | reiserfs_write_lock(inode->i_sb); | ||
2202 | make_cpu_key(&key, inode, byte_offset, TYPE_ANY, 3); | ||
2203 | |||
2204 | research: | ||
2205 | retval = search_for_position_by_key(inode->i_sb, &key, &path); | ||
2206 | if (retval != POSITION_FOUND) { | ||
2207 | use_get_block = 1; | ||
2208 | goto out; | ||
2209 | } | ||
2210 | |||
2211 | bh = get_last_bh(&path); | ||
2212 | ih = get_ih(&path); | ||
2213 | item = get_item(&path); | ||
2214 | pos_in_item = path.pos_in_item; | ||
2215 | |||
2216 | /* we've found an unformatted node */ | ||
2217 | if (indirect_item_found(retval, ih)) { | ||
2218 | if (bytes_copied > 0) { | ||
2219 | reiserfs_warning(inode->i_sb, | ||
2220 | "clm-6002: bytes_copied %d", | ||
2221 | bytes_copied); | ||
2222 | } | ||
2223 | if (!get_block_num(item, pos_in_item)) { | ||
2224 | /* crap, we are writing to a hole */ | ||
2225 | use_get_block = 1; | ||
2226 | goto out; | ||
2227 | } | ||
2228 | set_block_dev_mapped(bh_result, | ||
2229 | get_block_num(item, pos_in_item), inode); | ||
2230 | } else if (is_direct_le_ih(ih)) { | ||
2231 | char *p; | ||
2232 | p = page_address(bh_result->b_page); | ||
2233 | p += (byte_offset - 1) & (PAGE_CACHE_SIZE - 1); | ||
2234 | copy_size = ih_item_len(ih) - pos_in_item; | ||
2235 | |||
2236 | fs_gen = get_generation(inode->i_sb); | ||
2237 | copy_item_head(&tmp_ih, ih); | ||
2238 | |||
2239 | if (!trans_running) { | ||
2240 | /* vs-3050 is gone, no need to drop the path */ | ||
2241 | retval = journal_begin(&th, inode->i_sb, jbegin_count); | ||
2242 | if (retval) | ||
2243 | goto out; | ||
2244 | reiserfs_update_inode_transaction(inode); | ||
2245 | trans_running = 1; | ||
2246 | if (fs_changed(fs_gen, inode->i_sb) | ||
2247 | && item_moved(&tmp_ih, &path)) { | ||
2248 | reiserfs_restore_prepared_buffer(inode->i_sb, | ||
2249 | bh); | ||
2250 | goto research; | ||
2251 | } | ||
2252 | } | ||
2253 | |||
2254 | reiserfs_prepare_for_journal(inode->i_sb, bh, 1); | ||
2255 | |||
2256 | if (fs_changed(fs_gen, inode->i_sb) | ||
2257 | && item_moved(&tmp_ih, &path)) { | ||
2258 | reiserfs_restore_prepared_buffer(inode->i_sb, bh); | ||
2259 | goto research; | ||
2260 | } | ||
2261 | |||
2262 | memcpy(B_I_PITEM(bh, ih) + pos_in_item, p + bytes_copied, | ||
2263 | copy_size); | ||
2264 | |||
2265 | journal_mark_dirty(&th, inode->i_sb, bh); | ||
2266 | bytes_copied += copy_size; | ||
2267 | set_block_dev_mapped(bh_result, 0, inode); | ||
2268 | |||
2269 | /* are there still bytes left? */ | ||
2270 | if (bytes_copied < bh_result->b_size && | ||
2271 | (byte_offset + bytes_copied) < inode->i_size) { | ||
2272 | set_cpu_key_k_offset(&key, | ||
2273 | cpu_key_k_offset(&key) + | ||
2274 | copy_size); | ||
2275 | goto research; | ||
2276 | } | ||
2277 | } else { | ||
2278 | reiserfs_warning(inode->i_sb, | ||
2279 | "clm-6003: bad item inode %lu, device %s", | ||
2280 | inode->i_ino, reiserfs_bdevname(inode->i_sb)); | ||
2281 | retval = -EIO; | ||
2282 | goto out; | ||
2283 | } | ||
2284 | retval = 0; | ||
2285 | |||
2286 | out: | ||
2287 | pathrelse(&path); | ||
2288 | if (trans_running) { | ||
2289 | int err = journal_end(&th, inode->i_sb, jbegin_count); | ||
2290 | if (err) | ||
2291 | retval = err; | ||
2292 | trans_running = 0; | ||
2293 | } | ||
2294 | reiserfs_write_unlock(inode->i_sb); | ||
2295 | |||
2296 | /* this is where we fill in holes in the file. */ | ||
2297 | if (use_get_block) { | ||
2298 | retval = reiserfs_get_block(inode, block, bh_result, | ||
2299 | GET_BLOCK_CREATE | GET_BLOCK_NO_ISEM | ||
2300 | | GET_BLOCK_NO_DANGLE); | ||
2301 | if (!retval) { | ||
2302 | if (!buffer_mapped(bh_result) | ||
2303 | || bh_result->b_blocknr == 0) { | ||
2304 | /* get_block failed to find a mapped unformatted node. */ | ||
2305 | use_get_block = 0; | ||
2306 | goto start_over; | ||
2307 | } | ||
2308 | } | ||
2309 | } | ||
2310 | kunmap(bh_result->b_page); | ||
2311 | |||
2312 | if (!retval && buffer_mapped(bh_result) && bh_result->b_blocknr == 0) { | ||
2313 | /* we've copied data from the page into the direct item, so the | ||
2314 | * buffer in the page is now clean, mark it to reflect that. | ||
2315 | */ | ||
2316 | lock_buffer(bh_result); | ||
2317 | clear_buffer_dirty(bh_result); | ||
2318 | unlock_buffer(bh_result); | ||
2319 | } | ||
2320 | return retval; | ||
2211 | } | 2321 | } |
2212 | 2322 | ||
2213 | /* | 2323 | /* |
@@ -2215,383 +2325,390 @@ out: | |||
2215 | * start/recovery path as __block_write_full_page, along with special | 2325 | * start/recovery path as __block_write_full_page, along with special |
2216 | * code to handle reiserfs tails. | 2326 | * code to handle reiserfs tails. |
2217 | */ | 2327 | */ |
2218 | static int reiserfs_write_full_page(struct page *page, struct writeback_control *wbc) { | 2328 | static int reiserfs_write_full_page(struct page *page, |
2219 | struct inode *inode = page->mapping->host ; | 2329 | struct writeback_control *wbc) |
2220 | unsigned long end_index = inode->i_size >> PAGE_CACHE_SHIFT ; | 2330 | { |
2221 | int error = 0; | 2331 | struct inode *inode = page->mapping->host; |
2222 | unsigned long block ; | 2332 | unsigned long end_index = inode->i_size >> PAGE_CACHE_SHIFT; |
2223 | struct buffer_head *head, *bh; | 2333 | int error = 0; |
2224 | int partial = 0 ; | 2334 | unsigned long block; |
2225 | int nr = 0; | 2335 | struct buffer_head *head, *bh; |
2226 | int checked = PageChecked(page); | 2336 | int partial = 0; |
2227 | struct reiserfs_transaction_handle th; | 2337 | int nr = 0; |
2228 | struct super_block *s = inode->i_sb; | 2338 | int checked = PageChecked(page); |
2229 | int bh_per_page = PAGE_CACHE_SIZE / s->s_blocksize; | 2339 | struct reiserfs_transaction_handle th; |
2230 | th.t_trans_id = 0; | 2340 | struct super_block *s = inode->i_sb; |
2231 | 2341 | int bh_per_page = PAGE_CACHE_SIZE / s->s_blocksize; | |
2232 | /* The page dirty bit is cleared before writepage is called, which | 2342 | th.t_trans_id = 0; |
2233 | * means we have to tell create_empty_buffers to make dirty buffers | 2343 | |
2234 | * The page really should be up to date at this point, so tossing | 2344 | /* The page dirty bit is cleared before writepage is called, which |
2235 | * in the BH_Uptodate is just a sanity check. | 2345 | * means we have to tell create_empty_buffers to make dirty buffers |
2236 | */ | 2346 | * The page really should be up to date at this point, so tossing |
2237 | if (!page_has_buffers(page)) { | 2347 | * in the BH_Uptodate is just a sanity check. |
2238 | create_empty_buffers(page, s->s_blocksize, | 2348 | */ |
2239 | (1 << BH_Dirty) | (1 << BH_Uptodate)); | 2349 | if (!page_has_buffers(page)) { |
2240 | } | 2350 | create_empty_buffers(page, s->s_blocksize, |
2241 | head = page_buffers(page) ; | 2351 | (1 << BH_Dirty) | (1 << BH_Uptodate)); |
2242 | 2352 | } | |
2243 | /* last page in the file, zero out any contents past the | 2353 | head = page_buffers(page); |
2244 | ** last byte in the file | ||
2245 | */ | ||
2246 | if (page->index >= end_index) { | ||
2247 | char *kaddr; | ||
2248 | unsigned last_offset; | ||
2249 | |||
2250 | last_offset = inode->i_size & (PAGE_CACHE_SIZE - 1) ; | ||
2251 | /* no file contents in this page */ | ||
2252 | if (page->index >= end_index + 1 || !last_offset) { | ||
2253 | unlock_page(page); | ||
2254 | return 0; | ||
2255 | } | ||
2256 | kaddr = kmap_atomic(page, KM_USER0); | ||
2257 | memset(kaddr + last_offset, 0, PAGE_CACHE_SIZE-last_offset) ; | ||
2258 | flush_dcache_page(page) ; | ||
2259 | kunmap_atomic(kaddr, KM_USER0) ; | ||
2260 | } | ||
2261 | bh = head ; | ||
2262 | block = page->index << (PAGE_CACHE_SHIFT - s->s_blocksize_bits) ; | ||
2263 | /* first map all the buffers, logging any direct items we find */ | ||
2264 | do { | ||
2265 | if ((checked || buffer_dirty(bh)) && (!buffer_mapped(bh) || | ||
2266 | (buffer_mapped(bh) && bh->b_blocknr == 0))) { | ||
2267 | /* not mapped yet, or it points to a direct item, search | ||
2268 | * the btree for the mapping info, and log any direct | ||
2269 | * items found | ||
2270 | */ | ||
2271 | if ((error = map_block_for_writepage(inode, bh, block))) { | ||
2272 | goto fail ; | ||
2273 | } | ||
2274 | } | ||
2275 | bh = bh->b_this_page; | ||
2276 | block++; | ||
2277 | } while(bh != head) ; | ||
2278 | |||
2279 | /* | ||
2280 | * we start the transaction after map_block_for_writepage, | ||
2281 | * because it can create holes in the file (an unbounded operation). | ||
2282 | * starting it here, we can make a reliable estimate for how many | ||
2283 | * blocks we're going to log | ||
2284 | */ | ||
2285 | if (checked) { | ||
2286 | ClearPageChecked(page); | ||
2287 | reiserfs_write_lock(s); | ||
2288 | error = journal_begin(&th, s, bh_per_page + 1); | ||
2289 | if (error) { | ||
2290 | reiserfs_write_unlock(s); | ||
2291 | goto fail; | ||
2292 | } | ||
2293 | reiserfs_update_inode_transaction(inode); | ||
2294 | } | ||
2295 | /* now go through and lock any dirty buffers on the page */ | ||
2296 | do { | ||
2297 | get_bh(bh); | ||
2298 | if (!buffer_mapped(bh)) | ||
2299 | continue; | ||
2300 | if (buffer_mapped(bh) && bh->b_blocknr == 0) | ||
2301 | continue; | ||
2302 | 2354 | ||
2303 | if (checked) { | 2355 | /* last page in the file, zero out any contents past the |
2304 | reiserfs_prepare_for_journal(s, bh, 1); | 2356 | ** last byte in the file |
2305 | journal_mark_dirty(&th, s, bh); | 2357 | */ |
2306 | continue; | 2358 | if (page->index >= end_index) { |
2359 | char *kaddr; | ||
2360 | unsigned last_offset; | ||
2361 | |||
2362 | last_offset = inode->i_size & (PAGE_CACHE_SIZE - 1); | ||
2363 | /* no file contents in this page */ | ||
2364 | if (page->index >= end_index + 1 || !last_offset) { | ||
2365 | unlock_page(page); | ||
2366 | return 0; | ||
2367 | } | ||
2368 | kaddr = kmap_atomic(page, KM_USER0); | ||
2369 | memset(kaddr + last_offset, 0, PAGE_CACHE_SIZE - last_offset); | ||
2370 | flush_dcache_page(page); | ||
2371 | kunmap_atomic(kaddr, KM_USER0); | ||
2307 | } | 2372 | } |
2308 | /* from this point on, we know the buffer is mapped to a | 2373 | bh = head; |
2309 | * real block and not a direct item | 2374 | block = page->index << (PAGE_CACHE_SHIFT - s->s_blocksize_bits); |
2375 | /* first map all the buffers, logging any direct items we find */ | ||
2376 | do { | ||
2377 | if ((checked || buffer_dirty(bh)) && (!buffer_mapped(bh) || | ||
2378 | (buffer_mapped(bh) | ||
2379 | && bh->b_blocknr == | ||
2380 | 0))) { | ||
2381 | /* not mapped yet, or it points to a direct item, search | ||
2382 | * the btree for the mapping info, and log any direct | ||
2383 | * items found | ||
2384 | */ | ||
2385 | if ((error = map_block_for_writepage(inode, bh, block))) { | ||
2386 | goto fail; | ||
2387 | } | ||
2388 | } | ||
2389 | bh = bh->b_this_page; | ||
2390 | block++; | ||
2391 | } while (bh != head); | ||
2392 | |||
2393 | /* | ||
2394 | * we start the transaction after map_block_for_writepage, | ||
2395 | * because it can create holes in the file (an unbounded operation). | ||
2396 | * starting it here, we can make a reliable estimate for how many | ||
2397 | * blocks we're going to log | ||
2310 | */ | 2398 | */ |
2311 | if (wbc->sync_mode != WB_SYNC_NONE || !wbc->nonblocking) { | 2399 | if (checked) { |
2312 | lock_buffer(bh); | 2400 | ClearPageChecked(page); |
2313 | } else { | 2401 | reiserfs_write_lock(s); |
2314 | if (test_set_buffer_locked(bh)) { | 2402 | error = journal_begin(&th, s, bh_per_page + 1); |
2315 | redirty_page_for_writepage(wbc, page); | 2403 | if (error) { |
2316 | continue; | 2404 | reiserfs_write_unlock(s); |
2317 | } | 2405 | goto fail; |
2406 | } | ||
2407 | reiserfs_update_inode_transaction(inode); | ||
2318 | } | 2408 | } |
2319 | if (test_clear_buffer_dirty(bh)) { | 2409 | /* now go through and lock any dirty buffers on the page */ |
2320 | mark_buffer_async_write(bh); | 2410 | do { |
2321 | } else { | 2411 | get_bh(bh); |
2322 | unlock_buffer(bh); | 2412 | if (!buffer_mapped(bh)) |
2413 | continue; | ||
2414 | if (buffer_mapped(bh) && bh->b_blocknr == 0) | ||
2415 | continue; | ||
2416 | |||
2417 | if (checked) { | ||
2418 | reiserfs_prepare_for_journal(s, bh, 1); | ||
2419 | journal_mark_dirty(&th, s, bh); | ||
2420 | continue; | ||
2421 | } | ||
2422 | /* from this point on, we know the buffer is mapped to a | ||
2423 | * real block and not a direct item | ||
2424 | */ | ||
2425 | if (wbc->sync_mode != WB_SYNC_NONE || !wbc->nonblocking) { | ||
2426 | lock_buffer(bh); | ||
2427 | } else { | ||
2428 | if (test_set_buffer_locked(bh)) { | ||
2429 | redirty_page_for_writepage(wbc, page); | ||
2430 | continue; | ||
2431 | } | ||
2432 | } | ||
2433 | if (test_clear_buffer_dirty(bh)) { | ||
2434 | mark_buffer_async_write(bh); | ||
2435 | } else { | ||
2436 | unlock_buffer(bh); | ||
2437 | } | ||
2438 | } while ((bh = bh->b_this_page) != head); | ||
2439 | |||
2440 | if (checked) { | ||
2441 | error = journal_end(&th, s, bh_per_page + 1); | ||
2442 | reiserfs_write_unlock(s); | ||
2443 | if (error) | ||
2444 | goto fail; | ||
2323 | } | 2445 | } |
2324 | } while((bh = bh->b_this_page) != head); | 2446 | BUG_ON(PageWriteback(page)); |
2447 | set_page_writeback(page); | ||
2448 | unlock_page(page); | ||
2325 | 2449 | ||
2326 | if (checked) { | 2450 | /* |
2327 | error = journal_end(&th, s, bh_per_page + 1); | 2451 | * since any buffer might be the only dirty buffer on the page, |
2328 | reiserfs_write_unlock(s); | 2452 | * the first submit_bh can bring the page out of writeback. |
2329 | if (error) | 2453 | * be careful with the buffers. |
2330 | goto fail; | ||
2331 | } | ||
2332 | BUG_ON(PageWriteback(page)); | ||
2333 | set_page_writeback(page); | ||
2334 | unlock_page(page); | ||
2335 | |||
2336 | /* | ||
2337 | * since any buffer might be the only dirty buffer on the page, | ||
2338 | * the first submit_bh can bring the page out of writeback. | ||
2339 | * be careful with the buffers. | ||
2340 | */ | ||
2341 | do { | ||
2342 | struct buffer_head *next = bh->b_this_page; | ||
2343 | if (buffer_async_write(bh)) { | ||
2344 | submit_bh(WRITE, bh); | ||
2345 | nr++; | ||
2346 | } | ||
2347 | put_bh(bh); | ||
2348 | bh = next; | ||
2349 | } while(bh != head); | ||
2350 | |||
2351 | error = 0; | ||
2352 | done: | ||
2353 | if (nr == 0) { | ||
2354 | /* | ||
2355 | * if this page only had a direct item, it is very possible for | ||
2356 | * no io to be required without there being an error. Or, | ||
2357 | * someone else could have locked them and sent them down the | ||
2358 | * pipe without locking the page | ||
2359 | */ | 2454 | */ |
2360 | bh = head ; | ||
2361 | do { | 2455 | do { |
2362 | if (!buffer_uptodate(bh)) { | 2456 | struct buffer_head *next = bh->b_this_page; |
2363 | partial = 1; | 2457 | if (buffer_async_write(bh)) { |
2364 | break; | 2458 | submit_bh(WRITE, bh); |
2365 | } | 2459 | nr++; |
2366 | bh = bh->b_this_page; | 2460 | } |
2367 | } while(bh != head); | 2461 | put_bh(bh); |
2368 | if (!partial) | 2462 | bh = next; |
2369 | SetPageUptodate(page); | 2463 | } while (bh != head); |
2370 | end_page_writeback(page); | ||
2371 | } | ||
2372 | return error; | ||
2373 | |||
2374 | fail: | ||
2375 | /* catches various errors, we need to make sure any valid dirty blocks | ||
2376 | * get to the media. The page is currently locked and not marked for | ||
2377 | * writeback | ||
2378 | */ | ||
2379 | ClearPageUptodate(page); | ||
2380 | bh = head; | ||
2381 | do { | ||
2382 | get_bh(bh); | ||
2383 | if (buffer_mapped(bh) && buffer_dirty(bh) && bh->b_blocknr) { | ||
2384 | lock_buffer(bh); | ||
2385 | mark_buffer_async_write(bh); | ||
2386 | } else { | ||
2387 | /* | ||
2388 | * clear any dirty bits that might have come from getting | ||
2389 | * attached to a dirty page | ||
2390 | */ | ||
2391 | clear_buffer_dirty(bh); | ||
2392 | } | ||
2393 | bh = bh->b_this_page; | ||
2394 | } while(bh != head); | ||
2395 | SetPageError(page); | ||
2396 | BUG_ON(PageWriteback(page)); | ||
2397 | set_page_writeback(page); | ||
2398 | unlock_page(page); | ||
2399 | do { | ||
2400 | struct buffer_head *next = bh->b_this_page; | ||
2401 | if (buffer_async_write(bh)) { | ||
2402 | clear_buffer_dirty(bh); | ||
2403 | submit_bh(WRITE, bh); | ||
2404 | nr++; | ||
2405 | } | ||
2406 | put_bh(bh); | ||
2407 | bh = next; | ||
2408 | } while(bh != head); | ||
2409 | goto done; | ||
2410 | } | ||
2411 | 2464 | ||
2465 | error = 0; | ||
2466 | done: | ||
2467 | if (nr == 0) { | ||
2468 | /* | ||
2469 | * if this page only had a direct item, it is very possible for | ||
2470 | * no io to be required without there being an error. Or, | ||
2471 | * someone else could have locked them and sent them down the | ||
2472 | * pipe without locking the page | ||
2473 | */ | ||
2474 | bh = head; | ||
2475 | do { | ||
2476 | if (!buffer_uptodate(bh)) { | ||
2477 | partial = 1; | ||
2478 | break; | ||
2479 | } | ||
2480 | bh = bh->b_this_page; | ||
2481 | } while (bh != head); | ||
2482 | if (!partial) | ||
2483 | SetPageUptodate(page); | ||
2484 | end_page_writeback(page); | ||
2485 | } | ||
2486 | return error; | ||
2412 | 2487 | ||
2413 | static int reiserfs_readpage (struct file *f, struct page * page) | 2488 | fail: |
2414 | { | 2489 | /* catches various errors, we need to make sure any valid dirty blocks |
2415 | return block_read_full_page (page, reiserfs_get_block); | 2490 | * get to the media. The page is currently locked and not marked for |
2491 | * writeback | ||
2492 | */ | ||
2493 | ClearPageUptodate(page); | ||
2494 | bh = head; | ||
2495 | do { | ||
2496 | get_bh(bh); | ||
2497 | if (buffer_mapped(bh) && buffer_dirty(bh) && bh->b_blocknr) { | ||
2498 | lock_buffer(bh); | ||
2499 | mark_buffer_async_write(bh); | ||
2500 | } else { | ||
2501 | /* | ||
2502 | * clear any dirty bits that might have come from getting | ||
2503 | * attached to a dirty page | ||
2504 | */ | ||
2505 | clear_buffer_dirty(bh); | ||
2506 | } | ||
2507 | bh = bh->b_this_page; | ||
2508 | } while (bh != head); | ||
2509 | SetPageError(page); | ||
2510 | BUG_ON(PageWriteback(page)); | ||
2511 | set_page_writeback(page); | ||
2512 | unlock_page(page); | ||
2513 | do { | ||
2514 | struct buffer_head *next = bh->b_this_page; | ||
2515 | if (buffer_async_write(bh)) { | ||
2516 | clear_buffer_dirty(bh); | ||
2517 | submit_bh(WRITE, bh); | ||
2518 | nr++; | ||
2519 | } | ||
2520 | put_bh(bh); | ||
2521 | bh = next; | ||
2522 | } while (bh != head); | ||
2523 | goto done; | ||
2416 | } | 2524 | } |
2417 | 2525 | ||
2526 | static int reiserfs_readpage(struct file *f, struct page *page) | ||
2527 | { | ||
2528 | return block_read_full_page(page, reiserfs_get_block); | ||
2529 | } | ||
2418 | 2530 | ||
2419 | static int reiserfs_writepage (struct page * page, struct writeback_control *wbc) | 2531 | static int reiserfs_writepage(struct page *page, struct writeback_control *wbc) |
2420 | { | 2532 | { |
2421 | struct inode *inode = page->mapping->host ; | 2533 | struct inode *inode = page->mapping->host; |
2422 | reiserfs_wait_on_write_block(inode->i_sb) ; | 2534 | reiserfs_wait_on_write_block(inode->i_sb); |
2423 | return reiserfs_write_full_page(page, wbc) ; | 2535 | return reiserfs_write_full_page(page, wbc); |
2424 | } | 2536 | } |
2425 | 2537 | ||
2426 | static int reiserfs_prepare_write(struct file *f, struct page *page, | 2538 | static int reiserfs_prepare_write(struct file *f, struct page *page, |
2427 | unsigned from, unsigned to) { | 2539 | unsigned from, unsigned to) |
2428 | struct inode *inode = page->mapping->host ; | 2540 | { |
2429 | int ret; | 2541 | struct inode *inode = page->mapping->host; |
2430 | int old_ref = 0; | 2542 | int ret; |
2431 | 2543 | int old_ref = 0; | |
2432 | reiserfs_wait_on_write_block(inode->i_sb) ; | 2544 | |
2433 | fix_tail_page_for_writing(page) ; | 2545 | reiserfs_wait_on_write_block(inode->i_sb); |
2434 | if (reiserfs_transaction_running(inode->i_sb)) { | 2546 | fix_tail_page_for_writing(page); |
2435 | struct reiserfs_transaction_handle *th; | 2547 | if (reiserfs_transaction_running(inode->i_sb)) { |
2436 | th = (struct reiserfs_transaction_handle *)current->journal_info; | 2548 | struct reiserfs_transaction_handle *th; |
2437 | BUG_ON (!th->t_refcount); | 2549 | th = (struct reiserfs_transaction_handle *)current-> |
2438 | BUG_ON (!th->t_trans_id); | 2550 | journal_info; |
2439 | old_ref = th->t_refcount; | 2551 | BUG_ON(!th->t_refcount); |
2440 | th->t_refcount++; | 2552 | BUG_ON(!th->t_trans_id); |
2441 | } | 2553 | old_ref = th->t_refcount; |
2442 | 2554 | th->t_refcount++; | |
2443 | ret = block_prepare_write(page, from, to, reiserfs_get_block) ; | ||
2444 | if (ret && reiserfs_transaction_running(inode->i_sb)) { | ||
2445 | struct reiserfs_transaction_handle *th = current->journal_info; | ||
2446 | /* this gets a little ugly. If reiserfs_get_block returned an | ||
2447 | * error and left a transacstion running, we've got to close it, | ||
2448 | * and we've got to free handle if it was a persistent transaction. | ||
2449 | * | ||
2450 | * But, if we had nested into an existing transaction, we need | ||
2451 | * to just drop the ref count on the handle. | ||
2452 | * | ||
2453 | * If old_ref == 0, the transaction is from reiserfs_get_block, | ||
2454 | * and it was a persistent trans. Otherwise, it was nested above. | ||
2455 | */ | ||
2456 | if (th->t_refcount > old_ref) { | ||
2457 | if (old_ref) | ||
2458 | th->t_refcount--; | ||
2459 | else { | ||
2460 | int err; | ||
2461 | reiserfs_write_lock(inode->i_sb); | ||
2462 | err = reiserfs_end_persistent_transaction(th); | ||
2463 | reiserfs_write_unlock(inode->i_sb); | ||
2464 | if (err) | ||
2465 | ret = err; | ||
2466 | } | ||
2467 | } | 2555 | } |
2468 | } | ||
2469 | return ret; | ||
2470 | 2556 | ||
2471 | } | 2557 | ret = block_prepare_write(page, from, to, reiserfs_get_block); |
2558 | if (ret && reiserfs_transaction_running(inode->i_sb)) { | ||
2559 | struct reiserfs_transaction_handle *th = current->journal_info; | ||
2560 | /* this gets a little ugly. If reiserfs_get_block returned an | ||
2561 | * error and left a transacstion running, we've got to close it, | ||
2562 | * and we've got to free handle if it was a persistent transaction. | ||
2563 | * | ||
2564 | * But, if we had nested into an existing transaction, we need | ||
2565 | * to just drop the ref count on the handle. | ||
2566 | * | ||
2567 | * If old_ref == 0, the transaction is from reiserfs_get_block, | ||
2568 | * and it was a persistent trans. Otherwise, it was nested above. | ||
2569 | */ | ||
2570 | if (th->t_refcount > old_ref) { | ||
2571 | if (old_ref) | ||
2572 | th->t_refcount--; | ||
2573 | else { | ||
2574 | int err; | ||
2575 | reiserfs_write_lock(inode->i_sb); | ||
2576 | err = reiserfs_end_persistent_transaction(th); | ||
2577 | reiserfs_write_unlock(inode->i_sb); | ||
2578 | if (err) | ||
2579 | ret = err; | ||
2580 | } | ||
2581 | } | ||
2582 | } | ||
2583 | return ret; | ||
2472 | 2584 | ||
2585 | } | ||
2473 | 2586 | ||
2474 | static sector_t reiserfs_aop_bmap(struct address_space *as, sector_t block) { | 2587 | static sector_t reiserfs_aop_bmap(struct address_space *as, sector_t block) |
2475 | return generic_block_bmap(as, block, reiserfs_bmap) ; | 2588 | { |
2589 | return generic_block_bmap(as, block, reiserfs_bmap); | ||
2476 | } | 2590 | } |
2477 | 2591 | ||
2478 | static int reiserfs_commit_write(struct file *f, struct page *page, | 2592 | static int reiserfs_commit_write(struct file *f, struct page *page, |
2479 | unsigned from, unsigned to) { | 2593 | unsigned from, unsigned to) |
2480 | struct inode *inode = page->mapping->host ; | 2594 | { |
2481 | loff_t pos = ((loff_t)page->index << PAGE_CACHE_SHIFT) + to; | 2595 | struct inode *inode = page->mapping->host; |
2482 | int ret = 0; | 2596 | loff_t pos = ((loff_t) page->index << PAGE_CACHE_SHIFT) + to; |
2483 | int update_sd = 0; | 2597 | int ret = 0; |
2484 | struct reiserfs_transaction_handle *th = NULL; | 2598 | int update_sd = 0; |
2485 | 2599 | struct reiserfs_transaction_handle *th = NULL; | |
2486 | reiserfs_wait_on_write_block(inode->i_sb) ; | 2600 | |
2487 | if (reiserfs_transaction_running(inode->i_sb)) { | 2601 | reiserfs_wait_on_write_block(inode->i_sb); |
2488 | th = current->journal_info; | 2602 | if (reiserfs_transaction_running(inode->i_sb)) { |
2489 | } | 2603 | th = current->journal_info; |
2490 | reiserfs_commit_page(inode, page, from, to); | 2604 | } |
2491 | 2605 | reiserfs_commit_page(inode, page, from, to); | |
2492 | /* generic_commit_write does this for us, but does not update the | ||
2493 | ** transaction tracking stuff when the size changes. So, we have | ||
2494 | ** to do the i_size updates here. | ||
2495 | */ | ||
2496 | if (pos > inode->i_size) { | ||
2497 | struct reiserfs_transaction_handle myth ; | ||
2498 | reiserfs_write_lock(inode->i_sb); | ||
2499 | /* If the file have grown beyond the border where it | ||
2500 | can have a tail, unmark it as needing a tail | ||
2501 | packing */ | ||
2502 | if ( (have_large_tails (inode->i_sb) && inode->i_size > i_block_size (inode)*4) || | ||
2503 | (have_small_tails (inode->i_sb) && inode->i_size > i_block_size(inode)) ) | ||
2504 | REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask ; | ||
2505 | |||
2506 | ret = journal_begin(&myth, inode->i_sb, 1) ; | ||
2507 | if (ret) { | ||
2508 | reiserfs_write_unlock(inode->i_sb); | ||
2509 | goto journal_error; | ||
2510 | } | ||
2511 | reiserfs_update_inode_transaction(inode) ; | ||
2512 | inode->i_size = pos ; | ||
2513 | reiserfs_update_sd(&myth, inode) ; | ||
2514 | update_sd = 1; | ||
2515 | ret = journal_end(&myth, inode->i_sb, 1) ; | ||
2516 | reiserfs_write_unlock(inode->i_sb); | ||
2517 | if (ret) | ||
2518 | goto journal_error; | ||
2519 | } | ||
2520 | if (th) { | ||
2521 | reiserfs_write_lock(inode->i_sb); | ||
2522 | if (!update_sd) | ||
2523 | reiserfs_update_sd(th, inode) ; | ||
2524 | ret = reiserfs_end_persistent_transaction(th); | ||
2525 | reiserfs_write_unlock(inode->i_sb); | ||
2526 | if (ret) | ||
2527 | goto out; | ||
2528 | } | ||
2529 | |||
2530 | /* we test for O_SYNC here so we can commit the transaction | ||
2531 | ** for any packed tails the file might have had | ||
2532 | */ | ||
2533 | if (f && (f->f_flags & O_SYNC)) { | ||
2534 | reiserfs_write_lock(inode->i_sb); | ||
2535 | ret = reiserfs_commit_for_inode(inode) ; | ||
2536 | reiserfs_write_unlock(inode->i_sb); | ||
2537 | } | ||
2538 | out: | ||
2539 | return ret ; | ||
2540 | 2606 | ||
2541 | journal_error: | 2607 | /* generic_commit_write does this for us, but does not update the |
2542 | if (th) { | 2608 | ** transaction tracking stuff when the size changes. So, we have |
2543 | reiserfs_write_lock(inode->i_sb); | 2609 | ** to do the i_size updates here. |
2544 | if (!update_sd) | 2610 | */ |
2545 | reiserfs_update_sd(th, inode) ; | 2611 | if (pos > inode->i_size) { |
2546 | ret = reiserfs_end_persistent_transaction(th); | 2612 | struct reiserfs_transaction_handle myth; |
2547 | reiserfs_write_unlock(inode->i_sb); | 2613 | reiserfs_write_lock(inode->i_sb); |
2548 | } | 2614 | /* If the file have grown beyond the border where it |
2615 | can have a tail, unmark it as needing a tail | ||
2616 | packing */ | ||
2617 | if ((have_large_tails(inode->i_sb) | ||
2618 | && inode->i_size > i_block_size(inode) * 4) | ||
2619 | || (have_small_tails(inode->i_sb) | ||
2620 | && inode->i_size > i_block_size(inode))) | ||
2621 | REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask; | ||
2622 | |||
2623 | ret = journal_begin(&myth, inode->i_sb, 1); | ||
2624 | if (ret) { | ||
2625 | reiserfs_write_unlock(inode->i_sb); | ||
2626 | goto journal_error; | ||
2627 | } | ||
2628 | reiserfs_update_inode_transaction(inode); | ||
2629 | inode->i_size = pos; | ||
2630 | reiserfs_update_sd(&myth, inode); | ||
2631 | update_sd = 1; | ||
2632 | ret = journal_end(&myth, inode->i_sb, 1); | ||
2633 | reiserfs_write_unlock(inode->i_sb); | ||
2634 | if (ret) | ||
2635 | goto journal_error; | ||
2636 | } | ||
2637 | if (th) { | ||
2638 | reiserfs_write_lock(inode->i_sb); | ||
2639 | if (!update_sd) | ||
2640 | reiserfs_update_sd(th, inode); | ||
2641 | ret = reiserfs_end_persistent_transaction(th); | ||
2642 | reiserfs_write_unlock(inode->i_sb); | ||
2643 | if (ret) | ||
2644 | goto out; | ||
2645 | } | ||
2646 | |||
2647 | /* we test for O_SYNC here so we can commit the transaction | ||
2648 | ** for any packed tails the file might have had | ||
2649 | */ | ||
2650 | if (f && (f->f_flags & O_SYNC)) { | ||
2651 | reiserfs_write_lock(inode->i_sb); | ||
2652 | ret = reiserfs_commit_for_inode(inode); | ||
2653 | reiserfs_write_unlock(inode->i_sb); | ||
2654 | } | ||
2655 | out: | ||
2656 | return ret; | ||
2549 | 2657 | ||
2550 | return ret; | 2658 | journal_error: |
2659 | if (th) { | ||
2660 | reiserfs_write_lock(inode->i_sb); | ||
2661 | if (!update_sd) | ||
2662 | reiserfs_update_sd(th, inode); | ||
2663 | ret = reiserfs_end_persistent_transaction(th); | ||
2664 | reiserfs_write_unlock(inode->i_sb); | ||
2665 | } | ||
2666 | |||
2667 | return ret; | ||
2551 | } | 2668 | } |
2552 | 2669 | ||
2553 | void sd_attrs_to_i_attrs( __u16 sd_attrs, struct inode *inode ) | 2670 | void sd_attrs_to_i_attrs(__u16 sd_attrs, struct inode *inode) |
2554 | { | 2671 | { |
2555 | if( reiserfs_attrs( inode -> i_sb ) ) { | 2672 | if (reiserfs_attrs(inode->i_sb)) { |
2556 | if( sd_attrs & REISERFS_SYNC_FL ) | 2673 | if (sd_attrs & REISERFS_SYNC_FL) |
2557 | inode -> i_flags |= S_SYNC; | 2674 | inode->i_flags |= S_SYNC; |
2558 | else | 2675 | else |
2559 | inode -> i_flags &= ~S_SYNC; | 2676 | inode->i_flags &= ~S_SYNC; |
2560 | if( sd_attrs & REISERFS_IMMUTABLE_FL ) | 2677 | if (sd_attrs & REISERFS_IMMUTABLE_FL) |
2561 | inode -> i_flags |= S_IMMUTABLE; | 2678 | inode->i_flags |= S_IMMUTABLE; |
2562 | else | 2679 | else |
2563 | inode -> i_flags &= ~S_IMMUTABLE; | 2680 | inode->i_flags &= ~S_IMMUTABLE; |
2564 | if( sd_attrs & REISERFS_APPEND_FL ) | 2681 | if (sd_attrs & REISERFS_APPEND_FL) |
2565 | inode -> i_flags |= S_APPEND; | 2682 | inode->i_flags |= S_APPEND; |
2566 | else | 2683 | else |
2567 | inode -> i_flags &= ~S_APPEND; | 2684 | inode->i_flags &= ~S_APPEND; |
2568 | if( sd_attrs & REISERFS_NOATIME_FL ) | 2685 | if (sd_attrs & REISERFS_NOATIME_FL) |
2569 | inode -> i_flags |= S_NOATIME; | 2686 | inode->i_flags |= S_NOATIME; |
2570 | else | 2687 | else |
2571 | inode -> i_flags &= ~S_NOATIME; | 2688 | inode->i_flags &= ~S_NOATIME; |
2572 | if( sd_attrs & REISERFS_NOTAIL_FL ) | 2689 | if (sd_attrs & REISERFS_NOTAIL_FL) |
2573 | REISERFS_I(inode)->i_flags |= i_nopack_mask; | 2690 | REISERFS_I(inode)->i_flags |= i_nopack_mask; |
2574 | else | 2691 | else |
2575 | REISERFS_I(inode)->i_flags &= ~i_nopack_mask; | 2692 | REISERFS_I(inode)->i_flags &= ~i_nopack_mask; |
2576 | } | 2693 | } |
2577 | } | 2694 | } |
2578 | 2695 | ||
2579 | void i_attrs_to_sd_attrs( struct inode *inode, __u16 *sd_attrs ) | 2696 | void i_attrs_to_sd_attrs(struct inode *inode, __u16 * sd_attrs) |
2580 | { | 2697 | { |
2581 | if( reiserfs_attrs( inode -> i_sb ) ) { | 2698 | if (reiserfs_attrs(inode->i_sb)) { |
2582 | if( inode -> i_flags & S_IMMUTABLE ) | 2699 | if (inode->i_flags & S_IMMUTABLE) |
2583 | *sd_attrs |= REISERFS_IMMUTABLE_FL; | 2700 | *sd_attrs |= REISERFS_IMMUTABLE_FL; |
2584 | else | 2701 | else |
2585 | *sd_attrs &= ~REISERFS_IMMUTABLE_FL; | 2702 | *sd_attrs &= ~REISERFS_IMMUTABLE_FL; |
2586 | if( inode -> i_flags & S_SYNC ) | 2703 | if (inode->i_flags & S_SYNC) |
2587 | *sd_attrs |= REISERFS_SYNC_FL; | 2704 | *sd_attrs |= REISERFS_SYNC_FL; |
2588 | else | 2705 | else |
2589 | *sd_attrs &= ~REISERFS_SYNC_FL; | 2706 | *sd_attrs &= ~REISERFS_SYNC_FL; |
2590 | if( inode -> i_flags & S_NOATIME ) | 2707 | if (inode->i_flags & S_NOATIME) |
2591 | *sd_attrs |= REISERFS_NOATIME_FL; | 2708 | *sd_attrs |= REISERFS_NOATIME_FL; |
2592 | else | 2709 | else |
2593 | *sd_attrs &= ~REISERFS_NOATIME_FL; | 2710 | *sd_attrs &= ~REISERFS_NOATIME_FL; |
2594 | if( REISERFS_I(inode)->i_flags & i_nopack_mask ) | 2711 | if (REISERFS_I(inode)->i_flags & i_nopack_mask) |
2595 | *sd_attrs |= REISERFS_NOTAIL_FL; | 2712 | *sd_attrs |= REISERFS_NOTAIL_FL; |
2596 | else | 2713 | else |
2597 | *sd_attrs &= ~REISERFS_NOTAIL_FL; | 2714 | *sd_attrs &= ~REISERFS_NOTAIL_FL; |
@@ -2603,106 +2720,107 @@ void i_attrs_to_sd_attrs( struct inode *inode, __u16 *sd_attrs ) | |||
2603 | */ | 2720 | */ |
2604 | static int invalidatepage_can_drop(struct inode *inode, struct buffer_head *bh) | 2721 | static int invalidatepage_can_drop(struct inode *inode, struct buffer_head *bh) |
2605 | { | 2722 | { |
2606 | int ret = 1 ; | 2723 | int ret = 1; |
2607 | struct reiserfs_journal *j = SB_JOURNAL(inode->i_sb) ; | 2724 | struct reiserfs_journal *j = SB_JOURNAL(inode->i_sb); |
2608 | 2725 | ||
2609 | spin_lock(&j->j_dirty_buffers_lock) ; | 2726 | spin_lock(&j->j_dirty_buffers_lock); |
2610 | if (!buffer_mapped(bh)) { | 2727 | if (!buffer_mapped(bh)) { |
2611 | goto free_jh; | 2728 | goto free_jh; |
2612 | } | 2729 | } |
2613 | /* the page is locked, and the only places that log a data buffer | 2730 | /* the page is locked, and the only places that log a data buffer |
2614 | * also lock the page. | 2731 | * also lock the page. |
2615 | */ | ||
2616 | if (reiserfs_file_data_log(inode)) { | ||
2617 | /* | ||
2618 | * very conservative, leave the buffer pinned if | ||
2619 | * anyone might need it. | ||
2620 | */ | ||
2621 | if (buffer_journaled(bh) || buffer_journal_dirty(bh)) { | ||
2622 | ret = 0 ; | ||
2623 | } | ||
2624 | } else | ||
2625 | if (buffer_dirty(bh) || buffer_locked(bh)) { | ||
2626 | struct reiserfs_journal_list *jl; | ||
2627 | struct reiserfs_jh *jh = bh->b_private; | ||
2628 | |||
2629 | /* why is this safe? | ||
2630 | * reiserfs_setattr updates i_size in the on disk | ||
2631 | * stat data before allowing vmtruncate to be called. | ||
2632 | * | ||
2633 | * If buffer was put onto the ordered list for this | ||
2634 | * transaction, we know for sure either this transaction | ||
2635 | * or an older one already has updated i_size on disk, | ||
2636 | * and this ordered data won't be referenced in the file | ||
2637 | * if we crash. | ||
2638 | * | ||
2639 | * if the buffer was put onto the ordered list for an older | ||
2640 | * transaction, we need to leave it around | ||
2641 | */ | 2732 | */ |
2642 | if (jh && (jl = jh->jl) && jl != SB_JOURNAL(inode->i_sb)->j_current_jl) | 2733 | if (reiserfs_file_data_log(inode)) { |
2643 | ret = 0; | 2734 | /* |
2644 | } | 2735 | * very conservative, leave the buffer pinned if |
2645 | free_jh: | 2736 | * anyone might need it. |
2646 | if (ret && bh->b_private) { | 2737 | */ |
2647 | reiserfs_free_jh(bh); | 2738 | if (buffer_journaled(bh) || buffer_journal_dirty(bh)) { |
2648 | } | 2739 | ret = 0; |
2649 | spin_unlock(&j->j_dirty_buffers_lock) ; | 2740 | } |
2650 | return ret ; | 2741 | } else if (buffer_dirty(bh) || buffer_locked(bh)) { |
2742 | struct reiserfs_journal_list *jl; | ||
2743 | struct reiserfs_jh *jh = bh->b_private; | ||
2744 | |||
2745 | /* why is this safe? | ||
2746 | * reiserfs_setattr updates i_size in the on disk | ||
2747 | * stat data before allowing vmtruncate to be called. | ||
2748 | * | ||
2749 | * If buffer was put onto the ordered list for this | ||
2750 | * transaction, we know for sure either this transaction | ||
2751 | * or an older one already has updated i_size on disk, | ||
2752 | * and this ordered data won't be referenced in the file | ||
2753 | * if we crash. | ||
2754 | * | ||
2755 | * if the buffer was put onto the ordered list for an older | ||
2756 | * transaction, we need to leave it around | ||
2757 | */ | ||
2758 | if (jh && (jl = jh->jl) | ||
2759 | && jl != SB_JOURNAL(inode->i_sb)->j_current_jl) | ||
2760 | ret = 0; | ||
2761 | } | ||
2762 | free_jh: | ||
2763 | if (ret && bh->b_private) { | ||
2764 | reiserfs_free_jh(bh); | ||
2765 | } | ||
2766 | spin_unlock(&j->j_dirty_buffers_lock); | ||
2767 | return ret; | ||
2651 | } | 2768 | } |
2652 | 2769 | ||
2653 | /* clm -- taken from fs/buffer.c:block_invalidate_page */ | 2770 | /* clm -- taken from fs/buffer.c:block_invalidate_page */ |
2654 | static int reiserfs_invalidatepage(struct page *page, unsigned long offset) | 2771 | static int reiserfs_invalidatepage(struct page *page, unsigned long offset) |
2655 | { | 2772 | { |
2656 | struct buffer_head *head, *bh, *next; | 2773 | struct buffer_head *head, *bh, *next; |
2657 | struct inode *inode = page->mapping->host; | 2774 | struct inode *inode = page->mapping->host; |
2658 | unsigned int curr_off = 0; | 2775 | unsigned int curr_off = 0; |
2659 | int ret = 1; | 2776 | int ret = 1; |
2660 | 2777 | ||
2661 | BUG_ON(!PageLocked(page)); | 2778 | BUG_ON(!PageLocked(page)); |
2662 | 2779 | ||
2663 | if (offset == 0) | 2780 | if (offset == 0) |
2664 | ClearPageChecked(page); | 2781 | ClearPageChecked(page); |
2665 | 2782 | ||
2666 | if (!page_has_buffers(page)) | 2783 | if (!page_has_buffers(page)) |
2667 | goto out; | 2784 | goto out; |
2785 | |||
2786 | head = page_buffers(page); | ||
2787 | bh = head; | ||
2788 | do { | ||
2789 | unsigned int next_off = curr_off + bh->b_size; | ||
2790 | next = bh->b_this_page; | ||
2668 | 2791 | ||
2669 | head = page_buffers(page); | 2792 | /* |
2670 | bh = head; | 2793 | * is this block fully invalidated? |
2671 | do { | 2794 | */ |
2672 | unsigned int next_off = curr_off + bh->b_size; | 2795 | if (offset <= curr_off) { |
2673 | next = bh->b_this_page; | 2796 | if (invalidatepage_can_drop(inode, bh)) |
2797 | reiserfs_unmap_buffer(bh); | ||
2798 | else | ||
2799 | ret = 0; | ||
2800 | } | ||
2801 | curr_off = next_off; | ||
2802 | bh = next; | ||
2803 | } while (bh != head); | ||
2674 | 2804 | ||
2675 | /* | 2805 | /* |
2676 | * is this block fully invalidated? | 2806 | * We release buffers only if the entire page is being invalidated. |
2807 | * The get_block cached value has been unconditionally invalidated, | ||
2808 | * so real IO is not possible anymore. | ||
2677 | */ | 2809 | */ |
2678 | if (offset <= curr_off) { | 2810 | if (!offset && ret) |
2679 | if (invalidatepage_can_drop(inode, bh)) | 2811 | ret = try_to_release_page(page, 0); |
2680 | reiserfs_unmap_buffer(bh); | 2812 | out: |
2681 | else | 2813 | return ret; |
2682 | ret = 0; | ||
2683 | } | ||
2684 | curr_off = next_off; | ||
2685 | bh = next; | ||
2686 | } while (bh != head); | ||
2687 | |||
2688 | /* | ||
2689 | * We release buffers only if the entire page is being invalidated. | ||
2690 | * The get_block cached value has been unconditionally invalidated, | ||
2691 | * so real IO is not possible anymore. | ||
2692 | */ | ||
2693 | if (!offset && ret) | ||
2694 | ret = try_to_release_page(page, 0); | ||
2695 | out: | ||
2696 | return ret; | ||
2697 | } | 2814 | } |
2698 | 2815 | ||
2699 | static int reiserfs_set_page_dirty(struct page *page) { | 2816 | static int reiserfs_set_page_dirty(struct page *page) |
2700 | struct inode *inode = page->mapping->host; | 2817 | { |
2701 | if (reiserfs_file_data_log(inode)) { | 2818 | struct inode *inode = page->mapping->host; |
2702 | SetPageChecked(page); | 2819 | if (reiserfs_file_data_log(inode)) { |
2703 | return __set_page_dirty_nobuffers(page); | 2820 | SetPageChecked(page); |
2704 | } | 2821 | return __set_page_dirty_nobuffers(page); |
2705 | return __set_page_dirty_buffers(page); | 2822 | } |
2823 | return __set_page_dirty_buffers(page); | ||
2706 | } | 2824 | } |
2707 | 2825 | ||
2708 | /* | 2826 | /* |
@@ -2716,143 +2834,152 @@ static int reiserfs_set_page_dirty(struct page *page) { | |||
2716 | */ | 2834 | */ |
2717 | static int reiserfs_releasepage(struct page *page, int unused_gfp_flags) | 2835 | static int reiserfs_releasepage(struct page *page, int unused_gfp_flags) |
2718 | { | 2836 | { |
2719 | struct inode *inode = page->mapping->host ; | 2837 | struct inode *inode = page->mapping->host; |
2720 | struct reiserfs_journal *j = SB_JOURNAL(inode->i_sb) ; | 2838 | struct reiserfs_journal *j = SB_JOURNAL(inode->i_sb); |
2721 | struct buffer_head *head ; | 2839 | struct buffer_head *head; |
2722 | struct buffer_head *bh ; | 2840 | struct buffer_head *bh; |
2723 | int ret = 1 ; | 2841 | int ret = 1; |
2724 | 2842 | ||
2725 | WARN_ON(PageChecked(page)); | 2843 | WARN_ON(PageChecked(page)); |
2726 | spin_lock(&j->j_dirty_buffers_lock) ; | 2844 | spin_lock(&j->j_dirty_buffers_lock); |
2727 | head = page_buffers(page) ; | 2845 | head = page_buffers(page); |
2728 | bh = head ; | 2846 | bh = head; |
2729 | do { | 2847 | do { |
2730 | if (bh->b_private) { | 2848 | if (bh->b_private) { |
2731 | if (!buffer_dirty(bh) && !buffer_locked(bh)) { | 2849 | if (!buffer_dirty(bh) && !buffer_locked(bh)) { |
2732 | reiserfs_free_jh(bh); | 2850 | reiserfs_free_jh(bh); |
2733 | } else { | 2851 | } else { |
2734 | ret = 0 ; | 2852 | ret = 0; |
2735 | break ; | 2853 | break; |
2736 | } | 2854 | } |
2737 | } | 2855 | } |
2738 | bh = bh->b_this_page ; | 2856 | bh = bh->b_this_page; |
2739 | } while (bh != head) ; | 2857 | } while (bh != head); |
2740 | if (ret) | 2858 | if (ret) |
2741 | ret = try_to_free_buffers(page) ; | 2859 | ret = try_to_free_buffers(page); |
2742 | spin_unlock(&j->j_dirty_buffers_lock) ; | 2860 | spin_unlock(&j->j_dirty_buffers_lock); |
2743 | return ret ; | 2861 | return ret; |
2744 | } | 2862 | } |
2745 | 2863 | ||
2746 | /* We thank Mingming Cao for helping us understand in great detail what | 2864 | /* We thank Mingming Cao for helping us understand in great detail what |
2747 | to do in this section of the code. */ | 2865 | to do in this section of the code. */ |
2748 | static ssize_t reiserfs_direct_IO(int rw, struct kiocb *iocb, | 2866 | static ssize_t reiserfs_direct_IO(int rw, struct kiocb *iocb, |
2749 | const struct iovec *iov, loff_t offset, unsigned long nr_segs) | 2867 | const struct iovec *iov, loff_t offset, |
2868 | unsigned long nr_segs) | ||
2750 | { | 2869 | { |
2751 | struct file *file = iocb->ki_filp; | 2870 | struct file *file = iocb->ki_filp; |
2752 | struct inode *inode = file->f_mapping->host; | 2871 | struct inode *inode = file->f_mapping->host; |
2753 | 2872 | ||
2754 | return blockdev_direct_IO(rw, iocb, inode, inode->i_sb->s_bdev, iov, | 2873 | return blockdev_direct_IO(rw, iocb, inode, inode->i_sb->s_bdev, iov, |
2755 | offset, nr_segs, reiserfs_get_blocks_direct_io, NULL); | 2874 | offset, nr_segs, |
2875 | reiserfs_get_blocks_direct_io, NULL); | ||
2756 | } | 2876 | } |
2757 | 2877 | ||
2758 | int reiserfs_setattr(struct dentry *dentry, struct iattr *attr) { | 2878 | int reiserfs_setattr(struct dentry *dentry, struct iattr *attr) |
2759 | struct inode *inode = dentry->d_inode ; | 2879 | { |
2760 | int error ; | 2880 | struct inode *inode = dentry->d_inode; |
2761 | unsigned int ia_valid = attr->ia_valid; | 2881 | int error; |
2762 | reiserfs_write_lock(inode->i_sb); | 2882 | unsigned int ia_valid = attr->ia_valid; |
2763 | if (attr->ia_valid & ATTR_SIZE) { | 2883 | reiserfs_write_lock(inode->i_sb); |
2764 | /* version 2 items will be caught by the s_maxbytes check | 2884 | if (attr->ia_valid & ATTR_SIZE) { |
2765 | ** done for us in vmtruncate | 2885 | /* version 2 items will be caught by the s_maxbytes check |
2766 | */ | 2886 | ** done for us in vmtruncate |
2767 | if (get_inode_item_key_version(inode) == KEY_FORMAT_3_5 && | 2887 | */ |
2768 | attr->ia_size > MAX_NON_LFS) { | 2888 | if (get_inode_item_key_version(inode) == KEY_FORMAT_3_5 && |
2769 | error = -EFBIG ; | 2889 | attr->ia_size > MAX_NON_LFS) { |
2770 | goto out; | 2890 | error = -EFBIG; |
2771 | } | 2891 | goto out; |
2772 | /* fill in hole pointers in the expanding truncate case. */ | 2892 | } |
2773 | if (attr->ia_size > inode->i_size) { | 2893 | /* fill in hole pointers in the expanding truncate case. */ |
2774 | error = generic_cont_expand(inode, attr->ia_size) ; | 2894 | if (attr->ia_size > inode->i_size) { |
2775 | if (REISERFS_I(inode)->i_prealloc_count > 0) { | 2895 | error = generic_cont_expand(inode, attr->ia_size); |
2776 | int err; | 2896 | if (REISERFS_I(inode)->i_prealloc_count > 0) { |
2777 | struct reiserfs_transaction_handle th ; | 2897 | int err; |
2778 | /* we're changing at most 2 bitmaps, inode + super */ | 2898 | struct reiserfs_transaction_handle th; |
2779 | err = journal_begin(&th, inode->i_sb, 4) ; | 2899 | /* we're changing at most 2 bitmaps, inode + super */ |
2780 | if (!err) { | 2900 | err = journal_begin(&th, inode->i_sb, 4); |
2781 | reiserfs_discard_prealloc (&th, inode); | 2901 | if (!err) { |
2782 | err = journal_end(&th, inode->i_sb, 4) ; | 2902 | reiserfs_discard_prealloc(&th, inode); |
2903 | err = journal_end(&th, inode->i_sb, 4); | ||
2904 | } | ||
2905 | if (err) | ||
2906 | error = err; | ||
2907 | } | ||
2908 | if (error) | ||
2909 | goto out; | ||
2783 | } | 2910 | } |
2784 | if (err) | ||
2785 | error = err; | ||
2786 | } | ||
2787 | if (error) | ||
2788 | goto out; | ||
2789 | } | 2911 | } |
2790 | } | ||
2791 | 2912 | ||
2792 | if ((((attr->ia_valid & ATTR_UID) && (attr->ia_uid & ~0xffff)) || | 2913 | if ((((attr->ia_valid & ATTR_UID) && (attr->ia_uid & ~0xffff)) || |
2793 | ((attr->ia_valid & ATTR_GID) && (attr->ia_gid & ~0xffff))) && | 2914 | ((attr->ia_valid & ATTR_GID) && (attr->ia_gid & ~0xffff))) && |
2794 | (get_inode_sd_version (inode) == STAT_DATA_V1)) { | 2915 | (get_inode_sd_version(inode) == STAT_DATA_V1)) { |
2795 | /* stat data of format v3.5 has 16 bit uid and gid */ | 2916 | /* stat data of format v3.5 has 16 bit uid and gid */ |
2796 | error = -EINVAL; | 2917 | error = -EINVAL; |
2797 | goto out; | 2918 | goto out; |
2798 | } | 2919 | } |
2799 | |||
2800 | error = inode_change_ok(inode, attr) ; | ||
2801 | if (!error) { | ||
2802 | if ((ia_valid & ATTR_UID && attr->ia_uid != inode->i_uid) || | ||
2803 | (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid)) { | ||
2804 | error = reiserfs_chown_xattrs (inode, attr); | ||
2805 | |||
2806 | if (!error) { | ||
2807 | struct reiserfs_transaction_handle th; | ||
2808 | int jbegin_count = 2*(REISERFS_QUOTA_INIT_BLOCKS(inode->i_sb)+REISERFS_QUOTA_DEL_BLOCKS(inode->i_sb))+2; | ||
2809 | |||
2810 | /* (user+group)*(old+new) structure - we count quota info and , inode write (sb, inode) */ | ||
2811 | error = journal_begin(&th, inode->i_sb, jbegin_count); | ||
2812 | if (error) | ||
2813 | goto out; | ||
2814 | error = DQUOT_TRANSFER(inode, attr) ? -EDQUOT : 0; | ||
2815 | if (error) { | ||
2816 | journal_end(&th, inode->i_sb, jbegin_count); | ||
2817 | goto out; | ||
2818 | } | ||
2819 | /* Update corresponding info in inode so that everything is in | ||
2820 | * one transaction */ | ||
2821 | if (attr->ia_valid & ATTR_UID) | ||
2822 | inode->i_uid = attr->ia_uid; | ||
2823 | if (attr->ia_valid & ATTR_GID) | ||
2824 | inode->i_gid = attr->ia_gid; | ||
2825 | mark_inode_dirty(inode); | ||
2826 | error = journal_end(&th, inode->i_sb, jbegin_count); | ||
2827 | } | ||
2828 | } | ||
2829 | if (!error) | ||
2830 | error = inode_setattr(inode, attr) ; | ||
2831 | } | ||
2832 | 2920 | ||
2921 | error = inode_change_ok(inode, attr); | ||
2922 | if (!error) { | ||
2923 | if ((ia_valid & ATTR_UID && attr->ia_uid != inode->i_uid) || | ||
2924 | (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid)) { | ||
2925 | error = reiserfs_chown_xattrs(inode, attr); | ||
2926 | |||
2927 | if (!error) { | ||
2928 | struct reiserfs_transaction_handle th; | ||
2929 | int jbegin_count = | ||
2930 | 2 * | ||
2931 | (REISERFS_QUOTA_INIT_BLOCKS(inode->i_sb) + | ||
2932 | REISERFS_QUOTA_DEL_BLOCKS(inode->i_sb)) + | ||
2933 | 2; | ||
2934 | |||
2935 | /* (user+group)*(old+new) structure - we count quota info and , inode write (sb, inode) */ | ||
2936 | error = | ||
2937 | journal_begin(&th, inode->i_sb, | ||
2938 | jbegin_count); | ||
2939 | if (error) | ||
2940 | goto out; | ||
2941 | error = | ||
2942 | DQUOT_TRANSFER(inode, attr) ? -EDQUOT : 0; | ||
2943 | if (error) { | ||
2944 | journal_end(&th, inode->i_sb, | ||
2945 | jbegin_count); | ||
2946 | goto out; | ||
2947 | } | ||
2948 | /* Update corresponding info in inode so that everything is in | ||
2949 | * one transaction */ | ||
2950 | if (attr->ia_valid & ATTR_UID) | ||
2951 | inode->i_uid = attr->ia_uid; | ||
2952 | if (attr->ia_valid & ATTR_GID) | ||
2953 | inode->i_gid = attr->ia_gid; | ||
2954 | mark_inode_dirty(inode); | ||
2955 | error = | ||
2956 | journal_end(&th, inode->i_sb, jbegin_count); | ||
2957 | } | ||
2958 | } | ||
2959 | if (!error) | ||
2960 | error = inode_setattr(inode, attr); | ||
2961 | } | ||
2833 | 2962 | ||
2834 | if (!error && reiserfs_posixacl (inode->i_sb)) { | 2963 | if (!error && reiserfs_posixacl(inode->i_sb)) { |
2835 | if (attr->ia_valid & ATTR_MODE) | 2964 | if (attr->ia_valid & ATTR_MODE) |
2836 | error = reiserfs_acl_chmod (inode); | 2965 | error = reiserfs_acl_chmod(inode); |
2837 | } | 2966 | } |
2838 | 2967 | ||
2839 | out: | 2968 | out: |
2840 | reiserfs_write_unlock(inode->i_sb); | 2969 | reiserfs_write_unlock(inode->i_sb); |
2841 | return error ; | 2970 | return error; |
2842 | } | 2971 | } |
2843 | 2972 | ||
2844 | |||
2845 | |||
2846 | struct address_space_operations reiserfs_address_space_operations = { | 2973 | struct address_space_operations reiserfs_address_space_operations = { |
2847 | .writepage = reiserfs_writepage, | 2974 | .writepage = reiserfs_writepage, |
2848 | .readpage = reiserfs_readpage, | 2975 | .readpage = reiserfs_readpage, |
2849 | .readpages = reiserfs_readpages, | 2976 | .readpages = reiserfs_readpages, |
2850 | .releasepage = reiserfs_releasepage, | 2977 | .releasepage = reiserfs_releasepage, |
2851 | .invalidatepage = reiserfs_invalidatepage, | 2978 | .invalidatepage = reiserfs_invalidatepage, |
2852 | .sync_page = block_sync_page, | 2979 | .sync_page = block_sync_page, |
2853 | .prepare_write = reiserfs_prepare_write, | 2980 | .prepare_write = reiserfs_prepare_write, |
2854 | .commit_write = reiserfs_commit_write, | 2981 | .commit_write = reiserfs_commit_write, |
2855 | .bmap = reiserfs_aop_bmap, | 2982 | .bmap = reiserfs_aop_bmap, |
2856 | .direct_IO = reiserfs_direct_IO, | 2983 | .direct_IO = reiserfs_direct_IO, |
2857 | .set_page_dirty = reiserfs_set_page_dirty, | 2984 | .set_page_dirty = reiserfs_set_page_dirty, |
2858 | } ; | 2985 | }; |