diff options
Diffstat (limited to 'fs/f2fs/recovery.c')
-rw-r--r-- | fs/f2fs/recovery.c | 150 |
1 files changed, 97 insertions, 53 deletions
diff --git a/fs/f2fs/recovery.c b/fs/f2fs/recovery.c index 60c8a5097058..d56d951c2253 100644 --- a/fs/f2fs/recovery.c +++ b/fs/f2fs/recovery.c | |||
@@ -40,36 +40,54 @@ static struct fsync_inode_entry *get_fsync_inode(struct list_head *head, | |||
40 | 40 | ||
41 | static int recover_dentry(struct page *ipage, struct inode *inode) | 41 | static int recover_dentry(struct page *ipage, struct inode *inode) |
42 | { | 42 | { |
43 | struct f2fs_node *raw_node = (struct f2fs_node *)kmap(ipage); | 43 | void *kaddr = page_address(ipage); |
44 | struct f2fs_node *raw_node = (struct f2fs_node *)kaddr; | ||
44 | struct f2fs_inode *raw_inode = &(raw_node->i); | 45 | struct f2fs_inode *raw_inode = &(raw_node->i); |
45 | struct qstr name; | 46 | nid_t pino = le32_to_cpu(raw_inode->i_pino); |
46 | struct f2fs_dir_entry *de; | 47 | struct f2fs_dir_entry *de; |
48 | struct qstr name; | ||
47 | struct page *page; | 49 | struct page *page; |
48 | struct inode *dir; | 50 | struct inode *dir, *einode; |
49 | int err = 0; | 51 | int err = 0; |
50 | 52 | ||
51 | if (!is_dent_dnode(ipage)) | 53 | dir = check_dirty_dir_inode(F2FS_SB(inode->i_sb), pino); |
52 | goto out; | 54 | if (!dir) { |
53 | 55 | dir = f2fs_iget(inode->i_sb, pino); | |
54 | dir = f2fs_iget(inode->i_sb, le32_to_cpu(raw_inode->i_pino)); | 56 | if (IS_ERR(dir)) { |
55 | if (IS_ERR(dir)) { | 57 | err = PTR_ERR(dir); |
56 | err = PTR_ERR(dir); | 58 | goto out; |
57 | goto out; | 59 | } |
60 | set_inode_flag(F2FS_I(dir), FI_DELAY_IPUT); | ||
61 | add_dirty_dir_inode(dir); | ||
58 | } | 62 | } |
59 | 63 | ||
60 | name.len = le32_to_cpu(raw_inode->i_namelen); | 64 | name.len = le32_to_cpu(raw_inode->i_namelen); |
61 | name.name = raw_inode->i_name; | 65 | name.name = raw_inode->i_name; |
62 | 66 | retry: | |
63 | de = f2fs_find_entry(dir, &name, &page); | 67 | de = f2fs_find_entry(dir, &name, &page); |
64 | if (de) { | 68 | if (de && inode->i_ino == le32_to_cpu(de->ino)) { |
65 | kunmap(page); | 69 | kunmap(page); |
66 | f2fs_put_page(page, 0); | 70 | f2fs_put_page(page, 0); |
67 | } else { | 71 | goto out; |
68 | err = __f2fs_add_link(dir, &name, inode); | 72 | } |
73 | if (de) { | ||
74 | einode = f2fs_iget(inode->i_sb, le32_to_cpu(de->ino)); | ||
75 | if (IS_ERR(einode)) { | ||
76 | WARN_ON(1); | ||
77 | if (PTR_ERR(einode) == -ENOENT) | ||
78 | err = -EEXIST; | ||
79 | goto out; | ||
80 | } | ||
81 | f2fs_delete_entry(de, page, einode); | ||
82 | iput(einode); | ||
83 | goto retry; | ||
69 | } | 84 | } |
70 | iput(dir); | 85 | err = __f2fs_add_link(dir, &name, inode); |
71 | out: | 86 | out: |
72 | kunmap(ipage); | 87 | f2fs_msg(inode->i_sb, KERN_NOTICE, "recover_inode and its dentry: " |
88 | "ino = %x, name = %s, dir = %lx, err = %d", | ||
89 | ino_of_node(ipage), raw_inode->i_name, | ||
90 | IS_ERR(dir) ? 0 : dir->i_ino, err); | ||
73 | return err; | 91 | return err; |
74 | } | 92 | } |
75 | 93 | ||
@@ -79,6 +97,9 @@ static int recover_inode(struct inode *inode, struct page *node_page) | |||
79 | struct f2fs_node *raw_node = (struct f2fs_node *)kaddr; | 97 | struct f2fs_node *raw_node = (struct f2fs_node *)kaddr; |
80 | struct f2fs_inode *raw_inode = &(raw_node->i); | 98 | struct f2fs_inode *raw_inode = &(raw_node->i); |
81 | 99 | ||
100 | if (!IS_INODE(node_page)) | ||
101 | return 0; | ||
102 | |||
82 | inode->i_mode = le16_to_cpu(raw_inode->i_mode); | 103 | inode->i_mode = le16_to_cpu(raw_inode->i_mode); |
83 | i_size_write(inode, le64_to_cpu(raw_inode->i_size)); | 104 | i_size_write(inode, le64_to_cpu(raw_inode->i_size)); |
84 | inode->i_atime.tv_sec = le64_to_cpu(raw_inode->i_mtime); | 105 | inode->i_atime.tv_sec = le64_to_cpu(raw_inode->i_mtime); |
@@ -88,7 +109,12 @@ static int recover_inode(struct inode *inode, struct page *node_page) | |||
88 | inode->i_ctime.tv_nsec = le32_to_cpu(raw_inode->i_ctime_nsec); | 109 | inode->i_ctime.tv_nsec = le32_to_cpu(raw_inode->i_ctime_nsec); |
89 | inode->i_mtime.tv_nsec = le32_to_cpu(raw_inode->i_mtime_nsec); | 110 | inode->i_mtime.tv_nsec = le32_to_cpu(raw_inode->i_mtime_nsec); |
90 | 111 | ||
91 | return recover_dentry(node_page, inode); | 112 | if (is_dent_dnode(node_page)) |
113 | return recover_dentry(node_page, inode); | ||
114 | |||
115 | f2fs_msg(inode->i_sb, KERN_NOTICE, "recover_inode: ino = %x, name = %s", | ||
116 | ino_of_node(node_page), raw_inode->i_name); | ||
117 | return 0; | ||
92 | } | 118 | } |
93 | 119 | ||
94 | static int find_fsync_dnodes(struct f2fs_sb_info *sbi, struct list_head *head) | 120 | static int find_fsync_dnodes(struct f2fs_sb_info *sbi, struct list_head *head) |
@@ -119,14 +145,13 @@ static int find_fsync_dnodes(struct f2fs_sb_info *sbi, struct list_head *head) | |||
119 | lock_page(page); | 145 | lock_page(page); |
120 | 146 | ||
121 | if (cp_ver != cpver_of_node(page)) | 147 | if (cp_ver != cpver_of_node(page)) |
122 | goto unlock_out; | 148 | break; |
123 | 149 | ||
124 | if (!is_fsync_dnode(page)) | 150 | if (!is_fsync_dnode(page)) |
125 | goto next; | 151 | goto next; |
126 | 152 | ||
127 | entry = get_fsync_inode(head, ino_of_node(page)); | 153 | entry = get_fsync_inode(head, ino_of_node(page)); |
128 | if (entry) { | 154 | if (entry) { |
129 | entry->blkaddr = blkaddr; | ||
130 | if (IS_INODE(page) && is_dent_dnode(page)) | 155 | if (IS_INODE(page) && is_dent_dnode(page)) |
131 | set_inode_flag(F2FS_I(entry->inode), | 156 | set_inode_flag(F2FS_I(entry->inode), |
132 | FI_INC_LINK); | 157 | FI_INC_LINK); |
@@ -134,48 +159,40 @@ static int find_fsync_dnodes(struct f2fs_sb_info *sbi, struct list_head *head) | |||
134 | if (IS_INODE(page) && is_dent_dnode(page)) { | 159 | if (IS_INODE(page) && is_dent_dnode(page)) { |
135 | err = recover_inode_page(sbi, page); | 160 | err = recover_inode_page(sbi, page); |
136 | if (err) | 161 | if (err) |
137 | goto unlock_out; | 162 | break; |
138 | } | 163 | } |
139 | 164 | ||
140 | /* add this fsync inode to the list */ | 165 | /* add this fsync inode to the list */ |
141 | entry = kmem_cache_alloc(fsync_entry_slab, GFP_NOFS); | 166 | entry = kmem_cache_alloc(fsync_entry_slab, GFP_NOFS); |
142 | if (!entry) { | 167 | if (!entry) { |
143 | err = -ENOMEM; | 168 | err = -ENOMEM; |
144 | goto unlock_out; | 169 | break; |
145 | } | 170 | } |
146 | 171 | ||
147 | entry->inode = f2fs_iget(sbi->sb, ino_of_node(page)); | 172 | entry->inode = f2fs_iget(sbi->sb, ino_of_node(page)); |
148 | if (IS_ERR(entry->inode)) { | 173 | if (IS_ERR(entry->inode)) { |
149 | err = PTR_ERR(entry->inode); | 174 | err = PTR_ERR(entry->inode); |
150 | kmem_cache_free(fsync_entry_slab, entry); | 175 | kmem_cache_free(fsync_entry_slab, entry); |
151 | goto unlock_out; | 176 | break; |
152 | } | 177 | } |
153 | |||
154 | list_add_tail(&entry->list, head); | 178 | list_add_tail(&entry->list, head); |
155 | entry->blkaddr = blkaddr; | ||
156 | } | ||
157 | if (IS_INODE(page)) { | ||
158 | err = recover_inode(entry->inode, page); | ||
159 | if (err == -ENOENT) { | ||
160 | goto next; | ||
161 | } else if (err) { | ||
162 | err = -EINVAL; | ||
163 | goto unlock_out; | ||
164 | } | ||
165 | } | 179 | } |
180 | entry->blkaddr = blkaddr; | ||
181 | |||
182 | err = recover_inode(entry->inode, page); | ||
183 | if (err && err != -ENOENT) | ||
184 | break; | ||
166 | next: | 185 | next: |
167 | /* check next segment */ | 186 | /* check next segment */ |
168 | blkaddr = next_blkaddr_of_node(page); | 187 | blkaddr = next_blkaddr_of_node(page); |
169 | } | 188 | } |
170 | unlock_out: | ||
171 | unlock_page(page); | 189 | unlock_page(page); |
172 | out: | 190 | out: |
173 | __free_pages(page, 0); | 191 | __free_pages(page, 0); |
174 | return err; | 192 | return err; |
175 | } | 193 | } |
176 | 194 | ||
177 | static void destroy_fsync_dnodes(struct f2fs_sb_info *sbi, | 195 | static void destroy_fsync_dnodes(struct list_head *head) |
178 | struct list_head *head) | ||
179 | { | 196 | { |
180 | struct fsync_inode_entry *entry, *tmp; | 197 | struct fsync_inode_entry *entry, *tmp; |
181 | 198 | ||
@@ -186,15 +203,15 @@ static void destroy_fsync_dnodes(struct f2fs_sb_info *sbi, | |||
186 | } | 203 | } |
187 | } | 204 | } |
188 | 205 | ||
189 | static void check_index_in_prev_nodes(struct f2fs_sb_info *sbi, | 206 | static int check_index_in_prev_nodes(struct f2fs_sb_info *sbi, |
190 | block_t blkaddr) | 207 | block_t blkaddr, struct dnode_of_data *dn) |
191 | { | 208 | { |
192 | struct seg_entry *sentry; | 209 | struct seg_entry *sentry; |
193 | unsigned int segno = GET_SEGNO(sbi, blkaddr); | 210 | unsigned int segno = GET_SEGNO(sbi, blkaddr); |
194 | unsigned short blkoff = GET_SEGOFF_FROM_SEG0(sbi, blkaddr) & | 211 | unsigned short blkoff = GET_SEGOFF_FROM_SEG0(sbi, blkaddr) & |
195 | (sbi->blocks_per_seg - 1); | 212 | (sbi->blocks_per_seg - 1); |
196 | struct f2fs_summary sum; | 213 | struct f2fs_summary sum; |
197 | nid_t ino; | 214 | nid_t ino, nid; |
198 | void *kaddr; | 215 | void *kaddr; |
199 | struct inode *inode; | 216 | struct inode *inode; |
200 | struct page *node_page; | 217 | struct page *node_page; |
@@ -203,7 +220,7 @@ static void check_index_in_prev_nodes(struct f2fs_sb_info *sbi, | |||
203 | 220 | ||
204 | sentry = get_seg_entry(sbi, segno); | 221 | sentry = get_seg_entry(sbi, segno); |
205 | if (!f2fs_test_bit(blkoff, sentry->cur_valid_map)) | 222 | if (!f2fs_test_bit(blkoff, sentry->cur_valid_map)) |
206 | return; | 223 | return 0; |
207 | 224 | ||
208 | /* Get the previous summary */ | 225 | /* Get the previous summary */ |
209 | for (i = CURSEG_WARM_DATA; i <= CURSEG_COLD_DATA; i++) { | 226 | for (i = CURSEG_WARM_DATA; i <= CURSEG_COLD_DATA; i++) { |
@@ -222,20 +239,39 @@ static void check_index_in_prev_nodes(struct f2fs_sb_info *sbi, | |||
222 | f2fs_put_page(sum_page, 1); | 239 | f2fs_put_page(sum_page, 1); |
223 | } | 240 | } |
224 | 241 | ||
242 | /* Use the locked dnode page and inode */ | ||
243 | nid = le32_to_cpu(sum.nid); | ||
244 | if (dn->inode->i_ino == nid) { | ||
245 | struct dnode_of_data tdn = *dn; | ||
246 | tdn.nid = nid; | ||
247 | tdn.node_page = dn->inode_page; | ||
248 | tdn.ofs_in_node = le16_to_cpu(sum.ofs_in_node); | ||
249 | truncate_data_blocks_range(&tdn, 1); | ||
250 | return 0; | ||
251 | } else if (dn->nid == nid) { | ||
252 | struct dnode_of_data tdn = *dn; | ||
253 | tdn.ofs_in_node = le16_to_cpu(sum.ofs_in_node); | ||
254 | truncate_data_blocks_range(&tdn, 1); | ||
255 | return 0; | ||
256 | } | ||
257 | |||
225 | /* Get the node page */ | 258 | /* Get the node page */ |
226 | node_page = get_node_page(sbi, le32_to_cpu(sum.nid)); | 259 | node_page = get_node_page(sbi, nid); |
260 | if (IS_ERR(node_page)) | ||
261 | return PTR_ERR(node_page); | ||
227 | bidx = start_bidx_of_node(ofs_of_node(node_page)) + | 262 | bidx = start_bidx_of_node(ofs_of_node(node_page)) + |
228 | le16_to_cpu(sum.ofs_in_node); | 263 | le16_to_cpu(sum.ofs_in_node); |
229 | ino = ino_of_node(node_page); | 264 | ino = ino_of_node(node_page); |
230 | f2fs_put_page(node_page, 1); | 265 | f2fs_put_page(node_page, 1); |
231 | 266 | ||
232 | /* Deallocate previous index in the node page */ | 267 | /* Deallocate previous index in the node page */ |
233 | inode = f2fs_iget(sbi->sb, ino); | 268 | inode = f2fs_iget(sbi->sb, ino); |
234 | if (IS_ERR(inode)) | 269 | if (IS_ERR(inode)) |
235 | return; | 270 | return PTR_ERR(inode); |
236 | 271 | ||
237 | truncate_hole(inode, bidx, bidx + 1); | 272 | truncate_hole(inode, bidx, bidx + 1); |
238 | iput(inode); | 273 | iput(inode); |
274 | return 0; | ||
239 | } | 275 | } |
240 | 276 | ||
241 | static int do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, | 277 | static int do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, |
@@ -245,7 +281,7 @@ static int do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, | |||
245 | struct dnode_of_data dn; | 281 | struct dnode_of_data dn; |
246 | struct f2fs_summary sum; | 282 | struct f2fs_summary sum; |
247 | struct node_info ni; | 283 | struct node_info ni; |
248 | int err = 0; | 284 | int err = 0, recovered = 0; |
249 | int ilock; | 285 | int ilock; |
250 | 286 | ||
251 | start = start_bidx_of_node(ofs_of_node(page)); | 287 | start = start_bidx_of_node(ofs_of_node(page)); |
@@ -283,13 +319,16 @@ static int do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, | |||
283 | } | 319 | } |
284 | 320 | ||
285 | /* Check the previous node page having this index */ | 321 | /* Check the previous node page having this index */ |
286 | check_index_in_prev_nodes(sbi, dest); | 322 | err = check_index_in_prev_nodes(sbi, dest, &dn); |
323 | if (err) | ||
324 | goto err; | ||
287 | 325 | ||
288 | set_summary(&sum, dn.nid, dn.ofs_in_node, ni.version); | 326 | set_summary(&sum, dn.nid, dn.ofs_in_node, ni.version); |
289 | 327 | ||
290 | /* write dummy data page */ | 328 | /* write dummy data page */ |
291 | recover_data_page(sbi, NULL, &sum, src, dest); | 329 | recover_data_page(sbi, NULL, &sum, src, dest); |
292 | update_extent_cache(dest, &dn); | 330 | update_extent_cache(dest, &dn); |
331 | recovered++; | ||
293 | } | 332 | } |
294 | dn.ofs_in_node++; | 333 | dn.ofs_in_node++; |
295 | } | 334 | } |
@@ -305,9 +344,14 @@ static int do_recover_data(struct f2fs_sb_info *sbi, struct inode *inode, | |||
305 | set_page_dirty(dn.node_page); | 344 | set_page_dirty(dn.node_page); |
306 | 345 | ||
307 | recover_node_page(sbi, dn.node_page, &sum, &ni, blkaddr); | 346 | recover_node_page(sbi, dn.node_page, &sum, &ni, blkaddr); |
347 | err: | ||
308 | f2fs_put_dnode(&dn); | 348 | f2fs_put_dnode(&dn); |
309 | mutex_unlock_op(sbi, ilock); | 349 | mutex_unlock_op(sbi, ilock); |
310 | return 0; | 350 | |
351 | f2fs_msg(sbi->sb, KERN_NOTICE, "recover_data: ino = %lx, " | ||
352 | "recovered_data = %d blocks, err = %d", | ||
353 | inode->i_ino, recovered, err); | ||
354 | return err; | ||
311 | } | 355 | } |
312 | 356 | ||
313 | static int recover_data(struct f2fs_sb_info *sbi, | 357 | static int recover_data(struct f2fs_sb_info *sbi, |
@@ -340,7 +384,7 @@ static int recover_data(struct f2fs_sb_info *sbi, | |||
340 | lock_page(page); | 384 | lock_page(page); |
341 | 385 | ||
342 | if (cp_ver != cpver_of_node(page)) | 386 | if (cp_ver != cpver_of_node(page)) |
343 | goto unlock_out; | 387 | break; |
344 | 388 | ||
345 | entry = get_fsync_inode(head, ino_of_node(page)); | 389 | entry = get_fsync_inode(head, ino_of_node(page)); |
346 | if (!entry) | 390 | if (!entry) |
@@ -348,7 +392,7 @@ static int recover_data(struct f2fs_sb_info *sbi, | |||
348 | 392 | ||
349 | err = do_recover_data(sbi, entry->inode, page, blkaddr); | 393 | err = do_recover_data(sbi, entry->inode, page, blkaddr); |
350 | if (err) | 394 | if (err) |
351 | goto out; | 395 | break; |
352 | 396 | ||
353 | if (entry->blkaddr == blkaddr) { | 397 | if (entry->blkaddr == blkaddr) { |
354 | iput(entry->inode); | 398 | iput(entry->inode); |
@@ -359,7 +403,6 @@ next: | |||
359 | /* check next segment */ | 403 | /* check next segment */ |
360 | blkaddr = next_blkaddr_of_node(page); | 404 | blkaddr = next_blkaddr_of_node(page); |
361 | } | 405 | } |
362 | unlock_out: | ||
363 | unlock_page(page); | 406 | unlock_page(page); |
364 | out: | 407 | out: |
365 | __free_pages(page, 0); | 408 | __free_pages(page, 0); |
@@ -382,6 +425,7 @@ int recover_fsync_data(struct f2fs_sb_info *sbi) | |||
382 | INIT_LIST_HEAD(&inode_list); | 425 | INIT_LIST_HEAD(&inode_list); |
383 | 426 | ||
384 | /* step #1: find fsynced inode numbers */ | 427 | /* step #1: find fsynced inode numbers */ |
428 | sbi->por_doing = 1; | ||
385 | err = find_fsync_dnodes(sbi, &inode_list); | 429 | err = find_fsync_dnodes(sbi, &inode_list); |
386 | if (err) | 430 | if (err) |
387 | goto out; | 431 | goto out; |
@@ -390,13 +434,13 @@ int recover_fsync_data(struct f2fs_sb_info *sbi) | |||
390 | goto out; | 434 | goto out; |
391 | 435 | ||
392 | /* step #2: recover data */ | 436 | /* step #2: recover data */ |
393 | sbi->por_doing = 1; | ||
394 | err = recover_data(sbi, &inode_list, CURSEG_WARM_NODE); | 437 | err = recover_data(sbi, &inode_list, CURSEG_WARM_NODE); |
395 | sbi->por_doing = 0; | ||
396 | BUG_ON(!list_empty(&inode_list)); | 438 | BUG_ON(!list_empty(&inode_list)); |
397 | out: | 439 | out: |
398 | destroy_fsync_dnodes(sbi, &inode_list); | 440 | destroy_fsync_dnodes(&inode_list); |
399 | kmem_cache_destroy(fsync_entry_slab); | 441 | kmem_cache_destroy(fsync_entry_slab); |
400 | write_checkpoint(sbi, false); | 442 | sbi->por_doing = 0; |
443 | if (!err) | ||
444 | write_checkpoint(sbi, false); | ||
401 | return err; | 445 | return err; |
402 | } | 446 | } |