diff options
author | Josef Bacik <josef@redhat.com> | 2011-04-25 19:43:52 -0400 |
---|---|---|
committer | Chris Mason <chris.mason@oracle.com> | 2011-04-25 19:43:52 -0400 |
commit | a4f0162fd4490daf2c823c185fff79080d266a7c (patch) | |
tree | c2a7186103a4b31115eb10555864620a46a18ad9 /fs | |
parent | f789b684bdb96e7ec2fce79445555d4fd55fb94c (diff) |
Btrfs: free bitmaps properly when evicting the cache
If our space cache is wrong, we do the right thing and free up everything that
we loaded, however we don't reset the total_bitmaps counter or the thresholds or
anything. So in btrfs_remove_free_space_cache make sure to call free_bitmap()
if it's a bitmap, this will keep us from panicing when we check to make sure we
don't have too many bitmaps. Thanks,
Signed-off-by: Josef Bacik <josef@redhat.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
Diffstat (limited to 'fs')
-rw-r--r-- | fs/btrfs/free-space-cache.c | 11 |
1 files changed, 7 insertions, 4 deletions
diff --git a/fs/btrfs/free-space-cache.c b/fs/btrfs/free-space-cache.c index 13c29b12a213..63731a1fb0a1 100644 --- a/fs/btrfs/free-space-cache.c +++ b/fs/btrfs/free-space-cache.c | |||
@@ -1768,10 +1768,13 @@ void btrfs_remove_free_space_cache(struct btrfs_block_group_cache *block_group) | |||
1768 | 1768 | ||
1769 | while ((node = rb_last(&block_group->free_space_offset)) != NULL) { | 1769 | while ((node = rb_last(&block_group->free_space_offset)) != NULL) { |
1770 | info = rb_entry(node, struct btrfs_free_space, offset_index); | 1770 | info = rb_entry(node, struct btrfs_free_space, offset_index); |
1771 | unlink_free_space(block_group, info); | 1771 | if (!info->bitmap) { |
1772 | if (info->bitmap) | 1772 | unlink_free_space(block_group, info); |
1773 | kfree(info->bitmap); | 1773 | kmem_cache_free(btrfs_free_space_cachep, info); |
1774 | kmem_cache_free(btrfs_free_space_cachep, info); | 1774 | } else { |
1775 | free_bitmap(block_group, info); | ||
1776 | } | ||
1777 | |||
1775 | if (need_resched()) { | 1778 | if (need_resched()) { |
1776 | spin_unlock(&block_group->tree_lock); | 1779 | spin_unlock(&block_group->tree_lock); |
1777 | cond_resched(); | 1780 | cond_resched(); |