diff options
author | Nick Piggin <npiggin@kernel.dk> | 2011-01-07 01:49:49 -0500 |
---|---|---|
committer | Nick Piggin <npiggin@kernel.dk> | 2011-01-07 01:50:26 -0500 |
commit | fa0d7e3de6d6fc5004ad9dea0dd6b286af8f03e9 (patch) | |
tree | 203e0f73883e4c26b5597e36042386a1237dab35 /fs/inode.c | |
parent | 77812a1ef139d84270d27faacc0630c887411013 (diff) |
fs: icache RCU free inodes
RCU free the struct inode. This will allow:
- Subsequent store-free path walking patch. The inode must be consulted for
permissions when walking, so an RCU inode reference is a must.
- sb_inode_list_lock to be moved inside i_lock because sb list walkers who want
to take i_lock no longer need to take sb_inode_list_lock to walk the list in
the first place. This will simplify and optimize locking.
- Could remove some nested trylock loops in dcache code
- Could potentially simplify things a bit in VM land. Do not need to take the
page lock to follow page->mapping.
The downsides of this is the performance cost of using RCU. In a simple
creat/unlink microbenchmark, performance drops by about 10% due to inability to
reuse cache-hot slab objects. As iterations increase and RCU freeing starts
kicking over, this increases to about 20%.
In cases where inode lifetimes are longer (ie. many inodes may be allocated
during the average life span of a single inode), a lot of this cache reuse is
not applicable, so the regression caused by this patch is smaller.
The cache-hot regression could largely be avoided by using SLAB_DESTROY_BY_RCU,
however this adds some complexity to list walking and store-free path walking,
so I prefer to implement this at a later date, if it is shown to be a win in
real situations. I haven't found a regression in any non-micro benchmark so I
doubt it will be a problem.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Diffstat (limited to 'fs/inode.c')
-rw-r--r-- | fs/inode.c | 10 |
1 files changed, 9 insertions, 1 deletions
diff --git a/fs/inode.c b/fs/inode.c index 5a0a898f55d1..6751dfe8cc06 100644 --- a/fs/inode.c +++ b/fs/inode.c | |||
@@ -272,6 +272,13 @@ void __destroy_inode(struct inode *inode) | |||
272 | } | 272 | } |
273 | EXPORT_SYMBOL(__destroy_inode); | 273 | EXPORT_SYMBOL(__destroy_inode); |
274 | 274 | ||
275 | static void i_callback(struct rcu_head *head) | ||
276 | { | ||
277 | struct inode *inode = container_of(head, struct inode, i_rcu); | ||
278 | INIT_LIST_HEAD(&inode->i_dentry); | ||
279 | kmem_cache_free(inode_cachep, inode); | ||
280 | } | ||
281 | |||
275 | static void destroy_inode(struct inode *inode) | 282 | static void destroy_inode(struct inode *inode) |
276 | { | 283 | { |
277 | BUG_ON(!list_empty(&inode->i_lru)); | 284 | BUG_ON(!list_empty(&inode->i_lru)); |
@@ -279,7 +286,7 @@ static void destroy_inode(struct inode *inode) | |||
279 | if (inode->i_sb->s_op->destroy_inode) | 286 | if (inode->i_sb->s_op->destroy_inode) |
280 | inode->i_sb->s_op->destroy_inode(inode); | 287 | inode->i_sb->s_op->destroy_inode(inode); |
281 | else | 288 | else |
282 | kmem_cache_free(inode_cachep, (inode)); | 289 | call_rcu(&inode->i_rcu, i_callback); |
283 | } | 290 | } |
284 | 291 | ||
285 | /* | 292 | /* |
@@ -432,6 +439,7 @@ void end_writeback(struct inode *inode) | |||
432 | BUG_ON(!(inode->i_state & I_FREEING)); | 439 | BUG_ON(!(inode->i_state & I_FREEING)); |
433 | BUG_ON(inode->i_state & I_CLEAR); | 440 | BUG_ON(inode->i_state & I_CLEAR); |
434 | inode_sync_wait(inode); | 441 | inode_sync_wait(inode); |
442 | /* don't need i_lock here, no concurrent mods to i_state */ | ||
435 | inode->i_state = I_FREEING | I_CLEAR; | 443 | inode->i_state = I_FREEING | I_CLEAR; |
436 | } | 444 | } |
437 | EXPORT_SYMBOL(end_writeback); | 445 | EXPORT_SYMBOL(end_writeback); |