diff options
author | Wu Fengguang <fengguang.wu@intel.com> | 2009-09-16 05:50:14 -0400 |
---|---|---|
committer | Andi Kleen <ak@linux.intel.com> | 2009-09-16 05:50:14 -0400 |
commit | 6746aff74da293b5fd24e5c68b870b721e86cd5f (patch) | |
tree | 6d2b9e99d247e788c13665ce5e8a0753538ef641 /mm/shmem.c | |
parent | 257187362123f15d9d1e09918cf87cebbea4e786 (diff) |
HWPOISON: shmem: call set_page_dirty() with locked page
The dirtying of page and set_page_dirty() can be moved into the page lock.
- In shmem_write_end(), the page was dirtied while the page lock was held,
but it's being marked dirty just after dropping the page lock.
- In shmem_symlink(), both dirtying and marking can be moved into page lock.
It's valuable for the hwpoison code to know whether one bad page can be dropped
without losing data. It mainly judges by testing the PG_dirty bit after taking
the page lock. So it becomes important that the dirtying of page and the
marking of dirtiness are both done inside the page lock. Which is a common
practice, but sadly not a rule.
The noticeable exceptions are
- mapped pages
- pages with buffer_heads
The above pages could go dirty at any time. Fortunately the hwpoison will
unmap the page and release the buffer_heads beforehand anyway.
Many other types of pages (eg. metadata pages) can also be dirtied at will by
their owners, the hwpoison code cannot do meaningful things to them anyway.
Only the dirtiness of pagecache pages owned by regular files are interested.
v2: AK: Add comment about set_page_dirty rules (suggested by Peter Zijlstra)
Acked-by: Hugh Dickins <hugh.dickins@tiscali.co.uk>
Reviewed-by: WANG Cong <xiyou.wangcong@gmail.com>
Signed-off-by: Wu Fengguang <fengguang.wu@intel.com>
Signed-off-by: Andi Kleen <ak@linux.intel.com>
Diffstat (limited to 'mm/shmem.c')
-rw-r--r-- | mm/shmem.c | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/mm/shmem.c b/mm/shmem.c index 5a0b3d4055f..46936601e37 100644 --- a/mm/shmem.c +++ b/mm/shmem.c | |||
@@ -1630,8 +1630,8 @@ shmem_write_end(struct file *file, struct address_space *mapping, | |||
1630 | if (pos + copied > inode->i_size) | 1630 | if (pos + copied > inode->i_size) |
1631 | i_size_write(inode, pos + copied); | 1631 | i_size_write(inode, pos + copied); |
1632 | 1632 | ||
1633 | unlock_page(page); | ||
1634 | set_page_dirty(page); | 1633 | set_page_dirty(page); |
1634 | unlock_page(page); | ||
1635 | page_cache_release(page); | 1635 | page_cache_release(page); |
1636 | 1636 | ||
1637 | return copied; | 1637 | return copied; |
@@ -1968,13 +1968,13 @@ static int shmem_symlink(struct inode *dir, struct dentry *dentry, const char *s | |||
1968 | iput(inode); | 1968 | iput(inode); |
1969 | return error; | 1969 | return error; |
1970 | } | 1970 | } |
1971 | unlock_page(page); | ||
1972 | inode->i_mapping->a_ops = &shmem_aops; | 1971 | inode->i_mapping->a_ops = &shmem_aops; |
1973 | inode->i_op = &shmem_symlink_inode_operations; | 1972 | inode->i_op = &shmem_symlink_inode_operations; |
1974 | kaddr = kmap_atomic(page, KM_USER0); | 1973 | kaddr = kmap_atomic(page, KM_USER0); |
1975 | memcpy(kaddr, symname, len); | 1974 | memcpy(kaddr, symname, len); |
1976 | kunmap_atomic(kaddr, KM_USER0); | 1975 | kunmap_atomic(kaddr, KM_USER0); |
1977 | set_page_dirty(page); | 1976 | set_page_dirty(page); |
1977 | unlock_page(page); | ||
1978 | page_cache_release(page); | 1978 | page_cache_release(page); |
1979 | } | 1979 | } |
1980 | if (dir->i_mode & S_ISGID) | 1980 | if (dir->i_mode & S_ISGID) |