diff options
author | Hugh Dickins <hughd@google.com> | 2012-05-29 18:06:39 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2012-05-29 19:22:22 -0400 |
commit | ec9516fbc5fa814014991e1ae7f8860127122105 (patch) | |
tree | e719b64b939e791499e115a63f4e7d1dd0448b16 /mm | |
parent | 2f6e38f3cd17a7858112f538c1700c747170db1f (diff) |
tmpfs: optimize clearing when writing
Nick proposed years ago that tmpfs should avoid clearing its pages where
write will overwrite them with new data, as ramfs has long done. But I
messed it up and just got bad data. Tried again recently, it works
fine.
Here's time output for writing 4GiB 16 times on this Core i5 laptop:
before: real 0m21.169s user 0m0.028s sys 0m21.057s
real 0m21.382s user 0m0.016s sys 0m21.289s
real 0m21.311s user 0m0.020s sys 0m21.217s
after: real 0m18.273s user 0m0.032s sys 0m18.165s
real 0m18.354s user 0m0.020s sys 0m18.265s
real 0m18.440s user 0m0.032s sys 0m18.337s
ramfs: real 0m16.860s user 0m0.028s sys 0m16.765s
real 0m17.382s user 0m0.040s sys 0m17.273s
real 0m17.133s user 0m0.044s sys 0m17.021s
Yes, I have done perf reports, but they need more explanation than they
deserve: in summary, clear_page vanishes, its cache loading shifts into
copy_user_generic_unrolled; shmem_getpage_gfp goes down, and
surprisingly mark_page_accessed goes way up - I think because they are
respectively where the cache gets to be reloaded after being purged by
clear or copy.
Suggested-by: Nick Piggin <npiggin@gmail.com>
Signed-off-by: Hugh Dickins <hughd@google.com>
Cc: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/shmem.c | 20 |
1 files changed, 17 insertions, 3 deletions
diff --git a/mm/shmem.c b/mm/shmem.c index fe5ae6962ab3..45c26476f0fc 100644 --- a/mm/shmem.c +++ b/mm/shmem.c | |||
@@ -1095,9 +1095,14 @@ repeat: | |||
1095 | shmem_recalc_inode(inode); | 1095 | shmem_recalc_inode(inode); |
1096 | spin_unlock(&info->lock); | 1096 | spin_unlock(&info->lock); |
1097 | 1097 | ||
1098 | clear_highpage(page); | 1098 | /* |
1099 | flush_dcache_page(page); | 1099 | * Let SGP_WRITE caller clear ends if write does not fill page |
1100 | SetPageUptodate(page); | 1100 | */ |
1101 | if (sgp != SGP_WRITE) { | ||
1102 | clear_highpage(page); | ||
1103 | flush_dcache_page(page); | ||
1104 | SetPageUptodate(page); | ||
1105 | } | ||
1101 | if (sgp == SGP_DIRTY) | 1106 | if (sgp == SGP_DIRTY) |
1102 | set_page_dirty(page); | 1107 | set_page_dirty(page); |
1103 | } | 1108 | } |
@@ -1307,6 +1312,14 @@ shmem_write_end(struct file *file, struct address_space *mapping, | |||
1307 | if (pos + copied > inode->i_size) | 1312 | if (pos + copied > inode->i_size) |
1308 | i_size_write(inode, pos + copied); | 1313 | i_size_write(inode, pos + copied); |
1309 | 1314 | ||
1315 | if (!PageUptodate(page)) { | ||
1316 | if (copied < PAGE_CACHE_SIZE) { | ||
1317 | unsigned from = pos & (PAGE_CACHE_SIZE - 1); | ||
1318 | zero_user_segments(page, 0, from, | ||
1319 | from + copied, PAGE_CACHE_SIZE); | ||
1320 | } | ||
1321 | SetPageUptodate(page); | ||
1322 | } | ||
1310 | set_page_dirty(page); | 1323 | set_page_dirty(page); |
1311 | unlock_page(page); | 1324 | unlock_page(page); |
1312 | page_cache_release(page); | 1325 | page_cache_release(page); |
@@ -1768,6 +1781,7 @@ static int shmem_symlink(struct inode *dir, struct dentry *dentry, const char *s | |||
1768 | kaddr = kmap_atomic(page); | 1781 | kaddr = kmap_atomic(page); |
1769 | memcpy(kaddr, symname, len); | 1782 | memcpy(kaddr, symname, len); |
1770 | kunmap_atomic(kaddr); | 1783 | kunmap_atomic(kaddr); |
1784 | SetPageUptodate(page); | ||
1771 | set_page_dirty(page); | 1785 | set_page_dirty(page); |
1772 | unlock_page(page); | 1786 | unlock_page(page); |
1773 | page_cache_release(page); | 1787 | page_cache_release(page); |