diff options
author | Hugh Dickins <hugh@veritas.com> | 2005-10-29 21:16:18 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@g5.osdl.org> | 2005-10-30 00:40:39 -0400 |
commit | 365e9c87a982c03d0af3886e29d877f581b59611 (patch) | |
tree | d06c1918ca9fe6677d7e4e869555e095004274f7 /mm/mmap.c | |
parent | 861f2fb8e796022b4928cab9c74fca6681a1c557 (diff) |
[PATCH] mm: update_hiwaters just in time
update_mem_hiwater has attracted various criticisms, in particular from those
concerned with mm scalability. Originally it was called whenever rss or
total_vm got raised. Then many of those callsites were replaced by a timer
tick call from account_system_time. Now Frank van Maarseveen reports that to
be found inadequate. How about this? Works for Frank.
Replace update_mem_hiwater, a poor combination of two unrelated ops, by macros
update_hiwater_rss and update_hiwater_vm. Don't attempt to keep
mm->hiwater_rss up to date at timer tick, nor every time we raise rss (usually
by 1): those are hot paths. Do the opposite, update only when about to lower
rss (usually by many), or just before final accounting in do_exit. Handle
mm->hiwater_vm in the same way, though it's much less of an issue. Demand
that whoever collects these hiwater statistics do the work of taking the
maximum with rss or total_vm.
And there has been no collector of these hiwater statistics in the tree. The
new convention needs an example, so match Frank's usage by adding a VmPeak
line above VmSize to /proc/<pid>/status, and also a VmHWM line above VmRSS
(High-Water-Mark or High-Water-Memory).
There was a particular anomaly during mremap move, that hiwater_vm might be
captured too high. A fleeting such anomaly remains, but it's quickly
corrected now, whereas before it would stick.
What locking? None: if the app is racy then these statistics will be racy,
it's not worth any overhead to make them exact. But whenever it suits,
hiwater_vm is updated under exclusive mmap_sem, and hiwater_rss under
page_table_lock (for now) or with preemption disabled (later on): without
going to any trouble, minimize the time between reading current values and
updating, to minimize those occasions when a racing thread bumps a count up
and back down in between.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'mm/mmap.c')
-rw-r--r-- | mm/mmap.c | 4 |
1 files changed, 4 insertions, 0 deletions
@@ -1640,6 +1640,8 @@ find_extend_vma(struct mm_struct * mm, unsigned long addr) | |||
1640 | */ | 1640 | */ |
1641 | static void remove_vma_list(struct mm_struct *mm, struct vm_area_struct *vma) | 1641 | static void remove_vma_list(struct mm_struct *mm, struct vm_area_struct *vma) |
1642 | { | 1642 | { |
1643 | /* Update high watermark before we lower total_vm */ | ||
1644 | update_hiwater_vm(mm); | ||
1643 | do { | 1645 | do { |
1644 | long nrpages = vma_pages(vma); | 1646 | long nrpages = vma_pages(vma); |
1645 | 1647 | ||
@@ -1668,6 +1670,7 @@ static void unmap_region(struct mm_struct *mm, | |||
1668 | lru_add_drain(); | 1670 | lru_add_drain(); |
1669 | spin_lock(&mm->page_table_lock); | 1671 | spin_lock(&mm->page_table_lock); |
1670 | tlb = tlb_gather_mmu(mm, 0); | 1672 | tlb = tlb_gather_mmu(mm, 0); |
1673 | update_hiwater_rss(mm); | ||
1671 | unmap_vmas(&tlb, mm, vma, start, end, &nr_accounted, NULL); | 1674 | unmap_vmas(&tlb, mm, vma, start, end, &nr_accounted, NULL); |
1672 | vm_unacct_memory(nr_accounted); | 1675 | vm_unacct_memory(nr_accounted); |
1673 | free_pgtables(&tlb, vma, prev? prev->vm_end: FIRST_USER_ADDRESS, | 1676 | free_pgtables(&tlb, vma, prev? prev->vm_end: FIRST_USER_ADDRESS, |
@@ -1953,6 +1956,7 @@ void exit_mmap(struct mm_struct *mm) | |||
1953 | 1956 | ||
1954 | flush_cache_mm(mm); | 1957 | flush_cache_mm(mm); |
1955 | tlb = tlb_gather_mmu(mm, 1); | 1958 | tlb = tlb_gather_mmu(mm, 1); |
1959 | /* Don't update_hiwater_rss(mm) here, do_exit already did */ | ||
1956 | /* Use -1 here to ensure all VMAs in the mm are unmapped */ | 1960 | /* Use -1 here to ensure all VMAs in the mm are unmapped */ |
1957 | end = unmap_vmas(&tlb, mm, vma, 0, -1, &nr_accounted, NULL); | 1961 | end = unmap_vmas(&tlb, mm, vma, 0, -1, &nr_accounted, NULL); |
1958 | vm_unacct_memory(nr_accounted); | 1962 | vm_unacct_memory(nr_accounted); |