aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorKAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>2008-02-07 03:14:11 -0500
committerLinus Torvalds <torvalds@woody.linux-foundation.org>2008-02-07 11:42:20 -0500
commitff7283fa3a66823933991ad55a558a3a01d5ab27 (patch)
treee183c588497ce9f18fea1589fa66b6a000e2cbb8 /mm
parentae41be374293e70e1ed441d986afcc6e744ef9d9 (diff)
bugfix for memory cgroup controller: avoid !PageLRU page in mem_cgroup_isolate_pages
This patch makes mem_cgroup_isolate_pages() to be - ignore !PageLRU pages. - fixes the bug that isolation makes no progress if page_zone(page) != zone page once find. (just increment scan in this case.) kswapd and memory migration removes a page from list when it handles a page for reclaiming/migration. Because __isolate_lru_page() doesn't moves page !PageLRU pages, it will be safe to avoid touching !PageLRU() page and its page_cgroup. Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com> Cc: Balbir Singh <balbir@linux.vnet.ibm.com> Cc: Pavel Emelianov <xemul@openvz.org> Cc: Paul Menage <menage@google.com> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: "Eric W. Biederman" <ebiederm@xmission.com> Cc: Nick Piggin <nickpiggin@yahoo.com.au> Cc: Kirill Korotaev <dev@sw.ru> Cc: Herbert Poetzl <herbert@13thfloor.at> Cc: David Rientjes <rientjes@google.com> Cc: Vaidyanathan Srinivasan <svaidy@linux.vnet.ibm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/memcontrol.c13
1 files changed, 10 insertions, 3 deletions
diff --git a/mm/memcontrol.c b/mm/memcontrol.c
index 128f45c16fa6..e8493fb2d69e 100644
--- a/mm/memcontrol.c
+++ b/mm/memcontrol.c
@@ -250,7 +250,7 @@ unsigned long mem_cgroup_isolate_pages(unsigned long nr_to_scan,
250 unsigned long scan; 250 unsigned long scan;
251 LIST_HEAD(pc_list); 251 LIST_HEAD(pc_list);
252 struct list_head *src; 252 struct list_head *src;
253 struct page_cgroup *pc; 253 struct page_cgroup *pc, *tmp;
254 254
255 if (active) 255 if (active)
256 src = &mem_cont->active_list; 256 src = &mem_cont->active_list;
@@ -258,11 +258,18 @@ unsigned long mem_cgroup_isolate_pages(unsigned long nr_to_scan,
258 src = &mem_cont->inactive_list; 258 src = &mem_cont->inactive_list;
259 259
260 spin_lock(&mem_cont->lru_lock); 260 spin_lock(&mem_cont->lru_lock);
261 for (scan = 0; scan < nr_to_scan && !list_empty(src); scan++) { 261 scan = 0;
262 pc = list_entry(src->prev, struct page_cgroup, lru); 262 list_for_each_entry_safe_reverse(pc, tmp, src, lru) {
263 if (scan++ > nr_to_scan)
264 break;
263 page = pc->page; 265 page = pc->page;
264 VM_BUG_ON(!pc); 266 VM_BUG_ON(!pc);
265 267
268 if (unlikely(!PageLRU(page))) {
269 scan--;
270 continue;
271 }
272
266 if (PageActive(page) && !active) { 273 if (PageActive(page) && !active) {
267 __mem_cgroup_move_lists(pc, true); 274 __mem_cgroup_move_lists(pc, true);
268 scan--; 275 scan--;