aboutsummaryrefslogtreecommitdiffstats
path: root/mm/memcontrol.c
diff options
context:
space:
mode:
authorHugh Dickins <hugh@veritas.com>2008-03-04 17:29:08 -0500
committerLinus Torvalds <torvalds@woody.linux-foundation.org>2008-03-04 19:35:15 -0500
commit7e924aafa4b03ff71de34af8553d9a1ebc86c071 (patch)
tree21735b7369ede27f34c41184b0a686020b80e2c6 /mm/memcontrol.c
parent9442ec9df40d952b0de185ae5638a74970388e01 (diff)
memcg: mem_cgroup_charge never NULL
My memcgroup patch to fix hang with shmem/tmpfs added NULL page handling to mem_cgroup_charge_common. It seemed convenient at the time, but hard to justify now: there's a perfectly appropriate swappage to charge and uncharge instead, this is not on any hot path through shmem_getpage, and no performance hit was observed from the slight extra overhead. So revert that NULL page handling from mem_cgroup_charge_common; and make it clearer by bringing page_cgroup_assign_new_page_cgroup into its body - that was a helper I found more of a hindrance to understanding. Signed-off-by: Hugh Dickins <hugh@veritas.com> Cc: David Rientjes <rientjes@google.com> Acked-by: Balbir Singh <balbir@linux.vnet.ibm.com> Acked-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com> Cc: Hirokazu Takahashi <taka@valinux.co.jp> Cc: YAMAMOTO Takashi <yamamoto@valinux.co.jp> Cc: Paul Menage <menage@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/memcontrol.c')
-rw-r--r--mm/memcontrol.c61
1 files changed, 21 insertions, 40 deletions
diff --git a/mm/memcontrol.c b/mm/memcontrol.c
index 9e170d3c71e5..83ba13ad31e1 100644
--- a/mm/memcontrol.c
+++ b/mm/memcontrol.c
@@ -301,25 +301,6 @@ static void __always_inline unlock_page_cgroup(struct page *page)
301} 301}
302 302
303/* 303/*
304 * Tie new page_cgroup to struct page under lock_page_cgroup()
305 * This can fail if the page has been tied to a page_cgroup.
306 * If success, returns 0.
307 */
308static int page_cgroup_assign_new_page_cgroup(struct page *page,
309 struct page_cgroup *pc)
310{
311 int ret = 0;
312
313 lock_page_cgroup(page);
314 if (!page_get_page_cgroup(page))
315 page_assign_page_cgroup(page, pc);
316 else /* A page is tied to other pc. */
317 ret = 1;
318 unlock_page_cgroup(page);
319 return ret;
320}
321
322/*
323 * Clear page->page_cgroup member under lock_page_cgroup(). 304 * Clear page->page_cgroup member under lock_page_cgroup().
324 * If given "pc" value is different from one page->page_cgroup, 305 * If given "pc" value is different from one page->page_cgroup,
325 * page->cgroup is not cleared. 306 * page->cgroup is not cleared.
@@ -585,26 +566,24 @@ static int mem_cgroup_charge_common(struct page *page, struct mm_struct *mm,
585 * with it 566 * with it
586 */ 567 */
587retry: 568retry:
588 if (page) { 569 lock_page_cgroup(page);
589 lock_page_cgroup(page); 570 pc = page_get_page_cgroup(page);
590 pc = page_get_page_cgroup(page); 571 /*
591 /* 572 * The page_cgroup exists and
592 * The page_cgroup exists and 573 * the page has already been accounted.
593 * the page has already been accounted. 574 */
594 */ 575 if (pc) {
595 if (pc) { 576 if (unlikely(!atomic_inc_not_zero(&pc->ref_cnt))) {
596 if (unlikely(!atomic_inc_not_zero(&pc->ref_cnt))) { 577 /* this page is under being uncharged ? */
597 /* this page is under being uncharged ? */ 578 unlock_page_cgroup(page);
598 unlock_page_cgroup(page); 579 cpu_relax();
599 cpu_relax(); 580 goto retry;
600 goto retry; 581 } else {
601 } else { 582 unlock_page_cgroup(page);
602 unlock_page_cgroup(page); 583 goto done;
603 goto done;
604 }
605 } 584 }
606 unlock_page_cgroup(page);
607 } 585 }
586 unlock_page_cgroup(page);
608 587
609 pc = kzalloc(sizeof(struct page_cgroup), gfp_mask); 588 pc = kzalloc(sizeof(struct page_cgroup), gfp_mask);
610 if (pc == NULL) 589 if (pc == NULL)
@@ -663,7 +642,9 @@ retry:
663 if (ctype == MEM_CGROUP_CHARGE_TYPE_CACHE) 642 if (ctype == MEM_CGROUP_CHARGE_TYPE_CACHE)
664 pc->flags |= PAGE_CGROUP_FLAG_CACHE; 643 pc->flags |= PAGE_CGROUP_FLAG_CACHE;
665 644
666 if (!page || page_cgroup_assign_new_page_cgroup(page, pc)) { 645 lock_page_cgroup(page);
646 if (page_get_page_cgroup(page)) {
647 unlock_page_cgroup(page);
667 /* 648 /*
668 * Another charge has been added to this page already. 649 * Another charge has been added to this page already.
669 * We take lock_page_cgroup(page) again and read 650 * We take lock_page_cgroup(page) again and read
@@ -672,10 +653,10 @@ retry:
672 res_counter_uncharge(&mem->res, PAGE_SIZE); 653 res_counter_uncharge(&mem->res, PAGE_SIZE);
673 css_put(&mem->css); 654 css_put(&mem->css);
674 kfree(pc); 655 kfree(pc);
675 if (!page)
676 goto done;
677 goto retry; 656 goto retry;
678 } 657 }
658 page_assign_page_cgroup(page, pc);
659 unlock_page_cgroup(page);
679 660
680 mz = page_cgroup_zoneinfo(pc); 661 mz = page_cgroup_zoneinfo(pc);
681 spin_lock_irqsave(&mz->lru_lock, flags); 662 spin_lock_irqsave(&mz->lru_lock, flags);