aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorLi Zefan <lizefan@huawei.com>2013-09-09 23:43:37 -0400
committerPekka Enberg <penberg@kernel.org>2013-12-29 06:44:45 -0500
commit8afb1474db4701d1ab80cd8251137a3260e6913e (patch)
tree34f9ac6ef791ed6c00dbd0a6ddabb419c67bd813 /mm
parente1168c2cc4a5a5e495a53e067a6be6b4f29abfe0 (diff)
slub: Fix calculation of cpu slabs
/sys/kernel/slab/:t-0000048 # cat cpu_slabs 231 N0=16 N1=215 /sys/kernel/slab/:t-0000048 # cat slabs 145 N0=36 N1=109 See, the number of slabs is smaller than that of cpu slabs. The bug was introduced by commit 49e2258586b423684f03c278149ab46d8f8b6700 ("slub: per cpu cache for partial pages"). We should use page->pages instead of page->pobjects when calculating the number of cpu partial slabs. This also fixes the mapping of slabs and nodes. As there's no variable storing the number of total/active objects in cpu partial slabs, and we don't have user interfaces requiring those statistics, I just add WARN_ON for those cases. Cc: <stable@vger.kernel.org> # 3.2+ Acked-by: Christoph Lameter <cl@linux.com> Reviewed-by: Wanpeng Li <liwanp@linux.vnet.ibm.com> Signed-off-by: Li Zefan <lizefan@huawei.com> Signed-off-by: Pekka Enberg <penberg@kernel.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/slub.c8
1 files changed, 7 insertions, 1 deletions
diff --git a/mm/slub.c b/mm/slub.c
index 545a170ebf9f..89490d9d91e0 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -4299,7 +4299,13 @@ static ssize_t show_slab_objects(struct kmem_cache *s,
4299 4299
4300 page = ACCESS_ONCE(c->partial); 4300 page = ACCESS_ONCE(c->partial);
4301 if (page) { 4301 if (page) {
4302 x = page->pobjects; 4302 node = page_to_nid(page);
4303 if (flags & SO_TOTAL)
4304 WARN_ON_ONCE(1);
4305 else if (flags & SO_OBJECTS)
4306 WARN_ON_ONCE(1);
4307 else
4308 x = page->pages;
4303 total += x; 4309 total += x;
4304 nodes[node] += x; 4310 nodes[node] += x;
4305 } 4311 }