aboutsummaryrefslogtreecommitdiffstats
path: root/mm/memcontrol.c
diff options
context:
space:
mode:
authorGlauber Costa <glommer@parallels.com>2011-12-11 16:47:01 -0500
committerDavid S. Miller <davem@davemloft.net>2011-12-12 19:03:55 -0500
commite5671dfae59b165e2adfd4dfbdeab11ac8db5bda (patch)
treedaf5570e8da71934970daa4b2044c6f13ee98f9d /mm/memcontrol.c
parent08e34eb14fe4cfd934b5c169a7682a969457c4ea (diff)
Basic kernel memory functionality for the Memory Controller
This patch lays down the foundation for the kernel memory component of the Memory Controller. As of today, I am only laying down the following files: * memory.independent_kmem_limit * memory.kmem.limit_in_bytes (currently ignored) * memory.kmem.usage_in_bytes (always zero) Signed-off-by: Glauber Costa <glommer@parallels.com> CC: Kirill A. Shutemov <kirill@shutemov.name> CC: Paul Menage <paul@paulmenage.org> CC: Greg Thelen <gthelen@google.com> CC: Johannes Weiner <jweiner@redhat.com> CC: Michal Hocko <mhocko@suse.cz> Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'mm/memcontrol.c')
-rw-r--r--mm/memcontrol.c105
1 files changed, 100 insertions, 5 deletions
diff --git a/mm/memcontrol.c b/mm/memcontrol.c
index 6aff93c98aca..9fbcff71245e 100644
--- a/mm/memcontrol.c
+++ b/mm/memcontrol.c
@@ -227,6 +227,10 @@ struct mem_cgroup {
227 */ 227 */
228 struct res_counter memsw; 228 struct res_counter memsw;
229 /* 229 /*
230 * the counter to account for kmem usage.
231 */
232 struct res_counter kmem;
233 /*
230 * Per cgroup active and inactive list, similar to the 234 * Per cgroup active and inactive list, similar to the
231 * per zone LRU lists. 235 * per zone LRU lists.
232 */ 236 */
@@ -277,6 +281,11 @@ struct mem_cgroup {
277 */ 281 */
278 unsigned long move_charge_at_immigrate; 282 unsigned long move_charge_at_immigrate;
279 /* 283 /*
284 * Should kernel memory limits be stabilished independently
285 * from user memory ?
286 */
287 int kmem_independent_accounting;
288 /*
280 * percpu counter. 289 * percpu counter.
281 */ 290 */
282 struct mem_cgroup_stat_cpu *stat; 291 struct mem_cgroup_stat_cpu *stat;
@@ -344,9 +353,14 @@ enum charge_type {
344}; 353};
345 354
346/* for encoding cft->private value on file */ 355/* for encoding cft->private value on file */
347#define _MEM (0) 356
348#define _MEMSWAP (1) 357enum mem_type {
349#define _OOM_TYPE (2) 358 _MEM = 0,
359 _MEMSWAP,
360 _OOM_TYPE,
361 _KMEM,
362};
363
350#define MEMFILE_PRIVATE(x, val) (((x) << 16) | (val)) 364#define MEMFILE_PRIVATE(x, val) (((x) << 16) | (val))
351#define MEMFILE_TYPE(val) (((val) >> 16) & 0xffff) 365#define MEMFILE_TYPE(val) (((val) >> 16) & 0xffff)
352#define MEMFILE_ATTR(val) ((val) & 0xffff) 366#define MEMFILE_ATTR(val) ((val) & 0xffff)
@@ -3848,10 +3862,17 @@ static inline u64 mem_cgroup_usage(struct mem_cgroup *memcg, bool swap)
3848 u64 val; 3862 u64 val;
3849 3863
3850 if (!mem_cgroup_is_root(memcg)) { 3864 if (!mem_cgroup_is_root(memcg)) {
3865 val = 0;
3866#ifdef CONFIG_CGROUP_MEM_RES_CTLR_KMEM
3867 if (!memcg->kmem_independent_accounting)
3868 val = res_counter_read_u64(&memcg->kmem, RES_USAGE);
3869#endif
3851 if (!swap) 3870 if (!swap)
3852 return res_counter_read_u64(&memcg->res, RES_USAGE); 3871 val += res_counter_read_u64(&memcg->res, RES_USAGE);
3853 else 3872 else
3854 return res_counter_read_u64(&memcg->memsw, RES_USAGE); 3873 val += res_counter_read_u64(&memcg->memsw, RES_USAGE);
3874
3875 return val;
3855 } 3876 }
3856 3877
3857 val = mem_cgroup_recursive_stat(memcg, MEM_CGROUP_STAT_CACHE); 3878 val = mem_cgroup_recursive_stat(memcg, MEM_CGROUP_STAT_CACHE);
@@ -3884,6 +3905,11 @@ static u64 mem_cgroup_read(struct cgroup *cont, struct cftype *cft)
3884 else 3905 else
3885 val = res_counter_read_u64(&memcg->memsw, name); 3906 val = res_counter_read_u64(&memcg->memsw, name);
3886 break; 3907 break;
3908#ifdef CONFIG_CGROUP_MEM_RES_CTLR_KMEM
3909 case _KMEM:
3910 val = res_counter_read_u64(&memcg->kmem, name);
3911 break;
3912#endif
3887 default: 3913 default:
3888 BUG(); 3914 BUG();
3889 break; 3915 break;
@@ -4612,6 +4638,69 @@ static int mem_control_numa_stat_open(struct inode *unused, struct file *file)
4612} 4638}
4613#endif /* CONFIG_NUMA */ 4639#endif /* CONFIG_NUMA */
4614 4640
4641#ifdef CONFIG_CGROUP_MEM_RES_CTLR_KMEM
4642static u64 kmem_limit_independent_read(struct cgroup *cgroup, struct cftype *cft)
4643{
4644 return mem_cgroup_from_cont(cgroup)->kmem_independent_accounting;
4645}
4646
4647static int kmem_limit_independent_write(struct cgroup *cgroup, struct cftype *cft,
4648 u64 val)
4649{
4650 struct mem_cgroup *memcg = mem_cgroup_from_cont(cgroup);
4651 struct mem_cgroup *parent = parent_mem_cgroup(memcg);
4652
4653 val = !!val;
4654
4655 /*
4656 * This follows the same hierarchy restrictions than
4657 * mem_cgroup_hierarchy_write()
4658 */
4659 if (!parent || !parent->use_hierarchy) {
4660 if (list_empty(&cgroup->children))
4661 memcg->kmem_independent_accounting = val;
4662 else
4663 return -EBUSY;
4664 }
4665 else
4666 return -EINVAL;
4667
4668 return 0;
4669}
4670static struct cftype kmem_cgroup_files[] = {
4671 {
4672 .name = "independent_kmem_limit",
4673 .read_u64 = kmem_limit_independent_read,
4674 .write_u64 = kmem_limit_independent_write,
4675 },
4676 {
4677 .name = "kmem.usage_in_bytes",
4678 .private = MEMFILE_PRIVATE(_KMEM, RES_USAGE),
4679 .read_u64 = mem_cgroup_read,
4680 },
4681 {
4682 .name = "kmem.limit_in_bytes",
4683 .private = MEMFILE_PRIVATE(_KMEM, RES_LIMIT),
4684 .read_u64 = mem_cgroup_read,
4685 },
4686};
4687
4688static int register_kmem_files(struct cgroup *cont, struct cgroup_subsys *ss)
4689{
4690 int ret = 0;
4691
4692 ret = cgroup_add_files(cont, ss, kmem_cgroup_files,
4693 ARRAY_SIZE(kmem_cgroup_files));
4694 return ret;
4695};
4696
4697#else
4698static int register_kmem_files(struct cgroup *cont, struct cgroup_subsys *ss)
4699{
4700 return 0;
4701}
4702#endif
4703
4615static struct cftype mem_cgroup_files[] = { 4704static struct cftype mem_cgroup_files[] = {
4616 { 4705 {
4617 .name = "usage_in_bytes", 4706 .name = "usage_in_bytes",
@@ -4925,6 +5014,7 @@ mem_cgroup_create(struct cgroup_subsys *ss, struct cgroup *cont)
4925 if (parent && parent->use_hierarchy) { 5014 if (parent && parent->use_hierarchy) {
4926 res_counter_init(&memcg->res, &parent->res); 5015 res_counter_init(&memcg->res, &parent->res);
4927 res_counter_init(&memcg->memsw, &parent->memsw); 5016 res_counter_init(&memcg->memsw, &parent->memsw);
5017 res_counter_init(&memcg->kmem, &parent->kmem);
4928 /* 5018 /*
4929 * We increment refcnt of the parent to ensure that we can 5019 * We increment refcnt of the parent to ensure that we can
4930 * safely access it on res_counter_charge/uncharge. 5020 * safely access it on res_counter_charge/uncharge.
@@ -4935,6 +5025,7 @@ mem_cgroup_create(struct cgroup_subsys *ss, struct cgroup *cont)
4935 } else { 5025 } else {
4936 res_counter_init(&memcg->res, NULL); 5026 res_counter_init(&memcg->res, NULL);
4937 res_counter_init(&memcg->memsw, NULL); 5027 res_counter_init(&memcg->memsw, NULL);
5028 res_counter_init(&memcg->kmem, NULL);
4938 } 5029 }
4939 memcg->last_scanned_child = 0; 5030 memcg->last_scanned_child = 0;
4940 memcg->last_scanned_node = MAX_NUMNODES; 5031 memcg->last_scanned_node = MAX_NUMNODES;
@@ -4978,6 +5069,10 @@ static int mem_cgroup_populate(struct cgroup_subsys *ss,
4978 5069
4979 if (!ret) 5070 if (!ret)
4980 ret = register_memsw_files(cont, ss); 5071 ret = register_memsw_files(cont, ss);
5072
5073 if (!ret)
5074 ret = register_kmem_files(cont, ss);
5075
4981 return ret; 5076 return ret;
4982} 5077}
4983 5078