diff options
Diffstat (limited to 'include/linux')
-rw-r--r-- | include/linux/memcontrol.h | 5 | ||||
-rw-r--r-- | include/linux/mmzone.h | 15 |
2 files changed, 0 insertions, 20 deletions
diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index 9411d32840b0..9f1afd361583 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h | |||
@@ -98,11 +98,6 @@ extern void mem_cgroup_end_migration(struct mem_cgroup *mem, | |||
98 | /* | 98 | /* |
99 | * For memory reclaim. | 99 | * For memory reclaim. |
100 | */ | 100 | */ |
101 | extern int mem_cgroup_get_reclaim_priority(struct mem_cgroup *mem); | ||
102 | extern void mem_cgroup_note_reclaim_priority(struct mem_cgroup *mem, | ||
103 | int priority); | ||
104 | extern void mem_cgroup_record_reclaim_priority(struct mem_cgroup *mem, | ||
105 | int priority); | ||
106 | int mem_cgroup_inactive_anon_is_low(struct mem_cgroup *memcg); | 101 | int mem_cgroup_inactive_anon_is_low(struct mem_cgroup *memcg); |
107 | int mem_cgroup_inactive_file_is_low(struct mem_cgroup *memcg); | 102 | int mem_cgroup_inactive_file_is_low(struct mem_cgroup *memcg); |
108 | unsigned long mem_cgroup_zone_nr_pages(struct mem_cgroup *memcg, | 103 | unsigned long mem_cgroup_zone_nr_pages(struct mem_cgroup *memcg, |
diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h index 9ed9c459b14c..6e6e62648a4d 100644 --- a/include/linux/mmzone.h +++ b/include/linux/mmzone.h | |||
@@ -348,21 +348,6 @@ struct zone { | |||
348 | atomic_long_t vm_stat[NR_VM_ZONE_STAT_ITEMS]; | 348 | atomic_long_t vm_stat[NR_VM_ZONE_STAT_ITEMS]; |
349 | 349 | ||
350 | /* | 350 | /* |
351 | * prev_priority holds the scanning priority for this zone. It is | ||
352 | * defined as the scanning priority at which we achieved our reclaim | ||
353 | * target at the previous try_to_free_pages() or balance_pgdat() | ||
354 | * invocation. | ||
355 | * | ||
356 | * We use prev_priority as a measure of how much stress page reclaim is | ||
357 | * under - it drives the swappiness decision: whether to unmap mapped | ||
358 | * pages. | ||
359 | * | ||
360 | * Access to both this field is quite racy even on uniprocessor. But | ||
361 | * it is expected to average out OK. | ||
362 | */ | ||
363 | int prev_priority; | ||
364 | |||
365 | /* | ||
366 | * The target ratio of ACTIVE_ANON to INACTIVE_ANON pages on | 351 | * The target ratio of ACTIVE_ANON to INACTIVE_ANON pages on |
367 | * this zone's LRU. Maintained by the pageout code. | 352 | * this zone's LRU. Maintained by the pageout code. |
368 | */ | 353 | */ |