aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMel Gorman <mgorman@suse.de>2013-07-03 18:01:44 -0400
committerLinus Torvalds <torvalds@linux-foundation.org>2013-07-03 19:07:28 -0400
commite82e0561dae9f3ae5a21fc2d3d3ccbe69d90be46 (patch)
tree68b346634ac7dfcb99708862498f2e9cebd3cc75
parent75485363ce8552698bfb9970d901f755d5713cca (diff)
mm: vmscan: obey proportional scanning requirements for kswapd
Simplistically, the anon and file LRU lists are scanned proportionally depending on the value of vm.swappiness although there are other factors taken into account by get_scan_count(). The patch "mm: vmscan: Limit the number of pages kswapd reclaims" limits the number of pages kswapd reclaims but it breaks this proportional scanning and may evenly shrink anon/file LRUs regardless of vm.swappiness. This patch preserves the proportional scanning and reclaim. It does mean that kswapd will reclaim more than requested but the number of pages will be related to the high watermark. [mhocko@suse.cz: Correct proportional reclaim for memcg and simplify] [kamezawa.hiroyu@jp.fujitsu.com: Recalculate scan based on target] [hannes@cmpxchg.org: Account for already scanned pages properly] Signed-off-by: Mel Gorman <mgorman@suse.de> Acked-by: Rik van Riel <riel@redhat.com> Reviewed-by: Michal Hocko <mhocko@suse.cz> Cc: Johannes Weiner <hannes@cmpxchg.org> Acked-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com> Cc: Jiri Slaby <jslaby@suse.cz> Cc: Valdis Kletnieks <Valdis.Kletnieks@vt.edu> Tested-by: Zlatko Calusic <zcalusic@bitsync.net> Cc: dormando <dormando@rydia.net> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r--mm/vmscan.c67
1 files changed, 59 insertions, 8 deletions
diff --git a/mm/vmscan.c b/mm/vmscan.c
index cdbc0699ea21..26ad67f1962c 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -1822,17 +1822,25 @@ out:
1822static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc) 1822static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc)
1823{ 1823{
1824 unsigned long nr[NR_LRU_LISTS]; 1824 unsigned long nr[NR_LRU_LISTS];
1825 unsigned long targets[NR_LRU_LISTS];
1825 unsigned long nr_to_scan; 1826 unsigned long nr_to_scan;
1826 enum lru_list lru; 1827 enum lru_list lru;
1827 unsigned long nr_reclaimed = 0; 1828 unsigned long nr_reclaimed = 0;
1828 unsigned long nr_to_reclaim = sc->nr_to_reclaim; 1829 unsigned long nr_to_reclaim = sc->nr_to_reclaim;
1829 struct blk_plug plug; 1830 struct blk_plug plug;
1831 bool scan_adjusted = false;
1830 1832
1831 get_scan_count(lruvec, sc, nr); 1833 get_scan_count(lruvec, sc, nr);
1832 1834
1835 /* Record the original scan target for proportional adjustments later */
1836 memcpy(targets, nr, sizeof(nr));
1837
1833 blk_start_plug(&plug); 1838 blk_start_plug(&plug);
1834 while (nr[LRU_INACTIVE_ANON] || nr[LRU_ACTIVE_FILE] || 1839 while (nr[LRU_INACTIVE_ANON] || nr[LRU_ACTIVE_FILE] ||
1835 nr[LRU_INACTIVE_FILE]) { 1840 nr[LRU_INACTIVE_FILE]) {
1841 unsigned long nr_anon, nr_file, percentage;
1842 unsigned long nr_scanned;
1843
1836 for_each_evictable_lru(lru) { 1844 for_each_evictable_lru(lru) {
1837 if (nr[lru]) { 1845 if (nr[lru]) {
1838 nr_to_scan = min(nr[lru], SWAP_CLUSTER_MAX); 1846 nr_to_scan = min(nr[lru], SWAP_CLUSTER_MAX);
@@ -1842,17 +1850,60 @@ static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc)
1842 lruvec, sc); 1850 lruvec, sc);
1843 } 1851 }
1844 } 1852 }
1853
1854 if (nr_reclaimed < nr_to_reclaim || scan_adjusted)
1855 continue;
1856
1845 /* 1857 /*
1846 * On large memory systems, scan >> priority can become 1858 * For global direct reclaim, reclaim only the number of pages
1847 * really large. This is fine for the starting priority; 1859 * requested. Less care is taken to scan proportionally as it
1848 * we want to put equal scanning pressure on each zone. 1860 * is more important to minimise direct reclaim stall latency
1849 * However, if the VM has a harder time of freeing pages, 1861 * than it is to properly age the LRU lists.
1850 * with multiple processes reclaiming pages, the total
1851 * freeing target can get unreasonably large.
1852 */ 1862 */
1853 if (nr_reclaimed >= nr_to_reclaim && 1863 if (global_reclaim(sc) && !current_is_kswapd())
1854 sc->priority < DEF_PRIORITY)
1855 break; 1864 break;
1865
1866 /*
1867 * For kswapd and memcg, reclaim at least the number of pages
1868 * requested. Ensure that the anon and file LRUs shrink
1869 * proportionally what was requested by get_scan_count(). We
1870 * stop reclaiming one LRU and reduce the amount scanning
1871 * proportional to the original scan target.
1872 */
1873 nr_file = nr[LRU_INACTIVE_FILE] + nr[LRU_ACTIVE_FILE];
1874 nr_anon = nr[LRU_INACTIVE_ANON] + nr[LRU_ACTIVE_ANON];
1875
1876 if (nr_file > nr_anon) {
1877 unsigned long scan_target = targets[LRU_INACTIVE_ANON] +
1878 targets[LRU_ACTIVE_ANON] + 1;
1879 lru = LRU_BASE;
1880 percentage = nr_anon * 100 / scan_target;
1881 } else {
1882 unsigned long scan_target = targets[LRU_INACTIVE_FILE] +
1883 targets[LRU_ACTIVE_FILE] + 1;
1884 lru = LRU_FILE;
1885 percentage = nr_file * 100 / scan_target;
1886 }
1887
1888 /* Stop scanning the smaller of the LRU */
1889 nr[lru] = 0;
1890 nr[lru + LRU_ACTIVE] = 0;
1891
1892 /*
1893 * Recalculate the other LRU scan count based on its original
1894 * scan target and the percentage scanning already complete
1895 */
1896 lru = (lru == LRU_FILE) ? LRU_BASE : LRU_FILE;
1897 nr_scanned = targets[lru] - nr[lru];
1898 nr[lru] = targets[lru] * (100 - percentage) / 100;
1899 nr[lru] -= min(nr[lru], nr_scanned);
1900
1901 lru += LRU_ACTIVE;
1902 nr_scanned = targets[lru] - nr[lru];
1903 nr[lru] = targets[lru] * (100 - percentage) / 100;
1904 nr[lru] -= min(nr[lru], nr_scanned);
1905
1906 scan_adjusted = true;
1856 } 1907 }
1857 blk_finish_plug(&plug); 1908 blk_finish_plug(&plug);
1858 sc->nr_reclaimed += nr_reclaimed; 1909 sc->nr_reclaimed += nr_reclaimed;