aboutsummaryrefslogtreecommitdiffstats
path: root/mm/vmscan.c
diff options
context:
space:
mode:
Diffstat (limited to 'mm/vmscan.c')
-rw-r--r--mm/vmscan.c45
1 files changed, 13 insertions, 32 deletions
diff --git a/mm/vmscan.c b/mm/vmscan.c
index c82ee9a33cfc..9588973849d0 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -1064,7 +1064,6 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
1064 int pgdeactivate = 0; 1064 int pgdeactivate = 0;
1065 unsigned long pgscanned; 1065 unsigned long pgscanned;
1066 LIST_HEAD(l_hold); /* The pages which were snipped off */ 1066 LIST_HEAD(l_hold); /* The pages which were snipped off */
1067 LIST_HEAD(l_active);
1068 LIST_HEAD(l_inactive); 1067 LIST_HEAD(l_inactive);
1069 struct page *page; 1068 struct page *page;
1070 struct pagevec pvec; 1069 struct pagevec pvec;
@@ -1095,21 +1094,28 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
1095 cond_resched(); 1094 cond_resched();
1096 page = lru_to_page(&l_hold); 1095 page = lru_to_page(&l_hold);
1097 list_del(&page->lru); 1096 list_del(&page->lru);
1097
1098 /* page_referenced clears PageReferenced */
1099 if (page_mapping_inuse(page) &&
1100 page_referenced(page, 0, sc->mem_cgroup))
1101 pgmoved++;
1102
1098 list_add(&page->lru, &l_inactive); 1103 list_add(&page->lru, &l_inactive);
1099 } 1104 }
1100 1105
1101 /* 1106 /*
1102 * Count the referenced pages as rotated, even when they are moved 1107 * Count referenced pages from currently used mappings as
1103 * to the inactive list. This helps balance scan pressure between 1108 * rotated, even though they are moved to the inactive list.
1104 * file and anonymous pages in get_scan_ratio. 1109 * This helps balance scan pressure between file and anonymous
1105 */ 1110 * pages in get_scan_ratio.
1111 */
1106 zone->recent_rotated[!!file] += pgmoved; 1112 zone->recent_rotated[!!file] += pgmoved;
1107 1113
1108 /* 1114 /*
1109 * Now put the pages back on the appropriate [file or anon] inactive 1115 * Move the pages to the [file or anon] inactive list.
1110 * and active lists.
1111 */ 1116 */
1112 pagevec_init(&pvec, 1); 1117 pagevec_init(&pvec, 1);
1118
1113 pgmoved = 0; 1119 pgmoved = 0;
1114 lru = LRU_BASE + file * LRU_FILE; 1120 lru = LRU_BASE + file * LRU_FILE;
1115 spin_lock_irq(&zone->lru_lock); 1121 spin_lock_irq(&zone->lru_lock);
@@ -1142,31 +1148,6 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
1142 pagevec_strip(&pvec); 1148 pagevec_strip(&pvec);
1143 spin_lock_irq(&zone->lru_lock); 1149 spin_lock_irq(&zone->lru_lock);
1144 } 1150 }
1145
1146 pgmoved = 0;
1147 lru = LRU_ACTIVE + file * LRU_FILE;
1148 while (!list_empty(&l_active)) {
1149 page = lru_to_page(&l_active);
1150 prefetchw_prev_lru_page(page, &l_active, flags);
1151 VM_BUG_ON(PageLRU(page));
1152 SetPageLRU(page);
1153 VM_BUG_ON(!PageActive(page));
1154
1155 list_move(&page->lru, &zone->lru[lru].list);
1156 mem_cgroup_move_lists(page, true);
1157 pgmoved++;
1158 if (!pagevec_add(&pvec, page)) {
1159 __mod_zone_page_state(zone, NR_LRU_BASE + lru, pgmoved);
1160 pgmoved = 0;
1161 spin_unlock_irq(&zone->lru_lock);
1162 if (vm_swap_full())
1163 pagevec_swap_free(&pvec);
1164 __pagevec_release(&pvec);
1165 spin_lock_irq(&zone->lru_lock);
1166 }
1167 }
1168 __mod_zone_page_state(zone, NR_LRU_BASE + lru, pgmoved);
1169
1170 __count_zone_vm_events(PGREFILL, zone, pgscanned); 1151 __count_zone_vm_events(PGREFILL, zone, pgscanned);
1171 __count_vm_events(PGDEACTIVATE, pgdeactivate); 1152 __count_vm_events(PGDEACTIVATE, pgdeactivate);
1172 spin_unlock_irq(&zone->lru_lock); 1153 spin_unlock_irq(&zone->lru_lock);