aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--include/linux/swap.h2
-rw-r--r--mm/swap.c24
-rw-r--r--mm/truncate.c2
3 files changed, 14 insertions, 14 deletions
diff --git a/include/linux/swap.h b/include/linux/swap.h
index 7067eca501e2..cee108cbe2d5 100644
--- a/include/linux/swap.h
+++ b/include/linux/swap.h
@@ -307,7 +307,7 @@ extern void lru_add_drain(void);
307extern void lru_add_drain_cpu(int cpu); 307extern void lru_add_drain_cpu(int cpu);
308extern void lru_add_drain_all(void); 308extern void lru_add_drain_all(void);
309extern void rotate_reclaimable_page(struct page *page); 309extern void rotate_reclaimable_page(struct page *page);
310extern void deactivate_page(struct page *page); 310extern void deactivate_file_page(struct page *page);
311extern void swap_setup(void); 311extern void swap_setup(void);
312 312
313extern void add_page_to_unevictable_list(struct page *page); 313extern void add_page_to_unevictable_list(struct page *page);
diff --git a/mm/swap.c b/mm/swap.c
index cd3a5e64cea9..e3a4feac9b0e 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -42,7 +42,7 @@ int page_cluster;
42 42
43static DEFINE_PER_CPU(struct pagevec, lru_add_pvec); 43static DEFINE_PER_CPU(struct pagevec, lru_add_pvec);
44static DEFINE_PER_CPU(struct pagevec, lru_rotate_pvecs); 44static DEFINE_PER_CPU(struct pagevec, lru_rotate_pvecs);
45static DEFINE_PER_CPU(struct pagevec, lru_deactivate_pvecs); 45static DEFINE_PER_CPU(struct pagevec, lru_deactivate_file_pvecs);
46 46
47/* 47/*
48 * This path almost never happens for VM activity - pages are normally 48 * This path almost never happens for VM activity - pages are normally
@@ -743,7 +743,7 @@ void lru_cache_add_active_or_unevictable(struct page *page,
743 * be write it out by flusher threads as this is much more effective 743 * be write it out by flusher threads as this is much more effective
744 * than the single-page writeout from reclaim. 744 * than the single-page writeout from reclaim.
745 */ 745 */
746static void lru_deactivate_fn(struct page *page, struct lruvec *lruvec, 746static void lru_deactivate_file_fn(struct page *page, struct lruvec *lruvec,
747 void *arg) 747 void *arg)
748{ 748{
749 int lru, file; 749 int lru, file;
@@ -811,36 +811,36 @@ void lru_add_drain_cpu(int cpu)
811 local_irq_restore(flags); 811 local_irq_restore(flags);
812 } 812 }
813 813
814 pvec = &per_cpu(lru_deactivate_pvecs, cpu); 814 pvec = &per_cpu(lru_deactivate_file_pvecs, cpu);
815 if (pagevec_count(pvec)) 815 if (pagevec_count(pvec))
816 pagevec_lru_move_fn(pvec, lru_deactivate_fn, NULL); 816 pagevec_lru_move_fn(pvec, lru_deactivate_file_fn, NULL);
817 817
818 activate_page_drain(cpu); 818 activate_page_drain(cpu);
819} 819}
820 820
821/** 821/**
822 * deactivate_page - forcefully deactivate a page 822 * deactivate_file_page - forcefully deactivate a file page
823 * @page: page to deactivate 823 * @page: page to deactivate
824 * 824 *
825 * This function hints the VM that @page is a good reclaim candidate, 825 * This function hints the VM that @page is a good reclaim candidate,
826 * for example if its invalidation fails due to the page being dirty 826 * for example if its invalidation fails due to the page being dirty
827 * or under writeback. 827 * or under writeback.
828 */ 828 */
829void deactivate_page(struct page *page) 829void deactivate_file_page(struct page *page)
830{ 830{
831 /* 831 /*
832 * In a workload with many unevictable page such as mprotect, unevictable 832 * In a workload with many unevictable page such as mprotect,
833 * page deactivation for accelerating reclaim is pointless. 833 * unevictable page deactivation for accelerating reclaim is pointless.
834 */ 834 */
835 if (PageUnevictable(page)) 835 if (PageUnevictable(page))
836 return; 836 return;
837 837
838 if (likely(get_page_unless_zero(page))) { 838 if (likely(get_page_unless_zero(page))) {
839 struct pagevec *pvec = &get_cpu_var(lru_deactivate_pvecs); 839 struct pagevec *pvec = &get_cpu_var(lru_deactivate_file_pvecs);
840 840
841 if (!pagevec_add(pvec, page)) 841 if (!pagevec_add(pvec, page))
842 pagevec_lru_move_fn(pvec, lru_deactivate_fn, NULL); 842 pagevec_lru_move_fn(pvec, lru_deactivate_file_fn, NULL);
843 put_cpu_var(lru_deactivate_pvecs); 843 put_cpu_var(lru_deactivate_file_pvecs);
844 } 844 }
845} 845}
846 846
@@ -872,7 +872,7 @@ void lru_add_drain_all(void)
872 872
873 if (pagevec_count(&per_cpu(lru_add_pvec, cpu)) || 873 if (pagevec_count(&per_cpu(lru_add_pvec, cpu)) ||
874 pagevec_count(&per_cpu(lru_rotate_pvecs, cpu)) || 874 pagevec_count(&per_cpu(lru_rotate_pvecs, cpu)) ||
875 pagevec_count(&per_cpu(lru_deactivate_pvecs, cpu)) || 875 pagevec_count(&per_cpu(lru_deactivate_file_pvecs, cpu)) ||
876 need_activate_page_drain(cpu)) { 876 need_activate_page_drain(cpu)) {
877 INIT_WORK(work, lru_add_drain_per_cpu); 877 INIT_WORK(work, lru_add_drain_per_cpu);
878 schedule_work_on(cpu, work); 878 schedule_work_on(cpu, work);
diff --git a/mm/truncate.c b/mm/truncate.c
index 7a9d8a3cb143..66af9031fae8 100644
--- a/mm/truncate.c
+++ b/mm/truncate.c
@@ -490,7 +490,7 @@ unsigned long invalidate_mapping_pages(struct address_space *mapping,
490 * of interest and try to speed up its reclaim. 490 * of interest and try to speed up its reclaim.
491 */ 491 */
492 if (!ret) 492 if (!ret)
493 deactivate_page(page); 493 deactivate_file_page(page);
494 count += ret; 494 count += ret;
495 } 495 }
496 pagevec_remove_exceptionals(&pvec); 496 pagevec_remove_exceptionals(&pvec);