aboutsummaryrefslogtreecommitdiffstats
path: root/mm/swap.c
diff options
context:
space:
mode:
Diffstat (limited to 'mm/swap.c')
-rw-r--r--mm/swap.c14
1 files changed, 7 insertions, 7 deletions
diff --git a/mm/swap.c b/mm/swap.c
index 09fe5e97714a..a0bc206b4ac6 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -114,7 +114,7 @@ void put_pages_list(struct list_head *pages)
114 114
115 victim = list_entry(pages->prev, struct page, lru); 115 victim = list_entry(pages->prev, struct page, lru);
116 list_del(&victim->lru); 116 list_del(&victim->lru);
117 page_cache_release(victim); 117 put_page(victim);
118 } 118 }
119} 119}
120EXPORT_SYMBOL(put_pages_list); 120EXPORT_SYMBOL(put_pages_list);
@@ -142,7 +142,7 @@ int get_kernel_pages(const struct kvec *kiov, int nr_segs, int write,
142 return seg; 142 return seg;
143 143
144 pages[seg] = kmap_to_page(kiov[seg].iov_base); 144 pages[seg] = kmap_to_page(kiov[seg].iov_base);
145 page_cache_get(pages[seg]); 145 get_page(pages[seg]);
146 } 146 }
147 147
148 return seg; 148 return seg;
@@ -236,7 +236,7 @@ void rotate_reclaimable_page(struct page *page)
236 struct pagevec *pvec; 236 struct pagevec *pvec;
237 unsigned long flags; 237 unsigned long flags;
238 238
239 page_cache_get(page); 239 get_page(page);
240 local_irq_save(flags); 240 local_irq_save(flags);
241 pvec = this_cpu_ptr(&lru_rotate_pvecs); 241 pvec = this_cpu_ptr(&lru_rotate_pvecs);
242 if (!pagevec_add(pvec, page)) 242 if (!pagevec_add(pvec, page))
@@ -294,7 +294,7 @@ void activate_page(struct page *page)
294 if (PageLRU(page) && !PageActive(page) && !PageUnevictable(page)) { 294 if (PageLRU(page) && !PageActive(page) && !PageUnevictable(page)) {
295 struct pagevec *pvec = &get_cpu_var(activate_page_pvecs); 295 struct pagevec *pvec = &get_cpu_var(activate_page_pvecs);
296 296
297 page_cache_get(page); 297 get_page(page);
298 if (!pagevec_add(pvec, page)) 298 if (!pagevec_add(pvec, page))
299 pagevec_lru_move_fn(pvec, __activate_page, NULL); 299 pagevec_lru_move_fn(pvec, __activate_page, NULL);
300 put_cpu_var(activate_page_pvecs); 300 put_cpu_var(activate_page_pvecs);
@@ -389,7 +389,7 @@ static void __lru_cache_add(struct page *page)
389{ 389{
390 struct pagevec *pvec = &get_cpu_var(lru_add_pvec); 390 struct pagevec *pvec = &get_cpu_var(lru_add_pvec);
391 391
392 page_cache_get(page); 392 get_page(page);
393 if (!pagevec_space(pvec)) 393 if (!pagevec_space(pvec))
394 __pagevec_lru_add(pvec); 394 __pagevec_lru_add(pvec);
395 pagevec_add(pvec, page); 395 pagevec_add(pvec, page);
@@ -646,7 +646,7 @@ void deactivate_page(struct page *page)
646 if (PageLRU(page) && PageActive(page) && !PageUnevictable(page)) { 646 if (PageLRU(page) && PageActive(page) && !PageUnevictable(page)) {
647 struct pagevec *pvec = &get_cpu_var(lru_deactivate_pvecs); 647 struct pagevec *pvec = &get_cpu_var(lru_deactivate_pvecs);
648 648
649 page_cache_get(page); 649 get_page(page);
650 if (!pagevec_add(pvec, page)) 650 if (!pagevec_add(pvec, page))
651 pagevec_lru_move_fn(pvec, lru_deactivate_fn, NULL); 651 pagevec_lru_move_fn(pvec, lru_deactivate_fn, NULL);
652 put_cpu_var(lru_deactivate_pvecs); 652 put_cpu_var(lru_deactivate_pvecs);
@@ -698,7 +698,7 @@ void lru_add_drain_all(void)
698} 698}
699 699
700/** 700/**
701 * release_pages - batched page_cache_release() 701 * release_pages - batched put_page()
702 * @pages: array of pages to release 702 * @pages: array of pages to release
703 * @nr: number of pages 703 * @nr: number of pages
704 * @cold: whether the pages are cache cold 704 * @cold: whether the pages are cache cold