aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
Diffstat (limited to 'mm')
-rw-r--r--mm/filemap.c10
-rw-r--r--mm/highmem.c4
-rw-r--r--mm/internal.h2
-rw-r--r--mm/memory.c3
-rw-r--r--mm/page-writeback.c2
-rw-r--r--mm/page_alloc.c16
-rw-r--r--mm/swap.c10
7 files changed, 24 insertions, 23 deletions
diff --git a/mm/filemap.c b/mm/filemap.c
index 96920f840562..81fb9bff0d4f 100644
--- a/mm/filemap.c
+++ b/mm/filemap.c
@@ -527,7 +527,7 @@ static inline void wake_up_page(struct page *page, int bit)
527 __wake_up_bit(page_waitqueue(page), &page->flags, bit); 527 __wake_up_bit(page_waitqueue(page), &page->flags, bit);
528} 528}
529 529
530void fastcall wait_on_page_bit(struct page *page, int bit_nr) 530void wait_on_page_bit(struct page *page, int bit_nr)
531{ 531{
532 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); 532 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr);
533 533
@@ -551,7 +551,7 @@ EXPORT_SYMBOL(wait_on_page_bit);
551 * the clear_bit and the read of the waitqueue (to avoid SMP races with a 551 * the clear_bit and the read of the waitqueue (to avoid SMP races with a
552 * parallel wait_on_page_locked()). 552 * parallel wait_on_page_locked()).
553 */ 553 */
554void fastcall unlock_page(struct page *page) 554void unlock_page(struct page *page)
555{ 555{
556 smp_mb__before_clear_bit(); 556 smp_mb__before_clear_bit();
557 if (!TestClearPageLocked(page)) 557 if (!TestClearPageLocked(page))
@@ -585,7 +585,7 @@ EXPORT_SYMBOL(end_page_writeback);
585 * chances are that on the second loop, the block layer's plug list is empty, 585 * chances are that on the second loop, the block layer's plug list is empty,
586 * so sync_page() will then return in state TASK_UNINTERRUPTIBLE. 586 * so sync_page() will then return in state TASK_UNINTERRUPTIBLE.
587 */ 587 */
588void fastcall __lock_page(struct page *page) 588void __lock_page(struct page *page)
589{ 589{
590 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); 590 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked);
591 591
@@ -606,7 +606,7 @@ int fastcall __lock_page_killable(struct page *page)
606 * Variant of lock_page that does not require the caller to hold a reference 606 * Variant of lock_page that does not require the caller to hold a reference
607 * on the page's mapping. 607 * on the page's mapping.
608 */ 608 */
609void fastcall __lock_page_nosync(struct page *page) 609void __lock_page_nosync(struct page *page)
610{ 610{
611 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); 611 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked);
612 __wait_on_bit_lock(page_waitqueue(page), &wait, __sleep_on_page_lock, 612 __wait_on_bit_lock(page_waitqueue(page), &wait, __sleep_on_page_lock,
@@ -1276,7 +1276,7 @@ asmlinkage ssize_t sys_readahead(int fd, loff_t offset, size_t count)
1276 * This adds the requested page to the page cache if it isn't already there, 1276 * This adds the requested page to the page cache if it isn't already there,
1277 * and schedules an I/O to read in its contents from disk. 1277 * and schedules an I/O to read in its contents from disk.
1278 */ 1278 */
1279static int fastcall page_cache_read(struct file * file, pgoff_t offset) 1279static int page_cache_read(struct file *file, pgoff_t offset)
1280{ 1280{
1281 struct address_space *mapping = file->f_mapping; 1281 struct address_space *mapping = file->f_mapping;
1282 struct page *page; 1282 struct page *page;
diff --git a/mm/highmem.c b/mm/highmem.c
index 7a967bc35152..35d47733cde4 100644
--- a/mm/highmem.c
+++ b/mm/highmem.c
@@ -163,7 +163,7 @@ start:
163 return vaddr; 163 return vaddr;
164} 164}
165 165
166void fastcall *kmap_high(struct page *page) 166void *kmap_high(struct page *page)
167{ 167{
168 unsigned long vaddr; 168 unsigned long vaddr;
169 169
@@ -185,7 +185,7 @@ void fastcall *kmap_high(struct page *page)
185 185
186EXPORT_SYMBOL(kmap_high); 186EXPORT_SYMBOL(kmap_high);
187 187
188void fastcall kunmap_high(struct page *page) 188void kunmap_high(struct page *page)
189{ 189{
190 unsigned long vaddr; 190 unsigned long vaddr;
191 unsigned long nr; 191 unsigned long nr;
diff --git a/mm/internal.h b/mm/internal.h
index 953f941ea867..1e34d2462a48 100644
--- a/mm/internal.h
+++ b/mm/internal.h
@@ -34,7 +34,7 @@ static inline void __put_page(struct page *page)
34 atomic_dec(&page->_count); 34 atomic_dec(&page->_count);
35} 35}
36 36
37extern void fastcall __init __free_pages_bootmem(struct page *page, 37extern void __init __free_pages_bootmem(struct page *page,
38 unsigned int order); 38 unsigned int order);
39 39
40/* 40/*
diff --git a/mm/memory.c b/mm/memory.c
index 1c81fc2174cd..6a9c048f6012 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -1109,7 +1109,8 @@ int get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
1109} 1109}
1110EXPORT_SYMBOL(get_user_pages); 1110EXPORT_SYMBOL(get_user_pages);
1111 1111
1112pte_t * fastcall get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) 1112pte_t *get_locked_pte(struct mm_struct *mm, unsigned long addr,
1113 spinlock_t **ptl)
1113{ 1114{
1114 pgd_t * pgd = pgd_offset(mm, addr); 1115 pgd_t * pgd = pgd_offset(mm, addr);
1115 pud_t * pud = pud_alloc(mm, pgd, addr); 1116 pud_t * pud = pud_alloc(mm, pgd, addr);
diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index c689b60af000..a4ca162666c5 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -1073,7 +1073,7 @@ static int __set_page_dirty(struct page *page)
1073 return 0; 1073 return 0;
1074} 1074}
1075 1075
1076int fastcall set_page_dirty(struct page *page) 1076int set_page_dirty(struct page *page)
1077{ 1077{
1078 int ret = __set_page_dirty(page); 1078 int ret = __set_page_dirty(page);
1079 if (ret) 1079 if (ret)
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 55fe57cd99a1..d73c133fdbe1 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -537,7 +537,7 @@ static void __free_pages_ok(struct page *page, unsigned int order)
537/* 537/*
538 * permit the bootmem allocator to evade page validation on high-order frees 538 * permit the bootmem allocator to evade page validation on high-order frees
539 */ 539 */
540void fastcall __init __free_pages_bootmem(struct page *page, unsigned int order) 540void __init __free_pages_bootmem(struct page *page, unsigned int order)
541{ 541{
542 if (order == 0) { 542 if (order == 0) {
543 __ClearPageReserved(page); 543 __ClearPageReserved(page);
@@ -974,7 +974,7 @@ void mark_free_pages(struct zone *zone)
974/* 974/*
975 * Free a 0-order page 975 * Free a 0-order page
976 */ 976 */
977static void fastcall free_hot_cold_page(struct page *page, int cold) 977static void free_hot_cold_page(struct page *page, int cold)
978{ 978{
979 struct zone *zone = page_zone(page); 979 struct zone *zone = page_zone(page);
980 struct per_cpu_pages *pcp; 980 struct per_cpu_pages *pcp;
@@ -1007,12 +1007,12 @@ static void fastcall free_hot_cold_page(struct page *page, int cold)
1007 put_cpu(); 1007 put_cpu();
1008} 1008}
1009 1009
1010void fastcall free_hot_page(struct page *page) 1010void free_hot_page(struct page *page)
1011{ 1011{
1012 free_hot_cold_page(page, 0); 1012 free_hot_cold_page(page, 0);
1013} 1013}
1014 1014
1015void fastcall free_cold_page(struct page *page) 1015void free_cold_page(struct page *page)
1016{ 1016{
1017 free_hot_cold_page(page, 1); 1017 free_hot_cold_page(page, 1);
1018} 1018}
@@ -1641,7 +1641,7 @@ EXPORT_SYMBOL(__alloc_pages);
1641/* 1641/*
1642 * Common helper functions. 1642 * Common helper functions.
1643 */ 1643 */
1644fastcall unsigned long __get_free_pages(gfp_t gfp_mask, unsigned int order) 1644unsigned long __get_free_pages(gfp_t gfp_mask, unsigned int order)
1645{ 1645{
1646 struct page * page; 1646 struct page * page;
1647 page = alloc_pages(gfp_mask, order); 1647 page = alloc_pages(gfp_mask, order);
@@ -1652,7 +1652,7 @@ fastcall unsigned long __get_free_pages(gfp_t gfp_mask, unsigned int order)
1652 1652
1653EXPORT_SYMBOL(__get_free_pages); 1653EXPORT_SYMBOL(__get_free_pages);
1654 1654
1655fastcall unsigned long get_zeroed_page(gfp_t gfp_mask) 1655unsigned long get_zeroed_page(gfp_t gfp_mask)
1656{ 1656{
1657 struct page * page; 1657 struct page * page;
1658 1658
@@ -1678,7 +1678,7 @@ void __pagevec_free(struct pagevec *pvec)
1678 free_hot_cold_page(pvec->pages[i], pvec->cold); 1678 free_hot_cold_page(pvec->pages[i], pvec->cold);
1679} 1679}
1680 1680
1681fastcall void __free_pages(struct page *page, unsigned int order) 1681void __free_pages(struct page *page, unsigned int order)
1682{ 1682{
1683 if (put_page_testzero(page)) { 1683 if (put_page_testzero(page)) {
1684 if (order == 0) 1684 if (order == 0)
@@ -1690,7 +1690,7 @@ fastcall void __free_pages(struct page *page, unsigned int order)
1690 1690
1691EXPORT_SYMBOL(__free_pages); 1691EXPORT_SYMBOL(__free_pages);
1692 1692
1693fastcall void free_pages(unsigned long addr, unsigned int order) 1693void free_pages(unsigned long addr, unsigned int order)
1694{ 1694{
1695 if (addr != 0) { 1695 if (addr != 0) {
1696 VM_BUG_ON(!virt_addr_valid((void *)addr)); 1696 VM_BUG_ON(!virt_addr_valid((void *)addr));
diff --git a/mm/swap.c b/mm/swap.c
index 9ac88323d237..57b7e25a939c 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -41,7 +41,7 @@ static DEFINE_PER_CPU(struct pagevec, lru_rotate_pvecs) = { 0, };
41 * This path almost never happens for VM activity - pages are normally 41 * This path almost never happens for VM activity - pages are normally
42 * freed via pagevecs. But it gets used by networking. 42 * freed via pagevecs. But it gets used by networking.
43 */ 43 */
44static void fastcall __page_cache_release(struct page *page) 44static void __page_cache_release(struct page *page)
45{ 45{
46 if (PageLRU(page)) { 46 if (PageLRU(page)) {
47 unsigned long flags; 47 unsigned long flags;
@@ -165,7 +165,7 @@ int rotate_reclaimable_page(struct page *page)
165/* 165/*
166 * FIXME: speed this up? 166 * FIXME: speed this up?
167 */ 167 */
168void fastcall activate_page(struct page *page) 168void activate_page(struct page *page)
169{ 169{
170 struct zone *zone = page_zone(page); 170 struct zone *zone = page_zone(page);
171 171
@@ -186,7 +186,7 @@ void fastcall activate_page(struct page *page)
186 * inactive,referenced -> active,unreferenced 186 * inactive,referenced -> active,unreferenced
187 * active,unreferenced -> active,referenced 187 * active,unreferenced -> active,referenced
188 */ 188 */
189void fastcall mark_page_accessed(struct page *page) 189void mark_page_accessed(struct page *page)
190{ 190{
191 if (!PageActive(page) && PageReferenced(page) && PageLRU(page)) { 191 if (!PageActive(page) && PageReferenced(page) && PageLRU(page)) {
192 activate_page(page); 192 activate_page(page);
@@ -202,7 +202,7 @@ EXPORT_SYMBOL(mark_page_accessed);
202 * lru_cache_add: add a page to the page lists 202 * lru_cache_add: add a page to the page lists
203 * @page: the page to add 203 * @page: the page to add
204 */ 204 */
205void fastcall lru_cache_add(struct page *page) 205void lru_cache_add(struct page *page)
206{ 206{
207 struct pagevec *pvec = &get_cpu_var(lru_add_pvecs); 207 struct pagevec *pvec = &get_cpu_var(lru_add_pvecs);
208 208
@@ -212,7 +212,7 @@ void fastcall lru_cache_add(struct page *page)
212 put_cpu_var(lru_add_pvecs); 212 put_cpu_var(lru_add_pvecs);
213} 213}
214 214
215void fastcall lru_cache_add_active(struct page *page) 215void lru_cache_add_active(struct page *page)
216{ 216{
217 struct pagevec *pvec = &get_cpu_var(lru_add_active_pvecs); 217 struct pagevec *pvec = &get_cpu_var(lru_add_active_pvecs);
218 218