diff options
Diffstat (limited to 'mm')
| -rw-r--r-- | mm/allocpercpu.c | 2 | ||||
| -rw-r--r-- | mm/memory.c | 16 | ||||
| -rw-r--r-- | mm/mempolicy.c | 6 | ||||
| -rw-r--r-- | mm/migrate.c | 2 | ||||
| -rw-r--r-- | mm/page_alloc.c | 1 | ||||
| -rw-r--r-- | mm/slub.c | 14 | ||||
| -rw-r--r-- | mm/sparse-vmemmap.c | 2 |
7 files changed, 33 insertions, 10 deletions
diff --git a/mm/allocpercpu.c b/mm/allocpercpu.c index f4026bae6eed..05f2b4009ccc 100644 --- a/mm/allocpercpu.c +++ b/mm/allocpercpu.c | |||
| @@ -1,7 +1,7 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * linux/mm/allocpercpu.c | 2 | * linux/mm/allocpercpu.c |
| 3 | * | 3 | * |
| 4 | * Separated from slab.c August 11, 2006 Christoph Lameter <clameter@sgi.com> | 4 | * Separated from slab.c August 11, 2006 Christoph Lameter |
| 5 | */ | 5 | */ |
| 6 | #include <linux/mm.h> | 6 | #include <linux/mm.h> |
| 7 | #include <linux/module.h> | 7 | #include <linux/module.h> |
diff --git a/mm/memory.c b/mm/memory.c index d14b251a25a6..2302d228fe04 100644 --- a/mm/memory.c +++ b/mm/memory.c | |||
| @@ -1151,7 +1151,7 @@ int get_user_pages(struct task_struct *tsk, struct mm_struct *mm, | |||
| 1151 | * be processed until returning to user space. | 1151 | * be processed until returning to user space. |
| 1152 | */ | 1152 | */ |
| 1153 | if (unlikely(test_tsk_thread_flag(tsk, TIF_MEMDIE))) | 1153 | if (unlikely(test_tsk_thread_flag(tsk, TIF_MEMDIE))) |
| 1154 | return -ENOMEM; | 1154 | return i ? i : -ENOMEM; |
| 1155 | 1155 | ||
| 1156 | if (write) | 1156 | if (write) |
| 1157 | foll_flags |= FOLL_WRITE; | 1157 | foll_flags |= FOLL_WRITE; |
| @@ -1697,8 +1697,19 @@ static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
| 1697 | struct page *dirty_page = NULL; | 1697 | struct page *dirty_page = NULL; |
| 1698 | 1698 | ||
| 1699 | old_page = vm_normal_page(vma, address, orig_pte); | 1699 | old_page = vm_normal_page(vma, address, orig_pte); |
| 1700 | if (!old_page) | 1700 | if (!old_page) { |
| 1701 | /* | ||
| 1702 | * VM_MIXEDMAP !pfn_valid() case | ||
| 1703 | * | ||
| 1704 | * We should not cow pages in a shared writeable mapping. | ||
| 1705 | * Just mark the pages writable as we can't do any dirty | ||
| 1706 | * accounting on raw pfn maps. | ||
| 1707 | */ | ||
| 1708 | if ((vma->vm_flags & (VM_WRITE|VM_SHARED)) == | ||
| 1709 | (VM_WRITE|VM_SHARED)) | ||
| 1710 | goto reuse; | ||
| 1701 | goto gotten; | 1711 | goto gotten; |
| 1712 | } | ||
| 1702 | 1713 | ||
| 1703 | /* | 1714 | /* |
| 1704 | * Take out anonymous pages first, anonymous shared vmas are | 1715 | * Take out anonymous pages first, anonymous shared vmas are |
| @@ -1751,6 +1762,7 @@ static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
| 1751 | } | 1762 | } |
| 1752 | 1763 | ||
| 1753 | if (reuse) { | 1764 | if (reuse) { |
| 1765 | reuse: | ||
| 1754 | flush_cache_page(vma, address, pte_pfn(orig_pte)); | 1766 | flush_cache_page(vma, address, pte_pfn(orig_pte)); |
| 1755 | entry = pte_mkyoung(orig_pte); | 1767 | entry = pte_mkyoung(orig_pte); |
| 1756 | entry = maybe_mkwrite(pte_mkdirty(entry), vma); | 1768 | entry = maybe_mkwrite(pte_mkdirty(entry), vma); |
diff --git a/mm/mempolicy.c b/mm/mempolicy.c index a37a5034f63d..c94e58b192c3 100644 --- a/mm/mempolicy.c +++ b/mm/mempolicy.c | |||
| @@ -729,7 +729,11 @@ static long do_get_mempolicy(int *policy, nodemask_t *nmask, | |||
| 729 | } else { | 729 | } else { |
| 730 | *policy = pol == &default_policy ? MPOL_DEFAULT : | 730 | *policy = pol == &default_policy ? MPOL_DEFAULT : |
| 731 | pol->mode; | 731 | pol->mode; |
| 732 | *policy |= pol->flags; | 732 | /* |
| 733 | * Internal mempolicy flags must be masked off before exposing | ||
| 734 | * the policy to userspace. | ||
| 735 | */ | ||
| 736 | *policy |= (pol->flags & MPOL_MODE_FLAGS); | ||
| 733 | } | 737 | } |
| 734 | 738 | ||
| 735 | if (vma) { | 739 | if (vma) { |
diff --git a/mm/migrate.c b/mm/migrate.c index 112bcaeaa104..55bd355d170d 100644 --- a/mm/migrate.c +++ b/mm/migrate.c | |||
| @@ -9,7 +9,7 @@ | |||
| 9 | * IWAMOTO Toshihiro <iwamoto@valinux.co.jp> | 9 | * IWAMOTO Toshihiro <iwamoto@valinux.co.jp> |
| 10 | * Hirokazu Takahashi <taka@valinux.co.jp> | 10 | * Hirokazu Takahashi <taka@valinux.co.jp> |
| 11 | * Dave Hansen <haveblue@us.ibm.com> | 11 | * Dave Hansen <haveblue@us.ibm.com> |
| 12 | * Christoph Lameter <clameter@sgi.com> | 12 | * Christoph Lameter |
| 13 | */ | 13 | */ |
| 14 | 14 | ||
| 15 | #include <linux/migrate.h> | 15 | #include <linux/migrate.h> |
diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 2f552955a02f..f32fae3121f0 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c | |||
| @@ -2328,7 +2328,6 @@ static void build_zonelists(pg_data_t *pgdat) | |||
| 2328 | static void build_zonelist_cache(pg_data_t *pgdat) | 2328 | static void build_zonelist_cache(pg_data_t *pgdat) |
| 2329 | { | 2329 | { |
| 2330 | pgdat->node_zonelists[0].zlcache_ptr = NULL; | 2330 | pgdat->node_zonelists[0].zlcache_ptr = NULL; |
| 2331 | pgdat->node_zonelists[1].zlcache_ptr = NULL; | ||
| 2332 | } | 2331 | } |
| 2333 | 2332 | ||
| 2334 | #endif /* CONFIG_NUMA */ | 2333 | #endif /* CONFIG_NUMA */ |
| @@ -5,7 +5,7 @@ | |||
| 5 | * The allocator synchronizes using per slab locks and only | 5 | * The allocator synchronizes using per slab locks and only |
| 6 | * uses a centralized lock to manage a pool of partial slabs. | 6 | * uses a centralized lock to manage a pool of partial slabs. |
| 7 | * | 7 | * |
| 8 | * (C) 2007 SGI, Christoph Lameter <clameter@sgi.com> | 8 | * (C) 2007 SGI, Christoph Lameter |
| 9 | */ | 9 | */ |
| 10 | 10 | ||
| 11 | #include <linux/mm.h> | 11 | #include <linux/mm.h> |
| @@ -2995,8 +2995,6 @@ void __init kmem_cache_init(void) | |||
| 2995 | create_kmalloc_cache(&kmalloc_caches[1], | 2995 | create_kmalloc_cache(&kmalloc_caches[1], |
| 2996 | "kmalloc-96", 96, GFP_KERNEL); | 2996 | "kmalloc-96", 96, GFP_KERNEL); |
| 2997 | caches++; | 2997 | caches++; |
| 2998 | } | ||
| 2999 | if (KMALLOC_MIN_SIZE <= 128) { | ||
| 3000 | create_kmalloc_cache(&kmalloc_caches[2], | 2998 | create_kmalloc_cache(&kmalloc_caches[2], |
| 3001 | "kmalloc-192", 192, GFP_KERNEL); | 2999 | "kmalloc-192", 192, GFP_KERNEL); |
| 3002 | caches++; | 3000 | caches++; |
| @@ -3026,6 +3024,16 @@ void __init kmem_cache_init(void) | |||
| 3026 | for (i = 8; i < KMALLOC_MIN_SIZE; i += 8) | 3024 | for (i = 8; i < KMALLOC_MIN_SIZE; i += 8) |
| 3027 | size_index[(i - 1) / 8] = KMALLOC_SHIFT_LOW; | 3025 | size_index[(i - 1) / 8] = KMALLOC_SHIFT_LOW; |
| 3028 | 3026 | ||
| 3027 | if (KMALLOC_MIN_SIZE == 128) { | ||
| 3028 | /* | ||
| 3029 | * The 192 byte sized cache is not used if the alignment | ||
| 3030 | * is 128 byte. Redirect kmalloc to use the 256 byte cache | ||
| 3031 | * instead. | ||
| 3032 | */ | ||
| 3033 | for (i = 128 + 8; i <= 192; i += 8) | ||
| 3034 | size_index[(i - 1) / 8] = 8; | ||
| 3035 | } | ||
| 3036 | |||
| 3029 | slab_state = UP; | 3037 | slab_state = UP; |
| 3030 | 3038 | ||
| 3031 | /* Provide the correct kmalloc names now that the caches are up */ | 3039 | /* Provide the correct kmalloc names now that the caches are up */ |
diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c index 99c4f36eb8a3..a91b5f8fcaf6 100644 --- a/mm/sparse-vmemmap.c +++ b/mm/sparse-vmemmap.c | |||
| @@ -1,7 +1,7 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Virtual Memory Map support | 2 | * Virtual Memory Map support |
| 3 | * | 3 | * |
| 4 | * (C) 2007 sgi. Christoph Lameter <clameter@sgi.com>. | 4 | * (C) 2007 sgi. Christoph Lameter. |
| 5 | * | 5 | * |
| 6 | * Virtual memory maps allow VM primitives pfn_to_page, page_to_pfn, | 6 | * Virtual memory maps allow VM primitives pfn_to_page, page_to_pfn, |
| 7 | * virt_to_page, page_address() to be implemented as a base offset | 7 | * virt_to_page, page_address() to be implemented as a base offset |
