diff options
-rw-r--r-- | arch/x86/mm/pat.c | 5 | ||||
-rw-r--r-- | include/linux/mm.h | 15 | ||||
-rw-r--r-- | mm/memory.c | 6 |
3 files changed, 20 insertions, 6 deletions
diff --git a/arch/x86/mm/pat.c b/arch/x86/mm/pat.c index e0ab173b6974..21bc1f787ae2 100644 --- a/arch/x86/mm/pat.c +++ b/arch/x86/mm/pat.c | |||
@@ -641,10 +641,11 @@ static int reserve_pfn_range(u64 paddr, unsigned long size, pgprot_t *vma_prot, | |||
641 | is_ram = pat_pagerange_is_ram(paddr, paddr + size); | 641 | is_ram = pat_pagerange_is_ram(paddr, paddr + size); |
642 | 642 | ||
643 | /* | 643 | /* |
644 | * reserve_pfn_range() doesn't support RAM pages. | 644 | * reserve_pfn_range() doesn't support RAM pages. Maintain the current |
645 | * behavior with RAM pages by returning success. | ||
645 | */ | 646 | */ |
646 | if (is_ram != 0) | 647 | if (is_ram != 0) |
647 | return -EINVAL; | 648 | return 0; |
648 | 649 | ||
649 | ret = reserve_memtype(paddr, paddr + size, want_flags, &flags); | 650 | ret = reserve_memtype(paddr, paddr + size, want_flags, &flags); |
650 | if (ret) | 651 | if (ret) |
diff --git a/include/linux/mm.h b/include/linux/mm.h index 065cdf8c09fb..3daa05feed9f 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h | |||
@@ -98,7 +98,7 @@ extern unsigned int kobjsize(const void *objp); | |||
98 | #define VM_HUGETLB 0x00400000 /* Huge TLB Page VM */ | 98 | #define VM_HUGETLB 0x00400000 /* Huge TLB Page VM */ |
99 | #define VM_NONLINEAR 0x00800000 /* Is non-linear (remap_file_pages) */ | 99 | #define VM_NONLINEAR 0x00800000 /* Is non-linear (remap_file_pages) */ |
100 | #define VM_MAPPED_COPY 0x01000000 /* T if mapped copy of data (nommu mmap) */ | 100 | #define VM_MAPPED_COPY 0x01000000 /* T if mapped copy of data (nommu mmap) */ |
101 | #define VM_INSERTPAGE 0x02000000 /* The vma has had "vm_insert_page()" done on it */ | 101 | #define VM_INSERTPAGE 0x02000000 /* The vma has had "vm_insert_page()" done on it. Refer note in VM_PFNMAP_AT_MMAP below */ |
102 | #define VM_ALWAYSDUMP 0x04000000 /* Always include in core dumps */ | 102 | #define VM_ALWAYSDUMP 0x04000000 /* Always include in core dumps */ |
103 | 103 | ||
104 | #define VM_CAN_NONLINEAR 0x08000000 /* Has ->fault & does nonlinear pages */ | 104 | #define VM_CAN_NONLINEAR 0x08000000 /* Has ->fault & does nonlinear pages */ |
@@ -127,6 +127,17 @@ extern unsigned int kobjsize(const void *objp); | |||
127 | #define VM_SPECIAL (VM_IO | VM_DONTEXPAND | VM_RESERVED | VM_PFNMAP) | 127 | #define VM_SPECIAL (VM_IO | VM_DONTEXPAND | VM_RESERVED | VM_PFNMAP) |
128 | 128 | ||
129 | /* | 129 | /* |
130 | * pfnmap vmas that are fully mapped at mmap time (not mapped on fault). | ||
131 | * Used by x86 PAT to identify such PFNMAP mappings and optimize their handling. | ||
132 | * Note VM_INSERTPAGE flag is overloaded here. i.e, | ||
133 | * VM_INSERTPAGE && !VM_PFNMAP implies | ||
134 | * The vma has had "vm_insert_page()" done on it | ||
135 | * VM_INSERTPAGE && VM_PFNMAP implies | ||
136 | * The vma is PFNMAP with full mapping at mmap time | ||
137 | */ | ||
138 | #define VM_PFNMAP_AT_MMAP (VM_INSERTPAGE | VM_PFNMAP) | ||
139 | |||
140 | /* | ||
130 | * mapping from the currently active vm_flags protection bits (the | 141 | * mapping from the currently active vm_flags protection bits (the |
131 | * low four bits) to a page protection mask.. | 142 | * low four bits) to a page protection mask.. |
132 | */ | 143 | */ |
@@ -145,7 +156,7 @@ extern pgprot_t protection_map[16]; | |||
145 | */ | 156 | */ |
146 | static inline int is_linear_pfn_mapping(struct vm_area_struct *vma) | 157 | static inline int is_linear_pfn_mapping(struct vm_area_struct *vma) |
147 | { | 158 | { |
148 | return ((vma->vm_flags & VM_PFNMAP) && vma->vm_pgoff); | 159 | return ((vma->vm_flags & VM_PFNMAP_AT_MMAP) == VM_PFNMAP_AT_MMAP); |
149 | } | 160 | } |
150 | 161 | ||
151 | static inline int is_pfn_mapping(struct vm_area_struct *vma) | 162 | static inline int is_pfn_mapping(struct vm_area_struct *vma) |
diff --git a/mm/memory.c b/mm/memory.c index baa999e87cd2..d7df5babcba9 100644 --- a/mm/memory.c +++ b/mm/memory.c | |||
@@ -1665,9 +1665,10 @@ int remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, | |||
1665 | * behaviour that some programs depend on. We mark the "original" | 1665 | * behaviour that some programs depend on. We mark the "original" |
1666 | * un-COW'ed pages by matching them up with "vma->vm_pgoff". | 1666 | * un-COW'ed pages by matching them up with "vma->vm_pgoff". |
1667 | */ | 1667 | */ |
1668 | if (addr == vma->vm_start && end == vma->vm_end) | 1668 | if (addr == vma->vm_start && end == vma->vm_end) { |
1669 | vma->vm_pgoff = pfn; | 1669 | vma->vm_pgoff = pfn; |
1670 | else if (is_cow_mapping(vma->vm_flags)) | 1670 | vma->vm_flags |= VM_PFNMAP_AT_MMAP; |
1671 | } else if (is_cow_mapping(vma->vm_flags)) | ||
1671 | return -EINVAL; | 1672 | return -EINVAL; |
1672 | 1673 | ||
1673 | vma->vm_flags |= VM_IO | VM_RESERVED | VM_PFNMAP; | 1674 | vma->vm_flags |= VM_IO | VM_RESERVED | VM_PFNMAP; |
@@ -1679,6 +1680,7 @@ int remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, | |||
1679 | * needed from higher level routine calling unmap_vmas | 1680 | * needed from higher level routine calling unmap_vmas |
1680 | */ | 1681 | */ |
1681 | vma->vm_flags &= ~(VM_IO | VM_RESERVED | VM_PFNMAP); | 1682 | vma->vm_flags &= ~(VM_IO | VM_RESERVED | VM_PFNMAP); |
1683 | vma->vm_flags &= ~VM_PFNMAP_AT_MMAP; | ||
1682 | return -EINVAL; | 1684 | return -EINVAL; |
1683 | } | 1685 | } |
1684 | 1686 | ||