diff options
Diffstat (limited to 'arch/sh/kernel/sys_sh.c')
-rw-r--r-- | arch/sh/kernel/sys_sh.c | 92 |
1 files changed, 0 insertions, 92 deletions
diff --git a/arch/sh/kernel/sys_sh.c b/arch/sh/kernel/sys_sh.c index 38f098c9c72d..58dfc02c7af1 100644 --- a/arch/sh/kernel/sys_sh.c +++ b/arch/sh/kernel/sys_sh.c | |||
@@ -22,102 +22,10 @@ | |||
22 | #include <linux/module.h> | 22 | #include <linux/module.h> |
23 | #include <linux/fs.h> | 23 | #include <linux/fs.h> |
24 | #include <linux/ipc.h> | 24 | #include <linux/ipc.h> |
25 | #include <asm/cacheflush.h> | ||
26 | #include <asm/syscalls.h> | 25 | #include <asm/syscalls.h> |
27 | #include <asm/uaccess.h> | 26 | #include <asm/uaccess.h> |
28 | #include <asm/unistd.h> | 27 | #include <asm/unistd.h> |
29 | 28 | ||
30 | unsigned long shm_align_mask = PAGE_SIZE - 1; /* Sane caches */ | ||
31 | EXPORT_SYMBOL(shm_align_mask); | ||
32 | |||
33 | #ifdef CONFIG_MMU | ||
34 | /* | ||
35 | * To avoid cache aliases, we map the shared page with same color. | ||
36 | */ | ||
37 | #define COLOUR_ALIGN(addr, pgoff) \ | ||
38 | ((((addr) + shm_align_mask) & ~shm_align_mask) + \ | ||
39 | (((pgoff) << PAGE_SHIFT) & shm_align_mask)) | ||
40 | |||
41 | unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, | ||
42 | unsigned long len, unsigned long pgoff, unsigned long flags) | ||
43 | { | ||
44 | struct mm_struct *mm = current->mm; | ||
45 | struct vm_area_struct *vma; | ||
46 | unsigned long start_addr; | ||
47 | int do_colour_align; | ||
48 | |||
49 | if (flags & MAP_FIXED) { | ||
50 | /* We do not accept a shared mapping if it would violate | ||
51 | * cache aliasing constraints. | ||
52 | */ | ||
53 | if ((flags & MAP_SHARED) && (addr & shm_align_mask)) | ||
54 | return -EINVAL; | ||
55 | return addr; | ||
56 | } | ||
57 | |||
58 | if (unlikely(len > TASK_SIZE)) | ||
59 | return -ENOMEM; | ||
60 | |||
61 | do_colour_align = 0; | ||
62 | if (filp || (flags & MAP_SHARED)) | ||
63 | do_colour_align = 1; | ||
64 | |||
65 | if (addr) { | ||
66 | if (do_colour_align) | ||
67 | addr = COLOUR_ALIGN(addr, pgoff); | ||
68 | else | ||
69 | addr = PAGE_ALIGN(addr); | ||
70 | |||
71 | vma = find_vma(mm, addr); | ||
72 | if (TASK_SIZE - len >= addr && | ||
73 | (!vma || addr + len <= vma->vm_start)) | ||
74 | return addr; | ||
75 | } | ||
76 | |||
77 | if (len > mm->cached_hole_size) { | ||
78 | start_addr = addr = mm->free_area_cache; | ||
79 | } else { | ||
80 | mm->cached_hole_size = 0; | ||
81 | start_addr = addr = TASK_UNMAPPED_BASE; | ||
82 | } | ||
83 | |||
84 | full_search: | ||
85 | if (do_colour_align) | ||
86 | addr = COLOUR_ALIGN(addr, pgoff); | ||
87 | else | ||
88 | addr = PAGE_ALIGN(mm->free_area_cache); | ||
89 | |||
90 | for (vma = find_vma(mm, addr); ; vma = vma->vm_next) { | ||
91 | /* At this point: (!vma || addr < vma->vm_end). */ | ||
92 | if (unlikely(TASK_SIZE - len < addr)) { | ||
93 | /* | ||
94 | * Start a new search - just in case we missed | ||
95 | * some holes. | ||
96 | */ | ||
97 | if (start_addr != TASK_UNMAPPED_BASE) { | ||
98 | start_addr = addr = TASK_UNMAPPED_BASE; | ||
99 | mm->cached_hole_size = 0; | ||
100 | goto full_search; | ||
101 | } | ||
102 | return -ENOMEM; | ||
103 | } | ||
104 | if (likely(!vma || addr + len <= vma->vm_start)) { | ||
105 | /* | ||
106 | * Remember the place where we stopped the search: | ||
107 | */ | ||
108 | mm->free_area_cache = addr + len; | ||
109 | return addr; | ||
110 | } | ||
111 | if (addr + mm->cached_hole_size < vma->vm_start) | ||
112 | mm->cached_hole_size = vma->vm_start - addr; | ||
113 | |||
114 | addr = vma->vm_end; | ||
115 | if (do_colour_align) | ||
116 | addr = COLOUR_ALIGN(addr, pgoff); | ||
117 | } | ||
118 | } | ||
119 | #endif /* CONFIG_MMU */ | ||
120 | |||
121 | static inline long | 29 | static inline long |
122 | do_mmap2(unsigned long addr, unsigned long len, unsigned long prot, | 30 | do_mmap2(unsigned long addr, unsigned long len, unsigned long prot, |
123 | unsigned long flags, int fd, unsigned long pgoff) | 31 | unsigned long flags, int fd, unsigned long pgoff) |