diff options
| author | Linus Torvalds <torvalds@g5.osdl.org> | 2005-11-04 19:27:50 -0500 |
|---|---|---|
| committer | Linus Torvalds <torvalds@g5.osdl.org> | 2005-11-04 19:27:50 -0500 |
| commit | 602d4a7e2f4b843d1a67375d4d7104073495b758 (patch) | |
| tree | 0b9f184e54fa693c27bd5986c114bdcf6949f788 /include/asm-ppc | |
| parent | 0bbacc402e67abca8794a8401c1621dc0c0202e9 (diff) | |
| parent | c51e3a417bb0f295e13a5bad86302b5212eafdf3 (diff) | |
Merge master.kernel.org:/pub/scm/linux/kernel/git/paulus/powerpc-merge
Diffstat (limited to 'include/asm-ppc')
| -rw-r--r-- | include/asm-ppc/bitops.h | 460 | ||||
| -rw-r--r-- | include/asm-ppc/commproc.h | 2 | ||||
| -rw-r--r-- | include/asm-ppc/futex.h | 53 | ||||
| -rw-r--r-- | include/asm-ppc/ipcbuf.h | 29 | ||||
| -rw-r--r-- | include/asm-ppc/kexec.h | 40 | ||||
| -rw-r--r-- | include/asm-ppc/ptrace.h | 152 | ||||
| -rw-r--r-- | include/asm-ppc/sigcontext.h | 15 | ||||
| -rw-r--r-- | include/asm-ppc/stat.h | 69 | ||||
| -rw-r--r-- | include/asm-ppc/tlb.h | 57 | ||||
| -rw-r--r-- | include/asm-ppc/tlbflush.h | 115 | ||||
| -rw-r--r-- | include/asm-ppc/uaccess.h | 393 | ||||
| -rw-r--r-- | include/asm-ppc/ucontext.h | 27 |
12 files changed, 2 insertions, 1410 deletions
diff --git a/include/asm-ppc/bitops.h b/include/asm-ppc/bitops.h deleted file mode 100644 index e30f536fd830..000000000000 --- a/include/asm-ppc/bitops.h +++ /dev/null | |||
| @@ -1,460 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * bitops.h: Bit string operations on the ppc | ||
| 3 | */ | ||
| 4 | |||
| 5 | #ifdef __KERNEL__ | ||
| 6 | #ifndef _PPC_BITOPS_H | ||
| 7 | #define _PPC_BITOPS_H | ||
| 8 | |||
| 9 | #include <linux/config.h> | ||
| 10 | #include <linux/compiler.h> | ||
| 11 | #include <asm/byteorder.h> | ||
| 12 | #include <asm/atomic.h> | ||
| 13 | |||
| 14 | /* | ||
| 15 | * The test_and_*_bit operations are taken to imply a memory barrier | ||
| 16 | * on SMP systems. | ||
| 17 | */ | ||
| 18 | #ifdef CONFIG_SMP | ||
| 19 | #define SMP_WMB "eieio\n" | ||
| 20 | #define SMP_MB "\nsync" | ||
| 21 | #else | ||
| 22 | #define SMP_WMB | ||
| 23 | #define SMP_MB | ||
| 24 | #endif /* CONFIG_SMP */ | ||
| 25 | |||
| 26 | static __inline__ void set_bit(int nr, volatile unsigned long * addr) | ||
| 27 | { | ||
| 28 | unsigned long old; | ||
| 29 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 30 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 31 | |||
| 32 | __asm__ __volatile__("\n\ | ||
| 33 | 1: lwarx %0,0,%3 \n\ | ||
| 34 | or %0,%0,%2 \n" | ||
| 35 | PPC405_ERR77(0,%3) | ||
| 36 | " stwcx. %0,0,%3 \n\ | ||
| 37 | bne- 1b" | ||
| 38 | : "=&r" (old), "=m" (*p) | ||
| 39 | : "r" (mask), "r" (p), "m" (*p) | ||
| 40 | : "cc" ); | ||
| 41 | } | ||
| 42 | |||
| 43 | /* | ||
| 44 | * non-atomic version | ||
| 45 | */ | ||
| 46 | static __inline__ void __set_bit(int nr, volatile unsigned long *addr) | ||
| 47 | { | ||
| 48 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 49 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 50 | |||
| 51 | *p |= mask; | ||
| 52 | } | ||
| 53 | |||
| 54 | /* | ||
| 55 | * clear_bit doesn't imply a memory barrier | ||
| 56 | */ | ||
| 57 | #define smp_mb__before_clear_bit() smp_mb() | ||
| 58 | #define smp_mb__after_clear_bit() smp_mb() | ||
| 59 | |||
| 60 | static __inline__ void clear_bit(int nr, volatile unsigned long *addr) | ||
| 61 | { | ||
| 62 | unsigned long old; | ||
| 63 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 64 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 65 | |||
| 66 | __asm__ __volatile__("\n\ | ||
| 67 | 1: lwarx %0,0,%3 \n\ | ||
| 68 | andc %0,%0,%2 \n" | ||
| 69 | PPC405_ERR77(0,%3) | ||
| 70 | " stwcx. %0,0,%3 \n\ | ||
| 71 | bne- 1b" | ||
| 72 | : "=&r" (old), "=m" (*p) | ||
| 73 | : "r" (mask), "r" (p), "m" (*p) | ||
| 74 | : "cc"); | ||
| 75 | } | ||
| 76 | |||
| 77 | /* | ||
| 78 | * non-atomic version | ||
| 79 | */ | ||
| 80 | static __inline__ void __clear_bit(int nr, volatile unsigned long *addr) | ||
| 81 | { | ||
| 82 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 83 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 84 | |||
| 85 | *p &= ~mask; | ||
| 86 | } | ||
| 87 | |||
| 88 | static __inline__ void change_bit(int nr, volatile unsigned long *addr) | ||
| 89 | { | ||
| 90 | unsigned long old; | ||
| 91 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 92 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 93 | |||
| 94 | __asm__ __volatile__("\n\ | ||
| 95 | 1: lwarx %0,0,%3 \n\ | ||
| 96 | xor %0,%0,%2 \n" | ||
| 97 | PPC405_ERR77(0,%3) | ||
| 98 | " stwcx. %0,0,%3 \n\ | ||
| 99 | bne- 1b" | ||
| 100 | : "=&r" (old), "=m" (*p) | ||
| 101 | : "r" (mask), "r" (p), "m" (*p) | ||
| 102 | : "cc"); | ||
| 103 | } | ||
| 104 | |||
| 105 | /* | ||
| 106 | * non-atomic version | ||
| 107 | */ | ||
| 108 | static __inline__ void __change_bit(int nr, volatile unsigned long *addr) | ||
| 109 | { | ||
| 110 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 111 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 112 | |||
| 113 | *p ^= mask; | ||
| 114 | } | ||
| 115 | |||
| 116 | /* | ||
| 117 | * test_and_*_bit do imply a memory barrier (?) | ||
| 118 | */ | ||
| 119 | static __inline__ int test_and_set_bit(int nr, volatile unsigned long *addr) | ||
| 120 | { | ||
| 121 | unsigned int old, t; | ||
| 122 | unsigned int mask = 1 << (nr & 0x1f); | ||
| 123 | volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5); | ||
| 124 | |||
| 125 | __asm__ __volatile__(SMP_WMB "\n\ | ||
| 126 | 1: lwarx %0,0,%4 \n\ | ||
| 127 | or %1,%0,%3 \n" | ||
| 128 | PPC405_ERR77(0,%4) | ||
| 129 | " stwcx. %1,0,%4 \n\ | ||
| 130 | bne 1b" | ||
| 131 | SMP_MB | ||
| 132 | : "=&r" (old), "=&r" (t), "=m" (*p) | ||
| 133 | : "r" (mask), "r" (p), "m" (*p) | ||
| 134 | : "cc", "memory"); | ||
| 135 | |||
| 136 | return (old & mask) != 0; | ||
| 137 | } | ||
| 138 | |||
| 139 | /* | ||
| 140 | * non-atomic version | ||
| 141 | */ | ||
| 142 | static __inline__ int __test_and_set_bit(int nr, volatile unsigned long *addr) | ||
| 143 | { | ||
| 144 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 145 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 146 | unsigned long old = *p; | ||
| 147 | |||
| 148 | *p = old | mask; | ||
| 149 | return (old & mask) != 0; | ||
| 150 | } | ||
| 151 | |||
| 152 | static __inline__ int test_and_clear_bit(int nr, volatile unsigned long *addr) | ||
| 153 | { | ||
| 154 | unsigned int old, t; | ||
| 155 | unsigned int mask = 1 << (nr & 0x1f); | ||
| 156 | volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5); | ||
| 157 | |||
| 158 | __asm__ __volatile__(SMP_WMB "\n\ | ||
| 159 | 1: lwarx %0,0,%4 \n\ | ||
| 160 | andc %1,%0,%3 \n" | ||
| 161 | PPC405_ERR77(0,%4) | ||
| 162 | " stwcx. %1,0,%4 \n\ | ||
| 163 | bne 1b" | ||
| 164 | SMP_MB | ||
| 165 | : "=&r" (old), "=&r" (t), "=m" (*p) | ||
| 166 | : "r" (mask), "r" (p), "m" (*p) | ||
| 167 | : "cc", "memory"); | ||
| 168 | |||
| 169 | return (old & mask) != 0; | ||
| 170 | } | ||
| 171 | |||
| 172 | /* | ||
| 173 | * non-atomic version | ||
| 174 | */ | ||
| 175 | static __inline__ int __test_and_clear_bit(int nr, volatile unsigned long *addr) | ||
| 176 | { | ||
| 177 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 178 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 179 | unsigned long old = *p; | ||
| 180 | |||
| 181 | *p = old & ~mask; | ||
| 182 | return (old & mask) != 0; | ||
| 183 | } | ||
| 184 | |||
| 185 | static __inline__ int test_and_change_bit(int nr, volatile unsigned long *addr) | ||
| 186 | { | ||
| 187 | unsigned int old, t; | ||
| 188 | unsigned int mask = 1 << (nr & 0x1f); | ||
| 189 | volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5); | ||
| 190 | |||
| 191 | __asm__ __volatile__(SMP_WMB "\n\ | ||
| 192 | 1: lwarx %0,0,%4 \n\ | ||
| 193 | xor %1,%0,%3 \n" | ||
| 194 | PPC405_ERR77(0,%4) | ||
| 195 | " stwcx. %1,0,%4 \n\ | ||
| 196 | bne 1b" | ||
| 197 | SMP_MB | ||
| 198 | : "=&r" (old), "=&r" (t), "=m" (*p) | ||
| 199 | : "r" (mask), "r" (p), "m" (*p) | ||
| 200 | : "cc", "memory"); | ||
| 201 | |||
| 202 | return (old & mask) != 0; | ||
| 203 | } | ||
| 204 | |||
| 205 | /* | ||
| 206 | * non-atomic version | ||
| 207 | */ | ||
| 208 | static __inline__ int __test_and_change_bit(int nr, volatile unsigned long *addr) | ||
| 209 | { | ||
| 210 | unsigned long mask = 1 << (nr & 0x1f); | ||
| 211 | unsigned long *p = ((unsigned long *)addr) + (nr >> 5); | ||
| 212 | unsigned long old = *p; | ||
| 213 | |||
| 214 | *p = old ^ mask; | ||
| 215 | return (old & mask) != 0; | ||
| 216 | } | ||
| 217 | |||
| 218 | static __inline__ int test_bit(int nr, __const__ volatile unsigned long *addr) | ||
| 219 | { | ||
| 220 | return ((addr[nr >> 5] >> (nr & 0x1f)) & 1) != 0; | ||
| 221 | } | ||
| 222 | |||
| 223 | /* Return the bit position of the most significant 1 bit in a word */ | ||
| 224 | static __inline__ int __ilog2(unsigned long x) | ||
| 225 | { | ||
| 226 | int lz; | ||
| 227 | |||
| 228 | asm ("cntlzw %0,%1" : "=r" (lz) : "r" (x)); | ||
| 229 | return 31 - lz; | ||
| 230 | } | ||
| 231 | |||
| 232 | static __inline__ int ffz(unsigned long x) | ||
| 233 | { | ||
| 234 | if ((x = ~x) == 0) | ||
| 235 | return 32; | ||
| 236 | return __ilog2(x & -x); | ||
| 237 | } | ||
| 238 | |||
| 239 | static inline int __ffs(unsigned long x) | ||
| 240 | { | ||
| 241 | return __ilog2(x & -x); | ||
| 242 | } | ||
| 243 | |||
| 244 | /* | ||
| 245 | * ffs: find first bit set. This is defined the same way as | ||
| 246 | * the libc and compiler builtin ffs routines, therefore | ||
| 247 | * differs in spirit from the above ffz (man ffs). | ||
| 248 | */ | ||
| 249 | static __inline__ int ffs(int x) | ||
| 250 | { | ||
| 251 | return __ilog2(x & -x) + 1; | ||
| 252 | } | ||
| 253 | |||
| 254 | /* | ||
| 255 | * fls: find last (most-significant) bit set. | ||
| 256 | * Note fls(0) = 0, fls(1) = 1, fls(0x80000000) = 32. | ||
| 257 | */ | ||
| 258 | static __inline__ int fls(unsigned int x) | ||
| 259 | { | ||
| 260 | int lz; | ||
| 261 | |||
| 262 | asm ("cntlzw %0,%1" : "=r" (lz) : "r" (x)); | ||
| 263 | return 32 - lz; | ||
| 264 | } | ||
| 265 | |||
| 266 | /* | ||
| 267 | * hweightN: returns the hamming weight (i.e. the number | ||
| 268 | * of bits set) of a N-bit word | ||
| 269 | */ | ||
| 270 | |||
| 271 | #define hweight32(x) generic_hweight32(x) | ||
| 272 | #define hweight16(x) generic_hweight16(x) | ||
| 273 | #define hweight8(x) generic_hweight8(x) | ||
| 274 | |||
| 275 | /* | ||
| 276 | * Find the first bit set in a 140-bit bitmap. | ||
| 277 | * The first 100 bits are unlikely to be set. | ||
| 278 | */ | ||
| 279 | static inline int sched_find_first_bit(const unsigned long *b) | ||
| 280 | { | ||
| 281 | if (unlikely(b[0])) | ||
| 282 | return __ffs(b[0]); | ||
| 283 | if (unlikely(b[1])) | ||
| 284 | return __ffs(b[1]) + 32; | ||
| 285 | if (unlikely(b[2])) | ||
| 286 | return __ffs(b[2]) + 64; | ||
| 287 | if (b[3]) | ||
| 288 | return __ffs(b[3]) + 96; | ||
| 289 | return __ffs(b[4]) + 128; | ||
| 290 | } | ||
| 291 | |||
| 292 | /** | ||
| 293 | * find_next_bit - find the next set bit in a memory region | ||
| 294 | * @addr: The address to base the search on | ||
| 295 | * @offset: The bitnumber to start searching at | ||
| 296 | * @size: The maximum size to search | ||
| 297 | */ | ||
| 298 | static __inline__ unsigned long find_next_bit(const unsigned long *addr, | ||
| 299 | unsigned long size, unsigned long offset) | ||
| 300 | { | ||
| 301 | unsigned int *p = ((unsigned int *) addr) + (offset >> 5); | ||
| 302 | unsigned int result = offset & ~31UL; | ||
| 303 | unsigned int tmp; | ||
| 304 | |||
| 305 | if (offset >= size) | ||
| 306 | return size; | ||
| 307 | size -= result; | ||
| 308 | offset &= 31UL; | ||
| 309 | if (offset) { | ||
| 310 | tmp = *p++; | ||
| 311 | tmp &= ~0UL << offset; | ||
| 312 | if (size < 32) | ||
| 313 | goto found_first; | ||
| 314 | if (tmp) | ||
| 315 | goto found_middle; | ||
| 316 | size -= 32; | ||
| 317 | result += 32; | ||
| 318 | } | ||
| 319 | while (size >= 32) { | ||
| 320 | if ((tmp = *p++) != 0) | ||
| 321 | goto found_middle; | ||
| 322 | result += 32; | ||
| 323 | size -= 32; | ||
| 324 | } | ||
| 325 | if (!size) | ||
| 326 | return result; | ||
| 327 | tmp = *p; | ||
| 328 | |||
| 329 | found_first: | ||
| 330 | tmp &= ~0UL >> (32 - size); | ||
| 331 | if (tmp == 0UL) /* Are any bits set? */ | ||
| 332 | return result + size; /* Nope. */ | ||
| 333 | found_middle: | ||
| 334 | return result + __ffs(tmp); | ||
| 335 | } | ||
| 336 | |||
| 337 | /** | ||
| 338 | * find_first_bit - find the first set bit in a memory region | ||
| 339 | * @addr: The address to start the search at | ||
| 340 | * @size: The maximum size to search | ||
| 341 | * | ||
| 342 | * Returns the bit-number of the first set bit, not the number of the byte | ||
| 343 | * containing a bit. | ||
| 344 | */ | ||
| 345 | #define find_first_bit(addr, size) \ | ||
| 346 | find_next_bit((addr), (size), 0) | ||
| 347 | |||
| 348 | /* | ||
| 349 | * This implementation of find_{first,next}_zero_bit was stolen from | ||
| 350 | * Linus' asm-alpha/bitops.h. | ||
| 351 | */ | ||
| 352 | #define find_first_zero_bit(addr, size) \ | ||
| 353 | find_next_zero_bit((addr), (size), 0) | ||
| 354 | |||
| 355 | static __inline__ unsigned long find_next_zero_bit(const unsigned long *addr, | ||
| 356 | unsigned long size, unsigned long offset) | ||
| 357 | { | ||
| 358 | unsigned int * p = ((unsigned int *) addr) + (offset >> 5); | ||
| 359 | unsigned int result = offset & ~31UL; | ||
| 360 | unsigned int tmp; | ||
| 361 | |||
| 362 | if (offset >= size) | ||
| 363 | return size; | ||
| 364 | size -= result; | ||
| 365 | offset &= 31UL; | ||
| 366 | if (offset) { | ||
| 367 | tmp = *p++; | ||
| 368 | tmp |= ~0UL >> (32-offset); | ||
| 369 | if (size < 32) | ||
| 370 | goto found_first; | ||
| 371 | if (tmp != ~0U) | ||
| 372 | goto found_middle; | ||
| 373 | size -= 32; | ||
| 374 | result += 32; | ||
| 375 | } | ||
| 376 | while (size >= 32) { | ||
| 377 | if ((tmp = *p++) != ~0U) | ||
| 378 | goto found_middle; | ||
| 379 | result += 32; | ||
| 380 | size -= 32; | ||
| 381 | } | ||
| 382 | if (!size) | ||
| 383 | return result; | ||
| 384 | tmp = *p; | ||
| 385 | found_first: | ||
| 386 | tmp |= ~0UL << size; | ||
| 387 | if (tmp == ~0UL) /* Are any bits zero? */ | ||
| 388 | return result + size; /* Nope. */ | ||
| 389 | found_middle: | ||
| 390 | return result + ffz(tmp); | ||
| 391 | } | ||
| 392 | |||
| 393 | |||
| 394 | #define ext2_set_bit(nr, addr) __test_and_set_bit((nr) ^ 0x18, (unsigned long *)(addr)) | ||
| 395 | #define ext2_set_bit_atomic(lock, nr, addr) test_and_set_bit((nr) ^ 0x18, (unsigned long *)(addr)) | ||
| 396 | #define ext2_clear_bit(nr, addr) __test_and_clear_bit((nr) ^ 0x18, (unsigned long *)(addr)) | ||
| 397 | #define ext2_clear_bit_atomic(lock, nr, addr) test_and_clear_bit((nr) ^ 0x18, (unsigned long *)(addr)) | ||
| 398 | |||
| 399 | static __inline__ int ext2_test_bit(int nr, __const__ void * addr) | ||
| 400 | { | ||
| 401 | __const__ unsigned char *ADDR = (__const__ unsigned char *) addr; | ||
| 402 | |||
| 403 | return (ADDR[nr >> 3] >> (nr & 7)) & 1; | ||
| 404 | } | ||
| 405 | |||
| 406 | /* | ||
| 407 | * This implementation of ext2_find_{first,next}_zero_bit was stolen from | ||
| 408 | * Linus' asm-alpha/bitops.h and modified for a big-endian machine. | ||
| 409 | */ | ||
| 410 | |||
| 411 | #define ext2_find_first_zero_bit(addr, size) \ | ||
| 412 | ext2_find_next_zero_bit((addr), (size), 0) | ||
| 413 | |||
| 414 | static __inline__ unsigned long ext2_find_next_zero_bit(const void *addr, | ||
| 415 | unsigned long size, unsigned long offset) | ||
| 416 | { | ||
| 417 | unsigned int *p = ((unsigned int *) addr) + (offset >> 5); | ||
| 418 | unsigned int result = offset & ~31UL; | ||
| 419 | unsigned int tmp; | ||
| 420 | |||
| 421 | if (offset >= size) | ||
| 422 | return size; | ||
| 423 | size -= result; | ||
| 424 | offset &= 31UL; | ||
| 425 | if (offset) { | ||
| 426 | tmp = cpu_to_le32p(p++); | ||
| 427 | tmp |= ~0UL >> (32-offset); | ||
| 428 | if (size < 32) | ||
| 429 | goto found_first; | ||
| 430 | if (tmp != ~0U) | ||
| 431 | goto found_middle; | ||
| 432 | size -= 32; | ||
| 433 | result += 32; | ||
| 434 | } | ||
| 435 | while (size >= 32) { | ||
| 436 | if ((tmp = cpu_to_le32p(p++)) != ~0U) | ||
| 437 | goto found_middle; | ||
| 438 | result += 32; | ||
| 439 | size -= 32; | ||
| 440 | } | ||
| 441 | if (!size) | ||
| 442 | return result; | ||
| 443 | tmp = cpu_to_le32p(p); | ||
| 444 | found_first: | ||
| 445 | tmp |= ~0U << size; | ||
| 446 | if (tmp == ~0UL) /* Are any bits zero? */ | ||
| 447 | return result + size; /* Nope. */ | ||
| 448 | found_middle: | ||
| 449 | return result + ffz(tmp); | ||
| 450 | } | ||
| 451 | |||
| 452 | /* Bitmap functions for the minix filesystem. */ | ||
| 453 | #define minix_test_and_set_bit(nr,addr) ext2_set_bit(nr,addr) | ||
| 454 | #define minix_set_bit(nr,addr) ((void)ext2_set_bit(nr,addr)) | ||
| 455 | #define minix_test_and_clear_bit(nr,addr) ext2_clear_bit(nr,addr) | ||
| 456 | #define minix_test_bit(nr,addr) ext2_test_bit(nr,addr) | ||
| 457 | #define minix_find_first_zero_bit(addr,size) ext2_find_first_zero_bit(addr,size) | ||
| 458 | |||
| 459 | #endif /* _PPC_BITOPS_H */ | ||
| 460 | #endif /* __KERNEL__ */ | ||
diff --git a/include/asm-ppc/commproc.h b/include/asm-ppc/commproc.h index 5bbb8e2c1c6d..973e60908234 100644 --- a/include/asm-ppc/commproc.h +++ b/include/asm-ppc/commproc.h | |||
| @@ -83,6 +83,8 @@ extern uint m8xx_cpm_hostalloc(uint size); | |||
| 83 | extern int m8xx_cpm_hostfree(uint start); | 83 | extern int m8xx_cpm_hostfree(uint start); |
| 84 | extern void m8xx_cpm_hostdump(void); | 84 | extern void m8xx_cpm_hostdump(void); |
| 85 | 85 | ||
| 86 | extern void cpm_load_patch(volatile immap_t *immr); | ||
| 87 | |||
| 86 | /* Buffer descriptors used by many of the CPM protocols. | 88 | /* Buffer descriptors used by many of the CPM protocols. |
| 87 | */ | 89 | */ |
| 88 | typedef struct cpm_buf_desc { | 90 | typedef struct cpm_buf_desc { |
diff --git a/include/asm-ppc/futex.h b/include/asm-ppc/futex.h deleted file mode 100644 index 9feff4ce1424..000000000000 --- a/include/asm-ppc/futex.h +++ /dev/null | |||
| @@ -1,53 +0,0 @@ | |||
| 1 | #ifndef _ASM_FUTEX_H | ||
| 2 | #define _ASM_FUTEX_H | ||
| 3 | |||
| 4 | #ifdef __KERNEL__ | ||
| 5 | |||
| 6 | #include <linux/futex.h> | ||
| 7 | #include <asm/errno.h> | ||
| 8 | #include <asm/uaccess.h> | ||
| 9 | |||
| 10 | static inline int | ||
| 11 | futex_atomic_op_inuser (int encoded_op, int __user *uaddr) | ||
| 12 | { | ||
| 13 | int op = (encoded_op >> 28) & 7; | ||
| 14 | int cmp = (encoded_op >> 24) & 15; | ||
| 15 | int oparg = (encoded_op << 8) >> 20; | ||
| 16 | int cmparg = (encoded_op << 20) >> 20; | ||
| 17 | int oldval = 0, ret; | ||
| 18 | if (encoded_op & (FUTEX_OP_OPARG_SHIFT << 28)) | ||
| 19 | oparg = 1 << oparg; | ||
| 20 | |||
| 21 | if (! access_ok (VERIFY_WRITE, uaddr, sizeof(int))) | ||
| 22 | return -EFAULT; | ||
| 23 | |||
| 24 | inc_preempt_count(); | ||
| 25 | |||
| 26 | switch (op) { | ||
| 27 | case FUTEX_OP_SET: | ||
| 28 | case FUTEX_OP_ADD: | ||
| 29 | case FUTEX_OP_OR: | ||
| 30 | case FUTEX_OP_ANDN: | ||
| 31 | case FUTEX_OP_XOR: | ||
| 32 | default: | ||
| 33 | ret = -ENOSYS; | ||
| 34 | } | ||
| 35 | |||
| 36 | dec_preempt_count(); | ||
| 37 | |||
| 38 | if (!ret) { | ||
| 39 | switch (cmp) { | ||
| 40 | case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; | ||
| 41 | case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; | ||
| 42 | case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break; | ||
| 43 | case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break; | ||
| 44 | case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break; | ||
| 45 | case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break; | ||
| 46 | default: ret = -ENOSYS; | ||
| 47 | } | ||
| 48 | } | ||
| 49 | return ret; | ||
| 50 | } | ||
| 51 | |||
| 52 | #endif | ||
| 53 | #endif | ||
diff --git a/include/asm-ppc/ipcbuf.h b/include/asm-ppc/ipcbuf.h deleted file mode 100644 index fab6752c7480..000000000000 --- a/include/asm-ppc/ipcbuf.h +++ /dev/null | |||
| @@ -1,29 +0,0 @@ | |||
| 1 | #ifndef __PPC_IPCBUF_H__ | ||
| 2 | #define __PPC_IPCBUF_H__ | ||
| 3 | |||
| 4 | /* | ||
| 5 | * The ipc64_perm structure for PPC architecture. | ||
| 6 | * Note extra padding because this structure is passed back and forth | ||
| 7 | * between kernel and user space. | ||
| 8 | * | ||
| 9 | * Pad space is left for: | ||
| 10 | * - 1 32-bit value to fill up for 8-byte alignment | ||
| 11 | * - 2 miscellaneous 64-bit values (so that this structure matches | ||
| 12 | * PPC64 ipc64_perm) | ||
| 13 | */ | ||
| 14 | |||
| 15 | struct ipc64_perm | ||
| 16 | { | ||
| 17 | __kernel_key_t key; | ||
| 18 | __kernel_uid_t uid; | ||
| 19 | __kernel_gid_t gid; | ||
| 20 | __kernel_uid_t cuid; | ||
| 21 | __kernel_gid_t cgid; | ||
| 22 | __kernel_mode_t mode; | ||
| 23 | unsigned long seq; | ||
| 24 | unsigned int __pad2; | ||
| 25 | unsigned long long __unused1; | ||
| 26 | unsigned long long __unused2; | ||
| 27 | }; | ||
| 28 | |||
| 29 | #endif /* __PPC_IPCBUF_H__ */ | ||
diff --git a/include/asm-ppc/kexec.h b/include/asm-ppc/kexec.h deleted file mode 100644 index 6d2aa0aa4642..000000000000 --- a/include/asm-ppc/kexec.h +++ /dev/null | |||
| @@ -1,40 +0,0 @@ | |||
| 1 | #ifndef _PPC_KEXEC_H | ||
| 2 | #define _PPC_KEXEC_H | ||
| 3 | |||
| 4 | #ifdef CONFIG_KEXEC | ||
| 5 | |||
| 6 | /* | ||
| 7 | * KEXEC_SOURCE_MEMORY_LIMIT maximum page get_free_page can return. | ||
| 8 | * I.e. Maximum page that is mapped directly into kernel memory, | ||
| 9 | * and kmap is not required. | ||
| 10 | * | ||
| 11 | * Someone correct me if FIXADDR_START - PAGEOFFSET is not the correct | ||
| 12 | * calculation for the amount of memory directly mappable into the | ||
| 13 | * kernel memory space. | ||
| 14 | */ | ||
| 15 | |||
| 16 | /* Maximum physical address we can use pages from */ | ||
| 17 | #define KEXEC_SOURCE_MEMORY_LIMIT (-1UL) | ||
| 18 | /* Maximum address we can reach in physical address mode */ | ||
| 19 | #define KEXEC_DESTINATION_MEMORY_LIMIT (-1UL) | ||
| 20 | /* Maximum address we can use for the control code buffer */ | ||
| 21 | #define KEXEC_CONTROL_MEMORY_LIMIT TASK_SIZE | ||
| 22 | |||
| 23 | #define KEXEC_CONTROL_CODE_SIZE 4096 | ||
| 24 | |||
| 25 | /* The native architecture */ | ||
| 26 | #define KEXEC_ARCH KEXEC_ARCH_PPC | ||
| 27 | |||
| 28 | #ifndef __ASSEMBLY__ | ||
| 29 | |||
| 30 | extern void *crash_notes; | ||
| 31 | |||
| 32 | struct kimage; | ||
| 33 | |||
| 34 | extern void machine_kexec_simple(struct kimage *image); | ||
| 35 | |||
| 36 | #endif /* __ASSEMBLY__ */ | ||
| 37 | |||
| 38 | #endif /* CONFIG_KEXEC */ | ||
| 39 | |||
| 40 | #endif /* _PPC_KEXEC_H */ | ||
diff --git a/include/asm-ppc/ptrace.h b/include/asm-ppc/ptrace.h deleted file mode 100644 index c34fb4e37a97..000000000000 --- a/include/asm-ppc/ptrace.h +++ /dev/null | |||
| @@ -1,152 +0,0 @@ | |||
| 1 | #ifndef _PPC_PTRACE_H | ||
| 2 | #define _PPC_PTRACE_H | ||
| 3 | |||
| 4 | /* | ||
| 5 | * This struct defines the way the registers are stored on the | ||
| 6 | * kernel stack during a system call or other kernel entry. | ||
| 7 | * | ||
| 8 | * this should only contain volatile regs | ||
| 9 | * since we can keep non-volatile in the thread_struct | ||
| 10 | * should set this up when only volatiles are saved | ||
| 11 | * by intr code. | ||
| 12 | * | ||
| 13 | * Since this is going on the stack, *CARE MUST BE TAKEN* to insure | ||
| 14 | * that the overall structure is a multiple of 16 bytes in length. | ||
| 15 | * | ||
| 16 | * Note that the offsets of the fields in this struct correspond with | ||
| 17 | * the PT_* values below. This simplifies arch/ppc/kernel/ptrace.c. | ||
| 18 | */ | ||
| 19 | |||
| 20 | #ifndef __ASSEMBLY__ | ||
| 21 | struct pt_regs { | ||
| 22 | unsigned long gpr[32]; | ||
| 23 | unsigned long nip; | ||
| 24 | unsigned long msr; | ||
| 25 | unsigned long orig_gpr3; /* Used for restarting system calls */ | ||
| 26 | unsigned long ctr; | ||
| 27 | unsigned long link; | ||
| 28 | unsigned long xer; | ||
| 29 | unsigned long ccr; | ||
| 30 | unsigned long mq; /* 601 only (not used at present) */ | ||
| 31 | /* Used on APUS to hold IPL value. */ | ||
| 32 | unsigned long trap; /* Reason for being here */ | ||
| 33 | /* N.B. for critical exceptions on 4xx, the dar and dsisr | ||
| 34 | fields are overloaded to hold srr0 and srr1. */ | ||
| 35 | unsigned long dar; /* Fault registers */ | ||
| 36 | unsigned long dsisr; /* on 4xx/Book-E used for ESR */ | ||
| 37 | unsigned long result; /* Result of a system call */ | ||
| 38 | }; | ||
| 39 | |||
| 40 | #endif /* __ASSEMBLY__ */ | ||
| 41 | |||
| 42 | #ifdef __KERNEL__ | ||
| 43 | #define STACK_FRAME_OVERHEAD 16 /* size of minimum stack frame */ | ||
| 44 | |||
| 45 | /* Size of stack frame allocated when calling signal handler. */ | ||
| 46 | #define __SIGNAL_FRAMESIZE 64 | ||
| 47 | |||
| 48 | #ifndef __ASSEMBLY__ | ||
| 49 | #define instruction_pointer(regs) ((regs)->nip) | ||
| 50 | #ifdef CONFIG_SMP | ||
| 51 | extern unsigned long profile_pc(struct pt_regs *regs); | ||
| 52 | #else | ||
| 53 | #define profile_pc(regs) instruction_pointer(regs) | ||
| 54 | #endif | ||
| 55 | |||
| 56 | #define user_mode(regs) (((regs)->msr & MSR_PR) != 0) | ||
| 57 | |||
| 58 | #define force_successful_syscall_return() \ | ||
| 59 | do { \ | ||
| 60 | current_thread_info()->syscall_noerror = 1; \ | ||
| 61 | } while(0) | ||
| 62 | |||
| 63 | /* | ||
| 64 | * We use the least-significant bit of the trap field to indicate | ||
| 65 | * whether we have saved the full set of registers, or only a | ||
| 66 | * partial set. A 1 there means the partial set. | ||
| 67 | * On 4xx we use the next bit to indicate whether the exception | ||
| 68 | * is a critical exception (1 means it is). | ||
| 69 | */ | ||
| 70 | #define FULL_REGS(regs) (((regs)->trap & 1) == 0) | ||
| 71 | #define IS_CRITICAL_EXC(regs) (((regs)->trap & 2) == 0) | ||
| 72 | #define TRAP(regs) ((regs)->trap & ~0xF) | ||
| 73 | |||
| 74 | #define CHECK_FULL_REGS(regs) \ | ||
| 75 | do { \ | ||
| 76 | if ((regs)->trap & 1) \ | ||
| 77 | printk(KERN_CRIT "%s: partial register set\n", __FUNCTION__); \ | ||
| 78 | } while (0) | ||
| 79 | #endif /* __ASSEMBLY__ */ | ||
| 80 | |||
| 81 | #endif /* __KERNEL__ */ | ||
| 82 | |||
| 83 | /* | ||
| 84 | * Offsets used by 'ptrace' system call interface. | ||
| 85 | * These can't be changed without breaking binary compatibility | ||
| 86 | * with MkLinux, etc. | ||
| 87 | */ | ||
| 88 | #define PT_R0 0 | ||
| 89 | #define PT_R1 1 | ||
| 90 | #define PT_R2 2 | ||
| 91 | #define PT_R3 3 | ||
| 92 | #define PT_R4 4 | ||
| 93 | #define PT_R5 5 | ||
| 94 | #define PT_R6 6 | ||
| 95 | #define PT_R7 7 | ||
| 96 | #define PT_R8 8 | ||
| 97 | #define PT_R9 9 | ||
| 98 | #define PT_R10 10 | ||
| 99 | #define PT_R11 11 | ||
| 100 | #define PT_R12 12 | ||
| 101 | #define PT_R13 13 | ||
| 102 | #define PT_R14 14 | ||
| 103 | #define PT_R15 15 | ||
| 104 | #define PT_R16 16 | ||
| 105 | #define PT_R17 17 | ||
| 106 | #define PT_R18 18 | ||
| 107 | #define PT_R19 19 | ||
| 108 | #define PT_R20 20 | ||
| 109 | #define PT_R21 21 | ||
| 110 | #define PT_R22 22 | ||
| 111 | #define PT_R23 23 | ||
| 112 | #define PT_R24 24 | ||
| 113 | #define PT_R25 25 | ||
| 114 | #define PT_R26 26 | ||
| 115 | #define PT_R27 27 | ||
| 116 | #define PT_R28 28 | ||
| 117 | #define PT_R29 29 | ||
| 118 | #define PT_R30 30 | ||
| 119 | #define PT_R31 31 | ||
| 120 | |||
| 121 | #define PT_NIP 32 | ||
| 122 | #define PT_MSR 33 | ||
| 123 | #ifdef __KERNEL__ | ||
| 124 | #define PT_ORIG_R3 34 | ||
| 125 | #endif | ||
| 126 | #define PT_CTR 35 | ||
| 127 | #define PT_LNK 36 | ||
| 128 | #define PT_XER 37 | ||
| 129 | #define PT_CCR 38 | ||
| 130 | #define PT_MQ 39 | ||
| 131 | |||
| 132 | #define PT_FPR0 48 /* each FP reg occupies 2 slots in this space */ | ||
| 133 | #define PT_FPR31 (PT_FPR0 + 2*31) | ||
| 134 | #define PT_FPSCR (PT_FPR0 + 2*32 + 1) | ||
| 135 | |||
| 136 | /* Get/set all the altivec registers vr0..vr31, vscr, vrsave, in one go */ | ||
| 137 | #define PTRACE_GETVRREGS 18 | ||
| 138 | #define PTRACE_SETVRREGS 19 | ||
| 139 | |||
| 140 | /* Get/set all the upper 32-bits of the SPE registers, accumulator, and | ||
| 141 | * spefscr, in one go */ | ||
| 142 | #define PTRACE_GETEVRREGS 20 | ||
| 143 | #define PTRACE_SETEVRREGS 21 | ||
| 144 | |||
| 145 | /* | ||
| 146 | * Get or set a debug register. The first 16 are DABR registers and the | ||
| 147 | * second 16 are IABR registers. | ||
| 148 | */ | ||
| 149 | #define PTRACE_GET_DEBUGREG 25 | ||
| 150 | #define PTRACE_SET_DEBUGREG 26 | ||
| 151 | |||
| 152 | #endif | ||
diff --git a/include/asm-ppc/sigcontext.h b/include/asm-ppc/sigcontext.h deleted file mode 100644 index b7a417e0a921..000000000000 --- a/include/asm-ppc/sigcontext.h +++ /dev/null | |||
| @@ -1,15 +0,0 @@ | |||
| 1 | #ifndef _ASM_PPC_SIGCONTEXT_H | ||
| 2 | #define _ASM_PPC_SIGCONTEXT_H | ||
| 3 | |||
| 4 | #include <asm/ptrace.h> | ||
| 5 | #include <linux/compiler.h> | ||
| 6 | |||
| 7 | struct sigcontext { | ||
| 8 | unsigned long _unused[4]; | ||
| 9 | int signal; | ||
| 10 | unsigned long handler; | ||
| 11 | unsigned long oldmask; | ||
| 12 | struct pt_regs __user *regs; | ||
| 13 | }; | ||
| 14 | |||
| 15 | #endif | ||
diff --git a/include/asm-ppc/stat.h b/include/asm-ppc/stat.h deleted file mode 100644 index cadb34298496..000000000000 --- a/include/asm-ppc/stat.h +++ /dev/null | |||
| @@ -1,69 +0,0 @@ | |||
| 1 | #ifndef _PPC_STAT_H | ||
| 2 | #define _PPC_STAT_H | ||
| 3 | |||
| 4 | #ifdef __KERNEL__ | ||
| 5 | #include <linux/types.h> | ||
| 6 | #endif /* __KERNEL__ */ | ||
| 7 | |||
| 8 | struct __old_kernel_stat { | ||
| 9 | unsigned short st_dev; | ||
| 10 | unsigned short st_ino; | ||
| 11 | unsigned short st_mode; | ||
| 12 | unsigned short st_nlink; | ||
| 13 | unsigned short st_uid; | ||
| 14 | unsigned short st_gid; | ||
| 15 | unsigned short st_rdev; | ||
| 16 | unsigned long st_size; | ||
| 17 | unsigned long st_atime; | ||
| 18 | unsigned long st_mtime; | ||
| 19 | unsigned long st_ctime; | ||
| 20 | }; | ||
| 21 | |||
| 22 | #define STAT_HAVE_NSEC 1 | ||
| 23 | |||
| 24 | struct stat { | ||
| 25 | unsigned st_dev; | ||
| 26 | ino_t st_ino; | ||
| 27 | mode_t st_mode; | ||
| 28 | nlink_t st_nlink; | ||
| 29 | uid_t st_uid; | ||
| 30 | gid_t st_gid; | ||
| 31 | unsigned st_rdev; | ||
| 32 | off_t st_size; | ||
| 33 | unsigned long st_blksize; | ||
| 34 | unsigned long st_blocks; | ||
| 35 | unsigned long st_atime; | ||
| 36 | unsigned long st_atime_nsec; | ||
| 37 | unsigned long st_mtime; | ||
| 38 | unsigned long st_mtime_nsec; | ||
| 39 | unsigned long st_ctime; | ||
| 40 | unsigned long st_ctime_nsec; | ||
| 41 | unsigned long __unused4; | ||
| 42 | unsigned long __unused5; | ||
| 43 | }; | ||
| 44 | |||
| 45 | /* This matches struct stat64 in glibc2.1. | ||
| 46 | */ | ||
| 47 | struct stat64 { | ||
| 48 | unsigned long long st_dev; /* Device. */ | ||
| 49 | unsigned long long st_ino; /* File serial number. */ | ||
| 50 | unsigned int st_mode; /* File mode. */ | ||
| 51 | unsigned int st_nlink; /* Link count. */ | ||
| 52 | unsigned int st_uid; /* User ID of the file's owner. */ | ||
| 53 | unsigned int st_gid; /* Group ID of the file's group. */ | ||
| 54 | unsigned long long st_rdev; /* Device number, if device. */ | ||
| 55 | unsigned short int __pad2; | ||
| 56 | long long st_size; /* Size of file, in bytes. */ | ||
| 57 | long st_blksize; /* Optimal block size for I/O. */ | ||
| 58 | |||
| 59 | long long st_blocks; /* Number 512-byte blocks allocated. */ | ||
| 60 | long st_atime; /* Time of last access. */ | ||
| 61 | unsigned long st_atime_nsec; | ||
| 62 | long st_mtime; /* Time of last modification. */ | ||
| 63 | unsigned long int st_mtime_nsec; | ||
| 64 | long st_ctime; /* Time of last status change. */ | ||
| 65 | unsigned long int st_ctime_nsec; | ||
| 66 | unsigned long int __unused4; | ||
| 67 | unsigned long int __unused5; | ||
| 68 | }; | ||
| 69 | #endif | ||
diff --git a/include/asm-ppc/tlb.h b/include/asm-ppc/tlb.h deleted file mode 100644 index 2c142c5d8584..000000000000 --- a/include/asm-ppc/tlb.h +++ /dev/null | |||
| @@ -1,57 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * TLB shootdown specifics for PPC | ||
| 3 | * | ||
| 4 | * Copyright (C) 2002 Paul Mackerras, IBM Corp. | ||
| 5 | * | ||
| 6 | * This program is free software; you can redistribute it and/or | ||
| 7 | * modify it under the terms of the GNU General Public License | ||
| 8 | * as published by the Free Software Foundation; either version | ||
| 9 | * 2 of the License, or (at your option) any later version. | ||
| 10 | */ | ||
| 11 | #ifndef _PPC_TLB_H | ||
| 12 | #define _PPC_TLB_H | ||
| 13 | |||
| 14 | #include <linux/config.h> | ||
| 15 | #include <asm/pgtable.h> | ||
| 16 | #include <asm/pgalloc.h> | ||
| 17 | #include <asm/tlbflush.h> | ||
| 18 | #include <asm/page.h> | ||
| 19 | #include <asm/mmu.h> | ||
| 20 | |||
| 21 | #ifdef CONFIG_PPC_STD_MMU | ||
| 22 | /* Classic PPC with hash-table based MMU... */ | ||
| 23 | |||
| 24 | struct mmu_gather; | ||
| 25 | extern void tlb_flush(struct mmu_gather *tlb); | ||
| 26 | |||
| 27 | /* Get the generic bits... */ | ||
| 28 | #include <asm-generic/tlb.h> | ||
| 29 | |||
| 30 | /* Nothing needed here in fact... */ | ||
| 31 | #define tlb_start_vma(tlb, vma) do { } while (0) | ||
| 32 | #define tlb_end_vma(tlb, vma) do { } while (0) | ||
| 33 | |||
| 34 | extern void flush_hash_entry(struct mm_struct *mm, pte_t *ptep, | ||
| 35 | unsigned long address); | ||
| 36 | |||
| 37 | static inline void __tlb_remove_tlb_entry(struct mmu_gather *tlb, pte_t *ptep, | ||
| 38 | unsigned long address) | ||
| 39 | { | ||
| 40 | if (pte_val(*ptep) & _PAGE_HASHPTE) | ||
| 41 | flush_hash_entry(tlb->mm, ptep, address); | ||
| 42 | } | ||
| 43 | |||
| 44 | #else | ||
| 45 | /* Embedded PPC with software-loaded TLB, very simple... */ | ||
| 46 | |||
| 47 | #define tlb_start_vma(tlb, vma) do { } while (0) | ||
| 48 | #define tlb_end_vma(tlb, vma) do { } while (0) | ||
| 49 | #define __tlb_remove_tlb_entry(tlb, pte, address) do { } while (0) | ||
| 50 | #define tlb_flush(tlb) flush_tlb_mm((tlb)->mm) | ||
| 51 | |||
| 52 | /* Get the generic bits... */ | ||
| 53 | #include <asm-generic/tlb.h> | ||
| 54 | |||
| 55 | #endif /* CONFIG_PPC_STD_MMU */ | ||
| 56 | |||
| 57 | #endif /* __PPC_TLB_H */ | ||
diff --git a/include/asm-ppc/tlbflush.h b/include/asm-ppc/tlbflush.h deleted file mode 100644 index 9afee4ffc835..000000000000 --- a/include/asm-ppc/tlbflush.h +++ /dev/null | |||
| @@ -1,115 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * include/asm-ppc/tlbflush.h | ||
| 3 | * | ||
| 4 | * This program is free software; you can redistribute it and/or | ||
| 5 | * modify it under the terms of the GNU General Public License | ||
| 6 | * as published by the Free Software Foundation; either version | ||
| 7 | * 2 of the License, or (at your option) any later version. | ||
| 8 | */ | ||
| 9 | #ifdef __KERNEL__ | ||
| 10 | #ifndef _PPC_TLBFLUSH_H | ||
| 11 | #define _PPC_TLBFLUSH_H | ||
| 12 | |||
| 13 | #include <linux/config.h> | ||
| 14 | #include <linux/mm.h> | ||
| 15 | |||
| 16 | extern void _tlbie(unsigned long address); | ||
| 17 | extern void _tlbia(void); | ||
| 18 | |||
| 19 | #if defined(CONFIG_4xx) | ||
| 20 | |||
| 21 | #ifndef CONFIG_44x | ||
| 22 | #define __tlbia() asm volatile ("sync; tlbia; isync" : : : "memory") | ||
| 23 | #else | ||
| 24 | #define __tlbia _tlbia | ||
| 25 | #endif | ||
| 26 | |||
| 27 | static inline void flush_tlb_mm(struct mm_struct *mm) | ||
| 28 | { __tlbia(); } | ||
| 29 | static inline void flush_tlb_page(struct vm_area_struct *vma, | ||
| 30 | unsigned long vmaddr) | ||
| 31 | { _tlbie(vmaddr); } | ||
| 32 | static inline void flush_tlb_page_nohash(struct vm_area_struct *vma, | ||
| 33 | unsigned long vmaddr) | ||
| 34 | { _tlbie(vmaddr); } | ||
| 35 | static inline void flush_tlb_range(struct vm_area_struct *vma, | ||
| 36 | unsigned long start, unsigned long end) | ||
| 37 | { __tlbia(); } | ||
| 38 | static inline void flush_tlb_kernel_range(unsigned long start, | ||
| 39 | unsigned long end) | ||
| 40 | { __tlbia(); } | ||
| 41 | |||
| 42 | #elif defined(CONFIG_FSL_BOOKE) | ||
| 43 | |||
| 44 | /* TODO: determine if flush_tlb_range & flush_tlb_kernel_range | ||
| 45 | * are best implemented as tlbia vs specific tlbie's */ | ||
| 46 | |||
| 47 | #define __tlbia() _tlbia() | ||
| 48 | |||
| 49 | static inline void flush_tlb_mm(struct mm_struct *mm) | ||
| 50 | { __tlbia(); } | ||
| 51 | static inline void flush_tlb_page(struct vm_area_struct *vma, | ||
| 52 | unsigned long vmaddr) | ||
| 53 | { _tlbie(vmaddr); } | ||
| 54 | static inline void flush_tlb_page_nohash(struct vm_area_struct *vma, | ||
| 55 | unsigned long vmaddr) | ||
| 56 | { _tlbie(vmaddr); } | ||
| 57 | static inline void flush_tlb_range(struct vm_area_struct *vma, | ||
| 58 | unsigned long start, unsigned long end) | ||
| 59 | { __tlbia(); } | ||
| 60 | static inline void flush_tlb_kernel_range(unsigned long start, | ||
| 61 | unsigned long end) | ||
| 62 | { __tlbia(); } | ||
| 63 | |||
| 64 | #elif defined(CONFIG_8xx) | ||
| 65 | #define __tlbia() asm volatile ("tlbia; sync" : : : "memory") | ||
| 66 | |||
| 67 | static inline void flush_tlb_mm(struct mm_struct *mm) | ||
| 68 | { __tlbia(); } | ||
| 69 | static inline void flush_tlb_page(struct vm_area_struct *vma, | ||
| 70 | unsigned long vmaddr) | ||
| 71 | { _tlbie(vmaddr); } | ||
| 72 | static inline void flush_tlb_page_nohash(struct vm_area_struct *vma, | ||
| 73 | unsigned long vmaddr) | ||
| 74 | { _tlbie(vmaddr); } | ||
| 75 | static inline void flush_tlb_range(struct vm_area_struct *vma, | ||
| 76 | unsigned long start, unsigned long end) | ||
| 77 | { __tlbia(); } | ||
| 78 | static inline void flush_tlb_kernel_range(unsigned long start, | ||
| 79 | unsigned long end) | ||
| 80 | { __tlbia(); } | ||
| 81 | |||
| 82 | #else /* 6xx, 7xx, 7xxx cpus */ | ||
| 83 | struct mm_struct; | ||
| 84 | struct vm_area_struct; | ||
| 85 | extern void flush_tlb_mm(struct mm_struct *mm); | ||
| 86 | extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr); | ||
| 87 | extern void flush_tlb_page_nohash(struct vm_area_struct *vma, unsigned long addr); | ||
| 88 | extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, | ||
| 89 | unsigned long end); | ||
| 90 | extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); | ||
| 91 | #endif | ||
| 92 | |||
| 93 | /* | ||
| 94 | * This is called in munmap when we have freed up some page-table | ||
| 95 | * pages. We don't need to do anything here, there's nothing special | ||
| 96 | * about our page-table pages. -- paulus | ||
| 97 | */ | ||
| 98 | static inline void flush_tlb_pgtables(struct mm_struct *mm, | ||
| 99 | unsigned long start, unsigned long end) | ||
| 100 | { | ||
| 101 | } | ||
| 102 | |||
| 103 | /* | ||
| 104 | * This gets called at the end of handling a page fault, when | ||
| 105 | * the kernel has put a new PTE into the page table for the process. | ||
| 106 | * We use it to ensure coherency between the i-cache and d-cache | ||
| 107 | * for the page which has just been mapped in. | ||
| 108 | * On machines which use an MMU hash table, we use this to put a | ||
| 109 | * corresponding HPTE into the hash table ahead of time, instead of | ||
| 110 | * waiting for the inevitable extra hash-table miss exception. | ||
| 111 | */ | ||
| 112 | extern void update_mmu_cache(struct vm_area_struct *, unsigned long, pte_t); | ||
| 113 | |||
| 114 | #endif /* _PPC_TLBFLUSH_H */ | ||
| 115 | #endif /*__KERNEL__ */ | ||
diff --git a/include/asm-ppc/uaccess.h b/include/asm-ppc/uaccess.h deleted file mode 100644 index 63f56224da8c..000000000000 --- a/include/asm-ppc/uaccess.h +++ /dev/null | |||
| @@ -1,393 +0,0 @@ | |||
| 1 | #ifdef __KERNEL__ | ||
| 2 | #ifndef _PPC_UACCESS_H | ||
| 3 | #define _PPC_UACCESS_H | ||
| 4 | |||
| 5 | #ifndef __ASSEMBLY__ | ||
| 6 | #include <linux/sched.h> | ||
| 7 | #include <linux/errno.h> | ||
| 8 | #include <asm/processor.h> | ||
| 9 | |||
| 10 | #define VERIFY_READ 0 | ||
| 11 | #define VERIFY_WRITE 1 | ||
| 12 | |||
| 13 | /* | ||
| 14 | * The fs value determines whether argument validity checking should be | ||
| 15 | * performed or not. If get_fs() == USER_DS, checking is performed, with | ||
| 16 | * get_fs() == KERNEL_DS, checking is bypassed. | ||
| 17 | * | ||
| 18 | * For historical reasons, these macros are grossly misnamed. | ||
| 19 | * | ||
| 20 | * The fs/ds values are now the highest legal address in the "segment". | ||
| 21 | * This simplifies the checking in the routines below. | ||
| 22 | */ | ||
| 23 | |||
| 24 | #define KERNEL_DS ((mm_segment_t) { ~0UL }) | ||
| 25 | #define USER_DS ((mm_segment_t) { TASK_SIZE - 1 }) | ||
| 26 | |||
| 27 | #define get_ds() (KERNEL_DS) | ||
| 28 | #define get_fs() (current->thread.fs) | ||
| 29 | #define set_fs(val) (current->thread.fs = (val)) | ||
| 30 | |||
| 31 | #define segment_eq(a,b) ((a).seg == (b).seg) | ||
| 32 | |||
| 33 | #define __access_ok(addr,size) \ | ||
| 34 | ((addr) <= current->thread.fs.seg \ | ||
| 35 | && ((size) == 0 || (size) - 1 <= current->thread.fs.seg - (addr))) | ||
| 36 | |||
| 37 | #define access_ok(type, addr, size) \ | ||
| 38 | (__chk_user_ptr(addr),__access_ok((unsigned long)(addr),(size))) | ||
| 39 | |||
| 40 | /* | ||
| 41 | * The exception table consists of pairs of addresses: the first is the | ||
| 42 | * address of an instruction that is allowed to fault, and the second is | ||
| 43 | * the address at which the program should continue. No registers are | ||
| 44 | * modified, so it is entirely up to the continuation code to figure out | ||
| 45 | * what to do. | ||
| 46 | * | ||
| 47 | * All the routines below use bits of fixup code that are out of line | ||
| 48 | * with the main instruction path. This means when everything is well, | ||
| 49 | * we don't even have to jump over them. Further, they do not intrude | ||
| 50 | * on our cache or tlb entries. | ||
| 51 | */ | ||
| 52 | |||
| 53 | struct exception_table_entry | ||
| 54 | { | ||
| 55 | unsigned long insn, fixup; | ||
| 56 | }; | ||
| 57 | |||
| 58 | /* | ||
| 59 | * These are the main single-value transfer routines. They automatically | ||
| 60 | * use the right size if we just have the right pointer type. | ||
| 61 | * | ||
| 62 | * This gets kind of ugly. We want to return _two_ values in "get_user()" | ||
| 63 | * and yet we don't want to do any pointers, because that is too much | ||
| 64 | * of a performance impact. Thus we have a few rather ugly macros here, | ||
| 65 | * and hide all the ugliness from the user. | ||
| 66 | * | ||
| 67 | * The "__xxx" versions of the user access functions are versions that | ||
| 68 | * do not verify the address space, that must have been done previously | ||
| 69 | * with a separate "access_ok()" call (this is used when we do multiple | ||
| 70 | * accesses to the same area of user memory). | ||
| 71 | * | ||
| 72 | * As we use the same address space for kernel and user data on the | ||
| 73 | * PowerPC, we can just do these as direct assignments. (Of course, the | ||
| 74 | * exception handling means that it's no longer "just"...) | ||
| 75 | * | ||
| 76 | * The "user64" versions of the user access functions are versions that | ||
| 77 | * allow access of 64-bit data. The "get_user" functions do not | ||
| 78 | * properly handle 64-bit data because the value gets down cast to a long. | ||
| 79 | * The "put_user" functions already handle 64-bit data properly but we add | ||
| 80 | * "user64" versions for completeness | ||
| 81 | */ | ||
| 82 | #define get_user(x,ptr) \ | ||
| 83 | __get_user_check((x),(ptr),sizeof(*(ptr))) | ||
| 84 | #define get_user64(x,ptr) \ | ||
| 85 | __get_user64_check((x),(ptr),sizeof(*(ptr))) | ||
| 86 | #define put_user(x,ptr) \ | ||
| 87 | __put_user_check((__typeof__(*(ptr)))(x),(ptr),sizeof(*(ptr))) | ||
| 88 | #define put_user64(x,ptr) put_user(x,ptr) | ||
| 89 | |||
| 90 | #define __get_user(x,ptr) \ | ||
| 91 | __get_user_nocheck((x),(ptr),sizeof(*(ptr))) | ||
| 92 | #define __get_user64(x,ptr) \ | ||
| 93 | __get_user64_nocheck((x),(ptr),sizeof(*(ptr))) | ||
| 94 | #define __put_user(x,ptr) \ | ||
| 95 | __put_user_nocheck((__typeof__(*(ptr)))(x),(ptr),sizeof(*(ptr))) | ||
| 96 | #define __put_user64(x,ptr) __put_user(x,ptr) | ||
| 97 | |||
| 98 | extern long __put_user_bad(void); | ||
| 99 | |||
| 100 | #define __put_user_nocheck(x,ptr,size) \ | ||
| 101 | ({ \ | ||
| 102 | long __pu_err; \ | ||
| 103 | __chk_user_ptr(ptr); \ | ||
| 104 | __put_user_size((x),(ptr),(size),__pu_err); \ | ||
| 105 | __pu_err; \ | ||
| 106 | }) | ||
| 107 | |||
| 108 | #define __put_user_check(x,ptr,size) \ | ||
| 109 | ({ \ | ||
| 110 | long __pu_err = -EFAULT; \ | ||
| 111 | __typeof__(*(ptr)) __user *__pu_addr = (ptr); \ | ||
| 112 | if (access_ok(VERIFY_WRITE,__pu_addr,size)) \ | ||
| 113 | __put_user_size((x),__pu_addr,(size),__pu_err); \ | ||
| 114 | __pu_err; \ | ||
| 115 | }) | ||
| 116 | |||
| 117 | #define __put_user_size(x,ptr,size,retval) \ | ||
| 118 | do { \ | ||
| 119 | retval = 0; \ | ||
| 120 | switch (size) { \ | ||
| 121 | case 1: \ | ||
| 122 | __put_user_asm(x, ptr, retval, "stb"); \ | ||
| 123 | break; \ | ||
| 124 | case 2: \ | ||
| 125 | __put_user_asm(x, ptr, retval, "sth"); \ | ||
| 126 | break; \ | ||
| 127 | case 4: \ | ||
| 128 | __put_user_asm(x, ptr, retval, "stw"); \ | ||
| 129 | break; \ | ||
| 130 | case 8: \ | ||
| 131 | __put_user_asm2(x, ptr, retval); \ | ||
| 132 | break; \ | ||
| 133 | default: \ | ||
| 134 | __put_user_bad(); \ | ||
| 135 | } \ | ||
| 136 | } while (0) | ||
| 137 | |||
| 138 | /* | ||
| 139 | * We don't tell gcc that we are accessing memory, but this is OK | ||
| 140 | * because we do not write to any memory gcc knows about, so there | ||
| 141 | * are no aliasing issues. | ||
| 142 | */ | ||
| 143 | #define __put_user_asm(x, addr, err, op) \ | ||
| 144 | __asm__ __volatile__( \ | ||
| 145 | "1: "op" %1,0(%2)\n" \ | ||
| 146 | "2:\n" \ | ||
| 147 | ".section .fixup,\"ax\"\n" \ | ||
| 148 | "3: li %0,%3\n" \ | ||
| 149 | " b 2b\n" \ | ||
| 150 | ".previous\n" \ | ||
| 151 | ".section __ex_table,\"a\"\n" \ | ||
| 152 | " .align 2\n" \ | ||
| 153 | " .long 1b,3b\n" \ | ||
| 154 | ".previous" \ | ||
| 155 | : "=r" (err) \ | ||
| 156 | : "r" (x), "b" (addr), "i" (-EFAULT), "0" (err)) | ||
| 157 | |||
| 158 | #define __put_user_asm2(x, addr, err) \ | ||
| 159 | __asm__ __volatile__( \ | ||
| 160 | "1: stw %1,0(%2)\n" \ | ||
| 161 | "2: stw %1+1,4(%2)\n" \ | ||
| 162 | "3:\n" \ | ||
| 163 | ".section .fixup,\"ax\"\n" \ | ||
| 164 | "4: li %0,%3\n" \ | ||
| 165 | " b 3b\n" \ | ||
| 166 | ".previous\n" \ | ||
| 167 | ".section __ex_table,\"a\"\n" \ | ||
| 168 | " .align 2\n" \ | ||
| 169 | " .long 1b,4b\n" \ | ||
| 170 | " .long 2b,4b\n" \ | ||
| 171 | ".previous" \ | ||
| 172 | : "=r" (err) \ | ||
| 173 | : "r" (x), "b" (addr), "i" (-EFAULT), "0" (err)) | ||
| 174 | |||
| 175 | #define __get_user_nocheck(x, ptr, size) \ | ||
| 176 | ({ \ | ||
| 177 | long __gu_err; \ | ||
| 178 | unsigned long __gu_val; \ | ||
| 179 | __chk_user_ptr(ptr); \ | ||
| 180 | __get_user_size(__gu_val, (ptr), (size), __gu_err); \ | ||
| 181 | (x) = (__typeof__(*(ptr)))__gu_val; \ | ||
| 182 | __gu_err; \ | ||
| 183 | }) | ||
| 184 | |||
| 185 | #define __get_user64_nocheck(x, ptr, size) \ | ||
| 186 | ({ \ | ||
| 187 | long __gu_err; \ | ||
| 188 | long long __gu_val; \ | ||
| 189 | __chk_user_ptr(ptr); \ | ||
| 190 | __get_user_size64(__gu_val, (ptr), (size), __gu_err); \ | ||
| 191 | (x) = (__typeof__(*(ptr)))__gu_val; \ | ||
| 192 | __gu_err; \ | ||
| 193 | }) | ||
| 194 | |||
| 195 | #define __get_user_check(x, ptr, size) \ | ||
| 196 | ({ \ | ||
| 197 | long __gu_err = -EFAULT; \ | ||
| 198 | unsigned long __gu_val = 0; \ | ||
| 199 | const __typeof__(*(ptr)) __user *__gu_addr = (ptr); \ | ||
| 200 | if (access_ok(VERIFY_READ, __gu_addr, (size))) \ | ||
| 201 | __get_user_size(__gu_val, __gu_addr, (size), __gu_err); \ | ||
| 202 | (x) = (__typeof__(*(ptr)))__gu_val; \ | ||
| 203 | __gu_err; \ | ||
| 204 | }) | ||
| 205 | |||
| 206 | #define __get_user64_check(x, ptr, size) \ | ||
| 207 | ({ \ | ||
| 208 | long __gu_err = -EFAULT; \ | ||
| 209 | long long __gu_val = 0; \ | ||
| 210 | const __typeof__(*(ptr)) __user *__gu_addr = (ptr); \ | ||
| 211 | if (access_ok(VERIFY_READ, __gu_addr, (size))) \ | ||
| 212 | __get_user_size64(__gu_val, __gu_addr, (size), __gu_err); \ | ||
| 213 | (x) = (__typeof__(*(ptr)))__gu_val; \ | ||
| 214 | __gu_err; \ | ||
| 215 | }) | ||
| 216 | |||
| 217 | extern long __get_user_bad(void); | ||
| 218 | |||
| 219 | #define __get_user_size(x, ptr, size, retval) \ | ||
| 220 | do { \ | ||
| 221 | retval = 0; \ | ||
| 222 | switch (size) { \ | ||
| 223 | case 1: \ | ||
| 224 | __get_user_asm(x, ptr, retval, "lbz"); \ | ||
| 225 | break; \ | ||
| 226 | case 2: \ | ||
| 227 | __get_user_asm(x, ptr, retval, "lhz"); \ | ||
| 228 | break; \ | ||
| 229 | case 4: \ | ||
| 230 | __get_user_asm(x, ptr, retval, "lwz"); \ | ||
| 231 | break; \ | ||
| 232 | default: \ | ||
| 233 | x = __get_user_bad(); \ | ||
| 234 | } \ | ||
| 235 | } while (0) | ||
| 236 | |||
| 237 | #define __get_user_size64(x, ptr, size, retval) \ | ||
| 238 | do { \ | ||
| 239 | retval = 0; \ | ||
| 240 | switch (size) { \ | ||
| 241 | case 1: \ | ||
| 242 | __get_user_asm(x, ptr, retval, "lbz"); \ | ||
| 243 | break; \ | ||
| 244 | case 2: \ | ||
| 245 | __get_user_asm(x, ptr, retval, "lhz"); \ | ||
| 246 | break; \ | ||
| 247 | case 4: \ | ||
| 248 | __get_user_asm(x, ptr, retval, "lwz"); \ | ||
| 249 | break; \ | ||
| 250 | case 8: \ | ||
| 251 | __get_user_asm2(x, ptr, retval); \ | ||
| 252 | break; \ | ||
| 253 | default: \ | ||
| 254 | x = __get_user_bad(); \ | ||
| 255 | } \ | ||
| 256 | } while (0) | ||
| 257 | |||
| 258 | #define __get_user_asm(x, addr, err, op) \ | ||
| 259 | __asm__ __volatile__( \ | ||
| 260 | "1: "op" %1,0(%2)\n" \ | ||
| 261 | "2:\n" \ | ||
| 262 | ".section .fixup,\"ax\"\n" \ | ||
| 263 | "3: li %0,%3\n" \ | ||
| 264 | " li %1,0\n" \ | ||
| 265 | " b 2b\n" \ | ||
| 266 | ".previous\n" \ | ||
| 267 | ".section __ex_table,\"a\"\n" \ | ||
| 268 | " .align 2\n" \ | ||
| 269 | " .long 1b,3b\n" \ | ||
| 270 | ".previous" \ | ||
| 271 | : "=r"(err), "=r"(x) \ | ||
| 272 | : "b"(addr), "i"(-EFAULT), "0"(err)) | ||
| 273 | |||
| 274 | #define __get_user_asm2(x, addr, err) \ | ||
| 275 | __asm__ __volatile__( \ | ||
| 276 | "1: lwz %1,0(%2)\n" \ | ||
| 277 | "2: lwz %1+1,4(%2)\n" \ | ||
| 278 | "3:\n" \ | ||
| 279 | ".section .fixup,\"ax\"\n" \ | ||
| 280 | "4: li %0,%3\n" \ | ||
| 281 | " li %1,0\n" \ | ||
| 282 | " li %1+1,0\n" \ | ||
| 283 | " b 3b\n" \ | ||
| 284 | ".previous\n" \ | ||
| 285 | ".section __ex_table,\"a\"\n" \ | ||
| 286 | " .align 2\n" \ | ||
| 287 | " .long 1b,4b\n" \ | ||
| 288 | " .long 2b,4b\n" \ | ||
| 289 | ".previous" \ | ||
| 290 | : "=r"(err), "=&r"(x) \ | ||
| 291 | : "b"(addr), "i"(-EFAULT), "0"(err)) | ||
| 292 | |||
| 293 | /* more complex routines */ | ||
| 294 | |||
| 295 | extern int __copy_tofrom_user(void __user *to, const void __user *from, | ||
| 296 | unsigned long size); | ||
| 297 | |||
| 298 | extern inline unsigned long | ||
| 299 | copy_from_user(void *to, const void __user *from, unsigned long n) | ||
| 300 | { | ||
| 301 | unsigned long over; | ||
| 302 | |||
| 303 | if (access_ok(VERIFY_READ, from, n)) | ||
| 304 | return __copy_tofrom_user((__force void __user *)to, from, n); | ||
| 305 | if ((unsigned long)from < TASK_SIZE) { | ||
| 306 | over = (unsigned long)from + n - TASK_SIZE; | ||
| 307 | return __copy_tofrom_user((__force void __user *)to, from, n - over) + over; | ||
| 308 | } | ||
| 309 | return n; | ||
| 310 | } | ||
| 311 | |||
| 312 | extern inline unsigned long | ||
| 313 | copy_to_user(void __user *to, const void *from, unsigned long n) | ||
| 314 | { | ||
| 315 | unsigned long over; | ||
| 316 | |||
| 317 | if (access_ok(VERIFY_WRITE, to, n)) | ||
| 318 | return __copy_tofrom_user(to, (__force void __user *) from, n); | ||
| 319 | if ((unsigned long)to < TASK_SIZE) { | ||
| 320 | over = (unsigned long)to + n - TASK_SIZE; | ||
| 321 | return __copy_tofrom_user(to, (__force void __user *) from, n - over) + over; | ||
| 322 | } | ||
| 323 | return n; | ||
| 324 | } | ||
| 325 | |||
| 326 | static inline unsigned long __copy_from_user(void *to, const void __user *from, unsigned long size) | ||
| 327 | { | ||
| 328 | return __copy_tofrom_user((__force void __user *)to, from, size); | ||
| 329 | } | ||
| 330 | |||
| 331 | static inline unsigned long __copy_to_user(void __user *to, const void *from, unsigned long size) | ||
| 332 | { | ||
| 333 | return __copy_tofrom_user(to, (__force void __user *)from, size); | ||
| 334 | } | ||
| 335 | |||
| 336 | #define __copy_to_user_inatomic __copy_to_user | ||
| 337 | #define __copy_from_user_inatomic __copy_from_user | ||
| 338 | |||
| 339 | extern unsigned long __clear_user(void __user *addr, unsigned long size); | ||
| 340 | |||
| 341 | extern inline unsigned long | ||
| 342 | clear_user(void __user *addr, unsigned long size) | ||
| 343 | { | ||
| 344 | if (access_ok(VERIFY_WRITE, addr, size)) | ||
| 345 | return __clear_user(addr, size); | ||
| 346 | if ((unsigned long)addr < TASK_SIZE) { | ||
| 347 | unsigned long over = (unsigned long)addr + size - TASK_SIZE; | ||
| 348 | return __clear_user(addr, size - over) + over; | ||
| 349 | } | ||
| 350 | return size; | ||
| 351 | } | ||
| 352 | |||
| 353 | extern int __strncpy_from_user(char *dst, const char __user *src, long count); | ||
| 354 | |||
| 355 | extern inline long | ||
| 356 | strncpy_from_user(char *dst, const char __user *src, long count) | ||
| 357 | { | ||
| 358 | if (access_ok(VERIFY_READ, src, 1)) | ||
| 359 | return __strncpy_from_user(dst, src, count); | ||
| 360 | return -EFAULT; | ||
| 361 | } | ||
| 362 | |||
| 363 | /* | ||
| 364 | * Return the size of a string (including the ending 0) | ||
| 365 | * | ||
| 366 | * Return 0 for error | ||
| 367 | */ | ||
| 368 | |||
| 369 | extern int __strnlen_user(const char __user *str, long len, unsigned long top); | ||
| 370 | |||
| 371 | /* | ||
| 372 | * Returns the length of the string at str (including the null byte), | ||
| 373 | * or 0 if we hit a page we can't access, | ||
| 374 | * or something > len if we didn't find a null byte. | ||
| 375 | * | ||
| 376 | * The `top' parameter to __strnlen_user is to make sure that | ||
| 377 | * we can never overflow from the user area into kernel space. | ||
| 378 | */ | ||
| 379 | extern __inline__ int strnlen_user(const char __user *str, long len) | ||
| 380 | { | ||
| 381 | unsigned long top = current->thread.fs.seg; | ||
| 382 | |||
| 383 | if ((unsigned long)str > top) | ||
| 384 | return 0; | ||
| 385 | return __strnlen_user(str, len, top); | ||
| 386 | } | ||
| 387 | |||
| 388 | #define strlen_user(str) strnlen_user((str), 0x7ffffffe) | ||
| 389 | |||
| 390 | #endif /* __ASSEMBLY__ */ | ||
| 391 | |||
| 392 | #endif /* _PPC_UACCESS_H */ | ||
| 393 | #endif /* __KERNEL__ */ | ||
diff --git a/include/asm-ppc/ucontext.h b/include/asm-ppc/ucontext.h deleted file mode 100644 index 664bc984d51f..000000000000 --- a/include/asm-ppc/ucontext.h +++ /dev/null | |||
| @@ -1,27 +0,0 @@ | |||
| 1 | #ifndef _ASMPPC_UCONTEXT_H | ||
| 2 | #define _ASMPPC_UCONTEXT_H | ||
| 3 | |||
| 4 | #include <asm/elf.h> | ||
| 5 | #include <asm/signal.h> | ||
| 6 | |||
| 7 | struct mcontext { | ||
| 8 | elf_gregset_t mc_gregs; | ||
| 9 | elf_fpregset_t mc_fregs; | ||
| 10 | unsigned long mc_pad[2]; | ||
| 11 | elf_vrregset_t mc_vregs __attribute__((__aligned__(16))); | ||
| 12 | }; | ||
| 13 | |||
| 14 | struct ucontext { | ||
| 15 | unsigned long uc_flags; | ||
| 16 | struct ucontext __user *uc_link; | ||
| 17 | stack_t uc_stack; | ||
| 18 | int uc_pad[7]; | ||
| 19 | struct mcontext __user *uc_regs;/* points to uc_mcontext field */ | ||
| 20 | sigset_t uc_sigmask; | ||
| 21 | /* glibc has 1024-bit signal masks, ours are 64-bit */ | ||
| 22 | int uc_maskext[30]; | ||
| 23 | int uc_pad2[3]; | ||
| 24 | struct mcontext uc_mcontext; | ||
| 25 | }; | ||
| 26 | |||
| 27 | #endif /* !_ASMPPC_UCONTEXT_H */ | ||
