diff options
| author | Thomas Gleixner <tglx@linutronix.de> | 2009-03-23 09:50:03 -0400 |
|---|---|---|
| committer | Thomas Gleixner <tglx@linutronix.de> | 2009-03-23 16:20:20 -0400 |
| commit | 80c5520811d3805adcb15c570ea5e2d489fa5d0b (patch) | |
| tree | ae797a7f4af39f80e77526533d06ac23b439f0ab /arch/x86/include/asm/uaccess.h | |
| parent | b3e3b302cf6dc8d60b67f0e84d1fa5648889c038 (diff) | |
| parent | 8c083f081d0014057901c68a0a3e0f8ca7ac8d23 (diff) | |
Merge branch 'cpus4096' into irq/threaded
Conflicts:
arch/parisc/kernel/irq.c
kernel/irq/handle.c
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Diffstat (limited to 'arch/x86/include/asm/uaccess.h')
| -rw-r--r-- | arch/x86/include/asm/uaccess.h | 138 |
1 files changed, 130 insertions, 8 deletions
diff --git a/arch/x86/include/asm/uaccess.h b/arch/x86/include/asm/uaccess.h index 4340055b7559..b685ece89d5c 100644 --- a/arch/x86/include/asm/uaccess.h +++ b/arch/x86/include/asm/uaccess.h | |||
| @@ -121,7 +121,7 @@ extern int __get_user_bad(void); | |||
| 121 | 121 | ||
| 122 | #define __get_user_x(size, ret, x, ptr) \ | 122 | #define __get_user_x(size, ret, x, ptr) \ |
| 123 | asm volatile("call __get_user_" #size \ | 123 | asm volatile("call __get_user_" #size \ |
| 124 | : "=a" (ret),"=d" (x) \ | 124 | : "=a" (ret), "=d" (x) \ |
| 125 | : "0" (ptr)) \ | 125 | : "0" (ptr)) \ |
| 126 | 126 | ||
| 127 | /* Careful: we have to cast the result to the type of the pointer | 127 | /* Careful: we have to cast the result to the type of the pointer |
| @@ -181,12 +181,12 @@ extern int __get_user_bad(void); | |||
| 181 | 181 | ||
| 182 | #define __put_user_x(size, x, ptr, __ret_pu) \ | 182 | #define __put_user_x(size, x, ptr, __ret_pu) \ |
| 183 | asm volatile("call __put_user_" #size : "=a" (__ret_pu) \ | 183 | asm volatile("call __put_user_" #size : "=a" (__ret_pu) \ |
| 184 | :"0" ((typeof(*(ptr)))(x)), "c" (ptr) : "ebx") | 184 | : "0" ((typeof(*(ptr)))(x)), "c" (ptr) : "ebx") |
| 185 | 185 | ||
| 186 | 186 | ||
| 187 | 187 | ||
| 188 | #ifdef CONFIG_X86_32 | 188 | #ifdef CONFIG_X86_32 |
| 189 | #define __put_user_u64(x, addr, err) \ | 189 | #define __put_user_asm_u64(x, addr, err, errret) \ |
| 190 | asm volatile("1: movl %%eax,0(%2)\n" \ | 190 | asm volatile("1: movl %%eax,0(%2)\n" \ |
| 191 | "2: movl %%edx,4(%2)\n" \ | 191 | "2: movl %%edx,4(%2)\n" \ |
| 192 | "3:\n" \ | 192 | "3:\n" \ |
| @@ -197,14 +197,24 @@ extern int __get_user_bad(void); | |||
| 197 | _ASM_EXTABLE(1b, 4b) \ | 197 | _ASM_EXTABLE(1b, 4b) \ |
| 198 | _ASM_EXTABLE(2b, 4b) \ | 198 | _ASM_EXTABLE(2b, 4b) \ |
| 199 | : "=r" (err) \ | 199 | : "=r" (err) \ |
| 200 | : "A" (x), "r" (addr), "i" (-EFAULT), "0" (err)) | 200 | : "A" (x), "r" (addr), "i" (errret), "0" (err)) |
| 201 | |||
| 202 | #define __put_user_asm_ex_u64(x, addr) \ | ||
| 203 | asm volatile("1: movl %%eax,0(%1)\n" \ | ||
| 204 | "2: movl %%edx,4(%1)\n" \ | ||
| 205 | "3:\n" \ | ||
| 206 | _ASM_EXTABLE(1b, 2b - 1b) \ | ||
| 207 | _ASM_EXTABLE(2b, 3b - 2b) \ | ||
| 208 | : : "A" (x), "r" (addr)) | ||
| 201 | 209 | ||
| 202 | #define __put_user_x8(x, ptr, __ret_pu) \ | 210 | #define __put_user_x8(x, ptr, __ret_pu) \ |
| 203 | asm volatile("call __put_user_8" : "=a" (__ret_pu) \ | 211 | asm volatile("call __put_user_8" : "=a" (__ret_pu) \ |
| 204 | : "A" ((typeof(*(ptr)))(x)), "c" (ptr) : "ebx") | 212 | : "A" ((typeof(*(ptr)))(x)), "c" (ptr) : "ebx") |
| 205 | #else | 213 | #else |
| 206 | #define __put_user_u64(x, ptr, retval) \ | 214 | #define __put_user_asm_u64(x, ptr, retval, errret) \ |
| 207 | __put_user_asm(x, ptr, retval, "q", "", "Zr", -EFAULT) | 215 | __put_user_asm(x, ptr, retval, "q", "", "Zr", errret) |
| 216 | #define __put_user_asm_ex_u64(x, addr) \ | ||
| 217 | __put_user_asm_ex(x, addr, "q", "", "Zr") | ||
| 208 | #define __put_user_x8(x, ptr, __ret_pu) __put_user_x(8, x, ptr, __ret_pu) | 218 | #define __put_user_x8(x, ptr, __ret_pu) __put_user_x(8, x, ptr, __ret_pu) |
| 209 | #endif | 219 | #endif |
| 210 | 220 | ||
| @@ -276,10 +286,32 @@ do { \ | |||
| 276 | __put_user_asm(x, ptr, retval, "w", "w", "ir", errret); \ | 286 | __put_user_asm(x, ptr, retval, "w", "w", "ir", errret); \ |
| 277 | break; \ | 287 | break; \ |
| 278 | case 4: \ | 288 | case 4: \ |
| 279 | __put_user_asm(x, ptr, retval, "l", "k", "ir", errret);\ | 289 | __put_user_asm(x, ptr, retval, "l", "k", "ir", errret); \ |
| 280 | break; \ | 290 | break; \ |
| 281 | case 8: \ | 291 | case 8: \ |
| 282 | __put_user_u64((__typeof__(*ptr))(x), ptr, retval); \ | 292 | __put_user_asm_u64((__typeof__(*ptr))(x), ptr, retval, \ |
| 293 | errret); \ | ||
| 294 | break; \ | ||
| 295 | default: \ | ||
| 296 | __put_user_bad(); \ | ||
| 297 | } \ | ||
| 298 | } while (0) | ||
| 299 | |||
| 300 | #define __put_user_size_ex(x, ptr, size) \ | ||
| 301 | do { \ | ||
| 302 | __chk_user_ptr(ptr); \ | ||
| 303 | switch (size) { \ | ||
| 304 | case 1: \ | ||
| 305 | __put_user_asm_ex(x, ptr, "b", "b", "iq"); \ | ||
| 306 | break; \ | ||
| 307 | case 2: \ | ||
| 308 | __put_user_asm_ex(x, ptr, "w", "w", "ir"); \ | ||
| 309 | break; \ | ||
| 310 | case 4: \ | ||
| 311 | __put_user_asm_ex(x, ptr, "l", "k", "ir"); \ | ||
| 312 | break; \ | ||
| 313 | case 8: \ | ||
| 314 | __put_user_asm_ex_u64((__typeof__(*ptr))(x), ptr); \ | ||
| 283 | break; \ | 315 | break; \ |
| 284 | default: \ | 316 | default: \ |
| 285 | __put_user_bad(); \ | 317 | __put_user_bad(); \ |
| @@ -311,9 +343,12 @@ do { \ | |||
| 311 | 343 | ||
| 312 | #ifdef CONFIG_X86_32 | 344 | #ifdef CONFIG_X86_32 |
| 313 | #define __get_user_asm_u64(x, ptr, retval, errret) (x) = __get_user_bad() | 345 | #define __get_user_asm_u64(x, ptr, retval, errret) (x) = __get_user_bad() |
| 346 | #define __get_user_asm_ex_u64(x, ptr) (x) = __get_user_bad() | ||
| 314 | #else | 347 | #else |
| 315 | #define __get_user_asm_u64(x, ptr, retval, errret) \ | 348 | #define __get_user_asm_u64(x, ptr, retval, errret) \ |
| 316 | __get_user_asm(x, ptr, retval, "q", "", "=r", errret) | 349 | __get_user_asm(x, ptr, retval, "q", "", "=r", errret) |
| 350 | #define __get_user_asm_ex_u64(x, ptr) \ | ||
| 351 | __get_user_asm_ex(x, ptr, "q", "", "=r") | ||
| 317 | #endif | 352 | #endif |
| 318 | 353 | ||
| 319 | #define __get_user_size(x, ptr, size, retval, errret) \ | 354 | #define __get_user_size(x, ptr, size, retval, errret) \ |
| @@ -350,6 +385,33 @@ do { \ | |||
| 350 | : "=r" (err), ltype(x) \ | 385 | : "=r" (err), ltype(x) \ |
| 351 | : "m" (__m(addr)), "i" (errret), "0" (err)) | 386 | : "m" (__m(addr)), "i" (errret), "0" (err)) |
| 352 | 387 | ||
| 388 | #define __get_user_size_ex(x, ptr, size) \ | ||
| 389 | do { \ | ||
| 390 | __chk_user_ptr(ptr); \ | ||
| 391 | switch (size) { \ | ||
| 392 | case 1: \ | ||
| 393 | __get_user_asm_ex(x, ptr, "b", "b", "=q"); \ | ||
| 394 | break; \ | ||
| 395 | case 2: \ | ||
| 396 | __get_user_asm_ex(x, ptr, "w", "w", "=r"); \ | ||
| 397 | break; \ | ||
| 398 | case 4: \ | ||
| 399 | __get_user_asm_ex(x, ptr, "l", "k", "=r"); \ | ||
| 400 | break; \ | ||
| 401 | case 8: \ | ||
| 402 | __get_user_asm_ex_u64(x, ptr); \ | ||
| 403 | break; \ | ||
| 404 | default: \ | ||
| 405 | (x) = __get_user_bad(); \ | ||
| 406 | } \ | ||
| 407 | } while (0) | ||
| 408 | |||
| 409 | #define __get_user_asm_ex(x, addr, itype, rtype, ltype) \ | ||
| 410 | asm volatile("1: mov"itype" %1,%"rtype"0\n" \ | ||
| 411 | "2:\n" \ | ||
| 412 | _ASM_EXTABLE(1b, 2b - 1b) \ | ||
| 413 | : ltype(x) : "m" (__m(addr))) | ||
| 414 | |||
| 353 | #define __put_user_nocheck(x, ptr, size) \ | 415 | #define __put_user_nocheck(x, ptr, size) \ |
| 354 | ({ \ | 416 | ({ \ |
| 355 | int __pu_err; \ | 417 | int __pu_err; \ |
| @@ -385,6 +447,26 @@ struct __large_struct { unsigned long buf[100]; }; | |||
| 385 | _ASM_EXTABLE(1b, 3b) \ | 447 | _ASM_EXTABLE(1b, 3b) \ |
| 386 | : "=r"(err) \ | 448 | : "=r"(err) \ |
| 387 | : ltype(x), "m" (__m(addr)), "i" (errret), "0" (err)) | 449 | : ltype(x), "m" (__m(addr)), "i" (errret), "0" (err)) |
| 450 | |||
| 451 | #define __put_user_asm_ex(x, addr, itype, rtype, ltype) \ | ||
| 452 | asm volatile("1: mov"itype" %"rtype"0,%1\n" \ | ||
| 453 | "2:\n" \ | ||
| 454 | _ASM_EXTABLE(1b, 2b - 1b) \ | ||
| 455 | : : ltype(x), "m" (__m(addr))) | ||
| 456 | |||
| 457 | /* | ||
| 458 | * uaccess_try and catch | ||
| 459 | */ | ||
| 460 | #define uaccess_try do { \ | ||
| 461 | int prev_err = current_thread_info()->uaccess_err; \ | ||
| 462 | current_thread_info()->uaccess_err = 0; \ | ||
| 463 | barrier(); | ||
| 464 | |||
| 465 | #define uaccess_catch(err) \ | ||
| 466 | (err) |= current_thread_info()->uaccess_err; \ | ||
| 467 | current_thread_info()->uaccess_err = prev_err; \ | ||
| 468 | } while (0) | ||
| 469 | |||
| 388 | /** | 470 | /** |
| 389 | * __get_user: - Get a simple variable from user space, with less checking. | 471 | * __get_user: - Get a simple variable from user space, with less checking. |
| 390 | * @x: Variable to store result. | 472 | * @x: Variable to store result. |
| @@ -408,6 +490,7 @@ struct __large_struct { unsigned long buf[100]; }; | |||
| 408 | 490 | ||
| 409 | #define __get_user(x, ptr) \ | 491 | #define __get_user(x, ptr) \ |
| 410 | __get_user_nocheck((x), (ptr), sizeof(*(ptr))) | 492 | __get_user_nocheck((x), (ptr), sizeof(*(ptr))) |
| 493 | |||
| 411 | /** | 494 | /** |
| 412 | * __put_user: - Write a simple value into user space, with less checking. | 495 | * __put_user: - Write a simple value into user space, with less checking. |
| 413 | * @x: Value to copy to user space. | 496 | * @x: Value to copy to user space. |
| @@ -435,6 +518,45 @@ struct __large_struct { unsigned long buf[100]; }; | |||
| 435 | #define __put_user_unaligned __put_user | 518 | #define __put_user_unaligned __put_user |
| 436 | 519 | ||
| 437 | /* | 520 | /* |
| 521 | * {get|put}_user_try and catch | ||
| 522 | * | ||
| 523 | * get_user_try { | ||
| 524 | * get_user_ex(...); | ||
| 525 | * } get_user_catch(err) | ||
| 526 | */ | ||
| 527 | #define get_user_try uaccess_try | ||
| 528 | #define get_user_catch(err) uaccess_catch(err) | ||
| 529 | |||
| 530 | #define get_user_ex(x, ptr) do { \ | ||
| 531 | unsigned long __gue_val; \ | ||
| 532 | __get_user_size_ex((__gue_val), (ptr), (sizeof(*(ptr)))); \ | ||
| 533 | (x) = (__force __typeof__(*(ptr)))__gue_val; \ | ||
| 534 | } while (0) | ||
| 535 | |||
| 536 | #ifdef CONFIG_X86_WP_WORKS_OK | ||
| 537 | |||
| 538 | #define put_user_try uaccess_try | ||
| 539 | #define put_user_catch(err) uaccess_catch(err) | ||
| 540 | |||
| 541 | #define put_user_ex(x, ptr) \ | ||
| 542 | __put_user_size_ex((__typeof__(*(ptr)))(x), (ptr), sizeof(*(ptr))) | ||
| 543 | |||
| 544 | #else /* !CONFIG_X86_WP_WORKS_OK */ | ||
| 545 | |||
| 546 | #define put_user_try do { \ | ||
| 547 | int __uaccess_err = 0; | ||
| 548 | |||
| 549 | #define put_user_catch(err) \ | ||
| 550 | (err) |= __uaccess_err; \ | ||
| 551 | } while (0) | ||
| 552 | |||
| 553 | #define put_user_ex(x, ptr) do { \ | ||
| 554 | __uaccess_err |= __put_user(x, ptr); \ | ||
| 555 | } while (0) | ||
| 556 | |||
| 557 | #endif /* CONFIG_X86_WP_WORKS_OK */ | ||
| 558 | |||
| 559 | /* | ||
| 438 | * movsl can be slow when source and dest are not both 8-byte aligned | 560 | * movsl can be slow when source and dest are not both 8-byte aligned |
| 439 | */ | 561 | */ |
| 440 | #ifdef CONFIG_X86_INTEL_USERCOPY | 562 | #ifdef CONFIG_X86_INTEL_USERCOPY |
