diff options
| author | Vineet Gupta <vgupta@synopsys.com> | 2015-07-27 07:53:28 -0400 |
|---|---|---|
| committer | Vineet Gupta <vgupta@synopsys.com> | 2016-09-30 17:48:17 -0400 |
| commit | ce6365270ecd1216b48fb1440978e454ae0144de (patch) | |
| tree | 2e46017d408be15aa5834675cd11c06122fc353c /arch/arc/include/asm | |
| parent | 26c01c49d559268527d78f45a6818fae0c204a45 (diff) | |
ARCv2: Implement atomic64 based on LLOCKD/SCONDD instructions
ARCv2 ISA provides 64-bit exclusive load/stores so use them to implement
the 64-bit atomics and elide the spinlock based generic 64-bit atomics
boot tested with atomic64 self-test (and GOD bless the person who wrote
them, I realized my inline assmebly is sloppy as hell)
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Will Deacon <will.deacon@arm.com>
Cc: linux-snps-arc@lists.infradead.org
Cc: linux-kernel@vger.kernel.org
Signed-off-by: Vineet Gupta <vgupta@synopsys.com>
Diffstat (limited to 'arch/arc/include/asm')
| -rw-r--r-- | arch/arc/include/asm/atomic.h | 261 |
1 files changed, 259 insertions, 2 deletions
diff --git a/arch/arc/include/asm/atomic.h b/arch/arc/include/asm/atomic.h index 4e3c1b6b0806..d0e222e3776b 100644 --- a/arch/arc/include/asm/atomic.h +++ b/arch/arc/include/asm/atomic.h | |||
| @@ -20,6 +20,7 @@ | |||
| 20 | #ifndef CONFIG_ARC_PLAT_EZNPS | 20 | #ifndef CONFIG_ARC_PLAT_EZNPS |
| 21 | 21 | ||
| 22 | #define atomic_read(v) READ_ONCE((v)->counter) | 22 | #define atomic_read(v) READ_ONCE((v)->counter) |
| 23 | #define ATOMIC_INIT(i) { (i) } | ||
| 23 | 24 | ||
| 24 | #ifdef CONFIG_ARC_HAS_LLSC | 25 | #ifdef CONFIG_ARC_HAS_LLSC |
| 25 | 26 | ||
| @@ -343,10 +344,266 @@ ATOMIC_OPS(xor, ^=, CTOP_INST_AXOR_DI_R2_R2_R3) | |||
| 343 | 344 | ||
| 344 | #define atomic_add_negative(i, v) (atomic_add_return(i, v) < 0) | 345 | #define atomic_add_negative(i, v) (atomic_add_return(i, v) < 0) |
| 345 | 346 | ||
| 346 | #define ATOMIC_INIT(i) { (i) } | 347 | |
| 348 | #ifdef CONFIG_GENERIC_ATOMIC64 | ||
| 347 | 349 | ||
| 348 | #include <asm-generic/atomic64.h> | 350 | #include <asm-generic/atomic64.h> |
| 349 | 351 | ||
| 350 | #endif | 352 | #else /* Kconfig ensures this is only enabled with needed h/w assist */ |
| 353 | |||
| 354 | /* | ||
| 355 | * ARCv2 supports 64-bit exclusive load (LLOCKD) / store (SCONDD) | ||
| 356 | * - The address HAS to be 64-bit aligned | ||
| 357 | * - There are 2 semantics involved here: | ||
| 358 | * = exclusive implies no interim update between load/store to same addr | ||
| 359 | * = both words are observed/updated together: this is guaranteed even | ||
| 360 | * for regular 64-bit load (LDD) / store (STD). Thus atomic64_set() | ||
| 361 | * is NOT required to use LLOCKD+SCONDD, STD suffices | ||
| 362 | */ | ||
| 363 | |||
| 364 | typedef struct { | ||
| 365 | aligned_u64 counter; | ||
| 366 | } atomic64_t; | ||
| 367 | |||
| 368 | #define ATOMIC64_INIT(a) { (a) } | ||
| 369 | |||
| 370 | static inline long long atomic64_read(const atomic64_t *v) | ||
| 371 | { | ||
| 372 | unsigned long long val; | ||
| 373 | |||
| 374 | __asm__ __volatile__( | ||
| 375 | " ldd %0, [%1] \n" | ||
| 376 | : "=r"(val) | ||
| 377 | : "r"(&v->counter)); | ||
| 378 | |||
| 379 | return val; | ||
| 380 | } | ||
| 381 | |||
| 382 | static inline void atomic64_set(atomic64_t *v, long long a) | ||
| 383 | { | ||
| 384 | /* | ||
| 385 | * This could have been a simple assignment in "C" but would need | ||
| 386 | * explicit volatile. Otherwise gcc optimizers could elide the store | ||
| 387 | * which borked atomic64 self-test | ||
| 388 | * In the inline asm version, memory clobber needed for exact same | ||
| 389 | * reason, to tell gcc about the store. | ||
| 390 | * | ||
| 391 | * This however is not needed for sibling atomic64_add() etc since both | ||
| 392 | * load/store are explicitly done in inline asm. As long as API is used | ||
| 393 | * for each access, gcc has no way to optimize away any load/store | ||
| 394 | */ | ||
| 395 | __asm__ __volatile__( | ||
| 396 | " std %0, [%1] \n" | ||
| 397 | : | ||
| 398 | : "r"(a), "r"(&v->counter) | ||
| 399 | : "memory"); | ||
| 400 | } | ||
| 401 | |||
| 402 | #define ATOMIC64_OP(op, op1, op2) \ | ||
| 403 | static inline void atomic64_##op(long long a, atomic64_t *v) \ | ||
| 404 | { \ | ||
| 405 | unsigned long long val; \ | ||
| 406 | \ | ||
| 407 | __asm__ __volatile__( \ | ||
| 408 | "1: \n" \ | ||
| 409 | " llockd %0, [%1] \n" \ | ||
| 410 | " " #op1 " %L0, %L0, %L2 \n" \ | ||
| 411 | " " #op2 " %H0, %H0, %H2 \n" \ | ||
| 412 | " scondd %0, [%1] \n" \ | ||
| 413 | " bnz 1b \n" \ | ||
| 414 | : "=&r"(val) \ | ||
| 415 | : "r"(&v->counter), "ir"(a) \ | ||
| 416 | : "cc"); \ | ||
| 417 | } \ | ||
| 418 | |||
| 419 | #define ATOMIC64_OP_RETURN(op, op1, op2) \ | ||
| 420 | static inline long long atomic64_##op##_return(long long a, atomic64_t *v) \ | ||
| 421 | { \ | ||
| 422 | unsigned long long val; \ | ||
| 423 | \ | ||
| 424 | smp_mb(); \ | ||
| 425 | \ | ||
| 426 | __asm__ __volatile__( \ | ||
| 427 | "1: \n" \ | ||
| 428 | " llockd %0, [%1] \n" \ | ||
| 429 | " " #op1 " %L0, %L0, %L2 \n" \ | ||
| 430 | " " #op2 " %H0, %H0, %H2 \n" \ | ||
| 431 | " scondd %0, [%1] \n" \ | ||
| 432 | " bnz 1b \n" \ | ||
| 433 | : [val] "=&r"(val) \ | ||
| 434 | : "r"(&v->counter), "ir"(a) \ | ||
| 435 | : "cc"); /* memory clobber comes from smp_mb() */ \ | ||
| 436 | \ | ||
| 437 | smp_mb(); \ | ||
| 438 | \ | ||
| 439 | return val; \ | ||
| 440 | } | ||
| 441 | |||
| 442 | #define ATOMIC64_FETCH_OP(op, op1, op2) \ | ||
| 443 | static inline long long atomic64_fetch_##op(long long a, atomic64_t *v) \ | ||
| 444 | { \ | ||
| 445 | unsigned long long val, orig; \ | ||
| 446 | \ | ||
| 447 | smp_mb(); \ | ||
| 448 | \ | ||
| 449 | __asm__ __volatile__( \ | ||
| 450 | "1: \n" \ | ||
| 451 | " llockd %0, [%2] \n" \ | ||
| 452 | " " #op1 " %L1, %L0, %L3 \n" \ | ||
| 453 | " " #op2 " %H1, %H0, %H3 \n" \ | ||
| 454 | " scondd %1, [%2] \n" \ | ||
| 455 | " bnz 1b \n" \ | ||
| 456 | : "=&r"(orig), "=&r"(val) \ | ||
| 457 | : "r"(&v->counter), "ir"(a) \ | ||
| 458 | : "cc"); /* memory clobber comes from smp_mb() */ \ | ||
| 459 | \ | ||
| 460 | smp_mb(); \ | ||
| 461 | \ | ||
| 462 | return orig; \ | ||
| 463 | } | ||
| 464 | |||
| 465 | #define ATOMIC64_OPS(op, op1, op2) \ | ||
| 466 | ATOMIC64_OP(op, op1, op2) \ | ||
| 467 | ATOMIC64_OP_RETURN(op, op1, op2) \ | ||
| 468 | ATOMIC64_FETCH_OP(op, op1, op2) | ||
| 469 | |||
| 470 | #define atomic64_andnot atomic64_andnot | ||
| 471 | |||
| 472 | ATOMIC64_OPS(add, add.f, adc) | ||
| 473 | ATOMIC64_OPS(sub, sub.f, sbc) | ||
| 474 | ATOMIC64_OPS(and, and, and) | ||
| 475 | ATOMIC64_OPS(andnot, bic, bic) | ||
| 476 | ATOMIC64_OPS(or, or, or) | ||
| 477 | ATOMIC64_OPS(xor, xor, xor) | ||
| 478 | |||
| 479 | #undef ATOMIC64_OPS | ||
| 480 | #undef ATOMIC64_FETCH_OP | ||
| 481 | #undef ATOMIC64_OP_RETURN | ||
| 482 | #undef ATOMIC64_OP | ||
| 483 | |||
| 484 | static inline long long | ||
| 485 | atomic64_cmpxchg(atomic64_t *ptr, long long expected, long long new) | ||
| 486 | { | ||
| 487 | long long prev; | ||
| 488 | |||
| 489 | smp_mb(); | ||
| 490 | |||
| 491 | __asm__ __volatile__( | ||
| 492 | "1: llockd %0, [%1] \n" | ||
| 493 | " brne %L0, %L2, 2f \n" | ||
| 494 | " brne %H0, %H2, 2f \n" | ||
| 495 | " scondd %3, [%1] \n" | ||
| 496 | " bnz 1b \n" | ||
| 497 | "2: \n" | ||
| 498 | : "=&r"(prev) | ||
| 499 | : "r"(ptr), "ir"(expected), "r"(new) | ||
| 500 | : "cc"); /* memory clobber comes from smp_mb() */ | ||
| 501 | |||
| 502 | smp_mb(); | ||
| 503 | |||
| 504 | return prev; | ||
| 505 | } | ||
| 506 | |||
| 507 | static inline long long atomic64_xchg(atomic64_t *ptr, long long new) | ||
| 508 | { | ||
| 509 | long long prev; | ||
| 510 | |||
| 511 | smp_mb(); | ||
| 512 | |||
| 513 | __asm__ __volatile__( | ||
| 514 | "1: llockd %0, [%1] \n" | ||
| 515 | " scondd %2, [%1] \n" | ||
| 516 | " bnz 1b \n" | ||
| 517 | "2: \n" | ||
| 518 | : "=&r"(prev) | ||
| 519 | : "r"(ptr), "r"(new) | ||
| 520 | : "cc"); /* memory clobber comes from smp_mb() */ | ||
| 521 | |||
| 522 | smp_mb(); | ||
| 523 | |||
| 524 | return prev; | ||
| 525 | } | ||
| 526 | |||
| 527 | /** | ||
| 528 | * atomic64_dec_if_positive - decrement by 1 if old value positive | ||
| 529 | * @v: pointer of type atomic64_t | ||
| 530 | * | ||
| 531 | * The function returns the old value of *v minus 1, even if | ||
| 532 | * the atomic variable, v, was not decremented. | ||
| 533 | */ | ||
| 534 | |||
| 535 | static inline long long atomic64_dec_if_positive(atomic64_t *v) | ||
| 536 | { | ||
| 537 | long long val; | ||
| 538 | |||
| 539 | smp_mb(); | ||
| 540 | |||
| 541 | __asm__ __volatile__( | ||
| 542 | "1: llockd %0, [%1] \n" | ||
| 543 | " sub.f %L0, %L0, 1 # w0 - 1, set C on borrow\n" | ||
| 544 | " sub.c %H0, %H0, 1 # if C set, w1 - 1\n" | ||
| 545 | " brlt %H0, 0, 2f \n" | ||
| 546 | " scondd %0, [%1] \n" | ||
| 547 | " bnz 1b \n" | ||
| 548 | "2: \n" | ||
| 549 | : "=&r"(val) | ||
| 550 | : "r"(&v->counter) | ||
| 551 | : "cc"); /* memory clobber comes from smp_mb() */ | ||
| 552 | |||
| 553 | smp_mb(); | ||
| 554 | |||
| 555 | return val; | ||
| 556 | } | ||
| 557 | |||
| 558 | /** | ||
| 559 | * atomic64_add_unless - add unless the number is a given value | ||
| 560 | * @v: pointer of type atomic64_t | ||
| 561 | * @a: the amount to add to v... | ||
| 562 | * @u: ...unless v is equal to u. | ||
| 563 | * | ||
| 564 | * if (v != u) { v += a; ret = 1} else {ret = 0} | ||
| 565 | * Returns 1 iff @v was not @u (i.e. if add actually happened) | ||
| 566 | */ | ||
| 567 | static inline int atomic64_add_unless(atomic64_t *v, long long a, long long u) | ||
| 568 | { | ||
| 569 | long long val; | ||
| 570 | int op_done; | ||
| 571 | |||
| 572 | smp_mb(); | ||
| 573 | |||
| 574 | __asm__ __volatile__( | ||
| 575 | "1: llockd %0, [%2] \n" | ||
| 576 | " mov %1, 1 \n" | ||
| 577 | " brne %L0, %L4, 2f # continue to add since v != u \n" | ||
| 578 | " breq.d %H0, %H4, 3f # return since v == u \n" | ||
| 579 | " mov %1, 0 \n" | ||
| 580 | "2: \n" | ||
| 581 | " add.f %L0, %L0, %L3 \n" | ||
| 582 | " adc %H0, %H0, %H3 \n" | ||
| 583 | " scondd %0, [%2] \n" | ||
| 584 | " bnz 1b \n" | ||
| 585 | "3: \n" | ||
| 586 | : "=&r"(val), "=&r" (op_done) | ||
| 587 | : "r"(&v->counter), "r"(a), "r"(u) | ||
| 588 | : "cc"); /* memory clobber comes from smp_mb() */ | ||
| 589 | |||
| 590 | smp_mb(); | ||
| 591 | |||
| 592 | return op_done; | ||
| 593 | } | ||
| 594 | |||
| 595 | #define atomic64_add_negative(a, v) (atomic64_add_return((a), (v)) < 0) | ||
| 596 | #define atomic64_inc(v) atomic64_add(1LL, (v)) | ||
| 597 | #define atomic64_inc_return(v) atomic64_add_return(1LL, (v)) | ||
| 598 | #define atomic64_inc_and_test(v) (atomic64_inc_return(v) == 0) | ||
| 599 | #define atomic64_sub_and_test(a, v) (atomic64_sub_return((a), (v)) == 0) | ||
| 600 | #define atomic64_dec(v) atomic64_sub(1LL, (v)) | ||
| 601 | #define atomic64_dec_return(v) atomic64_sub_return(1LL, (v)) | ||
| 602 | #define atomic64_dec_and_test(v) (atomic64_dec_return((v)) == 0) | ||
| 603 | #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1LL, 0LL) | ||
| 604 | |||
| 605 | #endif /* !CONFIG_GENERIC_ATOMIC64 */ | ||
| 606 | |||
| 607 | #endif /* !__ASSEMBLY__ */ | ||
| 351 | 608 | ||
| 352 | #endif | 609 | #endif |
