diff options
Diffstat (limited to 'arch/x86/kernel')
-rw-r--r-- | arch/x86/kernel/tsc.c | 18 |
1 files changed, 16 insertions, 2 deletions
diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c index 1e6244202612..24249a5360b6 100644 --- a/arch/x86/kernel/tsc.c +++ b/arch/x86/kernel/tsc.c | |||
@@ -767,6 +767,7 @@ static cycle_t read_tsc(struct clocksource *cs) | |||
767 | static cycle_t __vsyscall_fn vread_tsc(void) | 767 | static cycle_t __vsyscall_fn vread_tsc(void) |
768 | { | 768 | { |
769 | cycle_t ret; | 769 | cycle_t ret; |
770 | u64 last; | ||
770 | 771 | ||
771 | /* | 772 | /* |
772 | * Empirically, a fence (of type that depends on the CPU) | 773 | * Empirically, a fence (of type that depends on the CPU) |
@@ -778,8 +779,21 @@ static cycle_t __vsyscall_fn vread_tsc(void) | |||
778 | rdtsc_barrier(); | 779 | rdtsc_barrier(); |
779 | ret = (cycle_t)vget_cycles(); | 780 | ret = (cycle_t)vget_cycles(); |
780 | 781 | ||
781 | return ret >= VVAR(vsyscall_gtod_data).clock.cycle_last ? | 782 | last = VVAR(vsyscall_gtod_data).clock.cycle_last; |
782 | ret : VVAR(vsyscall_gtod_data).clock.cycle_last; | 783 | |
784 | if (likely(ret >= last)) | ||
785 | return ret; | ||
786 | |||
787 | /* | ||
788 | * GCC likes to generate cmov here, but this branch is extremely | ||
789 | * predictable (it's just a funciton of time and the likely is | ||
790 | * very likely) and there's a data dependence, so force GCC | ||
791 | * to generate a branch instead. I don't barrier() because | ||
792 | * we don't actually need a barrier, and if this function | ||
793 | * ever gets inlined it will generate worse code. | ||
794 | */ | ||
795 | asm volatile (""); | ||
796 | return last; | ||
783 | } | 797 | } |
784 | #endif | 798 | #endif |
785 | 799 | ||