aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorCesar Eduardo Barros <cesarb@cesarb.eti.br>2013-11-25 19:00:41 -0500
committerHerbert Xu <herbert@gondor.apana.org.au>2013-12-05 08:28:41 -0500
commitfe8c8a126806fea4465c43d62a1f9d273a572bf5 (patch)
treeea40ae4afce934575065383d1de44d04cd6490fc
parentb62ffd8c72c2f04ac737ef3ff0f7ba37d2013633 (diff)
crypto: more robust crypto_memneq
Disabling compiler optimizations can be fragile, since a new optimization could be added to -O0 or -Os that breaks the assumptions the code is making. Instead of disabling compiler optimizations, use a dummy inline assembly (based on RELOC_HIDE) to block the problematic kinds of optimization, while still allowing other optimizations to be applied to the code. The dummy inline assembly is added after every OR, and has the accumulator variable as its input and output. The compiler is forced to assume that the dummy inline assembly could both depend on the accumulator variable and change the accumulator variable, so it is forced to compute the value correctly before the inline assembly, and cannot assume anything about its value after the inline assembly. This change should be enough to make crypto_memneq work correctly (with data-independent timing) even if it is inlined at its call sites. That can be done later in a followup patch. Compile-tested on x86_64. Signed-off-by: Cesar Eduardo Barros <cesarb@cesarb.eti.br> Acked-by: Daniel Borkmann <dborkman@redhat.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
-rw-r--r--crypto/Makefile5
-rw-r--r--crypto/memneq.c79
-rw-r--r--include/linux/compiler-gcc.h3
-rw-r--r--include/linux/compiler-intel.h7
-rw-r--r--include/linux/compiler.h4
5 files changed, 68 insertions, 30 deletions
diff --git a/crypto/Makefile b/crypto/Makefile
index 989c510da8cc..b29402a7b9b5 100644
--- a/crypto/Makefile
+++ b/crypto/Makefile
@@ -2,11 +2,6 @@
2# Cryptographic API 2# Cryptographic API
3# 3#
4 4
5# memneq MUST be built with -Os or -O0 to prevent early-return optimizations
6# that will defeat memneq's actual purpose to prevent timing attacks.
7CFLAGS_REMOVE_memneq.o := -O1 -O2 -O3
8CFLAGS_memneq.o := -Os
9
10obj-$(CONFIG_CRYPTO) += crypto.o 5obj-$(CONFIG_CRYPTO) += crypto.o
11crypto-y := api.o cipher.o compress.o memneq.o 6crypto-y := api.o cipher.o compress.o memneq.o
12 7
diff --git a/crypto/memneq.c b/crypto/memneq.c
index cd0162221c14..570f6f3401ce 100644
--- a/crypto/memneq.c
+++ b/crypto/memneq.c
@@ -72,6 +72,7 @@ __crypto_memneq_generic(const void *a, const void *b, size_t size)
72#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) 72#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
73 while (size >= sizeof(unsigned long)) { 73 while (size >= sizeof(unsigned long)) {
74 neq |= *(unsigned long *)a ^ *(unsigned long *)b; 74 neq |= *(unsigned long *)a ^ *(unsigned long *)b;
75 OPTIMIZER_HIDE_VAR(neq);
75 a += sizeof(unsigned long); 76 a += sizeof(unsigned long);
76 b += sizeof(unsigned long); 77 b += sizeof(unsigned long);
77 size -= sizeof(unsigned long); 78 size -= sizeof(unsigned long);
@@ -79,6 +80,7 @@ __crypto_memneq_generic(const void *a, const void *b, size_t size)
79#endif /* CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS */ 80#endif /* CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS */
80 while (size > 0) { 81 while (size > 0) {
81 neq |= *(unsigned char *)a ^ *(unsigned char *)b; 82 neq |= *(unsigned char *)a ^ *(unsigned char *)b;
83 OPTIMIZER_HIDE_VAR(neq);
82 a += 1; 84 a += 1;
83 b += 1; 85 b += 1;
84 size -= 1; 86 size -= 1;
@@ -89,33 +91,60 @@ __crypto_memneq_generic(const void *a, const void *b, size_t size)
89/* Loop-free fast-path for frequently used 16-byte size */ 91/* Loop-free fast-path for frequently used 16-byte size */
90static inline unsigned long __crypto_memneq_16(const void *a, const void *b) 92static inline unsigned long __crypto_memneq_16(const void *a, const void *b)
91{ 93{
94 unsigned long neq = 0;
95
92#ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS 96#ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
93 if (sizeof(unsigned long) == 8) 97 if (sizeof(unsigned long) == 8) {
94 return ((*(unsigned long *)(a) ^ *(unsigned long *)(b)) 98 neq |= *(unsigned long *)(a) ^ *(unsigned long *)(b);
95 | (*(unsigned long *)(a+8) ^ *(unsigned long *)(b+8))); 99 OPTIMIZER_HIDE_VAR(neq);
96 else if (sizeof(unsigned int) == 4) 100 neq |= *(unsigned long *)(a+8) ^ *(unsigned long *)(b+8);
97 return ((*(unsigned int *)(a) ^ *(unsigned int *)(b)) 101 OPTIMIZER_HIDE_VAR(neq);
98 | (*(unsigned int *)(a+4) ^ *(unsigned int *)(b+4)) 102 } else if (sizeof(unsigned int) == 4) {
99 | (*(unsigned int *)(a+8) ^ *(unsigned int *)(b+8)) 103 neq |= *(unsigned int *)(a) ^ *(unsigned int *)(b);
100 | (*(unsigned int *)(a+12) ^ *(unsigned int *)(b+12))); 104 OPTIMIZER_HIDE_VAR(neq);
101 else 105 neq |= *(unsigned int *)(a+4) ^ *(unsigned int *)(b+4);
106 OPTIMIZER_HIDE_VAR(neq);
107 neq |= *(unsigned int *)(a+8) ^ *(unsigned int *)(b+8);
108 OPTIMIZER_HIDE_VAR(neq);
109 neq |= *(unsigned int *)(a+12) ^ *(unsigned int *)(b+12);
110 OPTIMIZER_HIDE_VAR(neq);
111 } else {
102#endif /* CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS */ 112#endif /* CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS */
103 return ((*(unsigned char *)(a) ^ *(unsigned char *)(b)) 113 neq |= *(unsigned char *)(a) ^ *(unsigned char *)(b);
104 | (*(unsigned char *)(a+1) ^ *(unsigned char *)(b+1)) 114 OPTIMIZER_HIDE_VAR(neq);
105 | (*(unsigned char *)(a+2) ^ *(unsigned char *)(b+2)) 115 neq |= *(unsigned char *)(a+1) ^ *(unsigned char *)(b+1);
106 | (*(unsigned char *)(a+3) ^ *(unsigned char *)(b+3)) 116 OPTIMIZER_HIDE_VAR(neq);
107 | (*(unsigned char *)(a+4) ^ *(unsigned char *)(b+4)) 117 neq |= *(unsigned char *)(a+2) ^ *(unsigned char *)(b+2);
108 | (*(unsigned char *)(a+5) ^ *(unsigned char *)(b+5)) 118 OPTIMIZER_HIDE_VAR(neq);
109 | (*(unsigned char *)(a+6) ^ *(unsigned char *)(b+6)) 119 neq |= *(unsigned char *)(a+3) ^ *(unsigned char *)(b+3);
110 | (*(unsigned char *)(a+7) ^ *(unsigned char *)(b+7)) 120 OPTIMIZER_HIDE_VAR(neq);
111 | (*(unsigned char *)(a+8) ^ *(unsigned char *)(b+8)) 121 neq |= *(unsigned char *)(a+4) ^ *(unsigned char *)(b+4);
112 | (*(unsigned char *)(a+9) ^ *(unsigned char *)(b+9)) 122 OPTIMIZER_HIDE_VAR(neq);
113 | (*(unsigned char *)(a+10) ^ *(unsigned char *)(b+10)) 123 neq |= *(unsigned char *)(a+5) ^ *(unsigned char *)(b+5);
114 | (*(unsigned char *)(a+11) ^ *(unsigned char *)(b+11)) 124 OPTIMIZER_HIDE_VAR(neq);
115 | (*(unsigned char *)(a+12) ^ *(unsigned char *)(b+12)) 125 neq |= *(unsigned char *)(a+6) ^ *(unsigned char *)(b+6);
116 | (*(unsigned char *)(a+13) ^ *(unsigned char *)(b+13)) 126 OPTIMIZER_HIDE_VAR(neq);
117 | (*(unsigned char *)(a+14) ^ *(unsigned char *)(b+14)) 127 neq |= *(unsigned char *)(a+7) ^ *(unsigned char *)(b+7);
118 | (*(unsigned char *)(a+15) ^ *(unsigned char *)(b+15))); 128 OPTIMIZER_HIDE_VAR(neq);
129 neq |= *(unsigned char *)(a+8) ^ *(unsigned char *)(b+8);
130 OPTIMIZER_HIDE_VAR(neq);
131 neq |= *(unsigned char *)(a+9) ^ *(unsigned char *)(b+9);
132 OPTIMIZER_HIDE_VAR(neq);
133 neq |= *(unsigned char *)(a+10) ^ *(unsigned char *)(b+10);
134 OPTIMIZER_HIDE_VAR(neq);
135 neq |= *(unsigned char *)(a+11) ^ *(unsigned char *)(b+11);
136 OPTIMIZER_HIDE_VAR(neq);
137 neq |= *(unsigned char *)(a+12) ^ *(unsigned char *)(b+12);
138 OPTIMIZER_HIDE_VAR(neq);
139 neq |= *(unsigned char *)(a+13) ^ *(unsigned char *)(b+13);
140 OPTIMIZER_HIDE_VAR(neq);
141 neq |= *(unsigned char *)(a+14) ^ *(unsigned char *)(b+14);
142 OPTIMIZER_HIDE_VAR(neq);
143 neq |= *(unsigned char *)(a+15) ^ *(unsigned char *)(b+15);
144 OPTIMIZER_HIDE_VAR(neq);
145 }
146
147 return neq;
119} 148}
120 149
121/* Compare two areas of memory without leaking timing information, 150/* Compare two areas of memory without leaking timing information,
diff --git a/include/linux/compiler-gcc.h b/include/linux/compiler-gcc.h
index 24545cd90a25..02ae99e8e6d3 100644
--- a/include/linux/compiler-gcc.h
+++ b/include/linux/compiler-gcc.h
@@ -37,6 +37,9 @@
37 __asm__ ("" : "=r"(__ptr) : "0"(ptr)); \ 37 __asm__ ("" : "=r"(__ptr) : "0"(ptr)); \
38 (typeof(ptr)) (__ptr + (off)); }) 38 (typeof(ptr)) (__ptr + (off)); })
39 39
40/* Make the optimizer believe the variable can be manipulated arbitrarily. */
41#define OPTIMIZER_HIDE_VAR(var) __asm__ ("" : "=r" (var) : "0" (var))
42
40#ifdef __CHECKER__ 43#ifdef __CHECKER__
41#define __must_be_array(arr) 0 44#define __must_be_array(arr) 0
42#else 45#else
diff --git a/include/linux/compiler-intel.h b/include/linux/compiler-intel.h
index 973ce10c40b6..e784f5707749 100644
--- a/include/linux/compiler-intel.h
+++ b/include/linux/compiler-intel.h
@@ -15,6 +15,7 @@
15 */ 15 */
16#undef barrier 16#undef barrier
17#undef RELOC_HIDE 17#undef RELOC_HIDE
18#undef OPTIMIZER_HIDE_VAR
18 19
19#define barrier() __memory_barrier() 20#define barrier() __memory_barrier()
20 21
@@ -23,6 +24,12 @@
23 __ptr = (unsigned long) (ptr); \ 24 __ptr = (unsigned long) (ptr); \
24 (typeof(ptr)) (__ptr + (off)); }) 25 (typeof(ptr)) (__ptr + (off)); })
25 26
27/* This should act as an optimization barrier on var.
28 * Given that this compiler does not have inline assembly, a compiler barrier
29 * is the best we can do.
30 */
31#define OPTIMIZER_HIDE_VAR(var) barrier()
32
26/* Intel ECC compiler doesn't support __builtin_types_compatible_p() */ 33/* Intel ECC compiler doesn't support __builtin_types_compatible_p() */
27#define __must_be_array(a) 0 34#define __must_be_array(a) 0
28 35
diff --git a/include/linux/compiler.h b/include/linux/compiler.h
index 92669cd182a6..a2329c5e6206 100644
--- a/include/linux/compiler.h
+++ b/include/linux/compiler.h
@@ -170,6 +170,10 @@ void ftrace_likely_update(struct ftrace_branch_data *f, int val, int expect);
170 (typeof(ptr)) (__ptr + (off)); }) 170 (typeof(ptr)) (__ptr + (off)); })
171#endif 171#endif
172 172
173#ifndef OPTIMIZER_HIDE_VAR
174#define OPTIMIZER_HIDE_VAR(var) barrier()
175#endif
176
173/* Not-quite-unique ID. */ 177/* Not-quite-unique ID. */
174#ifndef __UNIQUE_ID 178#ifndef __UNIQUE_ID
175# define __UNIQUE_ID(prefix) __PASTE(__PASTE(__UNIQUE_ID_, prefix), __LINE__) 179# define __UNIQUE_ID(prefix) __PASTE(__PASTE(__UNIQUE_ID_, prefix), __LINE__)