diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2011-03-18 13:45:21 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2011-03-18 13:45:21 -0400 |
commit | f2e1fbb5f2177227f71c4fc0491e531dd7acd385 (patch) | |
tree | c45485e3cd8a04f78184ef3bd9ce6972c74ac834 /arch/x86/crypto | |
parent | 508996b6a0ef0c7aa7701995d137e71c56180752 (diff) | |
parent | 4981d01eada5354d81c8929d5b2836829ba3df7b (diff) |
Merge branch 'x86-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
* 'x86-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
x86: Flush TLB if PGD entry is changed in i386 PAE mode
x86, dumpstack: Correct stack dump info when frame pointer is available
x86: Clean up csum-copy_64.S a bit
x86: Fix common misspellings
x86: Fix misspelling and align params
x86: Use PentiumPro-optimized partial_csum() on VIA C7
Diffstat (limited to 'arch/x86/crypto')
-rw-r--r-- | arch/x86/crypto/aesni-intel_asm.S | 6 |
1 files changed, 3 insertions, 3 deletions
diff --git a/arch/x86/crypto/aesni-intel_asm.S b/arch/x86/crypto/aesni-intel_asm.S index 8fe2a4966b7a..adcf794b22e2 100644 --- a/arch/x86/crypto/aesni-intel_asm.S +++ b/arch/x86/crypto/aesni-intel_asm.S | |||
@@ -1346,7 +1346,7 @@ _zero_cipher_left_decrypt: | |||
1346 | and $15, %r13 # %r13 = arg4 (mod 16) | 1346 | and $15, %r13 # %r13 = arg4 (mod 16) |
1347 | je _multiple_of_16_bytes_decrypt | 1347 | je _multiple_of_16_bytes_decrypt |
1348 | 1348 | ||
1349 | # Handle the last <16 byte block seperately | 1349 | # Handle the last <16 byte block separately |
1350 | 1350 | ||
1351 | paddd ONE(%rip), %xmm0 # increment CNT to get Yn | 1351 | paddd ONE(%rip), %xmm0 # increment CNT to get Yn |
1352 | movdqa SHUF_MASK(%rip), %xmm10 | 1352 | movdqa SHUF_MASK(%rip), %xmm10 |
@@ -1355,7 +1355,7 @@ _zero_cipher_left_decrypt: | |||
1355 | ENCRYPT_SINGLE_BLOCK %xmm0, %xmm1 # E(K, Yn) | 1355 | ENCRYPT_SINGLE_BLOCK %xmm0, %xmm1 # E(K, Yn) |
1356 | sub $16, %r11 | 1356 | sub $16, %r11 |
1357 | add %r13, %r11 | 1357 | add %r13, %r11 |
1358 | movdqu (%arg3,%r11,1), %xmm1 # recieve the last <16 byte block | 1358 | movdqu (%arg3,%r11,1), %xmm1 # receive the last <16 byte block |
1359 | lea SHIFT_MASK+16(%rip), %r12 | 1359 | lea SHIFT_MASK+16(%rip), %r12 |
1360 | sub %r13, %r12 | 1360 | sub %r13, %r12 |
1361 | # adjust the shuffle mask pointer to be able to shift 16-%r13 bytes | 1361 | # adjust the shuffle mask pointer to be able to shift 16-%r13 bytes |
@@ -1607,7 +1607,7 @@ _zero_cipher_left_encrypt: | |||
1607 | and $15, %r13 # %r13 = arg4 (mod 16) | 1607 | and $15, %r13 # %r13 = arg4 (mod 16) |
1608 | je _multiple_of_16_bytes_encrypt | 1608 | je _multiple_of_16_bytes_encrypt |
1609 | 1609 | ||
1610 | # Handle the last <16 Byte block seperately | 1610 | # Handle the last <16 Byte block separately |
1611 | paddd ONE(%rip), %xmm0 # INCR CNT to get Yn | 1611 | paddd ONE(%rip), %xmm0 # INCR CNT to get Yn |
1612 | movdqa SHUF_MASK(%rip), %xmm10 | 1612 | movdqa SHUF_MASK(%rip), %xmm10 |
1613 | PSHUFB_XMM %xmm10, %xmm0 | 1613 | PSHUFB_XMM %xmm10, %xmm0 |