123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325 |
- /***************************************************************************
- * Copyright (C) 2006 by Joachim Fritschi, <jfritschi@freenet.de> *
- * *
- * This program is free software; you can redistribute it and/or modify *
- * it under the terms of the GNU General Public License as published by *
- * the Free Software Foundation; either version 2 of the License, or *
- * (at your option) any later version. *
- * *
- * This program is distributed in the hope that it will be useful, *
- * but WITHOUT ANY WARRANTY; without even the implied warranty of *
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
- * GNU General Public License for more details. *
- * *
- * You should have received a copy of the GNU General Public License *
- * along with this program; if not, write to the *
- * Free Software Foundation, Inc., *
- * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. *
- ***************************************************************************/
- .file "twofish-x86_64-asm.S"
- .text
- #include <asm/asm-offsets.h>
- #define a_offset 0
- #define b_offset 4
- #define c_offset 8
- #define d_offset 12
- /* Structure of the crypto context struct*/
- #define s0 0 /* S0 Array 256 Words each */
- #define s1 1024 /* S1 Array */
- #define s2 2048 /* S2 Array */
- #define s3 3072 /* S3 Array */
- #define w 4096 /* 8 whitening keys (word) */
- #define k 4128 /* key 1-32 ( word ) */
- /* define a few register aliases to allow macro substitution */
- #define R0 %rax
- #define R0D %eax
- #define R0B %al
- #define R0H %ah
- #define R1 %rbx
- #define R1D %ebx
- #define R1B %bl
- #define R1H %bh
- #define R2 %rcx
- #define R2D %ecx
- #define R2B %cl
- #define R2H %ch
- #define R3 %rdx
- #define R3D %edx
- #define R3B %dl
- #define R3H %dh
- /* performs input whitening */
- #define input_whitening(src,context,offset)\
- xor w+offset(context), src;
- /* performs input whitening */
- #define output_whitening(src,context,offset)\
- xor w+16+offset(context), src;
- /*
- * a input register containing a (rotated 16)
- * b input register containing b
- * c input register containing c
- * d input register containing d (already rol $1)
- * operations on a and b are interleaved to increase performance
- */
- #define encrypt_round(a,b,c,d,round)\
- movzx b ## B, %edi;\
- mov s1(%r11,%rdi,4),%r8d;\
- movzx a ## B, %edi;\
- mov s2(%r11,%rdi,4),%r9d;\
- movzx b ## H, %edi;\
- ror $16, b ## D;\
- xor s2(%r11,%rdi,4),%r8d;\
- movzx a ## H, %edi;\
- ror $16, a ## D;\
- xor s3(%r11,%rdi,4),%r9d;\
- movzx b ## B, %edi;\
- xor s3(%r11,%rdi,4),%r8d;\
- movzx a ## B, %edi;\
- xor (%r11,%rdi,4), %r9d;\
- movzx b ## H, %edi;\
- ror $15, b ## D;\
- xor (%r11,%rdi,4), %r8d;\
- movzx a ## H, %edi;\
- xor s1(%r11,%rdi,4),%r9d;\
- add %r8d, %r9d;\
- add %r9d, %r8d;\
- add k+round(%r11), %r9d;\
- xor %r9d, c ## D;\
- rol $15, c ## D;\
- add k+4+round(%r11),%r8d;\
- xor %r8d, d ## D;
- /*
- * a input register containing a(rotated 16)
- * b input register containing b
- * c input register containing c
- * d input register containing d (already rol $1)
- * operations on a and b are interleaved to increase performance
- * during the round a and b are prepared for the output whitening
- */
- #define encrypt_last_round(a,b,c,d,round)\
- mov b ## D, %r10d;\
- shl $32, %r10;\
- movzx b ## B, %edi;\
- mov s1(%r11,%rdi,4),%r8d;\
- movzx a ## B, %edi;\
- mov s2(%r11,%rdi,4),%r9d;\
- movzx b ## H, %edi;\
- ror $16, b ## D;\
- xor s2(%r11,%rdi,4),%r8d;\
- movzx a ## H, %edi;\
- ror $16, a ## D;\
- xor s3(%r11,%rdi,4),%r9d;\
- movzx b ## B, %edi;\
- xor s3(%r11,%rdi,4),%r8d;\
- movzx a ## B, %edi;\
- xor (%r11,%rdi,4), %r9d;\
- xor a, %r10;\
- movzx b ## H, %edi;\
- xor (%r11,%rdi,4), %r8d;\
- movzx a ## H, %edi;\
- xor s1(%r11,%rdi,4),%r9d;\
- add %r8d, %r9d;\
- add %r9d, %r8d;\
- add k+round(%r11), %r9d;\
- xor %r9d, c ## D;\
- ror $1, c ## D;\
- add k+4+round(%r11),%r8d;\
- xor %r8d, d ## D
- /*
- * a input register containing a
- * b input register containing b (rotated 16)
- * c input register containing c (already rol $1)
- * d input register containing d
- * operations on a and b are interleaved to increase performance
- */
- #define decrypt_round(a,b,c,d,round)\
- movzx a ## B, %edi;\
- mov (%r11,%rdi,4), %r9d;\
- movzx b ## B, %edi;\
- mov s3(%r11,%rdi,4),%r8d;\
- movzx a ## H, %edi;\
- ror $16, a ## D;\
- xor s1(%r11,%rdi,4),%r9d;\
- movzx b ## H, %edi;\
- ror $16, b ## D;\
- xor (%r11,%rdi,4), %r8d;\
- movzx a ## B, %edi;\
- xor s2(%r11,%rdi,4),%r9d;\
- movzx b ## B, %edi;\
- xor s1(%r11,%rdi,4),%r8d;\
- movzx a ## H, %edi;\
- ror $15, a ## D;\
- xor s3(%r11,%rdi,4),%r9d;\
- movzx b ## H, %edi;\
- xor s2(%r11,%rdi,4),%r8d;\
- add %r8d, %r9d;\
- add %r9d, %r8d;\
- add k+round(%r11), %r9d;\
- xor %r9d, c ## D;\
- add k+4+round(%r11),%r8d;\
- xor %r8d, d ## D;\
- rol $15, d ## D;
- /*
- * a input register containing a
- * b input register containing b
- * c input register containing c (already rol $1)
- * d input register containing d
- * operations on a and b are interleaved to increase performance
- * during the round a and b are prepared for the output whitening
- */
- #define decrypt_last_round(a,b,c,d,round)\
- movzx a ## B, %edi;\
- mov (%r11,%rdi,4), %r9d;\
- movzx b ## B, %edi;\
- mov s3(%r11,%rdi,4),%r8d;\
- movzx b ## H, %edi;\
- ror $16, b ## D;\
- xor (%r11,%rdi,4), %r8d;\
- movzx a ## H, %edi;\
- mov b ## D, %r10d;\
- shl $32, %r10;\
- xor a, %r10;\
- ror $16, a ## D;\
- xor s1(%r11,%rdi,4),%r9d;\
- movzx b ## B, %edi;\
- xor s1(%r11,%rdi,4),%r8d;\
- movzx a ## B, %edi;\
- xor s2(%r11,%rdi,4),%r9d;\
- movzx b ## H, %edi;\
- xor s2(%r11,%rdi,4),%r8d;\
- movzx a ## H, %edi;\
- xor s3(%r11,%rdi,4),%r9d;\
- add %r8d, %r9d;\
- add %r9d, %r8d;\
- add k+round(%r11), %r9d;\
- xor %r9d, c ## D;\
- add k+4+round(%r11),%r8d;\
- xor %r8d, d ## D;\
- ror $1, d ## D;
- .align 8
- .global twofish_enc_blk
- .global twofish_dec_blk
- twofish_enc_blk:
- pushq R1
- /* %rdi contains the crypto tfm address */
- /* %rsi contains the output address */
- /* %rdx contains the input address */
- add $crypto_tfm_ctx_offset, %rdi /* set ctx address */
- /* ctx address is moved to free one non-rex register
- as target for the 8bit high operations */
- mov %rdi, %r11
- movq (R3), R1
- movq 8(R3), R3
- input_whitening(R1,%r11,a_offset)
- input_whitening(R3,%r11,c_offset)
- mov R1D, R0D
- rol $16, R0D
- shr $32, R1
- mov R3D, R2D
- shr $32, R3
- rol $1, R3D
- encrypt_round(R0,R1,R2,R3,0);
- encrypt_round(R2,R3,R0,R1,8);
- encrypt_round(R0,R1,R2,R3,2*8);
- encrypt_round(R2,R3,R0,R1,3*8);
- encrypt_round(R0,R1,R2,R3,4*8);
- encrypt_round(R2,R3,R0,R1,5*8);
- encrypt_round(R0,R1,R2,R3,6*8);
- encrypt_round(R2,R3,R0,R1,7*8);
- encrypt_round(R0,R1,R2,R3,8*8);
- encrypt_round(R2,R3,R0,R1,9*8);
- encrypt_round(R0,R1,R2,R3,10*8);
- encrypt_round(R2,R3,R0,R1,11*8);
- encrypt_round(R0,R1,R2,R3,12*8);
- encrypt_round(R2,R3,R0,R1,13*8);
- encrypt_round(R0,R1,R2,R3,14*8);
- encrypt_last_round(R2,R3,R0,R1,15*8);
- output_whitening(%r10,%r11,a_offset)
- movq %r10, (%rsi)
- shl $32, R1
- xor R0, R1
- output_whitening(R1,%r11,c_offset)
- movq R1, 8(%rsi)
- popq R1
- movq $1,%rax
- ret
- twofish_dec_blk:
- pushq R1
- /* %rdi contains the crypto tfm address */
- /* %rsi contains the output address */
- /* %rdx contains the input address */
- add $crypto_tfm_ctx_offset, %rdi /* set ctx address */
- /* ctx address is moved to free one non-rex register
- as target for the 8bit high operations */
- mov %rdi, %r11
- movq (R3), R1
- movq 8(R3), R3
- output_whitening(R1,%r11,a_offset)
- output_whitening(R3,%r11,c_offset)
- mov R1D, R0D
- shr $32, R1
- rol $16, R1D
- mov R3D, R2D
- shr $32, R3
- rol $1, R2D
- decrypt_round(R0,R1,R2,R3,15*8);
- decrypt_round(R2,R3,R0,R1,14*8);
- decrypt_round(R0,R1,R2,R3,13*8);
- decrypt_round(R2,R3,R0,R1,12*8);
- decrypt_round(R0,R1,R2,R3,11*8);
- decrypt_round(R2,R3,R0,R1,10*8);
- decrypt_round(R0,R1,R2,R3,9*8);
- decrypt_round(R2,R3,R0,R1,8*8);
- decrypt_round(R0,R1,R2,R3,7*8);
- decrypt_round(R2,R3,R0,R1,6*8);
- decrypt_round(R0,R1,R2,R3,5*8);
- decrypt_round(R2,R3,R0,R1,4*8);
- decrypt_round(R0,R1,R2,R3,3*8);
- decrypt_round(R2,R3,R0,R1,2*8);
- decrypt_round(R0,R1,R2,R3,1*8);
- decrypt_last_round(R2,R3,R0,R1,0);
- input_whitening(%r10,%r11,a_offset)
- movq %r10, (%rsi)
- shl $32, R1
- xor R0, R1
- input_whitening(R1,%r11,c_offset)
- movq R1, 8(%rsi)
- popq R1
- movq $1,%rax
- ret
|