summaryrefslogtreecommitdiff
path: root/secure/lib/libcrypto/i386
diff options
context:
space:
mode:
authorcvs2svn <cvs2svn@FreeBSD.org>2004-02-23 15:32:56 +0000
committercvs2svn <cvs2svn@FreeBSD.org>2004-02-23 15:32:56 +0000
commit84a7b99b4e02a260bf7ff9b540032f4cb7bc612a (patch)
tree85059f7e88bcf124efae7c17bd99cfa18627a8c4 /secure/lib/libcrypto/i386
parenta28c3ad292f6f6cbe1c4ff05efb3d53eb45f17ff (diff)
Diffstat (limited to 'secure/lib/libcrypto/i386')
-rw-r--r--secure/lib/libcrypto/i386/bf-586.s932
-rw-r--r--secure/lib/libcrypto/i386/bf-686.s902
-rw-r--r--secure/lib/libcrypto/i386/bn-586.s890
-rw-r--r--secure/lib/libcrypto/i386/cast-586.s971
-rw-r--r--secure/lib/libcrypto/i386/co-586.s1270
-rw-r--r--secure/lib/libcrypto/i386/crypt586.s933
-rw-r--r--secure/lib/libcrypto/i386/des-586.s3154
-rw-r--r--secure/lib/libcrypto/i386/md5-586.s689
-rw-r--r--secure/lib/libcrypto/i386/rc4-586.s316
-rw-r--r--secure/lib/libcrypto/i386/rc5-586.s584
-rw-r--r--secure/lib/libcrypto/i386/rmd-586.s1975
-rw-r--r--secure/lib/libcrypto/i386/sha1-586.s1960
12 files changed, 0 insertions, 14576 deletions
diff --git a/secure/lib/libcrypto/i386/bf-586.s b/secure/lib/libcrypto/i386/bf-586.s
deleted file mode 100644
index 0965b1c33d02..000000000000
--- a/secure/lib/libcrypto/i386/bf-586.s
+++ /dev/null
@@ -1,932 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by bf-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "bf-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl BF_encrypt
- .type BF_encrypt,@function
-BF_encrypt:
-
- pushl %ebp
- pushl %ebx
- movl 12(%esp), %ebx
- movl 16(%esp), %ebp
- pushl %esi
- pushl %edi
- # Load the 2 words
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- xorl %eax, %eax
- movl (%ebp), %ebx
- xorl %ecx, %ecx
- xorl %ebx, %edi
-
- # Round 0
- movl 4(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 1
- movl 8(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 2
- movl 12(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 3
- movl 16(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 4
- movl 20(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 5
- movl 24(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 6
- movl 28(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 7
- movl 32(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 8
- movl 36(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 9
- movl 40(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 10
- movl 44(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 11
- movl 48(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 12
- movl 52(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 13
- movl 56(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 14
- movl 60(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 15
- movl 64(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- # Load parameter 0 (16) enc=1
- movl 20(%esp), %eax
- xorl %ebx, %edi
- movl 68(%ebp), %edx
- xorl %edx, %esi
- movl %edi, 4(%eax)
- movl %esi, (%eax)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.BF_encrypt_end:
- .size BF_encrypt,.BF_encrypt_end-BF_encrypt
-.ident "BF_encrypt"
-.text
- .align 16
-.globl BF_decrypt
- .type BF_decrypt,@function
-BF_decrypt:
-
- pushl %ebp
- pushl %ebx
- movl 12(%esp), %ebx
- movl 16(%esp), %ebp
- pushl %esi
- pushl %edi
- # Load the 2 words
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- xorl %eax, %eax
- movl 68(%ebp), %ebx
- xorl %ecx, %ecx
- xorl %ebx, %edi
-
- # Round 16
- movl 64(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 15
- movl 60(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 14
- movl 56(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 13
- movl 52(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 12
- movl 48(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 11
- movl 44(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 10
- movl 40(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 9
- movl 36(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 8
- movl 32(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 7
- movl 28(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 6
- movl 24(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 5
- movl 20(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 4
- movl 16(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 3
- movl 12(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %edi
-
- # Round 2
- movl 8(%ebp), %edx
- movl %edi, %ebx
- xorl %edx, %esi
- shrl $16, %ebx
- movl %edi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- xorl %eax, %eax
- xorl %ebx, %esi
-
- # Round 1
- movl 4(%ebp), %edx
- movl %esi, %ebx
- xorl %edx, %edi
- shrl $16, %ebx
- movl %esi, %edx
- movb %bh, %al
- andl $255, %ebx
- movb %dh, %cl
- andl $255, %edx
- movl 72(%ebp,%eax,4),%eax
- movl 1096(%ebp,%ebx,4),%ebx
- addl %eax, %ebx
- movl 2120(%ebp,%ecx,4),%eax
- xorl %eax, %ebx
- movl 3144(%ebp,%edx,4),%edx
- addl %edx, %ebx
- # Load parameter 0 (1) enc=0
- movl 20(%esp), %eax
- xorl %ebx, %edi
- movl (%ebp), %edx
- xorl %edx, %esi
- movl %edi, 4(%eax)
- movl %esi, (%eax)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.BF_decrypt_end:
- .size BF_decrypt,.BF_decrypt_end-BF_decrypt
-.ident "BF_decrypt"
-.text
- .align 16
-.globl BF_cbc_encrypt
- .type BF_cbc_encrypt,@function
-BF_cbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
- # getting iv ptr from parameter 4
- movl 36(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
- # getting encrypt flag from parameter 5
- movl 56(%esp), %ecx
- # get and push parameter 3
- movl 48(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L000decrypt
- andl $4294967288, %ebp
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- jz .L001encrypt_finish
-.L002encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L002encrypt_loop
-.L001encrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L003finish
- xorl %ecx, %ecx
- xorl %edx, %edx
- movl .L004cbc_enc_jmp_table(,%ebp,4),%ebp
- jmp *%ebp
-.L005ej7:
- movb 6(%esi), %dh
- sall $8, %edx
-.L006ej6:
- movb 5(%esi), %dh
-.L007ej5:
- movb 4(%esi), %dl
-.L008ej4:
- movl (%esi), %ecx
- jmp .L009ejend
-.L010ej3:
- movb 2(%esi), %ch
- sall $8, %ecx
-.L011ej2:
- movb 1(%esi), %ch
-.L012ej1:
- movb (%esi), %cl
-.L009ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L003finish
-.align 16
-.L000decrypt:
- andl $4294967288, %ebp
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- jz .L013decrypt_finish
-.L014decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L014decrypt_loop
-.L013decrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L003finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L015dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L016dj6:
- movb %dh, 5(%edi)
-.L017dj5:
- movb %dl, 4(%edi)
-.L018dj4:
- movl %ecx, (%edi)
- jmp .L019djend
-.L020dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L021dj2:
- movb %ch, 1(%esi)
-.L022dj1:
- movb %cl, (%esi)
-.L019djend:
- jmp .L003finish
-.align 16
-.L003finish:
- movl 60(%esp), %ecx
- addl $24, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L004cbc_enc_jmp_table:
- .long 0
- .long .L012ej1
- .long .L011ej2
- .long .L010ej3
- .long .L008ej4
- .long .L007ej5
- .long .L006ej6
- .long .L005ej7
-.align 16
-.L023cbc_dec_jmp_table:
- .long 0
- .long .L022dj1
- .long .L021dj2
- .long .L020dj3
- .long .L018dj4
- .long .L017dj5
- .long .L016dj6
- .long .L015dj7
-.L_BF_cbc_encrypt_end:
- .size BF_cbc_encrypt,.L_BF_cbc_encrypt_end-BF_cbc_encrypt
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/bf-686.s b/secure/lib/libcrypto/i386/bf-686.s
deleted file mode 100644
index bb3b9c75ddb7..000000000000
--- a/secure/lib/libcrypto/i386/bf-686.s
+++ /dev/null
@@ -1,902 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by bf-686.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "bf-686.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl BF_encrypt
- .type BF_encrypt,@function
-BF_encrypt:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- # Load the 2 words
- movl 20(%esp), %eax
- movl (%eax), %ecx
- movl 4(%eax), %edx
-
- # P pointer, s and enc flag
- movl 24(%esp), %edi
- xorl %eax, %eax
- xorl %ebx, %ebx
- xorl (%edi), %ecx
-
- # Round 0
- rorl $16, %ecx
- movl 4(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 1
- rorl $16, %edx
- movl 8(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 2
- rorl $16, %ecx
- movl 12(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 3
- rorl $16, %edx
- movl 16(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 4
- rorl $16, %ecx
- movl 20(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 5
- rorl $16, %edx
- movl 24(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 6
- rorl $16, %ecx
- movl 28(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 7
- rorl $16, %edx
- movl 32(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 8
- rorl $16, %ecx
- movl 36(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 9
- rorl $16, %edx
- movl 40(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 10
- rorl $16, %ecx
- movl 44(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 11
- rorl $16, %edx
- movl 48(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 12
- rorl $16, %ecx
- movl 52(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 13
- rorl $16, %edx
- movl 56(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 14
- rorl $16, %ecx
- movl 60(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 15
- rorl $16, %edx
- movl 64(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
- xorl 68(%edi), %edx
- movl 20(%esp), %eax
- movl %edx, (%eax)
- movl %ecx, 4(%eax)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.L_BF_encrypt_end:
- .size BF_encrypt,.L_BF_encrypt_end-BF_encrypt
-.ident "desasm.pl"
-.text
- .align 16
-.globl BF_decrypt
- .type BF_decrypt,@function
-BF_decrypt:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- # Load the 2 words
- movl 20(%esp), %eax
- movl (%eax), %ecx
- movl 4(%eax), %edx
-
- # P pointer, s and enc flag
- movl 24(%esp), %edi
- xorl %eax, %eax
- xorl %ebx, %ebx
- xorl 68(%edi), %ecx
-
- # Round 16
- rorl $16, %ecx
- movl 64(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 15
- rorl $16, %edx
- movl 60(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 14
- rorl $16, %ecx
- movl 56(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 13
- rorl $16, %edx
- movl 52(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 12
- rorl $16, %ecx
- movl 48(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 11
- rorl $16, %edx
- movl 44(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 10
- rorl $16, %ecx
- movl 40(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 9
- rorl $16, %edx
- movl 36(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 8
- rorl $16, %ecx
- movl 32(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 7
- rorl $16, %edx
- movl 28(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 6
- rorl $16, %ecx
- movl 24(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 5
- rorl $16, %edx
- movl 20(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 4
- rorl $16, %ecx
- movl 16(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 3
- rorl $16, %edx
- movl 12(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
-
- # Round 2
- rorl $16, %ecx
- movl 8(%edi), %esi
- movb %ch, %al
- movb %cl, %bl
- rorl $16, %ecx
- xorl %esi, %edx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %ch, %al
- movb %cl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %edx
-
- # Round 1
- rorl $16, %edx
- movl 4(%edi), %esi
- movb %dh, %al
- movb %dl, %bl
- rorl $16, %edx
- xorl %esi, %ecx
- movl 72(%edi,%eax,4),%esi
- movl 1096(%edi,%ebx,4),%ebp
- movb %dh, %al
- movb %dl, %bl
- addl %ebp, %esi
- movl 2120(%edi,%eax,4),%eax
- xorl %eax, %esi
- movl 3144(%edi,%ebx,4),%ebp
- addl %ebp, %esi
- xorl %eax, %eax
- xorl %esi, %ecx
- xorl (%edi), %edx
- movl 20(%esp), %eax
- movl %edx, (%eax)
- movl %ecx, 4(%eax)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.L_BF_decrypt_end:
- .size BF_decrypt,.L_BF_decrypt_end-BF_decrypt
-.ident "desasm.pl"
-.text
- .align 16
-.globl BF_cbc_encrypt
- .type BF_cbc_encrypt,@function
-BF_cbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
- # getting iv ptr from parameter 4
- movl 36(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
- # getting encrypt flag from parameter 5
- movl 56(%esp), %ecx
- # get and push parameter 3
- movl 48(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L000decrypt
- andl $4294967288, %ebp
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- jz .L001encrypt_finish
-.L002encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L002encrypt_loop
-.L001encrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L003finish
- xorl %ecx, %ecx
- xorl %edx, %edx
- movl .L004cbc_enc_jmp_table(,%ebp,4),%ebp
- jmp *%ebp
-.L005ej7:
- movb 6(%esi), %dh
- sall $8, %edx
-.L006ej6:
- movb 5(%esi), %dh
-.L007ej5:
- movb 4(%esi), %dl
-.L008ej4:
- movl (%esi), %ecx
- jmp .L009ejend
-.L010ej3:
- movb 2(%esi), %ch
- sall $8, %ecx
-.L011ej2:
- movb 1(%esi), %ch
-.L012ej1:
- movb (%esi), %cl
-.L009ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L003finish
-.align 16
-.L000decrypt:
- andl $4294967288, %ebp
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- jz .L013decrypt_finish
-.L014decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L014decrypt_loop
-.L013decrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L003finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call BF_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L015dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L016dj6:
- movb %dh, 5(%edi)
-.L017dj5:
- movb %dl, 4(%edi)
-.L018dj4:
- movl %ecx, (%edi)
- jmp .L019djend
-.L020dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L021dj2:
- movb %ch, 1(%esi)
-.L022dj1:
- movb %cl, (%esi)
-.L019djend:
- jmp .L003finish
-.align 16
-.L003finish:
- movl 60(%esp), %ecx
- addl $24, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L004cbc_enc_jmp_table:
- .long 0
- .long .L012ej1
- .long .L011ej2
- .long .L010ej3
- .long .L008ej4
- .long .L007ej5
- .long .L006ej6
- .long .L005ej7
-.align 16
-.L023cbc_dec_jmp_table:
- .long 0
- .long .L022dj1
- .long .L021dj2
- .long .L020dj3
- .long .L018dj4
- .long .L017dj5
- .long .L016dj6
- .long .L015dj7
-.L_BF_cbc_encrypt_end:
- .size BF_cbc_encrypt,.L_BF_cbc_encrypt_end-BF_cbc_encrypt
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/bn-586.s b/secure/lib/libcrypto/i386/bn-586.s
deleted file mode 100644
index 3ea4a8a959bc..000000000000
--- a/secure/lib/libcrypto/i386/bn-586.s
+++ /dev/null
@@ -1,890 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by bn-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "bn-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl bn_mul_add_words
- .type bn_mul_add_words,@function
-bn_mul_add_words:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- xorl %esi, %esi
- movl 20(%esp), %edi
- movl 28(%esp), %ecx
- movl 24(%esp), %ebx
- andl $4294967288, %ecx
- movl 32(%esp), %ebp
- pushl %ecx
- jz .L000maw_finish
-.L001maw_loop:
- movl %ecx, (%esp)
- # Round 0
- movl (%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl (%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, (%edi)
- movl %edx, %esi
- # Round 4
- movl 4(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 4(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 4(%edi)
- movl %edx, %esi
- # Round 8
- movl 8(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 8(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 8(%edi)
- movl %edx, %esi
- # Round 12
- movl 12(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 12(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 12(%edi)
- movl %edx, %esi
- # Round 16
- movl 16(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 16(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 16(%edi)
- movl %edx, %esi
- # Round 20
- movl 20(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 20(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 20(%edi)
- movl %edx, %esi
- # Round 24
- movl 24(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 24(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 24(%edi)
- movl %edx, %esi
- # Round 28
- movl 28(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 28(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 28(%edi)
- movl %edx, %esi
-
- movl (%esp), %ecx
- addl $32, %ebx
- addl $32, %edi
- subl $8, %ecx
- jnz .L001maw_loop
-.L000maw_finish:
- movl 32(%esp), %ecx
- andl $7, %ecx
- jnz .L002maw_finish2
- jmp .L003maw_end
-.align 16
-.L002maw_finish2:
- # Tail Round 0
- movl (%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl (%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- decl %ecx
- movl %eax, (%edi)
- movl %edx, %esi
- jz .L003maw_end
- # Tail Round 1
- movl 4(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 4(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- decl %ecx
- movl %eax, 4(%edi)
- movl %edx, %esi
- jz .L003maw_end
- # Tail Round 2
- movl 8(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 8(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- decl %ecx
- movl %eax, 8(%edi)
- movl %edx, %esi
- jz .L003maw_end
- # Tail Round 3
- movl 12(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 12(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- decl %ecx
- movl %eax, 12(%edi)
- movl %edx, %esi
- jz .L003maw_end
- # Tail Round 4
- movl 16(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 16(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- decl %ecx
- movl %eax, 16(%edi)
- movl %edx, %esi
- jz .L003maw_end
- # Tail Round 5
- movl 20(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 20(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- decl %ecx
- movl %eax, 20(%edi)
- movl %edx, %esi
- jz .L003maw_end
- # Tail Round 6
- movl 24(%ebx), %eax
- mull %ebp
- addl %esi, %eax
- movl 24(%edi), %esi
- adcl $0, %edx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 24(%edi)
- movl %edx, %esi
-.L003maw_end:
- movl %esi, %eax
- popl %ecx
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.bn_mul_add_words_end:
- .size bn_mul_add_words,.bn_mul_add_words_end-bn_mul_add_words
-.ident "bn_mul_add_words"
-.text
- .align 16
-.globl bn_mul_words
- .type bn_mul_words,@function
-bn_mul_words:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- xorl %esi, %esi
- movl 20(%esp), %edi
- movl 24(%esp), %ebx
- movl 28(%esp), %ebp
- movl 32(%esp), %ecx
- andl $4294967288, %ebp
- jz .L004mw_finish
-.L005mw_loop:
- # Round 0
- movl (%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, (%edi)
- movl %edx, %esi
- # Round 4
- movl 4(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 4(%edi)
- movl %edx, %esi
- # Round 8
- movl 8(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 8(%edi)
- movl %edx, %esi
- # Round 12
- movl 12(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 12(%edi)
- movl %edx, %esi
- # Round 16
- movl 16(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 16(%edi)
- movl %edx, %esi
- # Round 20
- movl 20(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 20(%edi)
- movl %edx, %esi
- # Round 24
- movl 24(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 24(%edi)
- movl %edx, %esi
- # Round 28
- movl 28(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 28(%edi)
- movl %edx, %esi
-
- addl $32, %ebx
- addl $32, %edi
- subl $8, %ebp
- jz .L004mw_finish
- jmp .L005mw_loop
-.L004mw_finish:
- movl 28(%esp), %ebp
- andl $7, %ebp
- jnz .L006mw_finish2
- jmp .L007mw_end
-.align 16
-.L006mw_finish2:
- # Tail Round 0
- movl (%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, (%edi)
- movl %edx, %esi
- decl %ebp
- jz .L007mw_end
- # Tail Round 1
- movl 4(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 4(%edi)
- movl %edx, %esi
- decl %ebp
- jz .L007mw_end
- # Tail Round 2
- movl 8(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 8(%edi)
- movl %edx, %esi
- decl %ebp
- jz .L007mw_end
- # Tail Round 3
- movl 12(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 12(%edi)
- movl %edx, %esi
- decl %ebp
- jz .L007mw_end
- # Tail Round 4
- movl 16(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 16(%edi)
- movl %edx, %esi
- decl %ebp
- jz .L007mw_end
- # Tail Round 5
- movl 20(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 20(%edi)
- movl %edx, %esi
- decl %ebp
- jz .L007mw_end
- # Tail Round 6
- movl 24(%ebx), %eax
- mull %ecx
- addl %esi, %eax
- adcl $0, %edx
- movl %eax, 24(%edi)
- movl %edx, %esi
-.L007mw_end:
- movl %esi, %eax
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.bn_mul_words_end:
- .size bn_mul_words,.bn_mul_words_end-bn_mul_words
-.ident "bn_mul_words"
-.text
- .align 16
-.globl bn_sqr_words
- .type bn_sqr_words,@function
-bn_sqr_words:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- movl 20(%esp), %esi
- movl 24(%esp), %edi
- movl 28(%esp), %ebx
- andl $4294967288, %ebx
- jz .L008sw_finish
-.L009sw_loop:
- # Round 0
- movl (%edi), %eax
- mull %eax
- movl %eax, (%esi)
- movl %edx, 4(%esi)
- # Round 4
- movl 4(%edi), %eax
- mull %eax
- movl %eax, 8(%esi)
- movl %edx, 12(%esi)
- # Round 8
- movl 8(%edi), %eax
- mull %eax
- movl %eax, 16(%esi)
- movl %edx, 20(%esi)
- # Round 12
- movl 12(%edi), %eax
- mull %eax
- movl %eax, 24(%esi)
- movl %edx, 28(%esi)
- # Round 16
- movl 16(%edi), %eax
- mull %eax
- movl %eax, 32(%esi)
- movl %edx, 36(%esi)
- # Round 20
- movl 20(%edi), %eax
- mull %eax
- movl %eax, 40(%esi)
- movl %edx, 44(%esi)
- # Round 24
- movl 24(%edi), %eax
- mull %eax
- movl %eax, 48(%esi)
- movl %edx, 52(%esi)
- # Round 28
- movl 28(%edi), %eax
- mull %eax
- movl %eax, 56(%esi)
- movl %edx, 60(%esi)
-
- addl $32, %edi
- addl $64, %esi
- subl $8, %ebx
- jnz .L009sw_loop
-.L008sw_finish:
- movl 28(%esp), %ebx
- andl $7, %ebx
- jz .L010sw_end
- # Tail Round 0
- movl (%edi), %eax
- mull %eax
- movl %eax, (%esi)
- decl %ebx
- movl %edx, 4(%esi)
- jz .L010sw_end
- # Tail Round 1
- movl 4(%edi), %eax
- mull %eax
- movl %eax, 8(%esi)
- decl %ebx
- movl %edx, 12(%esi)
- jz .L010sw_end
- # Tail Round 2
- movl 8(%edi), %eax
- mull %eax
- movl %eax, 16(%esi)
- decl %ebx
- movl %edx, 20(%esi)
- jz .L010sw_end
- # Tail Round 3
- movl 12(%edi), %eax
- mull %eax
- movl %eax, 24(%esi)
- decl %ebx
- movl %edx, 28(%esi)
- jz .L010sw_end
- # Tail Round 4
- movl 16(%edi), %eax
- mull %eax
- movl %eax, 32(%esi)
- decl %ebx
- movl %edx, 36(%esi)
- jz .L010sw_end
- # Tail Round 5
- movl 20(%edi), %eax
- mull %eax
- movl %eax, 40(%esi)
- decl %ebx
- movl %edx, 44(%esi)
- jz .L010sw_end
- # Tail Round 6
- movl 24(%edi), %eax
- mull %eax
- movl %eax, 48(%esi)
- movl %edx, 52(%esi)
-.L010sw_end:
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.bn_sqr_words_end:
- .size bn_sqr_words,.bn_sqr_words_end-bn_sqr_words
-.ident "bn_sqr_words"
-.text
- .align 16
-.globl bn_div_words
- .type bn_div_words,@function
-bn_div_words:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
- movl 20(%esp), %edx
- movl 24(%esp), %eax
- movl 28(%esp), %ebx
- divl %ebx
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.bn_div_words_end:
- .size bn_div_words,.bn_div_words_end-bn_div_words
-.ident "bn_div_words"
-.text
- .align 16
-.globl bn_add_words
- .type bn_add_words,@function
-bn_add_words:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- movl 20(%esp), %ebx
- movl 24(%esp), %esi
- movl 28(%esp), %edi
- movl 32(%esp), %ebp
- xorl %eax, %eax
- andl $4294967288, %ebp
- jz .L011aw_finish
-.L012aw_loop:
- # Round 0
- movl (%esi), %ecx
- movl (%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, (%ebx)
- # Round 1
- movl 4(%esi), %ecx
- movl 4(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 4(%ebx)
- # Round 2
- movl 8(%esi), %ecx
- movl 8(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 8(%ebx)
- # Round 3
- movl 12(%esi), %ecx
- movl 12(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 12(%ebx)
- # Round 4
- movl 16(%esi), %ecx
- movl 16(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 16(%ebx)
- # Round 5
- movl 20(%esi), %ecx
- movl 20(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 20(%ebx)
- # Round 6
- movl 24(%esi), %ecx
- movl 24(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 24(%ebx)
- # Round 7
- movl 28(%esi), %ecx
- movl 28(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 28(%ebx)
-
- addl $32, %esi
- addl $32, %edi
- addl $32, %ebx
- subl $8, %ebp
- jnz .L012aw_loop
-.L011aw_finish:
- movl 32(%esp), %ebp
- andl $7, %ebp
- jz .L013aw_end
- # Tail Round 0
- movl (%esi), %ecx
- movl (%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, (%ebx)
- jz .L013aw_end
- # Tail Round 1
- movl 4(%esi), %ecx
- movl 4(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 4(%ebx)
- jz .L013aw_end
- # Tail Round 2
- movl 8(%esi), %ecx
- movl 8(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 8(%ebx)
- jz .L013aw_end
- # Tail Round 3
- movl 12(%esi), %ecx
- movl 12(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 12(%ebx)
- jz .L013aw_end
- # Tail Round 4
- movl 16(%esi), %ecx
- movl 16(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 16(%ebx)
- jz .L013aw_end
- # Tail Round 5
- movl 20(%esi), %ecx
- movl 20(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 20(%ebx)
- jz .L013aw_end
- # Tail Round 6
- movl 24(%esi), %ecx
- movl 24(%edi), %edx
- addl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- addl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 24(%ebx)
-.L013aw_end:
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.bn_add_words_end:
- .size bn_add_words,.bn_add_words_end-bn_add_words
-.ident "bn_add_words"
-.text
- .align 16
-.globl bn_sub_words
- .type bn_sub_words,@function
-bn_sub_words:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- movl 20(%esp), %ebx
- movl 24(%esp), %esi
- movl 28(%esp), %edi
- movl 32(%esp), %ebp
- xorl %eax, %eax
- andl $4294967288, %ebp
- jz .L014aw_finish
-.L015aw_loop:
- # Round 0
- movl (%esi), %ecx
- movl (%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, (%ebx)
- # Round 1
- movl 4(%esi), %ecx
- movl 4(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 4(%ebx)
- # Round 2
- movl 8(%esi), %ecx
- movl 8(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 8(%ebx)
- # Round 3
- movl 12(%esi), %ecx
- movl 12(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 12(%ebx)
- # Round 4
- movl 16(%esi), %ecx
- movl 16(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 16(%ebx)
- # Round 5
- movl 20(%esi), %ecx
- movl 20(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 20(%ebx)
- # Round 6
- movl 24(%esi), %ecx
- movl 24(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 24(%ebx)
- # Round 7
- movl 28(%esi), %ecx
- movl 28(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 28(%ebx)
-
- addl $32, %esi
- addl $32, %edi
- addl $32, %ebx
- subl $8, %ebp
- jnz .L015aw_loop
-.L014aw_finish:
- movl 32(%esp), %ebp
- andl $7, %ebp
- jz .L016aw_end
- # Tail Round 0
- movl (%esi), %ecx
- movl (%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, (%ebx)
- jz .L016aw_end
- # Tail Round 1
- movl 4(%esi), %ecx
- movl 4(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 4(%ebx)
- jz .L016aw_end
- # Tail Round 2
- movl 8(%esi), %ecx
- movl 8(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 8(%ebx)
- jz .L016aw_end
- # Tail Round 3
- movl 12(%esi), %ecx
- movl 12(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 12(%ebx)
- jz .L016aw_end
- # Tail Round 4
- movl 16(%esi), %ecx
- movl 16(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 16(%ebx)
- jz .L016aw_end
- # Tail Round 5
- movl 20(%esi), %ecx
- movl 20(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- decl %ebp
- movl %ecx, 20(%ebx)
- jz .L016aw_end
- # Tail Round 6
- movl 24(%esi), %ecx
- movl 24(%edi), %edx
- subl %eax, %ecx
- movl $0, %eax
- adcl %eax, %eax
- subl %edx, %ecx
- adcl $0, %eax
- movl %ecx, 24(%ebx)
-.L016aw_end:
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.bn_sub_words_end:
- .size bn_sub_words,.bn_sub_words_end-bn_sub_words
-.ident "bn_sub_words"
diff --git a/secure/lib/libcrypto/i386/cast-586.s b/secure/lib/libcrypto/i386/cast-586.s
deleted file mode 100644
index b6da53bb4fda..000000000000
--- a/secure/lib/libcrypto/i386/cast-586.s
+++ /dev/null
@@ -1,971 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by cast-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "cast-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl CAST_encrypt
- .type CAST_encrypt,@function
-CAST_encrypt:
-
- pushl %ebp
- pushl %ebx
- movl 12(%esp), %ebx
- movl 16(%esp), %ebp
- pushl %esi
- pushl %edi
- # Load the 2 words
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- # Get short key flag
- movl 128(%ebp), %eax
- pushl %eax
- xorl %eax, %eax
- # round 0
- movl (%ebp), %edx
- movl 4(%ebp), %ecx
- addl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %edi
- # round 1
- movl 8(%ebp), %edx
- movl 12(%ebp), %ecx
- xorl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %esi
- # round 2
- movl 16(%ebp), %edx
- movl 20(%ebp), %ecx
- subl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %edi
- # round 3
- movl 24(%ebp), %edx
- movl 28(%ebp), %ecx
- addl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %esi
- # round 4
- movl 32(%ebp), %edx
- movl 36(%ebp), %ecx
- xorl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %edi
- # round 5
- movl 40(%ebp), %edx
- movl 44(%ebp), %ecx
- subl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %esi
- # round 6
- movl 48(%ebp), %edx
- movl 52(%ebp), %ecx
- addl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %edi
- # round 7
- movl 56(%ebp), %edx
- movl 60(%ebp), %ecx
- xorl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %esi
- # round 8
- movl 64(%ebp), %edx
- movl 68(%ebp), %ecx
- subl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %edi
- # round 9
- movl 72(%ebp), %edx
- movl 76(%ebp), %ecx
- addl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %esi
- # round 10
- movl 80(%ebp), %edx
- movl 84(%ebp), %ecx
- xorl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %edi
- # round 11
- movl 88(%ebp), %edx
- movl 92(%ebp), %ecx
- subl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %esi
- # test short key flag
- popl %edx
- orl %edx, %edx
- jnz .L000cast_enc_done
- # round 12
- movl 96(%ebp), %edx
- movl 100(%ebp), %ecx
- addl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %edi
- # round 13
- movl 104(%ebp), %edx
- movl 108(%ebp), %ecx
- xorl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %esi
- # round 14
- movl 112(%ebp), %edx
- movl 116(%ebp), %ecx
- subl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %edi
- # round 15
- movl 120(%ebp), %edx
- movl 124(%ebp), %ecx
- addl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %esi
-.L000cast_enc_done:
- nop
- movl 20(%esp), %eax
- movl %edi, 4(%eax)
- movl %esi, (%eax)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.CAST_encrypt_end:
- .size CAST_encrypt,.CAST_encrypt_end-CAST_encrypt
-.ident "CAST_encrypt"
-.text
- .align 16
-.globl CAST_decrypt
- .type CAST_decrypt,@function
-CAST_decrypt:
-
- pushl %ebp
- pushl %ebx
- movl 12(%esp), %ebx
- movl 16(%esp), %ebp
- pushl %esi
- pushl %edi
- # Load the 2 words
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- # Get short key flag
- movl 128(%ebp), %eax
- orl %eax, %eax
- jnz .L001cast_dec_skip
- xorl %eax, %eax
- # round 15
- movl 120(%ebp), %edx
- movl 124(%ebp), %ecx
- addl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %edi
- # round 14
- movl 112(%ebp), %edx
- movl 116(%ebp), %ecx
- subl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %esi
- # round 13
- movl 104(%ebp), %edx
- movl 108(%ebp), %ecx
- xorl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %edi
- # round 12
- movl 96(%ebp), %edx
- movl 100(%ebp), %ecx
- addl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %esi
-.L001cast_dec_skip:
- # round 11
- movl 88(%ebp), %edx
- movl 92(%ebp), %ecx
- subl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %edi
- # round 10
- movl 80(%ebp), %edx
- movl 84(%ebp), %ecx
- xorl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %esi
- # round 9
- movl 72(%ebp), %edx
- movl 76(%ebp), %ecx
- addl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %edi
- # round 8
- movl 64(%ebp), %edx
- movl 68(%ebp), %ecx
- subl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %esi
- # round 7
- movl 56(%ebp), %edx
- movl 60(%ebp), %ecx
- xorl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %edi
- # round 6
- movl 48(%ebp), %edx
- movl 52(%ebp), %ecx
- addl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %esi
- # round 5
- movl 40(%ebp), %edx
- movl 44(%ebp), %ecx
- subl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %edi
- # round 4
- movl 32(%ebp), %edx
- movl 36(%ebp), %ecx
- xorl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %esi
- # round 3
- movl 24(%ebp), %edx
- movl 28(%ebp), %ecx
- addl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %edi
- # round 2
- movl 16(%ebp), %edx
- movl 20(%ebp), %ecx
- subl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- subl %ebx, %ecx
- xorl %ecx, %esi
- # round 1
- movl 8(%ebp), %edx
- movl 12(%ebp), %ecx
- xorl %esi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- addl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- xorl %ebx, %ecx
- xorl %ecx, %edi
- # round 0
- movl (%ebp), %edx
- movl 4(%ebp), %ecx
- addl %edi, %edx
- roll %cl, %edx
- movl %edx, %ebx
- xorl %ecx, %ecx
- movb %dh, %cl
- andl $255, %ebx
- shrl $16, %edx
- xorl %eax, %eax
- movb %dh, %al
- andl $255, %edx
- movl CAST_S_table0(,%ecx,4),%ecx
- movl CAST_S_table1(,%ebx,4),%ebx
- xorl %ebx, %ecx
- movl CAST_S_table2(,%eax,4),%ebx
- subl %ebx, %ecx
- movl CAST_S_table3(,%edx,4),%ebx
- addl %ebx, %ecx
- xorl %ecx, %esi
- nop
- movl 20(%esp), %eax
- movl %edi, 4(%eax)
- movl %esi, (%eax)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.CAST_decrypt_end:
- .size CAST_decrypt,.CAST_decrypt_end-CAST_decrypt
-.ident "CAST_decrypt"
-.text
- .align 16
-.globl CAST_cbc_encrypt
- .type CAST_cbc_encrypt,@function
-CAST_cbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
- # getting iv ptr from parameter 4
- movl 36(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
- # getting encrypt flag from parameter 5
- movl 56(%esp), %ecx
- # get and push parameter 3
- movl 48(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L002decrypt
- andl $4294967288, %ebp
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- jz .L003encrypt_finish
-.L004encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call CAST_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L004encrypt_loop
-.L003encrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L005finish
- xorl %ecx, %ecx
- xorl %edx, %edx
- movl .L006cbc_enc_jmp_table(,%ebp,4),%ebp
- jmp *%ebp
-.L007ej7:
- xorl %edx, %edx
- movb 6(%esi), %dh
- sall $8, %edx
-.L008ej6:
- movb 5(%esi), %dh
-.L009ej5:
- movb 4(%esi), %dl
-.L010ej4:
- movl (%esi), %ecx
- jmp .L011ejend
-.L012ej3:
- movb 2(%esi), %ch
- xorl %ecx, %ecx
- sall $8, %ecx
-.L013ej2:
- movb 1(%esi), %ch
-.L014ej1:
- movb (%esi), %cl
-.L011ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call CAST_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L005finish
-.align 16
-.L002decrypt:
- andl $4294967288, %ebp
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- jz .L015decrypt_finish
-.L016decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call CAST_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L016decrypt_loop
-.L015decrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L005finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call CAST_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 203 # bswapl %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L017dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L018dj6:
- movb %dh, 5(%edi)
-.L019dj5:
- movb %dl, 4(%edi)
-.L020dj4:
- movl %ecx, (%edi)
- jmp .L021djend
-.L022dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L023dj2:
- movb %ch, 1(%esi)
-.L024dj1:
- movb %cl, (%esi)
-.L021djend:
- jmp .L005finish
-.align 16
-.L005finish:
- movl 60(%esp), %ecx
- addl $24, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L006cbc_enc_jmp_table:
- .long 0
- .long .L014ej1
- .long .L013ej2
- .long .L012ej3
- .long .L010ej4
- .long .L009ej5
- .long .L008ej6
- .long .L007ej7
-.align 16
-.L025cbc_dec_jmp_table:
- .long 0
- .long .L024dj1
- .long .L023dj2
- .long .L022dj3
- .long .L020dj4
- .long .L019dj5
- .long .L018dj6
- .long .L017dj7
-.L_CAST_cbc_encrypt_end:
- .size CAST_cbc_encrypt,.L_CAST_cbc_encrypt_end-CAST_cbc_encrypt
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/co-586.s b/secure/lib/libcrypto/i386/co-586.s
deleted file mode 100644
index 084f6fe1838b..000000000000
--- a/secure/lib/libcrypto/i386/co-586.s
+++ /dev/null
@@ -1,1270 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by co-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "co-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl bn_mul_comba8
- .type bn_mul_comba8,@function
-bn_mul_comba8:
- pushl %esi
- movl 12(%esp), %esi
- pushl %edi
- movl 20(%esp), %edi
- pushl %ebp
- pushl %ebx
- xorl %ebx, %ebx
- movl (%esi), %eax
- xorl %ecx, %ecx
- movl (%edi), %edx
- # ################## Calculate word 0
- xorl %ebp, %ebp
- # mul a[0]*b[0]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- movl (%edi), %edx
- adcl $0, %ebp
- movl %ebx, (%eax)
- movl 4(%esi), %eax
- # saved r[0]
- # ################## Calculate word 1
- xorl %ebx, %ebx
- # mul a[1]*b[0]
- mull %edx
- addl %eax, %ecx
- movl (%esi), %eax
- adcl %edx, %ebp
- movl 4(%edi), %edx
- adcl $0, %ebx
- # mul a[0]*b[1]
- mull %edx
- addl %eax, %ecx
- movl 20(%esp), %eax
- adcl %edx, %ebp
- movl (%edi), %edx
- adcl $0, %ebx
- movl %ecx, 4(%eax)
- movl 8(%esi), %eax
- # saved r[1]
- # ################## Calculate word 2
- xorl %ecx, %ecx
- # mul a[2]*b[0]
- mull %edx
- addl %eax, %ebp
- movl 4(%esi), %eax
- adcl %edx, %ebx
- movl 4(%edi), %edx
- adcl $0, %ecx
- # mul a[1]*b[1]
- mull %edx
- addl %eax, %ebp
- movl (%esi), %eax
- adcl %edx, %ebx
- movl 8(%edi), %edx
- adcl $0, %ecx
- # mul a[0]*b[2]
- mull %edx
- addl %eax, %ebp
- movl 20(%esp), %eax
- adcl %edx, %ebx
- movl (%edi), %edx
- adcl $0, %ecx
- movl %ebp, 8(%eax)
- movl 12(%esi), %eax
- # saved r[2]
- # ################## Calculate word 3
- xorl %ebp, %ebp
- # mul a[3]*b[0]
- mull %edx
- addl %eax, %ebx
- movl 8(%esi), %eax
- adcl %edx, %ecx
- movl 4(%edi), %edx
- adcl $0, %ebp
- # mul a[2]*b[1]
- mull %edx
- addl %eax, %ebx
- movl 4(%esi), %eax
- adcl %edx, %ecx
- movl 8(%edi), %edx
- adcl $0, %ebp
- # mul a[1]*b[2]
- mull %edx
- addl %eax, %ebx
- movl (%esi), %eax
- adcl %edx, %ecx
- movl 12(%edi), %edx
- adcl $0, %ebp
- # mul a[0]*b[3]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- movl (%edi), %edx
- adcl $0, %ebp
- movl %ebx, 12(%eax)
- movl 16(%esi), %eax
- # saved r[3]
- # ################## Calculate word 4
- xorl %ebx, %ebx
- # mul a[4]*b[0]
- mull %edx
- addl %eax, %ecx
- movl 12(%esi), %eax
- adcl %edx, %ebp
- movl 4(%edi), %edx
- adcl $0, %ebx
- # mul a[3]*b[1]
- mull %edx
- addl %eax, %ecx
- movl 8(%esi), %eax
- adcl %edx, %ebp
- movl 8(%edi), %edx
- adcl $0, %ebx
- # mul a[2]*b[2]
- mull %edx
- addl %eax, %ecx
- movl 4(%esi), %eax
- adcl %edx, %ebp
- movl 12(%edi), %edx
- adcl $0, %ebx
- # mul a[1]*b[3]
- mull %edx
- addl %eax, %ecx
- movl (%esi), %eax
- adcl %edx, %ebp
- movl 16(%edi), %edx
- adcl $0, %ebx
- # mul a[0]*b[4]
- mull %edx
- addl %eax, %ecx
- movl 20(%esp), %eax
- adcl %edx, %ebp
- movl (%edi), %edx
- adcl $0, %ebx
- movl %ecx, 16(%eax)
- movl 20(%esi), %eax
- # saved r[4]
- # ################## Calculate word 5
- xorl %ecx, %ecx
- # mul a[5]*b[0]
- mull %edx
- addl %eax, %ebp
- movl 16(%esi), %eax
- adcl %edx, %ebx
- movl 4(%edi), %edx
- adcl $0, %ecx
- # mul a[4]*b[1]
- mull %edx
- addl %eax, %ebp
- movl 12(%esi), %eax
- adcl %edx, %ebx
- movl 8(%edi), %edx
- adcl $0, %ecx
- # mul a[3]*b[2]
- mull %edx
- addl %eax, %ebp
- movl 8(%esi), %eax
- adcl %edx, %ebx
- movl 12(%edi), %edx
- adcl $0, %ecx
- # mul a[2]*b[3]
- mull %edx
- addl %eax, %ebp
- movl 4(%esi), %eax
- adcl %edx, %ebx
- movl 16(%edi), %edx
- adcl $0, %ecx
- # mul a[1]*b[4]
- mull %edx
- addl %eax, %ebp
- movl (%esi), %eax
- adcl %edx, %ebx
- movl 20(%edi), %edx
- adcl $0, %ecx
- # mul a[0]*b[5]
- mull %edx
- addl %eax, %ebp
- movl 20(%esp), %eax
- adcl %edx, %ebx
- movl (%edi), %edx
- adcl $0, %ecx
- movl %ebp, 20(%eax)
- movl 24(%esi), %eax
- # saved r[5]
- # ################## Calculate word 6
- xorl %ebp, %ebp
- # mul a[6]*b[0]
- mull %edx
- addl %eax, %ebx
- movl 20(%esi), %eax
- adcl %edx, %ecx
- movl 4(%edi), %edx
- adcl $0, %ebp
- # mul a[5]*b[1]
- mull %edx
- addl %eax, %ebx
- movl 16(%esi), %eax
- adcl %edx, %ecx
- movl 8(%edi), %edx
- adcl $0, %ebp
- # mul a[4]*b[2]
- mull %edx
- addl %eax, %ebx
- movl 12(%esi), %eax
- adcl %edx, %ecx
- movl 12(%edi), %edx
- adcl $0, %ebp
- # mul a[3]*b[3]
- mull %edx
- addl %eax, %ebx
- movl 8(%esi), %eax
- adcl %edx, %ecx
- movl 16(%edi), %edx
- adcl $0, %ebp
- # mul a[2]*b[4]
- mull %edx
- addl %eax, %ebx
- movl 4(%esi), %eax
- adcl %edx, %ecx
- movl 20(%edi), %edx
- adcl $0, %ebp
- # mul a[1]*b[5]
- mull %edx
- addl %eax, %ebx
- movl (%esi), %eax
- adcl %edx, %ecx
- movl 24(%edi), %edx
- adcl $0, %ebp
- # mul a[0]*b[6]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- movl (%edi), %edx
- adcl $0, %ebp
- movl %ebx, 24(%eax)
- movl 28(%esi), %eax
- # saved r[6]
- # ################## Calculate word 7
- xorl %ebx, %ebx
- # mul a[7]*b[0]
- mull %edx
- addl %eax, %ecx
- movl 24(%esi), %eax
- adcl %edx, %ebp
- movl 4(%edi), %edx
- adcl $0, %ebx
- # mul a[6]*b[1]
- mull %edx
- addl %eax, %ecx
- movl 20(%esi), %eax
- adcl %edx, %ebp
- movl 8(%edi), %edx
- adcl $0, %ebx
- # mul a[5]*b[2]
- mull %edx
- addl %eax, %ecx
- movl 16(%esi), %eax
- adcl %edx, %ebp
- movl 12(%edi), %edx
- adcl $0, %ebx
- # mul a[4]*b[3]
- mull %edx
- addl %eax, %ecx
- movl 12(%esi), %eax
- adcl %edx, %ebp
- movl 16(%edi), %edx
- adcl $0, %ebx
- # mul a[3]*b[4]
- mull %edx
- addl %eax, %ecx
- movl 8(%esi), %eax
- adcl %edx, %ebp
- movl 20(%edi), %edx
- adcl $0, %ebx
- # mul a[2]*b[5]
- mull %edx
- addl %eax, %ecx
- movl 4(%esi), %eax
- adcl %edx, %ebp
- movl 24(%edi), %edx
- adcl $0, %ebx
- # mul a[1]*b[6]
- mull %edx
- addl %eax, %ecx
- movl (%esi), %eax
- adcl %edx, %ebp
- movl 28(%edi), %edx
- adcl $0, %ebx
- # mul a[0]*b[7]
- mull %edx
- addl %eax, %ecx
- movl 20(%esp), %eax
- adcl %edx, %ebp
- movl 4(%edi), %edx
- adcl $0, %ebx
- movl %ecx, 28(%eax)
- movl 28(%esi), %eax
- # saved r[7]
- # ################## Calculate word 8
- xorl %ecx, %ecx
- # mul a[7]*b[1]
- mull %edx
- addl %eax, %ebp
- movl 24(%esi), %eax
- adcl %edx, %ebx
- movl 8(%edi), %edx
- adcl $0, %ecx
- # mul a[6]*b[2]
- mull %edx
- addl %eax, %ebp
- movl 20(%esi), %eax
- adcl %edx, %ebx
- movl 12(%edi), %edx
- adcl $0, %ecx
- # mul a[5]*b[3]
- mull %edx
- addl %eax, %ebp
- movl 16(%esi), %eax
- adcl %edx, %ebx
- movl 16(%edi), %edx
- adcl $0, %ecx
- # mul a[4]*b[4]
- mull %edx
- addl %eax, %ebp
- movl 12(%esi), %eax
- adcl %edx, %ebx
- movl 20(%edi), %edx
- adcl $0, %ecx
- # mul a[3]*b[5]
- mull %edx
- addl %eax, %ebp
- movl 8(%esi), %eax
- adcl %edx, %ebx
- movl 24(%edi), %edx
- adcl $0, %ecx
- # mul a[2]*b[6]
- mull %edx
- addl %eax, %ebp
- movl 4(%esi), %eax
- adcl %edx, %ebx
- movl 28(%edi), %edx
- adcl $0, %ecx
- # mul a[1]*b[7]
- mull %edx
- addl %eax, %ebp
- movl 20(%esp), %eax
- adcl %edx, %ebx
- movl 8(%edi), %edx
- adcl $0, %ecx
- movl %ebp, 32(%eax)
- movl 28(%esi), %eax
- # saved r[8]
- # ################## Calculate word 9
- xorl %ebp, %ebp
- # mul a[7]*b[2]
- mull %edx
- addl %eax, %ebx
- movl 24(%esi), %eax
- adcl %edx, %ecx
- movl 12(%edi), %edx
- adcl $0, %ebp
- # mul a[6]*b[3]
- mull %edx
- addl %eax, %ebx
- movl 20(%esi), %eax
- adcl %edx, %ecx
- movl 16(%edi), %edx
- adcl $0, %ebp
- # mul a[5]*b[4]
- mull %edx
- addl %eax, %ebx
- movl 16(%esi), %eax
- adcl %edx, %ecx
- movl 20(%edi), %edx
- adcl $0, %ebp
- # mul a[4]*b[5]
- mull %edx
- addl %eax, %ebx
- movl 12(%esi), %eax
- adcl %edx, %ecx
- movl 24(%edi), %edx
- adcl $0, %ebp
- # mul a[3]*b[6]
- mull %edx
- addl %eax, %ebx
- movl 8(%esi), %eax
- adcl %edx, %ecx
- movl 28(%edi), %edx
- adcl $0, %ebp
- # mul a[2]*b[7]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- movl 12(%edi), %edx
- adcl $0, %ebp
- movl %ebx, 36(%eax)
- movl 28(%esi), %eax
- # saved r[9]
- # ################## Calculate word 10
- xorl %ebx, %ebx
- # mul a[7]*b[3]
- mull %edx
- addl %eax, %ecx
- movl 24(%esi), %eax
- adcl %edx, %ebp
- movl 16(%edi), %edx
- adcl $0, %ebx
- # mul a[6]*b[4]
- mull %edx
- addl %eax, %ecx
- movl 20(%esi), %eax
- adcl %edx, %ebp
- movl 20(%edi), %edx
- adcl $0, %ebx
- # mul a[5]*b[5]
- mull %edx
- addl %eax, %ecx
- movl 16(%esi), %eax
- adcl %edx, %ebp
- movl 24(%edi), %edx
- adcl $0, %ebx
- # mul a[4]*b[6]
- mull %edx
- addl %eax, %ecx
- movl 12(%esi), %eax
- adcl %edx, %ebp
- movl 28(%edi), %edx
- adcl $0, %ebx
- # mul a[3]*b[7]
- mull %edx
- addl %eax, %ecx
- movl 20(%esp), %eax
- adcl %edx, %ebp
- movl 16(%edi), %edx
- adcl $0, %ebx
- movl %ecx, 40(%eax)
- movl 28(%esi), %eax
- # saved r[10]
- # ################## Calculate word 11
- xorl %ecx, %ecx
- # mul a[7]*b[4]
- mull %edx
- addl %eax, %ebp
- movl 24(%esi), %eax
- adcl %edx, %ebx
- movl 20(%edi), %edx
- adcl $0, %ecx
- # mul a[6]*b[5]
- mull %edx
- addl %eax, %ebp
- movl 20(%esi), %eax
- adcl %edx, %ebx
- movl 24(%edi), %edx
- adcl $0, %ecx
- # mul a[5]*b[6]
- mull %edx
- addl %eax, %ebp
- movl 16(%esi), %eax
- adcl %edx, %ebx
- movl 28(%edi), %edx
- adcl $0, %ecx
- # mul a[4]*b[7]
- mull %edx
- addl %eax, %ebp
- movl 20(%esp), %eax
- adcl %edx, %ebx
- movl 20(%edi), %edx
- adcl $0, %ecx
- movl %ebp, 44(%eax)
- movl 28(%esi), %eax
- # saved r[11]
- # ################## Calculate word 12
- xorl %ebp, %ebp
- # mul a[7]*b[5]
- mull %edx
- addl %eax, %ebx
- movl 24(%esi), %eax
- adcl %edx, %ecx
- movl 24(%edi), %edx
- adcl $0, %ebp
- # mul a[6]*b[6]
- mull %edx
- addl %eax, %ebx
- movl 20(%esi), %eax
- adcl %edx, %ecx
- movl 28(%edi), %edx
- adcl $0, %ebp
- # mul a[5]*b[7]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- movl 24(%edi), %edx
- adcl $0, %ebp
- movl %ebx, 48(%eax)
- movl 28(%esi), %eax
- # saved r[12]
- # ################## Calculate word 13
- xorl %ebx, %ebx
- # mul a[7]*b[6]
- mull %edx
- addl %eax, %ecx
- movl 24(%esi), %eax
- adcl %edx, %ebp
- movl 28(%edi), %edx
- adcl $0, %ebx
- # mul a[6]*b[7]
- mull %edx
- addl %eax, %ecx
- movl 20(%esp), %eax
- adcl %edx, %ebp
- movl 28(%edi), %edx
- adcl $0, %ebx
- movl %ecx, 52(%eax)
- movl 28(%esi), %eax
- # saved r[13]
- # ################## Calculate word 14
- xorl %ecx, %ecx
- # mul a[7]*b[7]
- mull %edx
- addl %eax, %ebp
- movl 20(%esp), %eax
- adcl %edx, %ebx
- adcl $0, %ecx
- movl %ebp, 56(%eax)
- # saved r[14]
- # save r[15]
- movl %ebx, 60(%eax)
- popl %ebx
- popl %ebp
- popl %edi
- popl %esi
- ret
-.L_bn_mul_comba8_end:
- .size bn_mul_comba8,.L_bn_mul_comba8_end-bn_mul_comba8
-.ident "desasm.pl"
-.text
- .align 16
-.globl bn_mul_comba4
- .type bn_mul_comba4,@function
-bn_mul_comba4:
- pushl %esi
- movl 12(%esp), %esi
- pushl %edi
- movl 20(%esp), %edi
- pushl %ebp
- pushl %ebx
- xorl %ebx, %ebx
- movl (%esi), %eax
- xorl %ecx, %ecx
- movl (%edi), %edx
- # ################## Calculate word 0
- xorl %ebp, %ebp
- # mul a[0]*b[0]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- movl (%edi), %edx
- adcl $0, %ebp
- movl %ebx, (%eax)
- movl 4(%esi), %eax
- # saved r[0]
- # ################## Calculate word 1
- xorl %ebx, %ebx
- # mul a[1]*b[0]
- mull %edx
- addl %eax, %ecx
- movl (%esi), %eax
- adcl %edx, %ebp
- movl 4(%edi), %edx
- adcl $0, %ebx
- # mul a[0]*b[1]
- mull %edx
- addl %eax, %ecx
- movl 20(%esp), %eax
- adcl %edx, %ebp
- movl (%edi), %edx
- adcl $0, %ebx
- movl %ecx, 4(%eax)
- movl 8(%esi), %eax
- # saved r[1]
- # ################## Calculate word 2
- xorl %ecx, %ecx
- # mul a[2]*b[0]
- mull %edx
- addl %eax, %ebp
- movl 4(%esi), %eax
- adcl %edx, %ebx
- movl 4(%edi), %edx
- adcl $0, %ecx
- # mul a[1]*b[1]
- mull %edx
- addl %eax, %ebp
- movl (%esi), %eax
- adcl %edx, %ebx
- movl 8(%edi), %edx
- adcl $0, %ecx
- # mul a[0]*b[2]
- mull %edx
- addl %eax, %ebp
- movl 20(%esp), %eax
- adcl %edx, %ebx
- movl (%edi), %edx
- adcl $0, %ecx
- movl %ebp, 8(%eax)
- movl 12(%esi), %eax
- # saved r[2]
- # ################## Calculate word 3
- xorl %ebp, %ebp
- # mul a[3]*b[0]
- mull %edx
- addl %eax, %ebx
- movl 8(%esi), %eax
- adcl %edx, %ecx
- movl 4(%edi), %edx
- adcl $0, %ebp
- # mul a[2]*b[1]
- mull %edx
- addl %eax, %ebx
- movl 4(%esi), %eax
- adcl %edx, %ecx
- movl 8(%edi), %edx
- adcl $0, %ebp
- # mul a[1]*b[2]
- mull %edx
- addl %eax, %ebx
- movl (%esi), %eax
- adcl %edx, %ecx
- movl 12(%edi), %edx
- adcl $0, %ebp
- # mul a[0]*b[3]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- movl 4(%edi), %edx
- adcl $0, %ebp
- movl %ebx, 12(%eax)
- movl 12(%esi), %eax
- # saved r[3]
- # ################## Calculate word 4
- xorl %ebx, %ebx
- # mul a[3]*b[1]
- mull %edx
- addl %eax, %ecx
- movl 8(%esi), %eax
- adcl %edx, %ebp
- movl 8(%edi), %edx
- adcl $0, %ebx
- # mul a[2]*b[2]
- mull %edx
- addl %eax, %ecx
- movl 4(%esi), %eax
- adcl %edx, %ebp
- movl 12(%edi), %edx
- adcl $0, %ebx
- # mul a[1]*b[3]
- mull %edx
- addl %eax, %ecx
- movl 20(%esp), %eax
- adcl %edx, %ebp
- movl 8(%edi), %edx
- adcl $0, %ebx
- movl %ecx, 16(%eax)
- movl 12(%esi), %eax
- # saved r[4]
- # ################## Calculate word 5
- xorl %ecx, %ecx
- # mul a[3]*b[2]
- mull %edx
- addl %eax, %ebp
- movl 8(%esi), %eax
- adcl %edx, %ebx
- movl 12(%edi), %edx
- adcl $0, %ecx
- # mul a[2]*b[3]
- mull %edx
- addl %eax, %ebp
- movl 20(%esp), %eax
- adcl %edx, %ebx
- movl 12(%edi), %edx
- adcl $0, %ecx
- movl %ebp, 20(%eax)
- movl 12(%esi), %eax
- # saved r[5]
- # ################## Calculate word 6
- xorl %ebp, %ebp
- # mul a[3]*b[3]
- mull %edx
- addl %eax, %ebx
- movl 20(%esp), %eax
- adcl %edx, %ecx
- adcl $0, %ebp
- movl %ebx, 24(%eax)
- # saved r[6]
- # save r[7]
- movl %ecx, 28(%eax)
- popl %ebx
- popl %ebp
- popl %edi
- popl %esi
- ret
-.L_bn_mul_comba4_end:
- .size bn_mul_comba4,.L_bn_mul_comba4_end-bn_mul_comba4
-.ident "desasm.pl"
-.text
- .align 16
-.globl bn_sqr_comba8
- .type bn_sqr_comba8,@function
-bn_sqr_comba8:
- pushl %esi
- pushl %edi
- pushl %ebp
- pushl %ebx
- movl 20(%esp), %edi
- movl 24(%esp), %esi
- xorl %ebx, %ebx
- xorl %ecx, %ecx
- movl (%esi), %eax
- # ############### Calculate word 0
- xorl %ebp, %ebp
- # sqr a[0]*a[0]
- mull %eax
- addl %eax, %ebx
- adcl %edx, %ecx
- movl (%esi), %edx
- adcl $0, %ebp
- movl %ebx, (%edi)
- movl 4(%esi), %eax
- # saved r[0]
- # ############### Calculate word 1
- xorl %ebx, %ebx
- # sqr a[1]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 8(%esi), %eax
- adcl $0, %ebx
- movl %ecx, 4(%edi)
- movl (%esi), %edx
- # saved r[1]
- # ############### Calculate word 2
- xorl %ecx, %ecx
- # sqr a[2]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 4(%esi), %eax
- adcl $0, %ecx
- # sqr a[1]*a[1]
- mull %eax
- addl %eax, %ebp
- adcl %edx, %ebx
- movl (%esi), %edx
- adcl $0, %ecx
- movl %ebp, 8(%edi)
- movl 12(%esi), %eax
- # saved r[2]
- # ############### Calculate word 3
- xorl %ebp, %ebp
- # sqr a[3]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 8(%esi), %eax
- adcl $0, %ebp
- movl 4(%esi), %edx
- # sqr a[2]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 16(%esi), %eax
- adcl $0, %ebp
- movl %ebx, 12(%edi)
- movl (%esi), %edx
- # saved r[3]
- # ############### Calculate word 4
- xorl %ebx, %ebx
- # sqr a[4]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 12(%esi), %eax
- adcl $0, %ebx
- movl 4(%esi), %edx
- # sqr a[3]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 8(%esi), %eax
- adcl $0, %ebx
- # sqr a[2]*a[2]
- mull %eax
- addl %eax, %ecx
- adcl %edx, %ebp
- movl (%esi), %edx
- adcl $0, %ebx
- movl %ecx, 16(%edi)
- movl 20(%esi), %eax
- # saved r[4]
- # ############### Calculate word 5
- xorl %ecx, %ecx
- # sqr a[5]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 16(%esi), %eax
- adcl $0, %ecx
- movl 4(%esi), %edx
- # sqr a[4]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 12(%esi), %eax
- adcl $0, %ecx
- movl 8(%esi), %edx
- # sqr a[3]*a[2]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 24(%esi), %eax
- adcl $0, %ecx
- movl %ebp, 20(%edi)
- movl (%esi), %edx
- # saved r[5]
- # ############### Calculate word 6
- xorl %ebp, %ebp
- # sqr a[6]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 20(%esi), %eax
- adcl $0, %ebp
- movl 4(%esi), %edx
- # sqr a[5]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 16(%esi), %eax
- adcl $0, %ebp
- movl 8(%esi), %edx
- # sqr a[4]*a[2]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 12(%esi), %eax
- adcl $0, %ebp
- # sqr a[3]*a[3]
- mull %eax
- addl %eax, %ebx
- adcl %edx, %ecx
- movl (%esi), %edx
- adcl $0, %ebp
- movl %ebx, 24(%edi)
- movl 28(%esi), %eax
- # saved r[6]
- # ############### Calculate word 7
- xorl %ebx, %ebx
- # sqr a[7]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 24(%esi), %eax
- adcl $0, %ebx
- movl 4(%esi), %edx
- # sqr a[6]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 20(%esi), %eax
- adcl $0, %ebx
- movl 8(%esi), %edx
- # sqr a[5]*a[2]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 16(%esi), %eax
- adcl $0, %ebx
- movl 12(%esi), %edx
- # sqr a[4]*a[3]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 28(%esi), %eax
- adcl $0, %ebx
- movl %ecx, 28(%edi)
- movl 4(%esi), %edx
- # saved r[7]
- # ############### Calculate word 8
- xorl %ecx, %ecx
- # sqr a[7]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 24(%esi), %eax
- adcl $0, %ecx
- movl 8(%esi), %edx
- # sqr a[6]*a[2]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 20(%esi), %eax
- adcl $0, %ecx
- movl 12(%esi), %edx
- # sqr a[5]*a[3]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 16(%esi), %eax
- adcl $0, %ecx
- # sqr a[4]*a[4]
- mull %eax
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 8(%esi), %edx
- adcl $0, %ecx
- movl %ebp, 32(%edi)
- movl 28(%esi), %eax
- # saved r[8]
- # ############### Calculate word 9
- xorl %ebp, %ebp
- # sqr a[7]*a[2]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 24(%esi), %eax
- adcl $0, %ebp
- movl 12(%esi), %edx
- # sqr a[6]*a[3]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 20(%esi), %eax
- adcl $0, %ebp
- movl 16(%esi), %edx
- # sqr a[5]*a[4]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 28(%esi), %eax
- adcl $0, %ebp
- movl %ebx, 36(%edi)
- movl 12(%esi), %edx
- # saved r[9]
- # ############### Calculate word 10
- xorl %ebx, %ebx
- # sqr a[7]*a[3]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 24(%esi), %eax
- adcl $0, %ebx
- movl 16(%esi), %edx
- # sqr a[6]*a[4]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 20(%esi), %eax
- adcl $0, %ebx
- # sqr a[5]*a[5]
- mull %eax
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 16(%esi), %edx
- adcl $0, %ebx
- movl %ecx, 40(%edi)
- movl 28(%esi), %eax
- # saved r[10]
- # ############### Calculate word 11
- xorl %ecx, %ecx
- # sqr a[7]*a[4]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 24(%esi), %eax
- adcl $0, %ecx
- movl 20(%esi), %edx
- # sqr a[6]*a[5]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 28(%esi), %eax
- adcl $0, %ecx
- movl %ebp, 44(%edi)
- movl 20(%esi), %edx
- # saved r[11]
- # ############### Calculate word 12
- xorl %ebp, %ebp
- # sqr a[7]*a[5]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 24(%esi), %eax
- adcl $0, %ebp
- # sqr a[6]*a[6]
- mull %eax
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 24(%esi), %edx
- adcl $0, %ebp
- movl %ebx, 48(%edi)
- movl 28(%esi), %eax
- # saved r[12]
- # ############### Calculate word 13
- xorl %ebx, %ebx
- # sqr a[7]*a[6]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 28(%esi), %eax
- adcl $0, %ebx
- movl %ecx, 52(%edi)
- # saved r[13]
- # ############### Calculate word 14
- xorl %ecx, %ecx
- # sqr a[7]*a[7]
- mull %eax
- addl %eax, %ebp
- adcl %edx, %ebx
- adcl $0, %ecx
- movl %ebp, 56(%edi)
- # saved r[14]
- movl %ebx, 60(%edi)
- popl %ebx
- popl %ebp
- popl %edi
- popl %esi
- ret
-.L_bn_sqr_comba8_end:
- .size bn_sqr_comba8,.L_bn_sqr_comba8_end-bn_sqr_comba8
-.ident "desasm.pl"
-.text
- .align 16
-.globl bn_sqr_comba4
- .type bn_sqr_comba4,@function
-bn_sqr_comba4:
- pushl %esi
- pushl %edi
- pushl %ebp
- pushl %ebx
- movl 20(%esp), %edi
- movl 24(%esp), %esi
- xorl %ebx, %ebx
- xorl %ecx, %ecx
- movl (%esi), %eax
- # ############### Calculate word 0
- xorl %ebp, %ebp
- # sqr a[0]*a[0]
- mull %eax
- addl %eax, %ebx
- adcl %edx, %ecx
- movl (%esi), %edx
- adcl $0, %ebp
- movl %ebx, (%edi)
- movl 4(%esi), %eax
- # saved r[0]
- # ############### Calculate word 1
- xorl %ebx, %ebx
- # sqr a[1]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 8(%esi), %eax
- adcl $0, %ebx
- movl %ecx, 4(%edi)
- movl (%esi), %edx
- # saved r[1]
- # ############### Calculate word 2
- xorl %ecx, %ecx
- # sqr a[2]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 4(%esi), %eax
- adcl $0, %ecx
- # sqr a[1]*a[1]
- mull %eax
- addl %eax, %ebp
- adcl %edx, %ebx
- movl (%esi), %edx
- adcl $0, %ecx
- movl %ebp, 8(%edi)
- movl 12(%esi), %eax
- # saved r[2]
- # ############### Calculate word 3
- xorl %ebp, %ebp
- # sqr a[3]*a[0]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 8(%esi), %eax
- adcl $0, %ebp
- movl 4(%esi), %edx
- # sqr a[2]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebp
- addl %eax, %ebx
- adcl %edx, %ecx
- movl 12(%esi), %eax
- adcl $0, %ebp
- movl %ebx, 12(%edi)
- movl 4(%esi), %edx
- # saved r[3]
- # ############### Calculate word 4
- xorl %ebx, %ebx
- # sqr a[3]*a[1]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ebx
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 8(%esi), %eax
- adcl $0, %ebx
- # sqr a[2]*a[2]
- mull %eax
- addl %eax, %ecx
- adcl %edx, %ebp
- movl 8(%esi), %edx
- adcl $0, %ebx
- movl %ecx, 16(%edi)
- movl 12(%esi), %eax
- # saved r[4]
- # ############### Calculate word 5
- xorl %ecx, %ecx
- # sqr a[3]*a[2]
- mull %edx
- addl %eax, %eax
- adcl %edx, %edx
- adcl $0, %ecx
- addl %eax, %ebp
- adcl %edx, %ebx
- movl 12(%esi), %eax
- adcl $0, %ecx
- movl %ebp, 20(%edi)
- # saved r[5]
- # ############### Calculate word 6
- xorl %ebp, %ebp
- # sqr a[3]*a[3]
- mull %eax
- addl %eax, %ebx
- adcl %edx, %ecx
- adcl $0, %ebp
- movl %ebx, 24(%edi)
- # saved r[6]
- movl %ecx, 28(%edi)
- popl %ebx
- popl %ebp
- popl %edi
- popl %esi
- ret
-.L_bn_sqr_comba4_end:
- .size bn_sqr_comba4,.L_bn_sqr_comba4_end-bn_sqr_comba4
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/crypt586.s b/secure/lib/libcrypto/i386/crypt586.s
deleted file mode 100644
index e80834ed6d77..000000000000
--- a/secure/lib/libcrypto/i386/crypt586.s
+++ /dev/null
@@ -1,933 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by crypt586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "crypt586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl fcrypt_body
- .type fcrypt_body,@function
-fcrypt_body:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
-
- # Load the 2 words
- xorl %edi, %edi
- xorl %esi, %esi
- movl 24(%esp), %ebp
- pushl $25
-.L000start:
-
- # Round 0
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl (%ebp), %ebx
- xorl %ebx, %eax
- movl 4(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 1
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 8(%ebp), %ebx
- xorl %ebx, %eax
- movl 12(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 2
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 16(%ebp), %ebx
- xorl %ebx, %eax
- movl 20(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 3
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 24(%ebp), %ebx
- xorl %ebx, %eax
- movl 28(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 4
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 32(%ebp), %ebx
- xorl %ebx, %eax
- movl 36(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 5
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 40(%ebp), %ebx
- xorl %ebx, %eax
- movl 44(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 6
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 48(%ebp), %ebx
- xorl %ebx, %eax
- movl 52(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 7
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 56(%ebp), %ebx
- xorl %ebx, %eax
- movl 60(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 8
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 64(%ebp), %ebx
- xorl %ebx, %eax
- movl 68(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 9
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 72(%ebp), %ebx
- xorl %ebx, %eax
- movl 76(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 10
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 80(%ebp), %ebx
- xorl %ebx, %eax
- movl 84(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 11
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 88(%ebp), %ebx
- xorl %ebx, %eax
- movl 92(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 12
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 96(%ebp), %ebx
- xorl %ebx, %eax
- movl 100(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 13
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 104(%ebp), %ebx
- xorl %ebx, %eax
- movl 108(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 14
- movl 32(%esp), %eax
- movl %esi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %esi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 112(%ebp), %ebx
- xorl %ebx, %eax
- movl 116(%ebp), %ecx
- xorl %esi, %eax
- xorl %esi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 15
- movl 32(%esp), %eax
- movl %edi, %edx
- shrl $16, %edx
- movl 36(%esp), %ecx
- xorl %edi, %edx
- andl %edx, %eax
- andl %ecx, %edx
- movl %eax, %ebx
- sall $16, %ebx
- movl %edx, %ecx
- sall $16, %ecx
- xorl %ebx, %eax
- xorl %ecx, %edx
- movl 120(%ebp), %ebx
- xorl %ebx, %eax
- movl 124(%ebp), %ecx
- xorl %edi, %eax
- xorl %edi, %edx
- xorl %ecx, %edx
- andl $0xfcfcfcfc, %eax
- xorl %ebx, %ebx
- andl $0xcfcfcfcf, %edx
- xorl %ecx, %ecx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
- movl (%esp), %ebx
- movl %edi, %eax
- decl %ebx
- movl %esi, %edi
- movl %eax, %esi
- movl %ebx, (%esp)
- jnz .L000start
-
- # FP
- movl 24(%esp), %edx
-.byte 209
-.byte 207 # rorl $1 %edi
- movl %esi, %eax
- xorl %edi, %esi
- andl $0xaaaaaaaa, %esi
- xorl %esi, %eax
- xorl %esi, %edi
-
- roll $23, %eax
- movl %eax, %esi
- xorl %edi, %eax
- andl $0x03fc03fc, %eax
- xorl %eax, %esi
- xorl %eax, %edi
-
- roll $10, %esi
- movl %esi, %eax
- xorl %edi, %esi
- andl $0x33333333, %esi
- xorl %esi, %eax
- xorl %esi, %edi
-
- roll $18, %edi
- movl %edi, %esi
- xorl %eax, %edi
- andl $0xfff0000f, %edi
- xorl %edi, %esi
- xorl %edi, %eax
-
- roll $12, %esi
- movl %esi, %edi
- xorl %eax, %esi
- andl $0xf0f0f0f0, %esi
- xorl %esi, %edi
- xorl %esi, %eax
-
- rorl $4, %eax
- movl %eax, (%edx)
- movl %edi, 4(%edx)
- popl %ecx
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.fcrypt_body_end:
- .size fcrypt_body,.fcrypt_body_end-fcrypt_body
-.ident "fcrypt_body"
diff --git a/secure/lib/libcrypto/i386/des-586.s b/secure/lib/libcrypto/i386/des-586.s
deleted file mode 100644
index 55ddd71d88ae..000000000000
--- a/secure/lib/libcrypto/i386/des-586.s
+++ /dev/null
@@ -1,3154 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by des-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "des-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl des_encrypt1
- .type des_encrypt1,@function
-des_encrypt1:
- pushl %esi
- pushl %edi
-
- # Load the 2 words
- movl 12(%esp), %esi
- xorl %ecx, %ecx
- pushl %ebx
- pushl %ebp
- movl (%esi), %eax
- movl 28(%esp), %ebx
- movl 4(%esi), %edi
-
- # IP
- roll $4, %eax
- movl %eax, %esi
- xorl %edi, %eax
- andl $0xf0f0f0f0, %eax
- xorl %eax, %esi
- xorl %eax, %edi
-
- roll $20, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xfff0000f, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $14, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x33333333, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $22, %esi
- movl %esi, %eax
- xorl %edi, %esi
- andl $0x03fc03fc, %esi
- xorl %esi, %eax
- xorl %esi, %edi
-
- roll $9, %eax
- movl %eax, %esi
- xorl %edi, %eax
- andl $0xaaaaaaaa, %eax
- xorl %eax, %esi
- xorl %eax, %edi
-
-.byte 209
-.byte 199 # roll $1 %edi
- movl 24(%esp), %ebp
- cmpl $0, %ebx
- je .L000start_decrypt
-
- # Round 0
- movl (%ebp), %eax
- xorl %ebx, %ebx
- movl 4(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 1
- movl 8(%ebp), %eax
- xorl %ebx, %ebx
- movl 12(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 2
- movl 16(%ebp), %eax
- xorl %ebx, %ebx
- movl 20(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 3
- movl 24(%ebp), %eax
- xorl %ebx, %ebx
- movl 28(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 4
- movl 32(%ebp), %eax
- xorl %ebx, %ebx
- movl 36(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 5
- movl 40(%ebp), %eax
- xorl %ebx, %ebx
- movl 44(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 6
- movl 48(%ebp), %eax
- xorl %ebx, %ebx
- movl 52(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 7
- movl 56(%ebp), %eax
- xorl %ebx, %ebx
- movl 60(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 8
- movl 64(%ebp), %eax
- xorl %ebx, %ebx
- movl 68(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 9
- movl 72(%ebp), %eax
- xorl %ebx, %ebx
- movl 76(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 10
- movl 80(%ebp), %eax
- xorl %ebx, %ebx
- movl 84(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 11
- movl 88(%ebp), %eax
- xorl %ebx, %ebx
- movl 92(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 12
- movl 96(%ebp), %eax
- xorl %ebx, %ebx
- movl 100(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 13
- movl 104(%ebp), %eax
- xorl %ebx, %ebx
- movl 108(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 14
- movl 112(%ebp), %eax
- xorl %ebx, %ebx
- movl 116(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 15
- movl 120(%ebp), %eax
- xorl %ebx, %ebx
- movl 124(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
- jmp .L001end
-.L000start_decrypt:
-
- # Round 15
- movl 120(%ebp), %eax
- xorl %ebx, %ebx
- movl 124(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 14
- movl 112(%ebp), %eax
- xorl %ebx, %ebx
- movl 116(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 13
- movl 104(%ebp), %eax
- xorl %ebx, %ebx
- movl 108(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 12
- movl 96(%ebp), %eax
- xorl %ebx, %ebx
- movl 100(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 11
- movl 88(%ebp), %eax
- xorl %ebx, %ebx
- movl 92(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 10
- movl 80(%ebp), %eax
- xorl %ebx, %ebx
- movl 84(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 9
- movl 72(%ebp), %eax
- xorl %ebx, %ebx
- movl 76(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 8
- movl 64(%ebp), %eax
- xorl %ebx, %ebx
- movl 68(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 7
- movl 56(%ebp), %eax
- xorl %ebx, %ebx
- movl 60(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 6
- movl 48(%ebp), %eax
- xorl %ebx, %ebx
- movl 52(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 5
- movl 40(%ebp), %eax
- xorl %ebx, %ebx
- movl 44(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 4
- movl 32(%ebp), %eax
- xorl %ebx, %ebx
- movl 36(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 3
- movl 24(%ebp), %eax
- xorl %ebx, %ebx
- movl 28(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 2
- movl 16(%ebp), %eax
- xorl %ebx, %ebx
- movl 20(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 1
- movl 8(%ebp), %eax
- xorl %ebx, %ebx
- movl 12(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 0
- movl (%ebp), %eax
- xorl %ebx, %ebx
- movl 4(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-.L001end:
-
- # FP
- movl 20(%esp), %edx
-.byte 209
-.byte 206 # rorl $1 %esi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $23, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x03fc03fc, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $10, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0x33333333, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $18, %esi
- movl %esi, %edi
- xorl %eax, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %eax
-
- roll $12, %edi
- movl %edi, %esi
- xorl %eax, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %esi
- xorl %edi, %eax
-
- rorl $4, %eax
- movl %eax, (%edx)
- movl %esi, 4(%edx)
- popl %ebp
- popl %ebx
- popl %edi
- popl %esi
- ret
-.L_des_encrypt1_end:
- .size des_encrypt1,.L_des_encrypt1_end-des_encrypt1
-.ident "desasm.pl"
-.text
- .align 16
-.globl des_encrypt2
- .type des_encrypt2,@function
-des_encrypt2:
- pushl %esi
- pushl %edi
-
- # Load the 2 words
- movl 12(%esp), %eax
- xorl %ecx, %ecx
- pushl %ebx
- pushl %ebp
- movl (%eax), %esi
- movl 28(%esp), %ebx
- roll $3, %esi
- movl 4(%eax), %edi
- roll $3, %edi
- movl 24(%esp), %ebp
- cmpl $0, %ebx
- je .L002start_decrypt
-
- # Round 0
- movl (%ebp), %eax
- xorl %ebx, %ebx
- movl 4(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 1
- movl 8(%ebp), %eax
- xorl %ebx, %ebx
- movl 12(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 2
- movl 16(%ebp), %eax
- xorl %ebx, %ebx
- movl 20(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 3
- movl 24(%ebp), %eax
- xorl %ebx, %ebx
- movl 28(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 4
- movl 32(%ebp), %eax
- xorl %ebx, %ebx
- movl 36(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 5
- movl 40(%ebp), %eax
- xorl %ebx, %ebx
- movl 44(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 6
- movl 48(%ebp), %eax
- xorl %ebx, %ebx
- movl 52(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 7
- movl 56(%ebp), %eax
- xorl %ebx, %ebx
- movl 60(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 8
- movl 64(%ebp), %eax
- xorl %ebx, %ebx
- movl 68(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 9
- movl 72(%ebp), %eax
- xorl %ebx, %ebx
- movl 76(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 10
- movl 80(%ebp), %eax
- xorl %ebx, %ebx
- movl 84(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 11
- movl 88(%ebp), %eax
- xorl %ebx, %ebx
- movl 92(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 12
- movl 96(%ebp), %eax
- xorl %ebx, %ebx
- movl 100(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 13
- movl 104(%ebp), %eax
- xorl %ebx, %ebx
- movl 108(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 14
- movl 112(%ebp), %eax
- xorl %ebx, %ebx
- movl 116(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 15
- movl 120(%ebp), %eax
- xorl %ebx, %ebx
- movl 124(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
- jmp .L003end
-.L002start_decrypt:
-
- # Round 15
- movl 120(%ebp), %eax
- xorl %ebx, %ebx
- movl 124(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 14
- movl 112(%ebp), %eax
- xorl %ebx, %ebx
- movl 116(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 13
- movl 104(%ebp), %eax
- xorl %ebx, %ebx
- movl 108(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 12
- movl 96(%ebp), %eax
- xorl %ebx, %ebx
- movl 100(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 11
- movl 88(%ebp), %eax
- xorl %ebx, %ebx
- movl 92(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 10
- movl 80(%ebp), %eax
- xorl %ebx, %ebx
- movl 84(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 9
- movl 72(%ebp), %eax
- xorl %ebx, %ebx
- movl 76(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 8
- movl 64(%ebp), %eax
- xorl %ebx, %ebx
- movl 68(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 7
- movl 56(%ebp), %eax
- xorl %ebx, %ebx
- movl 60(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 6
- movl 48(%ebp), %eax
- xorl %ebx, %ebx
- movl 52(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 5
- movl 40(%ebp), %eax
- xorl %ebx, %ebx
- movl 44(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 4
- movl 32(%ebp), %eax
- xorl %ebx, %ebx
- movl 36(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 3
- movl 24(%ebp), %eax
- xorl %ebx, %ebx
- movl 28(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 2
- movl 16(%ebp), %eax
- xorl %ebx, %ebx
- movl 20(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-
- # Round 1
- movl 8(%ebp), %eax
- xorl %ebx, %ebx
- movl 12(%ebp), %edx
- xorl %esi, %eax
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %edi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %edi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %edi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %edi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %edi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %edi
-
- # Round 0
- movl (%ebp), %eax
- xorl %ebx, %ebx
- movl 4(%ebp), %edx
- xorl %edi, %eax
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- movl des_SPtrans(%ebx),%ebp
- movb %dl, %bl
- xorl %ebp, %esi
- movl 0x200+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movb %dh, %cl
- shrl $16, %eax
- movl 0x100+des_SPtrans(%ebx),%ebp
- xorl %ebp, %esi
- movb %ah, %bl
- shrl $16, %edx
- movl 0x300+des_SPtrans(%ecx),%ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- movl 0x600+des_SPtrans(%ebx),%ebx
- xorl %ebx, %esi
- movl 0x700+des_SPtrans(%ecx),%ebx
- xorl %ebx, %esi
- movl 0x400+des_SPtrans(%eax),%ebx
- xorl %ebx, %esi
- movl 0x500+des_SPtrans(%edx),%ebx
- xorl %ebx, %esi
-.L003end:
-
- # Fixup
- rorl $3, %edi
- movl 20(%esp), %eax
- rorl $3, %esi
- movl %edi, (%eax)
- movl %esi, 4(%eax)
- popl %ebp
- popl %ebx
- popl %edi
- popl %esi
- ret
-.L_des_encrypt2_end:
- .size des_encrypt2,.L_des_encrypt2_end-des_encrypt2
-.ident "desasm.pl"
-.text
- .align 16
-.globl des_encrypt3
- .type des_encrypt3,@function
-des_encrypt3:
- pushl %ebx
- movl 8(%esp), %ebx
- pushl %ebp
- pushl %esi
- pushl %edi
-
- # Load the data words
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- subl $12, %esp
-
- # IP
- roll $4, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- roll $20, %esi
- movl %esi, %edi
- xorl %edx, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %edx
-
- roll $14, %edi
- movl %edi, %esi
- xorl %edx, %edi
- andl $0x33333333, %edi
- xorl %edi, %esi
- xorl %edi, %edx
-
- roll $22, %edx
- movl %edx, %edi
- xorl %esi, %edx
- andl $0x03fc03fc, %edx
- xorl %edx, %edi
- xorl %edx, %esi
-
- roll $9, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- rorl $3, %edx
- rorl $2, %esi
- movl %esi, 4(%ebx)
- movl 36(%esp), %eax
- movl %edx, (%ebx)
- movl 40(%esp), %edi
- movl 44(%esp), %esi
- movl $1, 8(%esp)
- movl %eax, 4(%esp)
- movl %ebx, (%esp)
- call des_encrypt2
- movl $0, 8(%esp)
- movl %edi, 4(%esp)
- movl %ebx, (%esp)
- call des_encrypt2
- movl $1, 8(%esp)
- movl %esi, 4(%esp)
- movl %ebx, (%esp)
- call des_encrypt2
- addl $12, %esp
- movl (%ebx), %edi
- movl 4(%ebx), %esi
-
- # FP
- roll $2, %esi
- roll $3, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $23, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x03fc03fc, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $10, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0x33333333, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $18, %esi
- movl %esi, %edi
- xorl %eax, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %eax
-
- roll $12, %edi
- movl %edi, %esi
- xorl %eax, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %esi
- xorl %edi, %eax
-
- rorl $4, %eax
- movl %eax, (%ebx)
- movl %esi, 4(%ebx)
- popl %edi
- popl %esi
- popl %ebp
- popl %ebx
- ret
-.L_des_encrypt3_end:
- .size des_encrypt3,.L_des_encrypt3_end-des_encrypt3
-.ident "desasm.pl"
-.text
- .align 16
-.globl des_decrypt3
- .type des_decrypt3,@function
-des_decrypt3:
- pushl %ebx
- movl 8(%esp), %ebx
- pushl %ebp
- pushl %esi
- pushl %edi
-
- # Load the data words
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- subl $12, %esp
-
- # IP
- roll $4, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- roll $20, %esi
- movl %esi, %edi
- xorl %edx, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %edx
-
- roll $14, %edi
- movl %edi, %esi
- xorl %edx, %edi
- andl $0x33333333, %edi
- xorl %edi, %esi
- xorl %edi, %edx
-
- roll $22, %edx
- movl %edx, %edi
- xorl %esi, %edx
- andl $0x03fc03fc, %edx
- xorl %edx, %edi
- xorl %edx, %esi
-
- roll $9, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- rorl $3, %edx
- rorl $2, %esi
- movl %esi, 4(%ebx)
- movl 36(%esp), %esi
- movl %edx, (%ebx)
- movl 40(%esp), %edi
- movl 44(%esp), %eax
- movl $0, 8(%esp)
- movl %eax, 4(%esp)
- movl %ebx, (%esp)
- call des_encrypt2
- movl $1, 8(%esp)
- movl %edi, 4(%esp)
- movl %ebx, (%esp)
- call des_encrypt2
- movl $0, 8(%esp)
- movl %esi, 4(%esp)
- movl %ebx, (%esp)
- call des_encrypt2
- addl $12, %esp
- movl (%ebx), %edi
- movl 4(%ebx), %esi
-
- # FP
- roll $2, %esi
- roll $3, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $23, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x03fc03fc, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $10, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0x33333333, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $18, %esi
- movl %esi, %edi
- xorl %eax, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %eax
-
- roll $12, %edi
- movl %edi, %esi
- xorl %eax, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %esi
- xorl %edi, %eax
-
- rorl $4, %eax
- movl %eax, (%ebx)
- movl %esi, 4(%ebx)
- popl %edi
- popl %esi
- popl %ebp
- popl %ebx
- ret
-.L_des_decrypt3_end:
- .size des_decrypt3,.L_des_decrypt3_end-des_decrypt3
-.ident "desasm.pl"
-.text
- .align 16
-.globl des_ncbc_encrypt
- .type des_ncbc_encrypt,@function
-des_ncbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
- # getting iv ptr from parameter 4
- movl 36(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
- # getting encrypt flag from parameter 5
- movl 56(%esp), %ecx
- # get and push parameter 5
- pushl %ecx
- # get and push parameter 3
- movl 52(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L004decrypt
- andl $4294967288, %ebp
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- jz .L005encrypt_finish
-.L006encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call des_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L006encrypt_loop
-.L005encrypt_finish:
- movl 56(%esp), %ebp
- andl $7, %ebp
- jz .L007finish
- xorl %ecx, %ecx
- xorl %edx, %edx
- movl .L008cbc_enc_jmp_table(,%ebp,4),%ebp
- jmp *%ebp
-.L009ej7:
- movb 6(%esi), %dh
- sall $8, %edx
-.L010ej6:
- movb 5(%esi), %dh
-.L011ej5:
- movb 4(%esi), %dl
-.L012ej4:
- movl (%esi), %ecx
- jmp .L013ejend
-.L014ej3:
- movb 2(%esi), %ch
- sall $8, %ecx
-.L015ej2:
- movb 1(%esi), %ch
-.L016ej1:
- movb (%esi), %cl
-.L013ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call des_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L007finish
-.align 16
-.L004decrypt:
- andl $4294967288, %ebp
- movl 20(%esp), %eax
- movl 24(%esp), %ebx
- jz .L017decrypt_finish
-.L018decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call des_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl 20(%esp), %ecx
- movl 24(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 20(%esp)
- movl %ebx, 24(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L018decrypt_loop
-.L017decrypt_finish:
- movl 56(%esp), %ebp
- andl $7, %ebp
- jz .L007finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call des_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl 20(%esp), %ecx
- movl 24(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L019dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L020dj6:
- movb %dh, 5(%edi)
-.L021dj5:
- movb %dl, 4(%edi)
-.L022dj4:
- movl %ecx, (%edi)
- jmp .L023djend
-.L024dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L025dj2:
- movb %ch, 1(%esi)
-.L026dj1:
- movb %cl, (%esi)
-.L023djend:
- jmp .L007finish
-.align 16
-.L007finish:
- movl 64(%esp), %ecx
- addl $28, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L008cbc_enc_jmp_table:
- .long 0
- .long .L016ej1
- .long .L015ej2
- .long .L014ej3
- .long .L012ej4
- .long .L011ej5
- .long .L010ej6
- .long .L009ej7
-.align 16
-.L027cbc_dec_jmp_table:
- .long 0
- .long .L026dj1
- .long .L025dj2
- .long .L024dj3
- .long .L022dj4
- .long .L021dj5
- .long .L020dj6
- .long .L019dj7
-.L_des_ncbc_encrypt_end:
- .size des_ncbc_encrypt,.L_des_ncbc_encrypt_end-des_ncbc_encrypt
-.ident "desasm.pl"
-.text
- .align 16
-.globl des_ede3_cbc_encrypt
- .type des_ede3_cbc_encrypt,@function
-des_ede3_cbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
- # getting iv ptr from parameter 6
- movl 44(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
- # getting encrypt flag from parameter 7
- movl 64(%esp), %ecx
- # get and push parameter 5
- movl 56(%esp), %eax
- pushl %eax
- # get and push parameter 4
- movl 56(%esp), %eax
- pushl %eax
- # get and push parameter 3
- movl 56(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L028decrypt
- andl $4294967288, %ebp
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- jz .L029encrypt_finish
-.L030encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call des_encrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L030encrypt_loop
-.L029encrypt_finish:
- movl 60(%esp), %ebp
- andl $7, %ebp
- jz .L031finish
- xorl %ecx, %ecx
- xorl %edx, %edx
- movl .L032cbc_enc_jmp_table(,%ebp,4),%ebp
- jmp *%ebp
-.L033ej7:
- movb 6(%esi), %dh
- sall $8, %edx
-.L034ej6:
- movb 5(%esi), %dh
-.L035ej5:
- movb 4(%esi), %dl
-.L036ej4:
- movl (%esi), %ecx
- jmp .L037ejend
-.L038ej3:
- movb 2(%esi), %ch
- sall $8, %ecx
-.L039ej2:
- movb 1(%esi), %ch
-.L040ej1:
- movb (%esi), %cl
-.L037ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call des_encrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L031finish
-.align 16
-.L028decrypt:
- andl $4294967288, %ebp
- movl 24(%esp), %eax
- movl 28(%esp), %ebx
- jz .L041decrypt_finish
-.L042decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call des_decrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl 24(%esp), %ecx
- movl 28(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 24(%esp)
- movl %ebx, 28(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L042decrypt_loop
-.L041decrypt_finish:
- movl 60(%esp), %ebp
- andl $7, %ebp
- jz .L031finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call des_decrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl 24(%esp), %ecx
- movl 28(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L043dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L044dj6:
- movb %dh, 5(%edi)
-.L045dj5:
- movb %dl, 4(%edi)
-.L046dj4:
- movl %ecx, (%edi)
- jmp .L047djend
-.L048dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L049dj2:
- movb %ch, 1(%esi)
-.L050dj1:
- movb %cl, (%esi)
-.L047djend:
- jmp .L031finish
-.align 16
-.L031finish:
- movl 76(%esp), %ecx
- addl $32, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L032cbc_enc_jmp_table:
- .long 0
- .long .L040ej1
- .long .L039ej2
- .long .L038ej3
- .long .L036ej4
- .long .L035ej5
- .long .L034ej6
- .long .L033ej7
-.align 16
-.L051cbc_dec_jmp_table:
- .long 0
- .long .L050dj1
- .long .L049dj2
- .long .L048dj3
- .long .L046dj4
- .long .L045dj5
- .long .L044dj6
- .long .L043dj7
-.L_des_ede3_cbc_encrypt_end:
- .size des_ede3_cbc_encrypt,.L_des_ede3_cbc_encrypt_end-des_ede3_cbc_encrypt
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/md5-586.s b/secure/lib/libcrypto/i386/md5-586.s
deleted file mode 100644
index 5816cc570c81..000000000000
--- a/secure/lib/libcrypto/i386/md5-586.s
+++ /dev/null
@@ -1,689 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by md5-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "md5-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl md5_block_asm_host_order
- .type md5_block_asm_host_order,@function
-md5_block_asm_host_order:
- pushl %esi
- pushl %edi
- movl 12(%esp), %edi
- movl 16(%esp), %esi
- movl 20(%esp), %ecx
- pushl %ebp
- sall $6, %ecx
- pushl %ebx
- addl %esi, %ecx
- subl $64, %ecx
- movl (%edi), %eax
- pushl %ecx
- movl 4(%edi), %ebx
- movl 8(%edi), %ecx
- movl 12(%edi), %edx
-.L000start:
-
- # R0 section
- movl %ecx, %edi
- movl (%esi), %ebp
- # R0 0
- xorl %edx, %edi
- andl %ebx, %edi
- leal 3614090360(%eax,%ebp,1),%eax
- xorl %edx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $7, %eax
- movl 4(%esi), %ebp
- addl %ebx, %eax
- # R0 1
- xorl %ecx, %edi
- andl %eax, %edi
- leal 3905402710(%edx,%ebp,1),%edx
- xorl %ecx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $12, %edx
- movl 8(%esi), %ebp
- addl %eax, %edx
- # R0 2
- xorl %ebx, %edi
- andl %edx, %edi
- leal 606105819(%ecx,%ebp,1),%ecx
- xorl %ebx, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $17, %ecx
- movl 12(%esi), %ebp
- addl %edx, %ecx
- # R0 3
- xorl %eax, %edi
- andl %ecx, %edi
- leal 3250441966(%ebx,%ebp,1),%ebx
- xorl %eax, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $22, %ebx
- movl 16(%esi), %ebp
- addl %ecx, %ebx
- # R0 4
- xorl %edx, %edi
- andl %ebx, %edi
- leal 4118548399(%eax,%ebp,1),%eax
- xorl %edx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $7, %eax
- movl 20(%esi), %ebp
- addl %ebx, %eax
- # R0 5
- xorl %ecx, %edi
- andl %eax, %edi
- leal 1200080426(%edx,%ebp,1),%edx
- xorl %ecx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $12, %edx
- movl 24(%esi), %ebp
- addl %eax, %edx
- # R0 6
- xorl %ebx, %edi
- andl %edx, %edi
- leal 2821735955(%ecx,%ebp,1),%ecx
- xorl %ebx, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $17, %ecx
- movl 28(%esi), %ebp
- addl %edx, %ecx
- # R0 7
- xorl %eax, %edi
- andl %ecx, %edi
- leal 4249261313(%ebx,%ebp,1),%ebx
- xorl %eax, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $22, %ebx
- movl 32(%esi), %ebp
- addl %ecx, %ebx
- # R0 8
- xorl %edx, %edi
- andl %ebx, %edi
- leal 1770035416(%eax,%ebp,1),%eax
- xorl %edx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $7, %eax
- movl 36(%esi), %ebp
- addl %ebx, %eax
- # R0 9
- xorl %ecx, %edi
- andl %eax, %edi
- leal 2336552879(%edx,%ebp,1),%edx
- xorl %ecx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $12, %edx
- movl 40(%esi), %ebp
- addl %eax, %edx
- # R0 10
- xorl %ebx, %edi
- andl %edx, %edi
- leal 4294925233(%ecx,%ebp,1),%ecx
- xorl %ebx, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $17, %ecx
- movl 44(%esi), %ebp
- addl %edx, %ecx
- # R0 11
- xorl %eax, %edi
- andl %ecx, %edi
- leal 2304563134(%ebx,%ebp,1),%ebx
- xorl %eax, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $22, %ebx
- movl 48(%esi), %ebp
- addl %ecx, %ebx
- # R0 12
- xorl %edx, %edi
- andl %ebx, %edi
- leal 1804603682(%eax,%ebp,1),%eax
- xorl %edx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $7, %eax
- movl 52(%esi), %ebp
- addl %ebx, %eax
- # R0 13
- xorl %ecx, %edi
- andl %eax, %edi
- leal 4254626195(%edx,%ebp,1),%edx
- xorl %ecx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $12, %edx
- movl 56(%esi), %ebp
- addl %eax, %edx
- # R0 14
- xorl %ebx, %edi
- andl %edx, %edi
- leal 2792965006(%ecx,%ebp,1),%ecx
- xorl %ebx, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $17, %ecx
- movl 60(%esi), %ebp
- addl %edx, %ecx
- # R0 15
- xorl %eax, %edi
- andl %ecx, %edi
- leal 1236535329(%ebx,%ebp,1),%ebx
- xorl %eax, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $22, %ebx
- movl 4(%esi), %ebp
- addl %ecx, %ebx
-
- # R1 section
- # R1 16
- leal 4129170786(%eax,%ebp,1),%eax
- xorl %ebx, %edi
- andl %edx, %edi
- movl 24(%esi), %ebp
- xorl %ecx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $5, %eax
- addl %ebx, %eax
- # R1 17
- leal 3225465664(%edx,%ebp,1),%edx
- xorl %eax, %edi
- andl %ecx, %edi
- movl 44(%esi), %ebp
- xorl %ebx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $9, %edx
- addl %eax, %edx
- # R1 18
- leal 643717713(%ecx,%ebp,1),%ecx
- xorl %edx, %edi
- andl %ebx, %edi
- movl (%esi), %ebp
- xorl %eax, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $14, %ecx
- addl %edx, %ecx
- # R1 19
- leal 3921069994(%ebx,%ebp,1),%ebx
- xorl %ecx, %edi
- andl %eax, %edi
- movl 20(%esi), %ebp
- xorl %edx, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $20, %ebx
- addl %ecx, %ebx
- # R1 20
- leal 3593408605(%eax,%ebp,1),%eax
- xorl %ebx, %edi
- andl %edx, %edi
- movl 40(%esi), %ebp
- xorl %ecx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $5, %eax
- addl %ebx, %eax
- # R1 21
- leal 38016083(%edx,%ebp,1),%edx
- xorl %eax, %edi
- andl %ecx, %edi
- movl 60(%esi), %ebp
- xorl %ebx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $9, %edx
- addl %eax, %edx
- # R1 22
- leal 3634488961(%ecx,%ebp,1),%ecx
- xorl %edx, %edi
- andl %ebx, %edi
- movl 16(%esi), %ebp
- xorl %eax, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $14, %ecx
- addl %edx, %ecx
- # R1 23
- leal 3889429448(%ebx,%ebp,1),%ebx
- xorl %ecx, %edi
- andl %eax, %edi
- movl 36(%esi), %ebp
- xorl %edx, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $20, %ebx
- addl %ecx, %ebx
- # R1 24
- leal 568446438(%eax,%ebp,1),%eax
- xorl %ebx, %edi
- andl %edx, %edi
- movl 56(%esi), %ebp
- xorl %ecx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $5, %eax
- addl %ebx, %eax
- # R1 25
- leal 3275163606(%edx,%ebp,1),%edx
- xorl %eax, %edi
- andl %ecx, %edi
- movl 12(%esi), %ebp
- xorl %ebx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $9, %edx
- addl %eax, %edx
- # R1 26
- leal 4107603335(%ecx,%ebp,1),%ecx
- xorl %edx, %edi
- andl %ebx, %edi
- movl 32(%esi), %ebp
- xorl %eax, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $14, %ecx
- addl %edx, %ecx
- # R1 27
- leal 1163531501(%ebx,%ebp,1),%ebx
- xorl %ecx, %edi
- andl %eax, %edi
- movl 52(%esi), %ebp
- xorl %edx, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $20, %ebx
- addl %ecx, %ebx
- # R1 28
- leal 2850285829(%eax,%ebp,1),%eax
- xorl %ebx, %edi
- andl %edx, %edi
- movl 8(%esi), %ebp
- xorl %ecx, %edi
- addl %edi, %eax
- movl %ebx, %edi
- roll $5, %eax
- addl %ebx, %eax
- # R1 29
- leal 4243563512(%edx,%ebp,1),%edx
- xorl %eax, %edi
- andl %ecx, %edi
- movl 28(%esi), %ebp
- xorl %ebx, %edi
- addl %edi, %edx
- movl %eax, %edi
- roll $9, %edx
- addl %eax, %edx
- # R1 30
- leal 1735328473(%ecx,%ebp,1),%ecx
- xorl %edx, %edi
- andl %ebx, %edi
- movl 48(%esi), %ebp
- xorl %eax, %edi
- addl %edi, %ecx
- movl %edx, %edi
- roll $14, %ecx
- addl %edx, %ecx
- # R1 31
- leal 2368359562(%ebx,%ebp,1),%ebx
- xorl %ecx, %edi
- andl %eax, %edi
- movl 20(%esi), %ebp
- xorl %edx, %edi
- addl %edi, %ebx
- movl %ecx, %edi
- roll $20, %ebx
- addl %ecx, %ebx
-
- # R2 section
- # R2 32
- xorl %edx, %edi
- xorl %ebx, %edi
- leal 4294588738(%eax,%ebp,1),%eax
- addl %edi, %eax
- roll $4, %eax
- movl 32(%esi), %ebp
- movl %ebx, %edi
- # R2 33
- leal 2272392833(%edx,%ebp,1),%edx
- addl %ebx, %eax
- xorl %ecx, %edi
- xorl %eax, %edi
- movl 44(%esi), %ebp
- addl %edi, %edx
- movl %eax, %edi
- roll $11, %edx
- addl %eax, %edx
- # R2 34
- xorl %ebx, %edi
- xorl %edx, %edi
- leal 1839030562(%ecx,%ebp,1),%ecx
- addl %edi, %ecx
- roll $16, %ecx
- movl 56(%esi), %ebp
- movl %edx, %edi
- # R2 35
- leal 4259657740(%ebx,%ebp,1),%ebx
- addl %edx, %ecx
- xorl %eax, %edi
- xorl %ecx, %edi
- movl 4(%esi), %ebp
- addl %edi, %ebx
- movl %ecx, %edi
- roll $23, %ebx
- addl %ecx, %ebx
- # R2 36
- xorl %edx, %edi
- xorl %ebx, %edi
- leal 2763975236(%eax,%ebp,1),%eax
- addl %edi, %eax
- roll $4, %eax
- movl 16(%esi), %ebp
- movl %ebx, %edi
- # R2 37
- leal 1272893353(%edx,%ebp,1),%edx
- addl %ebx, %eax
- xorl %ecx, %edi
- xorl %eax, %edi
- movl 28(%esi), %ebp
- addl %edi, %edx
- movl %eax, %edi
- roll $11, %edx
- addl %eax, %edx
- # R2 38
- xorl %ebx, %edi
- xorl %edx, %edi
- leal 4139469664(%ecx,%ebp,1),%ecx
- addl %edi, %ecx
- roll $16, %ecx
- movl 40(%esi), %ebp
- movl %edx, %edi
- # R2 39
- leal 3200236656(%ebx,%ebp,1),%ebx
- addl %edx, %ecx
- xorl %eax, %edi
- xorl %ecx, %edi
- movl 52(%esi), %ebp
- addl %edi, %ebx
- movl %ecx, %edi
- roll $23, %ebx
- addl %ecx, %ebx
- # R2 40
- xorl %edx, %edi
- xorl %ebx, %edi
- leal 681279174(%eax,%ebp,1),%eax
- addl %edi, %eax
- roll $4, %eax
- movl (%esi), %ebp
- movl %ebx, %edi
- # R2 41
- leal 3936430074(%edx,%ebp,1),%edx
- addl %ebx, %eax
- xorl %ecx, %edi
- xorl %eax, %edi
- movl 12(%esi), %ebp
- addl %edi, %edx
- movl %eax, %edi
- roll $11, %edx
- addl %eax, %edx
- # R2 42
- xorl %ebx, %edi
- xorl %edx, %edi
- leal 3572445317(%ecx,%ebp,1),%ecx
- addl %edi, %ecx
- roll $16, %ecx
- movl 24(%esi), %ebp
- movl %edx, %edi
- # R2 43
- leal 76029189(%ebx,%ebp,1),%ebx
- addl %edx, %ecx
- xorl %eax, %edi
- xorl %ecx, %edi
- movl 36(%esi), %ebp
- addl %edi, %ebx
- movl %ecx, %edi
- roll $23, %ebx
- addl %ecx, %ebx
- # R2 44
- xorl %edx, %edi
- xorl %ebx, %edi
- leal 3654602809(%eax,%ebp,1),%eax
- addl %edi, %eax
- roll $4, %eax
- movl 48(%esi), %ebp
- movl %ebx, %edi
- # R2 45
- leal 3873151461(%edx,%ebp,1),%edx
- addl %ebx, %eax
- xorl %ecx, %edi
- xorl %eax, %edi
- movl 60(%esi), %ebp
- addl %edi, %edx
- movl %eax, %edi
- roll $11, %edx
- addl %eax, %edx
- # R2 46
- xorl %ebx, %edi
- xorl %edx, %edi
- leal 530742520(%ecx,%ebp,1),%ecx
- addl %edi, %ecx
- roll $16, %ecx
- movl 8(%esi), %ebp
- movl %edx, %edi
- # R2 47
- leal 3299628645(%ebx,%ebp,1),%ebx
- addl %edx, %ecx
- xorl %eax, %edi
- xorl %ecx, %edi
- movl (%esi), %ebp
- addl %edi, %ebx
- movl $-1, %edi
- roll $23, %ebx
- addl %ecx, %ebx
-
- # R3 section
- # R3 48
- xorl %edx, %edi
- orl %ebx, %edi
- leal 4096336452(%eax,%ebp,1),%eax
- xorl %ecx, %edi
- movl 28(%esi), %ebp
- addl %edi, %eax
- movl $-1, %edi
- roll $6, %eax
- xorl %ecx, %edi
- addl %ebx, %eax
- # R3 49
- orl %eax, %edi
- leal 1126891415(%edx,%ebp,1),%edx
- xorl %ebx, %edi
- movl 56(%esi), %ebp
- addl %edi, %edx
- movl $-1, %edi
- roll $10, %edx
- xorl %ebx, %edi
- addl %eax, %edx
- # R3 50
- orl %edx, %edi
- leal 2878612391(%ecx,%ebp,1),%ecx
- xorl %eax, %edi
- movl 20(%esi), %ebp
- addl %edi, %ecx
- movl $-1, %edi
- roll $15, %ecx
- xorl %eax, %edi
- addl %edx, %ecx
- # R3 51
- orl %ecx, %edi
- leal 4237533241(%ebx,%ebp,1),%ebx
- xorl %edx, %edi
- movl 48(%esi), %ebp
- addl %edi, %ebx
- movl $-1, %edi
- roll $21, %ebx
- xorl %edx, %edi
- addl %ecx, %ebx
- # R3 52
- orl %ebx, %edi
- leal 1700485571(%eax,%ebp,1),%eax
- xorl %ecx, %edi
- movl 12(%esi), %ebp
- addl %edi, %eax
- movl $-1, %edi
- roll $6, %eax
- xorl %ecx, %edi
- addl %ebx, %eax
- # R3 53
- orl %eax, %edi
- leal 2399980690(%edx,%ebp,1),%edx
- xorl %ebx, %edi
- movl 40(%esi), %ebp
- addl %edi, %edx
- movl $-1, %edi
- roll $10, %edx
- xorl %ebx, %edi
- addl %eax, %edx
- # R3 54
- orl %edx, %edi
- leal 4293915773(%ecx,%ebp,1),%ecx
- xorl %eax, %edi
- movl 4(%esi), %ebp
- addl %edi, %ecx
- movl $-1, %edi
- roll $15, %ecx
- xorl %eax, %edi
- addl %edx, %ecx
- # R3 55
- orl %ecx, %edi
- leal 2240044497(%ebx,%ebp,1),%ebx
- xorl %edx, %edi
- movl 32(%esi), %ebp
- addl %edi, %ebx
- movl $-1, %edi
- roll $21, %ebx
- xorl %edx, %edi
- addl %ecx, %ebx
- # R3 56
- orl %ebx, %edi
- leal 1873313359(%eax,%ebp,1),%eax
- xorl %ecx, %edi
- movl 60(%esi), %ebp
- addl %edi, %eax
- movl $-1, %edi
- roll $6, %eax
- xorl %ecx, %edi
- addl %ebx, %eax
- # R3 57
- orl %eax, %edi
- leal 4264355552(%edx,%ebp,1),%edx
- xorl %ebx, %edi
- movl 24(%esi), %ebp
- addl %edi, %edx
- movl $-1, %edi
- roll $10, %edx
- xorl %ebx, %edi
- addl %eax, %edx
- # R3 58
- orl %edx, %edi
- leal 2734768916(%ecx,%ebp,1),%ecx
- xorl %eax, %edi
- movl 52(%esi), %ebp
- addl %edi, %ecx
- movl $-1, %edi
- roll $15, %ecx
- xorl %eax, %edi
- addl %edx, %ecx
- # R3 59
- orl %ecx, %edi
- leal 1309151649(%ebx,%ebp,1),%ebx
- xorl %edx, %edi
- movl 16(%esi), %ebp
- addl %edi, %ebx
- movl $-1, %edi
- roll $21, %ebx
- xorl %edx, %edi
- addl %ecx, %ebx
- # R3 60
- orl %ebx, %edi
- leal 4149444226(%eax,%ebp,1),%eax
- xorl %ecx, %edi
- movl 44(%esi), %ebp
- addl %edi, %eax
- movl $-1, %edi
- roll $6, %eax
- xorl %ecx, %edi
- addl %ebx, %eax
- # R3 61
- orl %eax, %edi
- leal 3174756917(%edx,%ebp,1),%edx
- xorl %ebx, %edi
- movl 8(%esi), %ebp
- addl %edi, %edx
- movl $-1, %edi
- roll $10, %edx
- xorl %ebx, %edi
- addl %eax, %edx
- # R3 62
- orl %edx, %edi
- leal 718787259(%ecx,%ebp,1),%ecx
- xorl %eax, %edi
- movl 36(%esi), %ebp
- addl %edi, %ecx
- movl $-1, %edi
- roll $15, %ecx
- xorl %eax, %edi
- addl %edx, %ecx
- # R3 63
- orl %ecx, %edi
- leal 3951481745(%ebx,%ebp,1),%ebx
- xorl %edx, %edi
- movl 24(%esp), %ebp
- addl %edi, %ebx
- addl $64, %esi
- roll $21, %ebx
- movl (%ebp), %edi
- addl %ecx, %ebx
- addl %edi, %eax
- movl 4(%ebp), %edi
- addl %edi, %ebx
- movl 8(%ebp), %edi
- addl %edi, %ecx
- movl 12(%ebp), %edi
- addl %edi, %edx
- movl %eax, (%ebp)
- movl %ebx, 4(%ebp)
- movl (%esp), %edi
- movl %ecx, 8(%ebp)
- movl %edx, 12(%ebp)
- cmpl %esi, %edi
- jge .L000start
- popl %eax
- popl %ebx
- popl %ebp
- popl %edi
- popl %esi
- ret
-.L_md5_block_asm_host_order_end:
- .size md5_block_asm_host_order,.L_md5_block_asm_host_order_end-md5_block_asm_host_order
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/rc4-586.s b/secure/lib/libcrypto/i386/rc4-586.s
deleted file mode 100644
index 996718c527aa..000000000000
--- a/secure/lib/libcrypto/i386/rc4-586.s
+++ /dev/null
@@ -1,316 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by rc4-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "rc4-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl RC4
- .type RC4,@function
-RC4:
-
- pushl %ebp
- pushl %ebx
- movl 12(%esp), %ebp
- movl 16(%esp), %ebx
- pushl %esi
- pushl %edi
- movl (%ebp), %ecx
- movl 4(%ebp), %edx
- movl 28(%esp), %esi
- incl %ecx
- subl $12, %esp
- addl $8, %ebp
- andl $255, %ecx
- leal -8(%ebx,%esi), %ebx
- movl 44(%esp), %edi
- movl %ebx, 8(%esp)
- movl (%ebp,%ecx,4), %eax
- cmpl %esi, %ebx
- jl .L000end
-.L001start:
- addl $8, %esi
- # Round 0
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb %bl, (%esp)
- # Round 1
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb %bl, 1(%esp)
- # Round 2
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb %bl, 2(%esp)
- # Round 3
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb %bl, 3(%esp)
- # Round 4
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb %bl, 4(%esp)
- # Round 5
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb %bl, 5(%esp)
- # Round 6
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb %bl, 6(%esp)
- # Round 7
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- addl $8, %edi
- movb %bl, 7(%esp)
- # apply the cipher text
- movl (%esp), %eax
- movl -8(%esi), %ebx
- xorl %ebx, %eax
- movl -4(%esi), %ebx
- movl %eax, -8(%edi)
- movl 4(%esp), %eax
- xorl %ebx, %eax
- movl 8(%esp), %ebx
- movl %eax, -4(%edi)
- movl (%ebp,%ecx,4), %eax
- cmpl %ebx, %esi
- jle .L001start
-.L000end:
- # Round 0
- addl $8, %ebx
- incl %esi
- cmpl %esi, %ebx
- jl .L002finished
- movl %ebx, 8(%esp)
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb -1(%esi), %bh
- xorb %bh, %bl
- movb %bl, (%edi)
- # Round 1
- movl 8(%esp), %ebx
- cmpl %esi, %ebx
- jle .L002finished
- incl %esi
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb -1(%esi), %bh
- xorb %bh, %bl
- movb %bl, 1(%edi)
- # Round 2
- movl 8(%esp), %ebx
- cmpl %esi, %ebx
- jle .L002finished
- incl %esi
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb -1(%esi), %bh
- xorb %bh, %bl
- movb %bl, 2(%edi)
- # Round 3
- movl 8(%esp), %ebx
- cmpl %esi, %ebx
- jle .L002finished
- incl %esi
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb -1(%esi), %bh
- xorb %bh, %bl
- movb %bl, 3(%edi)
- # Round 4
- movl 8(%esp), %ebx
- cmpl %esi, %ebx
- jle .L002finished
- incl %esi
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb -1(%esi), %bh
- xorb %bh, %bl
- movb %bl, 4(%edi)
- # Round 5
- movl 8(%esp), %ebx
- cmpl %esi, %ebx
- jle .L002finished
- incl %esi
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movl (%ebp,%ecx,4), %eax
- movb -1(%esi), %bh
- xorb %bh, %bl
- movb %bl, 5(%edi)
- # Round 6
- movl 8(%esp), %ebx
- cmpl %esi, %ebx
- jle .L002finished
- incl %esi
- addl %eax, %edx
- andl $255, %edx
- incl %ecx
- movl (%ebp,%edx,4), %ebx
- movl %ebx, -4(%ebp,%ecx,4)
- addl %eax, %ebx
- andl $255, %ecx
- andl $255, %ebx
- movl %eax, (%ebp,%edx,4)
- nop
- movl (%ebp,%ebx,4), %ebx
- movb -1(%esi), %bh
- xorb %bh, %bl
- movb %bl, 6(%edi)
-.L002finished:
- decl %ecx
- addl $12, %esp
- movl %edx, -4(%ebp)
- movb %cl, -8(%ebp)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.RC4_end:
- .size RC4,.RC4_end-RC4
-.ident "RC4"
diff --git a/secure/lib/libcrypto/i386/rc5-586.s b/secure/lib/libcrypto/i386/rc5-586.s
deleted file mode 100644
index 1a4c9d3c4397..000000000000
--- a/secure/lib/libcrypto/i386/rc5-586.s
+++ /dev/null
@@ -1,584 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by rc5-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "rc5-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl RC5_32_encrypt
- .type RC5_32_encrypt,@function
-RC5_32_encrypt:
-
- pushl %ebp
- pushl %esi
- pushl %edi
- movl 16(%esp), %edx
- movl 20(%esp), %ebp
- # Load the 2 words
- movl (%edx), %edi
- movl 4(%edx), %esi
- pushl %ebx
- movl (%ebp), %ebx
- addl 4(%ebp), %edi
- addl 8(%ebp), %esi
- xorl %esi, %edi
- movl 12(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 16(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 20(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 24(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 28(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 32(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 36(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 40(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 44(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 48(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 52(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 56(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 60(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 64(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 68(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 72(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- cmpl $8, %ebx
- je .L000rc5_exit
- xorl %esi, %edi
- movl 76(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 80(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 84(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 88(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 92(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 96(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 100(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 104(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- cmpl $12, %ebx
- je .L000rc5_exit
- xorl %esi, %edi
- movl 108(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 112(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 116(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 120(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 124(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 128(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
- xorl %esi, %edi
- movl 132(%ebp), %eax
- movl %esi, %ecx
- roll %cl, %edi
- addl %eax, %edi
- xorl %edi, %esi
- movl 136(%ebp), %eax
- movl %edi, %ecx
- roll %cl, %esi
- addl %eax, %esi
-.L000rc5_exit:
- movl %edi, (%edx)
- movl %esi, 4(%edx)
- popl %ebx
- popl %edi
- popl %esi
- popl %ebp
- ret
-.L_RC5_32_encrypt_end:
- .size RC5_32_encrypt,.L_RC5_32_encrypt_end-RC5_32_encrypt
-.ident "desasm.pl"
-.text
- .align 16
-.globl RC5_32_decrypt
- .type RC5_32_decrypt,@function
-RC5_32_decrypt:
-
- pushl %ebp
- pushl %esi
- pushl %edi
- movl 16(%esp), %edx
- movl 20(%esp), %ebp
- # Load the 2 words
- movl (%edx), %edi
- movl 4(%edx), %esi
- pushl %ebx
- movl (%ebp), %ebx
- cmpl $12, %ebx
- je .L001rc5_dec_12
- cmpl $8, %ebx
- je .L002rc5_dec_8
- movl 136(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 132(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 128(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 124(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 120(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 116(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 112(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 108(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
-.L001rc5_dec_12:
- movl 104(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 100(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 96(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 92(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 88(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 84(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 80(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 76(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
-.L002rc5_dec_8:
- movl 72(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 68(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 64(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 60(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 56(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 52(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 48(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 44(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 40(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 36(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 32(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 28(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 24(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 20(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- movl 16(%ebp), %eax
- subl %eax, %esi
- movl %edi, %ecx
- rorl %cl, %esi
- xorl %edi, %esi
- movl 12(%ebp), %eax
- subl %eax, %edi
- movl %esi, %ecx
- rorl %cl, %edi
- xorl %esi, %edi
- subl 8(%ebp), %esi
- subl 4(%ebp), %edi
-.L003rc5_exit:
- movl %edi, (%edx)
- movl %esi, 4(%edx)
- popl %ebx
- popl %edi
- popl %esi
- popl %ebp
- ret
-.L_RC5_32_decrypt_end:
- .size RC5_32_decrypt,.L_RC5_32_decrypt_end-RC5_32_decrypt
-.ident "desasm.pl"
-.text
- .align 16
-.globl RC5_32_cbc_encrypt
- .type RC5_32_cbc_encrypt,@function
-RC5_32_cbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
- # getting iv ptr from parameter 4
- movl 36(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
- # getting encrypt flag from parameter 5
- movl 56(%esp), %ecx
- # get and push parameter 3
- movl 48(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L004decrypt
- andl $4294967288, %ebp
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- jz .L005encrypt_finish
-.L006encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call RC5_32_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L006encrypt_loop
-.L005encrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L007finish
- xorl %ecx, %ecx
- xorl %edx, %edx
- movl .L008cbc_enc_jmp_table(,%ebp,4),%ebp
- jmp *%ebp
-.L009ej7:
- movb 6(%esi), %dh
- sall $8, %edx
-.L010ej6:
- movb 5(%esi), %dh
-.L011ej5:
- movb 4(%esi), %dl
-.L012ej4:
- movl (%esi), %ecx
- jmp .L013ejend
-.L014ej3:
- movb 2(%esi), %ch
- sall $8, %ecx
-.L015ej2:
- movb 1(%esi), %ch
-.L016ej1:
- movb (%esi), %cl
-.L013ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call RC5_32_encrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L007finish
-.align 16
-.L004decrypt:
- andl $4294967288, %ebp
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- jz .L017decrypt_finish
-.L018decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call RC5_32_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L018decrypt_loop
-.L017decrypt_finish:
- movl 52(%esp), %ebp
- andl $7, %ebp
- jz .L007finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 8(%esp)
- movl %ebx, 12(%esp)
- call RC5_32_decrypt
- movl 8(%esp), %eax
- movl 12(%esp), %ebx
- movl 16(%esp), %ecx
- movl 20(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L019dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L020dj6:
- movb %dh, 5(%edi)
-.L021dj5:
- movb %dl, 4(%edi)
-.L022dj4:
- movl %ecx, (%edi)
- jmp .L023djend
-.L024dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L025dj2:
- movb %ch, 1(%esi)
-.L026dj1:
- movb %cl, (%esi)
-.L023djend:
- jmp .L007finish
-.align 16
-.L007finish:
- movl 60(%esp), %ecx
- addl $24, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L008cbc_enc_jmp_table:
- .long 0
- .long .L016ej1
- .long .L015ej2
- .long .L014ej3
- .long .L012ej4
- .long .L011ej5
- .long .L010ej6
- .long .L009ej7
-.align 16
-.L027cbc_dec_jmp_table:
- .long 0
- .long .L026dj1
- .long .L025dj2
- .long .L024dj3
- .long .L022dj4
- .long .L021dj5
- .long .L020dj6
- .long .L019dj7
-.L_RC5_32_cbc_encrypt_end:
- .size RC5_32_cbc_encrypt,.L_RC5_32_cbc_encrypt_end-RC5_32_cbc_encrypt
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/rmd-586.s b/secure/lib/libcrypto/i386/rmd-586.s
deleted file mode 100644
index 96f49282fbc1..000000000000
--- a/secure/lib/libcrypto/i386/rmd-586.s
+++ /dev/null
@@ -1,1975 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by rmd-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "rmd-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl ripemd160_block_asm_host_order
- .type ripemd160_block_asm_host_order,@function
-ripemd160_block_asm_host_order:
- movl 4(%esp), %edx
- movl 8(%esp), %eax
- pushl %esi
- movl (%edx), %ecx
- pushl %edi
- movl 4(%edx), %esi
- pushl %ebp
- movl 8(%edx), %edi
- pushl %ebx
- subl $108, %esp
-.L000start:
-
- movl (%eax), %ebx
- movl 4(%eax), %ebp
- movl %ebx, (%esp)
- movl %ebp, 4(%esp)
- movl 8(%eax), %ebx
- movl 12(%eax), %ebp
- movl %ebx, 8(%esp)
- movl %ebp, 12(%esp)
- movl 16(%eax), %ebx
- movl 20(%eax), %ebp
- movl %ebx, 16(%esp)
- movl %ebp, 20(%esp)
- movl 24(%eax), %ebx
- movl 28(%eax), %ebp
- movl %ebx, 24(%esp)
- movl %ebp, 28(%esp)
- movl 32(%eax), %ebx
- movl 36(%eax), %ebp
- movl %ebx, 32(%esp)
- movl %ebp, 36(%esp)
- movl 40(%eax), %ebx
- movl 44(%eax), %ebp
- movl %ebx, 40(%esp)
- movl %ebp, 44(%esp)
- movl 48(%eax), %ebx
- movl 52(%eax), %ebp
- movl %ebx, 48(%esp)
- movl %ebp, 52(%esp)
- movl 56(%eax), %ebx
- movl 60(%eax), %ebp
- movl %ebx, 56(%esp)
- movl %ebp, 60(%esp)
- movl %edi, %eax
- movl 12(%edx), %ebx
- movl 16(%edx), %ebp
- # 0
- xorl %ebx, %eax
- movl (%esp), %edx
- xorl %esi, %eax
- addl %edx, %ecx
- roll $10, %edi
- addl %eax, %ecx
- movl %esi, %eax
- roll $11, %ecx
- addl %ebp, %ecx
- # 1
- xorl %edi, %eax
- movl 4(%esp), %edx
- xorl %ecx, %eax
- addl %eax, %ebp
- movl %ecx, %eax
- roll $10, %esi
- addl %edx, %ebp
- xorl %esi, %eax
- roll $14, %ebp
- addl %ebx, %ebp
- # 2
- movl 8(%esp), %edx
- xorl %ebp, %eax
- addl %edx, %ebx
- roll $10, %ecx
- addl %eax, %ebx
- movl %ebp, %eax
- roll $15, %ebx
- addl %edi, %ebx
- # 3
- xorl %ecx, %eax
- movl 12(%esp), %edx
- xorl %ebx, %eax
- addl %eax, %edi
- movl %ebx, %eax
- roll $10, %ebp
- addl %edx, %edi
- xorl %ebp, %eax
- roll $12, %edi
- addl %esi, %edi
- # 4
- movl 16(%esp), %edx
- xorl %edi, %eax
- addl %edx, %esi
- roll $10, %ebx
- addl %eax, %esi
- movl %edi, %eax
- roll $5, %esi
- addl %ecx, %esi
- # 5
- xorl %ebx, %eax
- movl 20(%esp), %edx
- xorl %esi, %eax
- addl %eax, %ecx
- movl %esi, %eax
- roll $10, %edi
- addl %edx, %ecx
- xorl %edi, %eax
- roll $8, %ecx
- addl %ebp, %ecx
- # 6
- movl 24(%esp), %edx
- xorl %ecx, %eax
- addl %edx, %ebp
- roll $10, %esi
- addl %eax, %ebp
- movl %ecx, %eax
- roll $7, %ebp
- addl %ebx, %ebp
- # 7
- xorl %esi, %eax
- movl 28(%esp), %edx
- xorl %ebp, %eax
- addl %eax, %ebx
- movl %ebp, %eax
- roll $10, %ecx
- addl %edx, %ebx
- xorl %ecx, %eax
- roll $9, %ebx
- addl %edi, %ebx
- # 8
- movl 32(%esp), %edx
- xorl %ebx, %eax
- addl %edx, %edi
- roll $10, %ebp
- addl %eax, %edi
- movl %ebx, %eax
- roll $11, %edi
- addl %esi, %edi
- # 9
- xorl %ebp, %eax
- movl 36(%esp), %edx
- xorl %edi, %eax
- addl %eax, %esi
- movl %edi, %eax
- roll $10, %ebx
- addl %edx, %esi
- xorl %ebx, %eax
- roll $13, %esi
- addl %ecx, %esi
- # 10
- movl 40(%esp), %edx
- xorl %esi, %eax
- addl %edx, %ecx
- roll $10, %edi
- addl %eax, %ecx
- movl %esi, %eax
- roll $14, %ecx
- addl %ebp, %ecx
- # 11
- xorl %edi, %eax
- movl 44(%esp), %edx
- xorl %ecx, %eax
- addl %eax, %ebp
- movl %ecx, %eax
- roll $10, %esi
- addl %edx, %ebp
- xorl %esi, %eax
- roll $15, %ebp
- addl %ebx, %ebp
- # 12
- movl 48(%esp), %edx
- xorl %ebp, %eax
- addl %edx, %ebx
- roll $10, %ecx
- addl %eax, %ebx
- movl %ebp, %eax
- roll $6, %ebx
- addl %edi, %ebx
- # 13
- xorl %ecx, %eax
- movl 52(%esp), %edx
- xorl %ebx, %eax
- addl %eax, %edi
- movl %ebx, %eax
- roll $10, %ebp
- addl %edx, %edi
- xorl %ebp, %eax
- roll $7, %edi
- addl %esi, %edi
- # 14
- movl 56(%esp), %edx
- xorl %edi, %eax
- addl %edx, %esi
- roll $10, %ebx
- addl %eax, %esi
- movl %edi, %eax
- roll $9, %esi
- addl %ecx, %esi
- # 15
- xorl %ebx, %eax
- movl 60(%esp), %edx
- xorl %esi, %eax
- addl %eax, %ecx
- movl $-1, %eax
- roll $10, %edi
- addl %edx, %ecx
- movl 28(%esp), %edx
- roll $8, %ecx
- addl %ebp, %ecx
- # 16
- addl %edx, %ebp
- movl %esi, %edx
- subl %ecx, %eax
- andl %ecx, %edx
- andl %edi, %eax
- orl %eax, %edx
- movl 16(%esp), %eax
- roll $10, %esi
- leal 1518500249(%ebp,%edx,1),%ebp
- movl $-1, %edx
- roll $7, %ebp
- addl %ebx, %ebp
- # 17
- addl %eax, %ebx
- movl %ecx, %eax
- subl %ebp, %edx
- andl %ebp, %eax
- andl %esi, %edx
- orl %edx, %eax
- movl 52(%esp), %edx
- roll $10, %ecx
- leal 1518500249(%ebx,%eax,1),%ebx
- movl $-1, %eax
- roll $6, %ebx
- addl %edi, %ebx
- # 18
- addl %edx, %edi
- movl %ebp, %edx
- subl %ebx, %eax
- andl %ebx, %edx
- andl %ecx, %eax
- orl %eax, %edx
- movl 4(%esp), %eax
- roll $10, %ebp
- leal 1518500249(%edi,%edx,1),%edi
- movl $-1, %edx
- roll $8, %edi
- addl %esi, %edi
- # 19
- addl %eax, %esi
- movl %ebx, %eax
- subl %edi, %edx
- andl %edi, %eax
- andl %ebp, %edx
- orl %edx, %eax
- movl 40(%esp), %edx
- roll $10, %ebx
- leal 1518500249(%esi,%eax,1),%esi
- movl $-1, %eax
- roll $13, %esi
- addl %ecx, %esi
- # 20
- addl %edx, %ecx
- movl %edi, %edx
- subl %esi, %eax
- andl %esi, %edx
- andl %ebx, %eax
- orl %eax, %edx
- movl 24(%esp), %eax
- roll $10, %edi
- leal 1518500249(%ecx,%edx,1),%ecx
- movl $-1, %edx
- roll $11, %ecx
- addl %ebp, %ecx
- # 21
- addl %eax, %ebp
- movl %esi, %eax
- subl %ecx, %edx
- andl %ecx, %eax
- andl %edi, %edx
- orl %edx, %eax
- movl 60(%esp), %edx
- roll $10, %esi
- leal 1518500249(%ebp,%eax,1),%ebp
- movl $-1, %eax
- roll $9, %ebp
- addl %ebx, %ebp
- # 22
- addl %edx, %ebx
- movl %ecx, %edx
- subl %ebp, %eax
- andl %ebp, %edx
- andl %esi, %eax
- orl %eax, %edx
- movl 12(%esp), %eax
- roll $10, %ecx
- leal 1518500249(%ebx,%edx,1),%ebx
- movl $-1, %edx
- roll $7, %ebx
- addl %edi, %ebx
- # 23
- addl %eax, %edi
- movl %ebp, %eax
- subl %ebx, %edx
- andl %ebx, %eax
- andl %ecx, %edx
- orl %edx, %eax
- movl 48(%esp), %edx
- roll $10, %ebp
- leal 1518500249(%edi,%eax,1),%edi
- movl $-1, %eax
- roll $15, %edi
- addl %esi, %edi
- # 24
- addl %edx, %esi
- movl %ebx, %edx
- subl %edi, %eax
- andl %edi, %edx
- andl %ebp, %eax
- orl %eax, %edx
- movl (%esp), %eax
- roll $10, %ebx
- leal 1518500249(%esi,%edx,1),%esi
- movl $-1, %edx
- roll $7, %esi
- addl %ecx, %esi
- # 25
- addl %eax, %ecx
- movl %edi, %eax
- subl %esi, %edx
- andl %esi, %eax
- andl %ebx, %edx
- orl %edx, %eax
- movl 36(%esp), %edx
- roll $10, %edi
- leal 1518500249(%ecx,%eax,1),%ecx
- movl $-1, %eax
- roll $12, %ecx
- addl %ebp, %ecx
- # 26
- addl %edx, %ebp
- movl %esi, %edx
- subl %ecx, %eax
- andl %ecx, %edx
- andl %edi, %eax
- orl %eax, %edx
- movl 20(%esp), %eax
- roll $10, %esi
- leal 1518500249(%ebp,%edx,1),%ebp
- movl $-1, %edx
- roll $15, %ebp
- addl %ebx, %ebp
- # 27
- addl %eax, %ebx
- movl %ecx, %eax
- subl %ebp, %edx
- andl %ebp, %eax
- andl %esi, %edx
- orl %edx, %eax
- movl 8(%esp), %edx
- roll $10, %ecx
- leal 1518500249(%ebx,%eax,1),%ebx
- movl $-1, %eax
- roll $9, %ebx
- addl %edi, %ebx
- # 28
- addl %edx, %edi
- movl %ebp, %edx
- subl %ebx, %eax
- andl %ebx, %edx
- andl %ecx, %eax
- orl %eax, %edx
- movl 56(%esp), %eax
- roll $10, %ebp
- leal 1518500249(%edi,%edx,1),%edi
- movl $-1, %edx
- roll $11, %edi
- addl %esi, %edi
- # 29
- addl %eax, %esi
- movl %ebx, %eax
- subl %edi, %edx
- andl %edi, %eax
- andl %ebp, %edx
- orl %edx, %eax
- movl 44(%esp), %edx
- roll $10, %ebx
- leal 1518500249(%esi,%eax,1),%esi
- movl $-1, %eax
- roll $7, %esi
- addl %ecx, %esi
- # 30
- addl %edx, %ecx
- movl %edi, %edx
- subl %esi, %eax
- andl %esi, %edx
- andl %ebx, %eax
- orl %eax, %edx
- movl 32(%esp), %eax
- roll $10, %edi
- leal 1518500249(%ecx,%edx,1),%ecx
- movl $-1, %edx
- roll $13, %ecx
- addl %ebp, %ecx
- # 31
- addl %eax, %ebp
- movl %esi, %eax
- subl %ecx, %edx
- andl %ecx, %eax
- andl %edi, %edx
- orl %edx, %eax
- movl $-1, %edx
- roll $10, %esi
- leal 1518500249(%ebp,%eax,1),%ebp
- subl %ecx, %edx
- roll $12, %ebp
- addl %ebx, %ebp
- # 32
- movl 12(%esp), %eax
- orl %ebp, %edx
- addl %eax, %ebx
- xorl %esi, %edx
- movl $-1, %eax
- roll $10, %ecx
- leal 1859775393(%ebx,%edx,1),%ebx
- subl %ebp, %eax
- roll $11, %ebx
- addl %edi, %ebx
- # 33
- movl 40(%esp), %edx
- orl %ebx, %eax
- addl %edx, %edi
- xorl %ecx, %eax
- movl $-1, %edx
- roll $10, %ebp
- leal 1859775393(%edi,%eax,1),%edi
- subl %ebx, %edx
- roll $13, %edi
- addl %esi, %edi
- # 34
- movl 56(%esp), %eax
- orl %edi, %edx
- addl %eax, %esi
- xorl %ebp, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 1859775393(%esi,%edx,1),%esi
- subl %edi, %eax
- roll $6, %esi
- addl %ecx, %esi
- # 35
- movl 16(%esp), %edx
- orl %esi, %eax
- addl %edx, %ecx
- xorl %ebx, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 1859775393(%ecx,%eax,1),%ecx
- subl %esi, %edx
- roll $7, %ecx
- addl %ebp, %ecx
- # 36
- movl 36(%esp), %eax
- orl %ecx, %edx
- addl %eax, %ebp
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %esi
- leal 1859775393(%ebp,%edx,1),%ebp
- subl %ecx, %eax
- roll $14, %ebp
- addl %ebx, %ebp
- # 37
- movl 60(%esp), %edx
- orl %ebp, %eax
- addl %edx, %ebx
- xorl %esi, %eax
- movl $-1, %edx
- roll $10, %ecx
- leal 1859775393(%ebx,%eax,1),%ebx
- subl %ebp, %edx
- roll $9, %ebx
- addl %edi, %ebx
- # 38
- movl 32(%esp), %eax
- orl %ebx, %edx
- addl %eax, %edi
- xorl %ecx, %edx
- movl $-1, %eax
- roll $10, %ebp
- leal 1859775393(%edi,%edx,1),%edi
- subl %ebx, %eax
- roll $13, %edi
- addl %esi, %edi
- # 39
- movl 4(%esp), %edx
- orl %edi, %eax
- addl %edx, %esi
- xorl %ebp, %eax
- movl $-1, %edx
- roll $10, %ebx
- leal 1859775393(%esi,%eax,1),%esi
- subl %edi, %edx
- roll $15, %esi
- addl %ecx, %esi
- # 40
- movl 8(%esp), %eax
- orl %esi, %edx
- addl %eax, %ecx
- xorl %ebx, %edx
- movl $-1, %eax
- roll $10, %edi
- leal 1859775393(%ecx,%edx,1),%ecx
- subl %esi, %eax
- roll $14, %ecx
- addl %ebp, %ecx
- # 41
- movl 28(%esp), %edx
- orl %ecx, %eax
- addl %edx, %ebp
- xorl %edi, %eax
- movl $-1, %edx
- roll $10, %esi
- leal 1859775393(%ebp,%eax,1),%ebp
- subl %ecx, %edx
- roll $8, %ebp
- addl %ebx, %ebp
- # 42
- movl (%esp), %eax
- orl %ebp, %edx
- addl %eax, %ebx
- xorl %esi, %edx
- movl $-1, %eax
- roll $10, %ecx
- leal 1859775393(%ebx,%edx,1),%ebx
- subl %ebp, %eax
- roll $13, %ebx
- addl %edi, %ebx
- # 43
- movl 24(%esp), %edx
- orl %ebx, %eax
- addl %edx, %edi
- xorl %ecx, %eax
- movl $-1, %edx
- roll $10, %ebp
- leal 1859775393(%edi,%eax,1),%edi
- subl %ebx, %edx
- roll $6, %edi
- addl %esi, %edi
- # 44
- movl 52(%esp), %eax
- orl %edi, %edx
- addl %eax, %esi
- xorl %ebp, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 1859775393(%esi,%edx,1),%esi
- subl %edi, %eax
- roll $5, %esi
- addl %ecx, %esi
- # 45
- movl 44(%esp), %edx
- orl %esi, %eax
- addl %edx, %ecx
- xorl %ebx, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 1859775393(%ecx,%eax,1),%ecx
- subl %esi, %edx
- roll $12, %ecx
- addl %ebp, %ecx
- # 46
- movl 20(%esp), %eax
- orl %ecx, %edx
- addl %eax, %ebp
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %esi
- leal 1859775393(%ebp,%edx,1),%ebp
- subl %ecx, %eax
- roll $7, %ebp
- addl %ebx, %ebp
- # 47
- movl 48(%esp), %edx
- orl %ebp, %eax
- addl %edx, %ebx
- xorl %esi, %eax
- movl $-1, %edx
- roll $10, %ecx
- leal 1859775393(%ebx,%eax,1),%ebx
- movl %ecx, %eax
- roll $5, %ebx
- addl %edi, %ebx
- # 48
- subl %ecx, %edx
- andl %ebx, %eax
- andl %ebp, %edx
- orl %eax, %edx
- movl 4(%esp), %eax
- roll $10, %ebp
- leal 2400959708(%edi,%edx),%edi
- movl $-1, %edx
- addl %eax, %edi
- movl %ebp, %eax
- roll $11, %edi
- addl %esi, %edi
- # 49
- subl %ebp, %edx
- andl %edi, %eax
- andl %ebx, %edx
- orl %eax, %edx
- movl 36(%esp), %eax
- roll $10, %ebx
- leal 2400959708(%esi,%edx),%esi
- movl $-1, %edx
- addl %eax, %esi
- movl %ebx, %eax
- roll $12, %esi
- addl %ecx, %esi
- # 50
- subl %ebx, %edx
- andl %esi, %eax
- andl %edi, %edx
- orl %eax, %edx
- movl 44(%esp), %eax
- roll $10, %edi
- leal 2400959708(%ecx,%edx),%ecx
- movl $-1, %edx
- addl %eax, %ecx
- movl %edi, %eax
- roll $14, %ecx
- addl %ebp, %ecx
- # 51
- subl %edi, %edx
- andl %ecx, %eax
- andl %esi, %edx
- orl %eax, %edx
- movl 40(%esp), %eax
- roll $10, %esi
- leal 2400959708(%ebp,%edx),%ebp
- movl $-1, %edx
- addl %eax, %ebp
- movl %esi, %eax
- roll $15, %ebp
- addl %ebx, %ebp
- # 52
- subl %esi, %edx
- andl %ebp, %eax
- andl %ecx, %edx
- orl %eax, %edx
- movl (%esp), %eax
- roll $10, %ecx
- leal 2400959708(%ebx,%edx),%ebx
- movl $-1, %edx
- addl %eax, %ebx
- movl %ecx, %eax
- roll $14, %ebx
- addl %edi, %ebx
- # 53
- subl %ecx, %edx
- andl %ebx, %eax
- andl %ebp, %edx
- orl %eax, %edx
- movl 32(%esp), %eax
- roll $10, %ebp
- leal 2400959708(%edi,%edx),%edi
- movl $-1, %edx
- addl %eax, %edi
- movl %ebp, %eax
- roll $15, %edi
- addl %esi, %edi
- # 54
- subl %ebp, %edx
- andl %edi, %eax
- andl %ebx, %edx
- orl %eax, %edx
- movl 48(%esp), %eax
- roll $10, %ebx
- leal 2400959708(%esi,%edx),%esi
- movl $-1, %edx
- addl %eax, %esi
- movl %ebx, %eax
- roll $9, %esi
- addl %ecx, %esi
- # 55
- subl %ebx, %edx
- andl %esi, %eax
- andl %edi, %edx
- orl %eax, %edx
- movl 16(%esp), %eax
- roll $10, %edi
- leal 2400959708(%ecx,%edx),%ecx
- movl $-1, %edx
- addl %eax, %ecx
- movl %edi, %eax
- roll $8, %ecx
- addl %ebp, %ecx
- # 56
- subl %edi, %edx
- andl %ecx, %eax
- andl %esi, %edx
- orl %eax, %edx
- movl 52(%esp), %eax
- roll $10, %esi
- leal 2400959708(%ebp,%edx),%ebp
- movl $-1, %edx
- addl %eax, %ebp
- movl %esi, %eax
- roll $9, %ebp
- addl %ebx, %ebp
- # 57
- subl %esi, %edx
- andl %ebp, %eax
- andl %ecx, %edx
- orl %eax, %edx
- movl 12(%esp), %eax
- roll $10, %ecx
- leal 2400959708(%ebx,%edx),%ebx
- movl $-1, %edx
- addl %eax, %ebx
- movl %ecx, %eax
- roll $14, %ebx
- addl %edi, %ebx
- # 58
- subl %ecx, %edx
- andl %ebx, %eax
- andl %ebp, %edx
- orl %eax, %edx
- movl 28(%esp), %eax
- roll $10, %ebp
- leal 2400959708(%edi,%edx),%edi
- movl $-1, %edx
- addl %eax, %edi
- movl %ebp, %eax
- roll $5, %edi
- addl %esi, %edi
- # 59
- subl %ebp, %edx
- andl %edi, %eax
- andl %ebx, %edx
- orl %eax, %edx
- movl 60(%esp), %eax
- roll $10, %ebx
- leal 2400959708(%esi,%edx),%esi
- movl $-1, %edx
- addl %eax, %esi
- movl %ebx, %eax
- roll $6, %esi
- addl %ecx, %esi
- # 60
- subl %ebx, %edx
- andl %esi, %eax
- andl %edi, %edx
- orl %eax, %edx
- movl 56(%esp), %eax
- roll $10, %edi
- leal 2400959708(%ecx,%edx),%ecx
- movl $-1, %edx
- addl %eax, %ecx
- movl %edi, %eax
- roll $8, %ecx
- addl %ebp, %ecx
- # 61
- subl %edi, %edx
- andl %ecx, %eax
- andl %esi, %edx
- orl %eax, %edx
- movl 20(%esp), %eax
- roll $10, %esi
- leal 2400959708(%ebp,%edx),%ebp
- movl $-1, %edx
- addl %eax, %ebp
- movl %esi, %eax
- roll $6, %ebp
- addl %ebx, %ebp
- # 62
- subl %esi, %edx
- andl %ebp, %eax
- andl %ecx, %edx
- orl %eax, %edx
- movl 24(%esp), %eax
- roll $10, %ecx
- leal 2400959708(%ebx,%edx),%ebx
- movl $-1, %edx
- addl %eax, %ebx
- movl %ecx, %eax
- roll $5, %ebx
- addl %edi, %ebx
- # 63
- subl %ecx, %edx
- andl %ebx, %eax
- andl %ebp, %edx
- orl %eax, %edx
- movl 8(%esp), %eax
- roll $10, %ebp
- leal 2400959708(%edi,%edx),%edi
- movl $-1, %edx
- addl %eax, %edi
- subl %ebp, %edx
- roll $12, %edi
- addl %esi, %edi
- # 64
- movl 16(%esp), %eax
- orl %ebx, %edx
- addl %eax, %esi
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 2840853838(%esi,%edx,1),%esi
- subl %ebx, %eax
- roll $9, %esi
- addl %ecx, %esi
- # 65
- movl (%esp), %edx
- orl %edi, %eax
- addl %edx, %ecx
- xorl %esi, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 2840853838(%ecx,%eax,1),%ecx
- subl %edi, %edx
- roll $15, %ecx
- addl %ebp, %ecx
- # 66
- movl 20(%esp), %eax
- orl %esi, %edx
- addl %eax, %ebp
- xorl %ecx, %edx
- movl $-1, %eax
- roll $10, %esi
- leal 2840853838(%ebp,%edx,1),%ebp
- subl %esi, %eax
- roll $5, %ebp
- addl %ebx, %ebp
- # 67
- movl 36(%esp), %edx
- orl %ecx, %eax
- addl %edx, %ebx
- xorl %ebp, %eax
- movl $-1, %edx
- roll $10, %ecx
- leal 2840853838(%ebx,%eax,1),%ebx
- subl %ecx, %edx
- roll $11, %ebx
- addl %edi, %ebx
- # 68
- movl 28(%esp), %eax
- orl %ebp, %edx
- addl %eax, %edi
- xorl %ebx, %edx
- movl $-1, %eax
- roll $10, %ebp
- leal 2840853838(%edi,%edx,1),%edi
- subl %ebp, %eax
- roll $6, %edi
- addl %esi, %edi
- # 69
- movl 48(%esp), %edx
- orl %ebx, %eax
- addl %edx, %esi
- xorl %edi, %eax
- movl $-1, %edx
- roll $10, %ebx
- leal 2840853838(%esi,%eax,1),%esi
- subl %ebx, %edx
- roll $8, %esi
- addl %ecx, %esi
- # 70
- movl 8(%esp), %eax
- orl %edi, %edx
- addl %eax, %ecx
- xorl %esi, %edx
- movl $-1, %eax
- roll $10, %edi
- leal 2840853838(%ecx,%edx,1),%ecx
- subl %edi, %eax
- roll $13, %ecx
- addl %ebp, %ecx
- # 71
- movl 40(%esp), %edx
- orl %esi, %eax
- addl %edx, %ebp
- xorl %ecx, %eax
- movl $-1, %edx
- roll $10, %esi
- leal 2840853838(%ebp,%eax,1),%ebp
- subl %esi, %edx
- roll $12, %ebp
- addl %ebx, %ebp
- # 72
- movl 56(%esp), %eax
- orl %ecx, %edx
- addl %eax, %ebx
- xorl %ebp, %edx
- movl $-1, %eax
- roll $10, %ecx
- leal 2840853838(%ebx,%edx,1),%ebx
- subl %ecx, %eax
- roll $5, %ebx
- addl %edi, %ebx
- # 73
- movl 4(%esp), %edx
- orl %ebp, %eax
- addl %edx, %edi
- xorl %ebx, %eax
- movl $-1, %edx
- roll $10, %ebp
- leal 2840853838(%edi,%eax,1),%edi
- subl %ebp, %edx
- roll $12, %edi
- addl %esi, %edi
- # 74
- movl 12(%esp), %eax
- orl %ebx, %edx
- addl %eax, %esi
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 2840853838(%esi,%edx,1),%esi
- subl %ebx, %eax
- roll $13, %esi
- addl %ecx, %esi
- # 75
- movl 32(%esp), %edx
- orl %edi, %eax
- addl %edx, %ecx
- xorl %esi, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 2840853838(%ecx,%eax,1),%ecx
- subl %edi, %edx
- roll $14, %ecx
- addl %ebp, %ecx
- # 76
- movl 44(%esp), %eax
- orl %esi, %edx
- addl %eax, %ebp
- xorl %ecx, %edx
- movl $-1, %eax
- roll $10, %esi
- leal 2840853838(%ebp,%edx,1),%ebp
- subl %esi, %eax
- roll $11, %ebp
- addl %ebx, %ebp
- # 77
- movl 24(%esp), %edx
- orl %ecx, %eax
- addl %edx, %ebx
- xorl %ebp, %eax
- movl $-1, %edx
- roll $10, %ecx
- leal 2840853838(%ebx,%eax,1),%ebx
- subl %ecx, %edx
- roll $8, %ebx
- addl %edi, %ebx
- # 78
- movl 60(%esp), %eax
- orl %ebp, %edx
- addl %eax, %edi
- xorl %ebx, %edx
- movl $-1, %eax
- roll $10, %ebp
- leal 2840853838(%edi,%edx,1),%edi
- subl %ebp, %eax
- roll $5, %edi
- addl %esi, %edi
- # 79
- movl 52(%esp), %edx
- orl %ebx, %eax
- addl %edx, %esi
- xorl %edi, %eax
- movl 128(%esp), %edx
- roll $10, %ebx
- leal 2840853838(%esi,%eax,1),%esi
- movl %ecx, 64(%esp)
- roll $6, %esi
- addl %ecx, %esi
- movl (%edx), %ecx
- movl %esi, 68(%esp)
- movl %edi, 72(%esp)
- movl 4(%edx), %esi
- movl %ebx, 76(%esp)
- movl 8(%edx), %edi
- movl %ebp, 80(%esp)
- movl 12(%edx), %ebx
- movl 16(%edx), %ebp
- # 80
- movl $-1, %edx
- subl %ebx, %edx
- movl 20(%esp), %eax
- orl %edi, %edx
- addl %eax, %ecx
- xorl %esi, %edx
- movl $-1, %eax
- roll $10, %edi
- leal 1352829926(%ecx,%edx,1),%ecx
- subl %edi, %eax
- roll $8, %ecx
- addl %ebp, %ecx
- # 81
- movl 56(%esp), %edx
- orl %esi, %eax
- addl %edx, %ebp
- xorl %ecx, %eax
- movl $-1, %edx
- roll $10, %esi
- leal 1352829926(%ebp,%eax,1),%ebp
- subl %esi, %edx
- roll $9, %ebp
- addl %ebx, %ebp
- # 82
- movl 28(%esp), %eax
- orl %ecx, %edx
- addl %eax, %ebx
- xorl %ebp, %edx
- movl $-1, %eax
- roll $10, %ecx
- leal 1352829926(%ebx,%edx,1),%ebx
- subl %ecx, %eax
- roll $9, %ebx
- addl %edi, %ebx
- # 83
- movl (%esp), %edx
- orl %ebp, %eax
- addl %edx, %edi
- xorl %ebx, %eax
- movl $-1, %edx
- roll $10, %ebp
- leal 1352829926(%edi,%eax,1),%edi
- subl %ebp, %edx
- roll $11, %edi
- addl %esi, %edi
- # 84
- movl 36(%esp), %eax
- orl %ebx, %edx
- addl %eax, %esi
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 1352829926(%esi,%edx,1),%esi
- subl %ebx, %eax
- roll $13, %esi
- addl %ecx, %esi
- # 85
- movl 8(%esp), %edx
- orl %edi, %eax
- addl %edx, %ecx
- xorl %esi, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 1352829926(%ecx,%eax,1),%ecx
- subl %edi, %edx
- roll $15, %ecx
- addl %ebp, %ecx
- # 86
- movl 44(%esp), %eax
- orl %esi, %edx
- addl %eax, %ebp
- xorl %ecx, %edx
- movl $-1, %eax
- roll $10, %esi
- leal 1352829926(%ebp,%edx,1),%ebp
- subl %esi, %eax
- roll $15, %ebp
- addl %ebx, %ebp
- # 87
- movl 16(%esp), %edx
- orl %ecx, %eax
- addl %edx, %ebx
- xorl %ebp, %eax
- movl $-1, %edx
- roll $10, %ecx
- leal 1352829926(%ebx,%eax,1),%ebx
- subl %ecx, %edx
- roll $5, %ebx
- addl %edi, %ebx
- # 88
- movl 52(%esp), %eax
- orl %ebp, %edx
- addl %eax, %edi
- xorl %ebx, %edx
- movl $-1, %eax
- roll $10, %ebp
- leal 1352829926(%edi,%edx,1),%edi
- subl %ebp, %eax
- roll $7, %edi
- addl %esi, %edi
- # 89
- movl 24(%esp), %edx
- orl %ebx, %eax
- addl %edx, %esi
- xorl %edi, %eax
- movl $-1, %edx
- roll $10, %ebx
- leal 1352829926(%esi,%eax,1),%esi
- subl %ebx, %edx
- roll $7, %esi
- addl %ecx, %esi
- # 90
- movl 60(%esp), %eax
- orl %edi, %edx
- addl %eax, %ecx
- xorl %esi, %edx
- movl $-1, %eax
- roll $10, %edi
- leal 1352829926(%ecx,%edx,1),%ecx
- subl %edi, %eax
- roll $8, %ecx
- addl %ebp, %ecx
- # 91
- movl 32(%esp), %edx
- orl %esi, %eax
- addl %edx, %ebp
- xorl %ecx, %eax
- movl $-1, %edx
- roll $10, %esi
- leal 1352829926(%ebp,%eax,1),%ebp
- subl %esi, %edx
- roll $11, %ebp
- addl %ebx, %ebp
- # 92
- movl 4(%esp), %eax
- orl %ecx, %edx
- addl %eax, %ebx
- xorl %ebp, %edx
- movl $-1, %eax
- roll $10, %ecx
- leal 1352829926(%ebx,%edx,1),%ebx
- subl %ecx, %eax
- roll $14, %ebx
- addl %edi, %ebx
- # 93
- movl 40(%esp), %edx
- orl %ebp, %eax
- addl %edx, %edi
- xorl %ebx, %eax
- movl $-1, %edx
- roll $10, %ebp
- leal 1352829926(%edi,%eax,1),%edi
- subl %ebp, %edx
- roll $14, %edi
- addl %esi, %edi
- # 94
- movl 12(%esp), %eax
- orl %ebx, %edx
- addl %eax, %esi
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 1352829926(%esi,%edx,1),%esi
- subl %ebx, %eax
- roll $12, %esi
- addl %ecx, %esi
- # 95
- movl 48(%esp), %edx
- orl %edi, %eax
- addl %edx, %ecx
- xorl %esi, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 1352829926(%ecx,%eax,1),%ecx
- movl %edi, %eax
- roll $6, %ecx
- addl %ebp, %ecx
- # 96
- subl %edi, %edx
- andl %ecx, %eax
- andl %esi, %edx
- orl %eax, %edx
- movl 24(%esp), %eax
- roll $10, %esi
- leal 1548603684(%ebp,%edx),%ebp
- movl $-1, %edx
- addl %eax, %ebp
- movl %esi, %eax
- roll $9, %ebp
- addl %ebx, %ebp
- # 97
- subl %esi, %edx
- andl %ebp, %eax
- andl %ecx, %edx
- orl %eax, %edx
- movl 44(%esp), %eax
- roll $10, %ecx
- leal 1548603684(%ebx,%edx),%ebx
- movl $-1, %edx
- addl %eax, %ebx
- movl %ecx, %eax
- roll $13, %ebx
- addl %edi, %ebx
- # 98
- subl %ecx, %edx
- andl %ebx, %eax
- andl %ebp, %edx
- orl %eax, %edx
- movl 12(%esp), %eax
- roll $10, %ebp
- leal 1548603684(%edi,%edx),%edi
- movl $-1, %edx
- addl %eax, %edi
- movl %ebp, %eax
- roll $15, %edi
- addl %esi, %edi
- # 99
- subl %ebp, %edx
- andl %edi, %eax
- andl %ebx, %edx
- orl %eax, %edx
- movl 28(%esp), %eax
- roll $10, %ebx
- leal 1548603684(%esi,%edx),%esi
- movl $-1, %edx
- addl %eax, %esi
- movl %ebx, %eax
- roll $7, %esi
- addl %ecx, %esi
- # 100
- subl %ebx, %edx
- andl %esi, %eax
- andl %edi, %edx
- orl %eax, %edx
- movl (%esp), %eax
- roll $10, %edi
- leal 1548603684(%ecx,%edx),%ecx
- movl $-1, %edx
- addl %eax, %ecx
- movl %edi, %eax
- roll $12, %ecx
- addl %ebp, %ecx
- # 101
- subl %edi, %edx
- andl %ecx, %eax
- andl %esi, %edx
- orl %eax, %edx
- movl 52(%esp), %eax
- roll $10, %esi
- leal 1548603684(%ebp,%edx),%ebp
- movl $-1, %edx
- addl %eax, %ebp
- movl %esi, %eax
- roll $8, %ebp
- addl %ebx, %ebp
- # 102
- subl %esi, %edx
- andl %ebp, %eax
- andl %ecx, %edx
- orl %eax, %edx
- movl 20(%esp), %eax
- roll $10, %ecx
- leal 1548603684(%ebx,%edx),%ebx
- movl $-1, %edx
- addl %eax, %ebx
- movl %ecx, %eax
- roll $9, %ebx
- addl %edi, %ebx
- # 103
- subl %ecx, %edx
- andl %ebx, %eax
- andl %ebp, %edx
- orl %eax, %edx
- movl 40(%esp), %eax
- roll $10, %ebp
- leal 1548603684(%edi,%edx),%edi
- movl $-1, %edx
- addl %eax, %edi
- movl %ebp, %eax
- roll $11, %edi
- addl %esi, %edi
- # 104
- subl %ebp, %edx
- andl %edi, %eax
- andl %ebx, %edx
- orl %eax, %edx
- movl 56(%esp), %eax
- roll $10, %ebx
- leal 1548603684(%esi,%edx),%esi
- movl $-1, %edx
- addl %eax, %esi
- movl %ebx, %eax
- roll $7, %esi
- addl %ecx, %esi
- # 105
- subl %ebx, %edx
- andl %esi, %eax
- andl %edi, %edx
- orl %eax, %edx
- movl 60(%esp), %eax
- roll $10, %edi
- leal 1548603684(%ecx,%edx),%ecx
- movl $-1, %edx
- addl %eax, %ecx
- movl %edi, %eax
- roll $7, %ecx
- addl %ebp, %ecx
- # 106
- subl %edi, %edx
- andl %ecx, %eax
- andl %esi, %edx
- orl %eax, %edx
- movl 32(%esp), %eax
- roll $10, %esi
- leal 1548603684(%ebp,%edx),%ebp
- movl $-1, %edx
- addl %eax, %ebp
- movl %esi, %eax
- roll $12, %ebp
- addl %ebx, %ebp
- # 107
- subl %esi, %edx
- andl %ebp, %eax
- andl %ecx, %edx
- orl %eax, %edx
- movl 48(%esp), %eax
- roll $10, %ecx
- leal 1548603684(%ebx,%edx),%ebx
- movl $-1, %edx
- addl %eax, %ebx
- movl %ecx, %eax
- roll $7, %ebx
- addl %edi, %ebx
- # 108
- subl %ecx, %edx
- andl %ebx, %eax
- andl %ebp, %edx
- orl %eax, %edx
- movl 16(%esp), %eax
- roll $10, %ebp
- leal 1548603684(%edi,%edx),%edi
- movl $-1, %edx
- addl %eax, %edi
- movl %ebp, %eax
- roll $6, %edi
- addl %esi, %edi
- # 109
- subl %ebp, %edx
- andl %edi, %eax
- andl %ebx, %edx
- orl %eax, %edx
- movl 36(%esp), %eax
- roll $10, %ebx
- leal 1548603684(%esi,%edx),%esi
- movl $-1, %edx
- addl %eax, %esi
- movl %ebx, %eax
- roll $15, %esi
- addl %ecx, %esi
- # 110
- subl %ebx, %edx
- andl %esi, %eax
- andl %edi, %edx
- orl %eax, %edx
- movl 4(%esp), %eax
- roll $10, %edi
- leal 1548603684(%ecx,%edx),%ecx
- movl $-1, %edx
- addl %eax, %ecx
- movl %edi, %eax
- roll $13, %ecx
- addl %ebp, %ecx
- # 111
- subl %edi, %edx
- andl %ecx, %eax
- andl %esi, %edx
- orl %eax, %edx
- movl 8(%esp), %eax
- roll $10, %esi
- leal 1548603684(%ebp,%edx),%ebp
- movl $-1, %edx
- addl %eax, %ebp
- subl %ecx, %edx
- roll $11, %ebp
- addl %ebx, %ebp
- # 112
- movl 60(%esp), %eax
- orl %ebp, %edx
- addl %eax, %ebx
- xorl %esi, %edx
- movl $-1, %eax
- roll $10, %ecx
- leal 1836072691(%ebx,%edx,1),%ebx
- subl %ebp, %eax
- roll $9, %ebx
- addl %edi, %ebx
- # 113
- movl 20(%esp), %edx
- orl %ebx, %eax
- addl %edx, %edi
- xorl %ecx, %eax
- movl $-1, %edx
- roll $10, %ebp
- leal 1836072691(%edi,%eax,1),%edi
- subl %ebx, %edx
- roll $7, %edi
- addl %esi, %edi
- # 114
- movl 4(%esp), %eax
- orl %edi, %edx
- addl %eax, %esi
- xorl %ebp, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 1836072691(%esi,%edx,1),%esi
- subl %edi, %eax
- roll $15, %esi
- addl %ecx, %esi
- # 115
- movl 12(%esp), %edx
- orl %esi, %eax
- addl %edx, %ecx
- xorl %ebx, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 1836072691(%ecx,%eax,1),%ecx
- subl %esi, %edx
- roll $11, %ecx
- addl %ebp, %ecx
- # 116
- movl 28(%esp), %eax
- orl %ecx, %edx
- addl %eax, %ebp
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %esi
- leal 1836072691(%ebp,%edx,1),%ebp
- subl %ecx, %eax
- roll $8, %ebp
- addl %ebx, %ebp
- # 117
- movl 56(%esp), %edx
- orl %ebp, %eax
- addl %edx, %ebx
- xorl %esi, %eax
- movl $-1, %edx
- roll $10, %ecx
- leal 1836072691(%ebx,%eax,1),%ebx
- subl %ebp, %edx
- roll $6, %ebx
- addl %edi, %ebx
- # 118
- movl 24(%esp), %eax
- orl %ebx, %edx
- addl %eax, %edi
- xorl %ecx, %edx
- movl $-1, %eax
- roll $10, %ebp
- leal 1836072691(%edi,%edx,1),%edi
- subl %ebx, %eax
- roll $6, %edi
- addl %esi, %edi
- # 119
- movl 36(%esp), %edx
- orl %edi, %eax
- addl %edx, %esi
- xorl %ebp, %eax
- movl $-1, %edx
- roll $10, %ebx
- leal 1836072691(%esi,%eax,1),%esi
- subl %edi, %edx
- roll $14, %esi
- addl %ecx, %esi
- # 120
- movl 44(%esp), %eax
- orl %esi, %edx
- addl %eax, %ecx
- xorl %ebx, %edx
- movl $-1, %eax
- roll $10, %edi
- leal 1836072691(%ecx,%edx,1),%ecx
- subl %esi, %eax
- roll $12, %ecx
- addl %ebp, %ecx
- # 121
- movl 32(%esp), %edx
- orl %ecx, %eax
- addl %edx, %ebp
- xorl %edi, %eax
- movl $-1, %edx
- roll $10, %esi
- leal 1836072691(%ebp,%eax,1),%ebp
- subl %ecx, %edx
- roll $13, %ebp
- addl %ebx, %ebp
- # 122
- movl 48(%esp), %eax
- orl %ebp, %edx
- addl %eax, %ebx
- xorl %esi, %edx
- movl $-1, %eax
- roll $10, %ecx
- leal 1836072691(%ebx,%edx,1),%ebx
- subl %ebp, %eax
- roll $5, %ebx
- addl %edi, %ebx
- # 123
- movl 8(%esp), %edx
- orl %ebx, %eax
- addl %edx, %edi
- xorl %ecx, %eax
- movl $-1, %edx
- roll $10, %ebp
- leal 1836072691(%edi,%eax,1),%edi
- subl %ebx, %edx
- roll $14, %edi
- addl %esi, %edi
- # 124
- movl 40(%esp), %eax
- orl %edi, %edx
- addl %eax, %esi
- xorl %ebp, %edx
- movl $-1, %eax
- roll $10, %ebx
- leal 1836072691(%esi,%edx,1),%esi
- subl %edi, %eax
- roll $13, %esi
- addl %ecx, %esi
- # 125
- movl (%esp), %edx
- orl %esi, %eax
- addl %edx, %ecx
- xorl %ebx, %eax
- movl $-1, %edx
- roll $10, %edi
- leal 1836072691(%ecx,%eax,1),%ecx
- subl %esi, %edx
- roll $13, %ecx
- addl %ebp, %ecx
- # 126
- movl 16(%esp), %eax
- orl %ecx, %edx
- addl %eax, %ebp
- xorl %edi, %edx
- movl $-1, %eax
- roll $10, %esi
- leal 1836072691(%ebp,%edx,1),%ebp
- subl %ecx, %eax
- roll $7, %ebp
- addl %ebx, %ebp
- # 127
- movl 52(%esp), %edx
- orl %ebp, %eax
- addl %edx, %ebx
- xorl %esi, %eax
- movl 32(%esp), %edx
- roll $10, %ecx
- leal 1836072691(%ebx,%eax,1),%ebx
- movl $-1, %eax
- roll $5, %ebx
- addl %edi, %ebx
- # 128
- addl %edx, %edi
- movl %ebp, %edx
- subl %ebx, %eax
- andl %ebx, %edx
- andl %ecx, %eax
- orl %eax, %edx
- movl 24(%esp), %eax
- roll $10, %ebp
- leal 2053994217(%edi,%edx,1),%edi
- movl $-1, %edx
- roll $15, %edi
- addl %esi, %edi
- # 129
- addl %eax, %esi
- movl %ebx, %eax
- subl %edi, %edx
- andl %edi, %eax
- andl %ebp, %edx
- orl %edx, %eax
- movl 16(%esp), %edx
- roll $10, %ebx
- leal 2053994217(%esi,%eax,1),%esi
- movl $-1, %eax
- roll $5, %esi
- addl %ecx, %esi
- # 130
- addl %edx, %ecx
- movl %edi, %edx
- subl %esi, %eax
- andl %esi, %edx
- andl %ebx, %eax
- orl %eax, %edx
- movl 4(%esp), %eax
- roll $10, %edi
- leal 2053994217(%ecx,%edx,1),%ecx
- movl $-1, %edx
- roll $8, %ecx
- addl %ebp, %ecx
- # 131
- addl %eax, %ebp
- movl %esi, %eax
- subl %ecx, %edx
- andl %ecx, %eax
- andl %edi, %edx
- orl %edx, %eax
- movl 12(%esp), %edx
- roll $10, %esi
- leal 2053994217(%ebp,%eax,1),%ebp
- movl $-1, %eax
- roll $11, %ebp
- addl %ebx, %ebp
- # 132
- addl %edx, %ebx
- movl %ecx, %edx
- subl %ebp, %eax
- andl %ebp, %edx
- andl %esi, %eax
- orl %eax, %edx
- movl 44(%esp), %eax
- roll $10, %ecx
- leal 2053994217(%ebx,%edx,1),%ebx
- movl $-1, %edx
- roll $14, %ebx
- addl %edi, %ebx
- # 133
- addl %eax, %edi
- movl %ebp, %eax
- subl %ebx, %edx
- andl %ebx, %eax
- andl %ecx, %edx
- orl %edx, %eax
- movl 60(%esp), %edx
- roll $10, %ebp
- leal 2053994217(%edi,%eax,1),%edi
- movl $-1, %eax
- roll $14, %edi
- addl %esi, %edi
- # 134
- addl %edx, %esi
- movl %ebx, %edx
- subl %edi, %eax
- andl %edi, %edx
- andl %ebp, %eax
- orl %eax, %edx
- movl (%esp), %eax
- roll $10, %ebx
- leal 2053994217(%esi,%edx,1),%esi
- movl $-1, %edx
- roll $6, %esi
- addl %ecx, %esi
- # 135
- addl %eax, %ecx
- movl %edi, %eax
- subl %esi, %edx
- andl %esi, %eax
- andl %ebx, %edx
- orl %edx, %eax
- movl 20(%esp), %edx
- roll $10, %edi
- leal 2053994217(%ecx,%eax,1),%ecx
- movl $-1, %eax
- roll $14, %ecx
- addl %ebp, %ecx
- # 136
- addl %edx, %ebp
- movl %esi, %edx
- subl %ecx, %eax
- andl %ecx, %edx
- andl %edi, %eax
- orl %eax, %edx
- movl 48(%esp), %eax
- roll $10, %esi
- leal 2053994217(%ebp,%edx,1),%ebp
- movl $-1, %edx
- roll $6, %ebp
- addl %ebx, %ebp
- # 137
- addl %eax, %ebx
- movl %ecx, %eax
- subl %ebp, %edx
- andl %ebp, %eax
- andl %esi, %edx
- orl %edx, %eax
- movl 8(%esp), %edx
- roll $10, %ecx
- leal 2053994217(%ebx,%eax,1),%ebx
- movl $-1, %eax
- roll $9, %ebx
- addl %edi, %ebx
- # 138
- addl %edx, %edi
- movl %ebp, %edx
- subl %ebx, %eax
- andl %ebx, %edx
- andl %ecx, %eax
- orl %eax, %edx
- movl 52(%esp), %eax
- roll $10, %ebp
- leal 2053994217(%edi,%edx,1),%edi
- movl $-1, %edx
- roll $12, %edi
- addl %esi, %edi
- # 139
- addl %eax, %esi
- movl %ebx, %eax
- subl %edi, %edx
- andl %edi, %eax
- andl %ebp, %edx
- orl %edx, %eax
- movl 36(%esp), %edx
- roll $10, %ebx
- leal 2053994217(%esi,%eax,1),%esi
- movl $-1, %eax
- roll $9, %esi
- addl %ecx, %esi
- # 140
- addl %edx, %ecx
- movl %edi, %edx
- subl %esi, %eax
- andl %esi, %edx
- andl %ebx, %eax
- orl %eax, %edx
- movl 28(%esp), %eax
- roll $10, %edi
- leal 2053994217(%ecx,%edx,1),%ecx
- movl $-1, %edx
- roll $12, %ecx
- addl %ebp, %ecx
- # 141
- addl %eax, %ebp
- movl %esi, %eax
- subl %ecx, %edx
- andl %ecx, %eax
- andl %edi, %edx
- orl %edx, %eax
- movl 40(%esp), %edx
- roll $10, %esi
- leal 2053994217(%ebp,%eax,1),%ebp
- movl $-1, %eax
- roll $5, %ebp
- addl %ebx, %ebp
- # 142
- addl %edx, %ebx
- movl %ecx, %edx
- subl %ebp, %eax
- andl %ebp, %edx
- andl %esi, %eax
- orl %eax, %edx
- movl 56(%esp), %eax
- roll $10, %ecx
- leal 2053994217(%ebx,%edx,1),%ebx
- movl $-1, %edx
- roll $15, %ebx
- addl %edi, %ebx
- # 143
- addl %eax, %edi
- movl %ebp, %eax
- subl %ebx, %edx
- andl %ebx, %eax
- andl %ecx, %edx
- orl %eax, %edx
- movl %ebx, %eax
- roll $10, %ebp
- leal 2053994217(%edi,%edx,1),%edi
- xorl %ebp, %eax
- roll $8, %edi
- addl %esi, %edi
- # 144
- movl 48(%esp), %edx
- xorl %edi, %eax
- addl %edx, %esi
- roll $10, %ebx
- addl %eax, %esi
- movl %edi, %eax
- roll $8, %esi
- addl %ecx, %esi
- # 145
- xorl %ebx, %eax
- movl 60(%esp), %edx
- xorl %esi, %eax
- addl %eax, %ecx
- movl %esi, %eax
- roll $10, %edi
- addl %edx, %ecx
- xorl %edi, %eax
- roll $5, %ecx
- addl %ebp, %ecx
- # 146
- movl 40(%esp), %edx
- xorl %ecx, %eax
- addl %edx, %ebp
- roll $10, %esi
- addl %eax, %ebp
- movl %ecx, %eax
- roll $12, %ebp
- addl %ebx, %ebp
- # 147
- xorl %esi, %eax
- movl 16(%esp), %edx
- xorl %ebp, %eax
- addl %eax, %ebx
- movl %ebp, %eax
- roll $10, %ecx
- addl %edx, %ebx
- xorl %ecx, %eax
- roll $9, %ebx
- addl %edi, %ebx
- # 148
- movl 4(%esp), %edx
- xorl %ebx, %eax
- addl %edx, %edi
- roll $10, %ebp
- addl %eax, %edi
- movl %ebx, %eax
- roll $12, %edi
- addl %esi, %edi
- # 149
- xorl %ebp, %eax
- movl 20(%esp), %edx
- xorl %edi, %eax
- addl %eax, %esi
- movl %edi, %eax
- roll $10, %ebx
- addl %edx, %esi
- xorl %ebx, %eax
- roll $5, %esi
- addl %ecx, %esi
- # 150
- movl 32(%esp), %edx
- xorl %esi, %eax
- addl %edx, %ecx
- roll $10, %edi
- addl %eax, %ecx
- movl %esi, %eax
- roll $14, %ecx
- addl %ebp, %ecx
- # 151
- xorl %edi, %eax
- movl 28(%esp), %edx
- xorl %ecx, %eax
- addl %eax, %ebp
- movl %ecx, %eax
- roll $10, %esi
- addl %edx, %ebp
- xorl %esi, %eax
- roll $6, %ebp
- addl %ebx, %ebp
- # 152
- movl 24(%esp), %edx
- xorl %ebp, %eax
- addl %edx, %ebx
- roll $10, %ecx
- addl %eax, %ebx
- movl %ebp, %eax
- roll $8, %ebx
- addl %edi, %ebx
- # 153
- xorl %ecx, %eax
- movl 8(%esp), %edx
- xorl %ebx, %eax
- addl %eax, %edi
- movl %ebx, %eax
- roll $10, %ebp
- addl %edx, %edi
- xorl %ebp, %eax
- roll $13, %edi
- addl %esi, %edi
- # 154
- movl 52(%esp), %edx
- xorl %edi, %eax
- addl %edx, %esi
- roll $10, %ebx
- addl %eax, %esi
- movl %edi, %eax
- roll $6, %esi
- addl %ecx, %esi
- # 155
- xorl %ebx, %eax
- movl 56(%esp), %edx
- xorl %esi, %eax
- addl %eax, %ecx
- movl %esi, %eax
- roll $10, %edi
- addl %edx, %ecx
- xorl %edi, %eax
- roll $5, %ecx
- addl %ebp, %ecx
- # 156
- movl (%esp), %edx
- xorl %ecx, %eax
- addl %edx, %ebp
- roll $10, %esi
- addl %eax, %ebp
- movl %ecx, %eax
- roll $15, %ebp
- addl %ebx, %ebp
- # 157
- xorl %esi, %eax
- movl 12(%esp), %edx
- xorl %ebp, %eax
- addl %eax, %ebx
- movl %ebp, %eax
- roll $10, %ecx
- addl %edx, %ebx
- xorl %ecx, %eax
- roll $13, %ebx
- addl %edi, %ebx
- # 158
- movl 36(%esp), %edx
- xorl %ebx, %eax
- addl %edx, %edi
- roll $10, %ebp
- addl %eax, %edi
- movl %ebx, %eax
- roll $11, %edi
- addl %esi, %edi
- # 159
- xorl %ebp, %eax
- movl 44(%esp), %edx
- xorl %edi, %eax
- addl %eax, %esi
- roll $10, %ebx
- addl %edx, %esi
- movl 128(%esp), %edx
- roll $11, %esi
- addl %ecx, %esi
- movl 4(%edx), %eax
- addl %eax, %ebx
- movl 72(%esp), %eax
- addl %eax, %ebx
- movl 8(%edx), %eax
- addl %eax, %ebp
- movl 76(%esp), %eax
- addl %eax, %ebp
- movl 12(%edx), %eax
- addl %eax, %ecx
- movl 80(%esp), %eax
- addl %eax, %ecx
- movl 16(%edx), %eax
- addl %eax, %esi
- movl 64(%esp), %eax
- addl %eax, %esi
- movl (%edx), %eax
- addl %eax, %edi
- movl 68(%esp), %eax
- addl %eax, %edi
- movl 136(%esp), %eax
- movl %ebx, (%edx)
- movl %ebp, 4(%edx)
- movl %ecx, 8(%edx)
- subl $1, %eax
- movl %esi, 12(%edx)
- movl %edi, 16(%edx)
- jle .L001get_out
- movl %eax, 136(%esp)
- movl %ecx, %edi
- movl 132(%esp), %eax
- movl %ebx, %ecx
- addl $64, %eax
- movl %ebp, %esi
- movl %eax, 132(%esp)
- jmp .L000start
-.L001get_out:
- addl $108, %esp
- popl %ebx
- popl %ebp
- popl %edi
- popl %esi
- ret
-.L_ripemd160_block_asm_host_order_end:
- .size ripemd160_block_asm_host_order,.L_ripemd160_block_asm_host_order_end-ripemd160_block_asm_host_order
-.ident "desasm.pl"
diff --git a/secure/lib/libcrypto/i386/sha1-586.s b/secure/lib/libcrypto/i386/sha1-586.s
deleted file mode 100644
index b13c6658edf0..000000000000
--- a/secure/lib/libcrypto/i386/sha1-586.s
+++ /dev/null
@@ -1,1960 +0,0 @@
- # $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by sha1-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
-
- .file "sha1-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl sha1_block_asm_data_order
- .type sha1_block_asm_data_order,@function
-sha1_block_asm_data_order:
- movl 12(%esp), %ecx
- pushl %esi
- sall $6, %ecx
- movl 12(%esp), %esi
- pushl %ebp
- addl %esi, %ecx
- pushl %ebx
- movl 16(%esp), %ebp
- pushl %edi
- movl 12(%ebp), %edx
- subl $108, %esp
- movl 16(%ebp), %edi
- movl 8(%ebp), %ebx
- movl %ecx, 68(%esp)
- # First we need to setup the X array
-.L000start:
- # First, load the words onto the stack in network byte order
- movl (%esi), %eax
- movl 4(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, (%esp)
- movl %ecx, 4(%esp)
- movl 8(%esi), %eax
- movl 12(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, 8(%esp)
- movl %ecx, 12(%esp)
- movl 16(%esi), %eax
- movl 20(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, 16(%esp)
- movl %ecx, 20(%esp)
- movl 24(%esi), %eax
- movl 28(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, 24(%esp)
- movl %ecx, 28(%esp)
- movl 32(%esi), %eax
- movl 36(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, 32(%esp)
- movl %ecx, 36(%esp)
- movl 40(%esi), %eax
- movl 44(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, 40(%esp)
- movl %ecx, 44(%esp)
- movl 48(%esi), %eax
- movl 52(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, 48(%esp)
- movl %ecx, 52(%esp)
- movl 56(%esi), %eax
- movl 60(%esi), %ecx
-.byte 15
-.byte 200 # bswapl %eax
-.byte 15
-.byte 201 # bswapl %ecx
- movl %eax, 56(%esp)
- movl %ecx, 60(%esp)
- # We now have the X array on the stack
- # starting at sp-4
- movl %esi, 132(%esp)
-.L001shortcut:
-
- # Start processing
- movl (%ebp), %eax
- movl 4(%ebp), %ecx
- # 00_15 0
- movl %ebx, %esi
- movl %eax, %ebp
- xorl %edx, %esi
- roll $5, %ebp
- andl %ecx, %esi
- addl %edi, %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- movl (%esp), %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %edx, %esi
- leal 1518500249(%ebp,%edi,1),%ebp
- movl %ecx, %edi
- addl %ebp, %esi
- xorl %ebx, %edi
- movl %esi, %ebp
- andl %eax, %edi
- roll $5, %ebp
- addl %edx, %ebp
- movl 4(%esp), %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- xorl %ebx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
- leal 1518500249(%ebp,%edx,1),%ebp
- addl %ebp, %edi
- # 00_15 2
- movl %eax, %edx
- movl %edi, %ebp
- xorl %ecx, %edx
- roll $5, %ebp
- andl %esi, %edx
- addl %ebx, %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- movl 8(%esp), %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ecx, %edx
- leal 1518500249(%ebp,%ebx,1),%ebp
- movl %esi, %ebx
- addl %ebp, %edx
- xorl %eax, %ebx
- movl %edx, %ebp
- andl %edi, %ebx
- roll $5, %ebp
- addl %ecx, %ebp
- movl 12(%esp), %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- xorl %eax, %ebx
-.byte 209
-.byte 207 # rorl $1 %edi
- leal 1518500249(%ebp,%ecx,1),%ebp
- addl %ebp, %ebx
- # 00_15 4
- movl %edi, %ecx
- movl %ebx, %ebp
- xorl %esi, %ecx
- roll $5, %ebp
- andl %edx, %ecx
- addl %eax, %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- movl 16(%esp), %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %esi, %ecx
- leal 1518500249(%ebp,%eax,1),%ebp
- movl %edx, %eax
- addl %ebp, %ecx
- xorl %edi, %eax
- movl %ecx, %ebp
- andl %ebx, %eax
- roll $5, %ebp
- addl %esi, %ebp
- movl 20(%esp), %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- xorl %edi, %eax
-.byte 209
-.byte 203 # rorl $1 %ebx
- leal 1518500249(%ebp,%esi,1),%ebp
- addl %ebp, %eax
- # 00_15 6
- movl %ebx, %esi
- movl %eax, %ebp
- xorl %edx, %esi
- roll $5, %ebp
- andl %ecx, %esi
- addl %edi, %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- movl 24(%esp), %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %edx, %esi
- leal 1518500249(%ebp,%edi,1),%ebp
- movl %ecx, %edi
- addl %ebp, %esi
- xorl %ebx, %edi
- movl %esi, %ebp
- andl %eax, %edi
- roll $5, %ebp
- addl %edx, %ebp
- movl 28(%esp), %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- xorl %ebx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
- leal 1518500249(%ebp,%edx,1),%ebp
- addl %ebp, %edi
- # 00_15 8
- movl %eax, %edx
- movl %edi, %ebp
- xorl %ecx, %edx
- roll $5, %ebp
- andl %esi, %edx
- addl %ebx, %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- movl 32(%esp), %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ecx, %edx
- leal 1518500249(%ebp,%ebx,1),%ebp
- movl %esi, %ebx
- addl %ebp, %edx
- xorl %eax, %ebx
- movl %edx, %ebp
- andl %edi, %ebx
- roll $5, %ebp
- addl %ecx, %ebp
- movl 36(%esp), %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- xorl %eax, %ebx
-.byte 209
-.byte 207 # rorl $1 %edi
- leal 1518500249(%ebp,%ecx,1),%ebp
- addl %ebp, %ebx
- # 00_15 10
- movl %edi, %ecx
- movl %ebx, %ebp
- xorl %esi, %ecx
- roll $5, %ebp
- andl %edx, %ecx
- addl %eax, %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- movl 40(%esp), %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %esi, %ecx
- leal 1518500249(%ebp,%eax,1),%ebp
- movl %edx, %eax
- addl %ebp, %ecx
- xorl %edi, %eax
- movl %ecx, %ebp
- andl %ebx, %eax
- roll $5, %ebp
- addl %esi, %ebp
- movl 44(%esp), %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- xorl %edi, %eax
-.byte 209
-.byte 203 # rorl $1 %ebx
- leal 1518500249(%ebp,%esi,1),%ebp
- addl %ebp, %eax
- # 00_15 12
- movl %ebx, %esi
- movl %eax, %ebp
- xorl %edx, %esi
- roll $5, %ebp
- andl %ecx, %esi
- addl %edi, %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- movl 48(%esp), %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %edx, %esi
- leal 1518500249(%ebp,%edi,1),%ebp
- movl %ecx, %edi
- addl %ebp, %esi
- xorl %ebx, %edi
- movl %esi, %ebp
- andl %eax, %edi
- roll $5, %ebp
- addl %edx, %ebp
- movl 52(%esp), %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- xorl %ebx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
- leal 1518500249(%ebp,%edx,1),%ebp
- addl %ebp, %edi
- # 00_15 14
- movl %eax, %edx
- movl %edi, %ebp
- xorl %ecx, %edx
- roll $5, %ebp
- andl %esi, %edx
- addl %ebx, %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- movl 56(%esp), %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ecx, %edx
- leal 1518500249(%ebp,%ebx,1),%ebp
- movl %esi, %ebx
- addl %ebp, %edx
- xorl %eax, %ebx
- movl %edx, %ebp
- andl %edi, %ebx
- roll $5, %ebp
- addl %ecx, %ebp
- movl 60(%esp), %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- xorl %eax, %ebx
-.byte 209
-.byte 207 # rorl $1 %edi
- leal 1518500249(%ebp,%ecx,1),%ebp
- addl %ebp, %ebx
- # 16_19 16
- nop
- movl (%esp), %ebp
- movl 8(%esp), %ecx
- xorl %ebp, %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 52(%esp), %ebp
- xorl %ebp, %ecx
- movl %edi, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- xorl %esi, %ebp
- movl %ecx, (%esp)
- andl %edx, %ebp
- leal 1518500249(%ecx,%eax,1),%ecx
- xorl %esi, %ebp
- movl %ebx, %eax
- addl %ebp, %ecx
- roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %eax, %ecx
- movl 4(%esp), %eax
- movl 12(%esp), %ebp
- xorl %ebp, %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 56(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
- movl %edx, %ebp
- xorl %edi, %ebp
- movl %eax, 4(%esp)
- andl %ebx, %ebp
- leal 1518500249(%eax,%esi,1),%eax
- xorl %edi, %ebp
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %eax
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %eax
- # 16_19 18
- movl 8(%esp), %ebp
- movl 16(%esp), %esi
- xorl %ebp, %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl 60(%esp), %ebp
- xorl %ebp, %esi
- movl %ebx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %edx, %ebp
- movl %esi, 8(%esp)
- andl %ecx, %ebp
- leal 1518500249(%esi,%edi,1),%esi
- xorl %edx, %ebp
- movl %eax, %edi
- addl %ebp, %esi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- movl 12(%esp), %edi
- movl 20(%esp), %ebp
- xorl %ebp, %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl (%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
- movl %ecx, %ebp
- xorl %ebx, %ebp
- movl %edi, 12(%esp)
- andl %eax, %ebp
- leal 1518500249(%edi,%edx,1),%edi
- xorl %ebx, %ebp
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edi
- # 20_39 20
- movl 16(%esp), %edx
- movl 24(%esp), %ebp
- xorl %ebp, %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 4(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- xorl %eax, %ebp
- movl %edx, 16(%esp)
- xorl %ecx, %ebp
- leal 1859775393(%edx,%ebx,1),%edx
- movl %edi, %ebx
- roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebx, %edx
- # 20_39 21
- movl 20(%esp), %ebx
- movl 28(%esp), %ebp
- xorl %ebp, %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 8(%esp), %ebp
- xorl %ebp, %ebx
- movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
- xorl %esi, %ebp
- movl %ebx, 20(%esp)
- xorl %eax, %ebp
- leal 1859775393(%ebx,%ecx,1),%ebx
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebx
- # 20_39 22
- movl 24(%esp), %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 12(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- xorl %edi, %ebp
- movl %ecx, 24(%esp)
- xorl %esi, %ebp
- leal 1859775393(%ecx,%eax,1),%ecx
- movl %ebx, %eax
- roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %eax, %ecx
- # 20_39 23
- movl 28(%esp), %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 16(%esp), %ebp
- xorl %ebp, %eax
- movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
- xorl %edx, %ebp
- movl %eax, 28(%esp)
- xorl %edi, %ebp
- leal 1859775393(%eax,%esi,1),%eax
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %eax
- # 20_39 24
- movl 32(%esp), %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 20(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %ebx, %ebp
- movl %esi, 32(%esp)
- xorl %edx, %ebp
- leal 1859775393(%esi,%edi,1),%esi
- movl %eax, %edi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- # 20_39 25
- movl 36(%esp), %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- xorl %ebp, %edi
- movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
- xorl %ecx, %ebp
- movl %edi, 36(%esp)
- xorl %ebx, %ebp
- leal 1859775393(%edi,%edx,1),%edi
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
- # 20_39 26
- movl 40(%esp), %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 8(%esp), %ebp
- xorl %ebp, %edx
- movl 28(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- xorl %eax, %ebp
- movl %edx, 40(%esp)
- xorl %ecx, %ebp
- leal 1859775393(%edx,%ebx,1),%edx
- movl %edi, %ebx
- roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebx, %edx
- # 20_39 27
- movl 44(%esp), %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 12(%esp), %ebp
- xorl %ebp, %ebx
- movl 32(%esp), %ebp
- xorl %ebp, %ebx
- movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
- xorl %esi, %ebp
- movl %ebx, 44(%esp)
- xorl %eax, %ebp
- leal 1859775393(%ebx,%ecx,1),%ebx
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebx
- # 20_39 28
- movl 48(%esp), %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 16(%esp), %ebp
- xorl %ebp, %ecx
- movl 36(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- xorl %edi, %ebp
- movl %ecx, 48(%esp)
- xorl %esi, %ebp
- leal 1859775393(%ecx,%eax,1),%ecx
- movl %ebx, %eax
- roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %eax, %ecx
- # 20_39 29
- movl 52(%esp), %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 20(%esp), %ebp
- xorl %ebp, %eax
- movl 40(%esp), %ebp
- xorl %ebp, %eax
- movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
- xorl %edx, %ebp
- movl %eax, 52(%esp)
- xorl %edi, %ebp
- leal 1859775393(%eax,%esi,1),%eax
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %eax
- # 20_39 30
- movl 56(%esp), %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- xorl %ebp, %esi
- movl 44(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %ebx, %ebp
- movl %esi, 56(%esp)
- xorl %edx, %ebp
- leal 1859775393(%esi,%edi,1),%esi
- movl %eax, %edi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- # 20_39 31
- movl 60(%esp), %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- xorl %ebp, %edi
- movl 48(%esp), %ebp
- xorl %ebp, %edi
- movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
- xorl %ecx, %ebp
- movl %edi, 60(%esp)
- xorl %ebx, %ebp
- leal 1859775393(%edi,%edx,1),%edi
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
- # 20_39 32
- movl (%esp), %edx
- movl 8(%esp), %ebp
- xorl %ebp, %edx
- movl 32(%esp), %ebp
- xorl %ebp, %edx
- movl 52(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- xorl %eax, %ebp
- movl %edx, (%esp)
- xorl %ecx, %ebp
- leal 1859775393(%edx,%ebx,1),%edx
- movl %edi, %ebx
- roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebx, %edx
- # 20_39 33
- movl 4(%esp), %ebx
- movl 12(%esp), %ebp
- xorl %ebp, %ebx
- movl 36(%esp), %ebp
- xorl %ebp, %ebx
- movl 56(%esp), %ebp
- xorl %ebp, %ebx
- movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
- xorl %esi, %ebp
- movl %ebx, 4(%esp)
- xorl %eax, %ebp
- leal 1859775393(%ebx,%ecx,1),%ebx
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebx
- # 20_39 34
- movl 8(%esp), %ecx
- movl 16(%esp), %ebp
- xorl %ebp, %ecx
- movl 40(%esp), %ebp
- xorl %ebp, %ecx
- movl 60(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- xorl %edi, %ebp
- movl %ecx, 8(%esp)
- xorl %esi, %ebp
- leal 1859775393(%ecx,%eax,1),%ecx
- movl %ebx, %eax
- roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %eax, %ecx
- # 20_39 35
- movl 12(%esp), %eax
- movl 20(%esp), %ebp
- xorl %ebp, %eax
- movl 44(%esp), %ebp
- xorl %ebp, %eax
- movl (%esp), %ebp
- xorl %ebp, %eax
- movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
- xorl %edx, %ebp
- movl %eax, 12(%esp)
- xorl %edi, %ebp
- leal 1859775393(%eax,%esi,1),%eax
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %eax
- # 20_39 36
- movl 16(%esp), %esi
- movl 24(%esp), %ebp
- xorl %ebp, %esi
- movl 48(%esp), %ebp
- xorl %ebp, %esi
- movl 4(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %ebx, %ebp
- movl %esi, 16(%esp)
- xorl %edx, %ebp
- leal 1859775393(%esi,%edi,1),%esi
- movl %eax, %edi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- # 20_39 37
- movl 20(%esp), %edi
- movl 28(%esp), %ebp
- xorl %ebp, %edi
- movl 52(%esp), %ebp
- xorl %ebp, %edi
- movl 8(%esp), %ebp
- xorl %ebp, %edi
- movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
- xorl %ecx, %ebp
- movl %edi, 20(%esp)
- xorl %ebx, %ebp
- leal 1859775393(%edi,%edx,1),%edi
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
- # 20_39 38
- movl 24(%esp), %edx
- movl 32(%esp), %ebp
- xorl %ebp, %edx
- movl 56(%esp), %ebp
- xorl %ebp, %edx
- movl 12(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- xorl %eax, %ebp
- movl %edx, 24(%esp)
- xorl %ecx, %ebp
- leal 1859775393(%edx,%ebx,1),%edx
- movl %edi, %ebx
- roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebx, %edx
- # 20_39 39
- movl 28(%esp), %ebx
- movl 36(%esp), %ebp
- xorl %ebp, %ebx
- movl 60(%esp), %ebp
- xorl %ebp, %ebx
- movl 16(%esp), %ebp
- xorl %ebp, %ebx
- movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
- xorl %esi, %ebp
- movl %ebx, 28(%esp)
- xorl %eax, %ebp
- leal 1859775393(%ebx,%ecx,1),%ebx
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebx
- # 40_59 40
- movl 32(%esp), %ecx
- movl 40(%esp), %ebp
- xorl %ebp, %ecx
- movl (%esp), %ebp
- xorl %ebp, %ecx
- movl 20(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- orl %edi, %ebp
- movl %ecx, 32(%esp)
- andl %esi, %ebp
- leal 2400959708(%ecx,%eax,1),%ecx
- movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- andl %edi, %eax
- orl %eax, %ebp
- movl %ebx, %eax
- roll $5, %eax
- addl %eax, %ebp
- movl 36(%esp), %eax
- addl %ebp, %ecx
- movl 44(%esp), %ebp
- xorl %ebp, %eax
- movl 4(%esp), %ebp
- xorl %ebp, %eax
- movl 24(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
- movl %ebx, %ebp
- movl %eax, 36(%esp)
- orl %edx, %ebp
- leal 2400959708(%eax,%esi,1),%eax
- movl %ebx, %esi
- andl %edi, %ebp
- andl %edx, %esi
- orl %esi, %ebp
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %eax
- # 40_59 41
- # 40_59 42
- movl 40(%esp), %esi
- movl 48(%esp), %ebp
- xorl %ebp, %esi
- movl 8(%esp), %ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- orl %ebx, %ebp
- movl %esi, 40(%esp)
- andl %edx, %ebp
- leal 2400959708(%esi,%edi,1),%esi
- movl %ecx, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- andl %ebx, %edi
- orl %edi, %ebp
- movl %eax, %edi
- roll $5, %edi
- addl %edi, %ebp
- movl 44(%esp), %edi
- addl %ebp, %esi
- movl 52(%esp), %ebp
- xorl %ebp, %edi
- movl 12(%esp), %ebp
- xorl %ebp, %edi
- movl 32(%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
- movl %eax, %ebp
- movl %edi, 44(%esp)
- orl %ecx, %ebp
- leal 2400959708(%edi,%edx,1),%edi
- movl %eax, %edx
- andl %ebx, %ebp
- andl %ecx, %edx
- orl %edx, %ebp
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edi
- # 40_59 43
- # 40_59 44
- movl 48(%esp), %edx
- movl 56(%esp), %ebp
- xorl %ebp, %edx
- movl 16(%esp), %ebp
- xorl %ebp, %edx
- movl 36(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- orl %eax, %ebp
- movl %edx, 48(%esp)
- andl %ecx, %ebp
- leal 2400959708(%edx,%ebx,1),%edx
- movl %esi, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- andl %eax, %ebx
- orl %ebx, %ebp
- movl %edi, %ebx
- roll $5, %ebx
- addl %ebx, %ebp
- movl 52(%esp), %ebx
- addl %ebp, %edx
- movl 60(%esp), %ebp
- xorl %ebp, %ebx
- movl 20(%esp), %ebp
- xorl %ebp, %ebx
- movl 40(%esp), %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ebp, %ebx
-.byte 209
-.byte 195 # roll $1 %ebx
- movl %edi, %ebp
- movl %ebx, 52(%esp)
- orl %esi, %ebp
- leal 2400959708(%ebx,%ecx,1),%ebx
- movl %edi, %ecx
- andl %eax, %ebp
- andl %esi, %ecx
- orl %ecx, %ebp
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebp
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ebx
- # 40_59 45
- # 40_59 46
- movl 56(%esp), %ecx
- movl (%esp), %ebp
- xorl %ebp, %ecx
- movl 24(%esp), %ebp
- xorl %ebp, %ecx
- movl 44(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- orl %edi, %ebp
- movl %ecx, 56(%esp)
- andl %esi, %ebp
- leal 2400959708(%ecx,%eax,1),%ecx
- movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- andl %edi, %eax
- orl %eax, %ebp
- movl %ebx, %eax
- roll $5, %eax
- addl %eax, %ebp
- movl 60(%esp), %eax
- addl %ebp, %ecx
- movl 4(%esp), %ebp
- xorl %ebp, %eax
- movl 28(%esp), %ebp
- xorl %ebp, %eax
- movl 48(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
- movl %ebx, %ebp
- movl %eax, 60(%esp)
- orl %edx, %ebp
- leal 2400959708(%eax,%esi,1),%eax
- movl %ebx, %esi
- andl %edi, %ebp
- andl %edx, %esi
- orl %esi, %ebp
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %eax
- # 40_59 47
- # 40_59 48
- movl (%esp), %esi
- movl 8(%esp), %ebp
- xorl %ebp, %esi
- movl 32(%esp), %ebp
- xorl %ebp, %esi
- movl 52(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- orl %ebx, %ebp
- movl %esi, (%esp)
- andl %edx, %ebp
- leal 2400959708(%esi,%edi,1),%esi
- movl %ecx, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- andl %ebx, %edi
- orl %edi, %ebp
- movl %eax, %edi
- roll $5, %edi
- addl %edi, %ebp
- movl 4(%esp), %edi
- addl %ebp, %esi
- movl 12(%esp), %ebp
- xorl %ebp, %edi
- movl 36(%esp), %ebp
- xorl %ebp, %edi
- movl 56(%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
- movl %eax, %ebp
- movl %edi, 4(%esp)
- orl %ecx, %ebp
- leal 2400959708(%edi,%edx,1),%edi
- movl %eax, %edx
- andl %ebx, %ebp
- andl %ecx, %edx
- orl %edx, %ebp
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edi
- # 40_59 49
- # 40_59 50
- movl 8(%esp), %edx
- movl 16(%esp), %ebp
- xorl %ebp, %edx
- movl 40(%esp), %ebp
- xorl %ebp, %edx
- movl 60(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- orl %eax, %ebp
- movl %edx, 8(%esp)
- andl %ecx, %ebp
- leal 2400959708(%edx,%ebx,1),%edx
- movl %esi, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- andl %eax, %ebx
- orl %ebx, %ebp
- movl %edi, %ebx
- roll $5, %ebx
- addl %ebx, %ebp
- movl 12(%esp), %ebx
- addl %ebp, %edx
- movl 20(%esp), %ebp
- xorl %ebp, %ebx
- movl 44(%esp), %ebp
- xorl %ebp, %ebx
- movl (%esp), %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ebp, %ebx
-.byte 209
-.byte 195 # roll $1 %ebx
- movl %edi, %ebp
- movl %ebx, 12(%esp)
- orl %esi, %ebp
- leal 2400959708(%ebx,%ecx,1),%ebx
- movl %edi, %ecx
- andl %eax, %ebp
- andl %esi, %ecx
- orl %ecx, %ebp
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebp
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ebx
- # 40_59 51
- # 40_59 52
- movl 16(%esp), %ecx
- movl 24(%esp), %ebp
- xorl %ebp, %ecx
- movl 48(%esp), %ebp
- xorl %ebp, %ecx
- movl 4(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- orl %edi, %ebp
- movl %ecx, 16(%esp)
- andl %esi, %ebp
- leal 2400959708(%ecx,%eax,1),%ecx
- movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- andl %edi, %eax
- orl %eax, %ebp
- movl %ebx, %eax
- roll $5, %eax
- addl %eax, %ebp
- movl 20(%esp), %eax
- addl %ebp, %ecx
- movl 28(%esp), %ebp
- xorl %ebp, %eax
- movl 52(%esp), %ebp
- xorl %ebp, %eax
- movl 8(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
- movl %ebx, %ebp
- movl %eax, 20(%esp)
- orl %edx, %ebp
- leal 2400959708(%eax,%esi,1),%eax
- movl %ebx, %esi
- andl %edi, %ebp
- andl %edx, %esi
- orl %esi, %ebp
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %eax
- # 40_59 53
- # 40_59 54
- movl 24(%esp), %esi
- movl 32(%esp), %ebp
- xorl %ebp, %esi
- movl 56(%esp), %ebp
- xorl %ebp, %esi
- movl 12(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- orl %ebx, %ebp
- movl %esi, 24(%esp)
- andl %edx, %ebp
- leal 2400959708(%esi,%edi,1),%esi
- movl %ecx, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- andl %ebx, %edi
- orl %edi, %ebp
- movl %eax, %edi
- roll $5, %edi
- addl %edi, %ebp
- movl 28(%esp), %edi
- addl %ebp, %esi
- movl 36(%esp), %ebp
- xorl %ebp, %edi
- movl 60(%esp), %ebp
- xorl %ebp, %edi
- movl 16(%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
- movl %eax, %ebp
- movl %edi, 28(%esp)
- orl %ecx, %ebp
- leal 2400959708(%edi,%edx,1),%edi
- movl %eax, %edx
- andl %ebx, %ebp
- andl %ecx, %edx
- orl %edx, %ebp
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edi
- # 40_59 55
- # 40_59 56
- movl 32(%esp), %edx
- movl 40(%esp), %ebp
- xorl %ebp, %edx
- movl (%esp), %ebp
- xorl %ebp, %edx
- movl 20(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- orl %eax, %ebp
- movl %edx, 32(%esp)
- andl %ecx, %ebp
- leal 2400959708(%edx,%ebx,1),%edx
- movl %esi, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- andl %eax, %ebx
- orl %ebx, %ebp
- movl %edi, %ebx
- roll $5, %ebx
- addl %ebx, %ebp
- movl 36(%esp), %ebx
- addl %ebp, %edx
- movl 44(%esp), %ebp
- xorl %ebp, %ebx
- movl 4(%esp), %ebp
- xorl %ebp, %ebx
- movl 24(%esp), %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ebp, %ebx
-.byte 209
-.byte 195 # roll $1 %ebx
- movl %edi, %ebp
- movl %ebx, 36(%esp)
- orl %esi, %ebp
- leal 2400959708(%ebx,%ecx,1),%ebx
- movl %edi, %ecx
- andl %eax, %ebp
- andl %esi, %ecx
- orl %ecx, %ebp
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebp
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ebx
- # 40_59 57
- # 40_59 58
- movl 40(%esp), %ecx
- movl 48(%esp), %ebp
- xorl %ebp, %ecx
- movl 8(%esp), %ebp
- xorl %ebp, %ecx
- movl 28(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- orl %edi, %ebp
- movl %ecx, 40(%esp)
- andl %esi, %ebp
- leal 2400959708(%ecx,%eax,1),%ecx
- movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- andl %edi, %eax
- orl %eax, %ebp
- movl %ebx, %eax
- roll $5, %eax
- addl %eax, %ebp
- movl 44(%esp), %eax
- addl %ebp, %ecx
- movl 52(%esp), %ebp
- xorl %ebp, %eax
- movl 12(%esp), %ebp
- xorl %ebp, %eax
- movl 32(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
- movl %ebx, %ebp
- movl %eax, 44(%esp)
- orl %edx, %ebp
- leal 2400959708(%eax,%esi,1),%eax
- movl %ebx, %esi
- andl %edi, %ebp
- andl %edx, %esi
- orl %esi, %ebp
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %eax
- # 40_59 59
- # 20_39 60
- movl 48(%esp), %esi
- movl 56(%esp), %ebp
- xorl %ebp, %esi
- movl 16(%esp), %ebp
- xorl %ebp, %esi
- movl 36(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %ebx, %ebp
- movl %esi, 48(%esp)
- xorl %edx, %ebp
- leal 3395469782(%esi,%edi,1),%esi
- movl %eax, %edi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- # 20_39 61
- movl 52(%esp), %edi
- movl 60(%esp), %ebp
- xorl %ebp, %edi
- movl 20(%esp), %ebp
- xorl %ebp, %edi
- movl 40(%esp), %ebp
- xorl %ebp, %edi
- movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
- xorl %ecx, %ebp
- movl %edi, 52(%esp)
- xorl %ebx, %ebp
- leal 3395469782(%edi,%edx,1),%edi
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
- # 20_39 62
- movl 56(%esp), %edx
- movl (%esp), %ebp
- xorl %ebp, %edx
- movl 24(%esp), %ebp
- xorl %ebp, %edx
- movl 44(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- xorl %eax, %ebp
- movl %edx, 56(%esp)
- xorl %ecx, %ebp
- leal 3395469782(%edx,%ebx,1),%edx
- movl %edi, %ebx
- roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebx, %edx
- # 20_39 63
- movl 60(%esp), %ebx
- movl 4(%esp), %ebp
- xorl %ebp, %ebx
- movl 28(%esp), %ebp
- xorl %ebp, %ebx
- movl 48(%esp), %ebp
- xorl %ebp, %ebx
- movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
- xorl %esi, %ebp
- movl %ebx, 60(%esp)
- xorl %eax, %ebp
- leal 3395469782(%ebx,%ecx,1),%ebx
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebx
- # 20_39 64
- movl (%esp), %ecx
- movl 8(%esp), %ebp
- xorl %ebp, %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 52(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- xorl %edi, %ebp
- movl %ecx, (%esp)
- xorl %esi, %ebp
- leal 3395469782(%ecx,%eax,1),%ecx
- movl %ebx, %eax
- roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %eax, %ecx
- # 20_39 65
- movl 4(%esp), %eax
- movl 12(%esp), %ebp
- xorl %ebp, %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 56(%esp), %ebp
- xorl %ebp, %eax
- movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
- xorl %edx, %ebp
- movl %eax, 4(%esp)
- xorl %edi, %ebp
- leal 3395469782(%eax,%esi,1),%eax
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %eax
- # 20_39 66
- movl 8(%esp), %esi
- movl 16(%esp), %ebp
- xorl %ebp, %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl 60(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %ebx, %ebp
- movl %esi, 8(%esp)
- xorl %edx, %ebp
- leal 3395469782(%esi,%edi,1),%esi
- movl %eax, %edi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- # 20_39 67
- movl 12(%esp), %edi
- movl 20(%esp), %ebp
- xorl %ebp, %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl (%esp), %ebp
- xorl %ebp, %edi
- movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
- xorl %ecx, %ebp
- movl %edi, 12(%esp)
- xorl %ebx, %ebp
- leal 3395469782(%edi,%edx,1),%edi
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
- # 20_39 68
- movl 16(%esp), %edx
- movl 24(%esp), %ebp
- xorl %ebp, %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 4(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- xorl %eax, %ebp
- movl %edx, 16(%esp)
- xorl %ecx, %ebp
- leal 3395469782(%edx,%ebx,1),%edx
- movl %edi, %ebx
- roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebx, %edx
- # 20_39 69
- movl 20(%esp), %ebx
- movl 28(%esp), %ebp
- xorl %ebp, %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 8(%esp), %ebp
- xorl %ebp, %ebx
- movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
- xorl %esi, %ebp
- movl %ebx, 20(%esp)
- xorl %eax, %ebp
- leal 3395469782(%ebx,%ecx,1),%ebx
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebx
- # 20_39 70
- movl 24(%esp), %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 12(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- xorl %edi, %ebp
- movl %ecx, 24(%esp)
- xorl %esi, %ebp
- leal 3395469782(%ecx,%eax,1),%ecx
- movl %ebx, %eax
- roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %eax, %ecx
- # 20_39 71
- movl 28(%esp), %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 16(%esp), %ebp
- xorl %ebp, %eax
- movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
- xorl %edx, %ebp
- movl %eax, 28(%esp)
- xorl %edi, %ebp
- leal 3395469782(%eax,%esi,1),%eax
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %eax
- # 20_39 72
- movl 32(%esp), %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 20(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %ebx, %ebp
- movl %esi, 32(%esp)
- xorl %edx, %ebp
- leal 3395469782(%esi,%edi,1),%esi
- movl %eax, %edi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- # 20_39 73
- movl 36(%esp), %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- xorl %ebp, %edi
- movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
- xorl %ecx, %ebp
- movl %edi, 36(%esp)
- xorl %ebx, %ebp
- leal 3395469782(%edi,%edx,1),%edi
- movl %esi, %edx
- roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
- # 20_39 74
- movl 40(%esp), %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 8(%esp), %ebp
- xorl %ebp, %edx
- movl 28(%esp), %ebp
- xorl %ebp, %edx
- movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
- xorl %eax, %ebp
- movl %edx, 40(%esp)
- xorl %ecx, %ebp
- leal 3395469782(%edx,%ebx,1),%edx
- movl %edi, %ebx
- roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebx, %edx
- # 20_39 75
- movl 44(%esp), %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 12(%esp), %ebp
- xorl %ebp, %ebx
- movl 32(%esp), %ebp
- xorl %ebp, %ebx
- movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
- xorl %esi, %ebp
- movl %ebx, 44(%esp)
- xorl %eax, %ebp
- leal 3395469782(%ebx,%ecx,1),%ebx
- movl %edx, %ecx
- roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ecx, %ebx
- # 20_39 76
- movl 48(%esp), %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 16(%esp), %ebp
- xorl %ebp, %ecx
- movl 36(%esp), %ebp
- xorl %ebp, %ecx
- movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
- xorl %edi, %ebp
- movl %ecx, 48(%esp)
- xorl %esi, %ebp
- leal 3395469782(%ecx,%eax,1),%ecx
- movl %ebx, %eax
- roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %eax, %ecx
- # 20_39 77
- movl 52(%esp), %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 20(%esp), %ebp
- xorl %ebp, %eax
- movl 40(%esp), %ebp
- xorl %ebp, %eax
- movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
- xorl %edx, %ebp
- movl %eax, 52(%esp)
- xorl %edi, %ebp
- leal 3395469782(%eax,%esi,1),%eax
- movl %ecx, %esi
- roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %esi, %eax
- # 20_39 78
- movl 56(%esp), %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- xorl %ebp, %esi
- movl 44(%esp), %ebp
- xorl %ebp, %esi
- movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
- xorl %ebx, %ebp
- movl %esi, 56(%esp)
- xorl %edx, %ebp
- leal 3395469782(%esi,%edi,1),%esi
- movl %eax, %edi
- roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %edi, %esi
- # 20_39 79
- movl 60(%esp), %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- xorl %ebp, %edi
- movl 48(%esp), %ebp
- xorl %ebp, %edi
- movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
- xorl %ecx, %ebp
- movl %edi, 60(%esp)
- xorl %ebx, %ebp
- leal 3395469782(%edi,%edx,1),%edi
- movl %esi, %edx
- roll $5, %edx
- addl %ebp, %edx
- movl 128(%esp), %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %edx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
- # End processing
-
- movl 12(%ebp), %edx
- addl %ecx, %edx
- movl 4(%ebp), %ecx
- addl %esi, %ecx
- movl %eax, %esi
- movl (%ebp), %eax
- movl %edx, 12(%ebp)
- addl %edi, %eax
- movl 16(%ebp), %edi
- addl %ebx, %edi
- movl 8(%ebp), %ebx
- addl %esi, %ebx
- movl %eax, (%ebp)
- movl 132(%esp), %esi
- movl %ebx, 8(%ebp)
- addl $64, %esi
- movl 68(%esp), %eax
- movl %edi, 16(%ebp)
- cmpl %eax, %esi
- movl %ecx, 4(%ebp)
- jl .L000start
- addl $108, %esp
- popl %edi
- popl %ebx
- popl %ebp
- popl %esi
- ret
-.L_sha1_block_asm_data_order_end:
- .size sha1_block_asm_data_order,.L_sha1_block_asm_data_order_end-sha1_block_asm_data_order
-.ident "desasm.pl"
-.text
- .align 16
-.globl sha1_block_asm_host_order
- .type sha1_block_asm_host_order,@function
-sha1_block_asm_host_order:
- movl 12(%esp), %ecx
- pushl %esi
- sall $6, %ecx
- movl 12(%esp), %esi
- pushl %ebp
- addl %esi, %ecx
- pushl %ebx
- movl 16(%esp), %ebp
- pushl %edi
- movl 12(%ebp), %edx
- subl $108, %esp
- movl 16(%ebp), %edi
- movl 8(%ebp), %ebx
- movl %ecx, 68(%esp)
- # First we need to setup the X array
- movl (%esi), %eax
- movl 4(%esi), %ecx
- movl %eax, (%esp)
- movl %ecx, 4(%esp)
- movl 8(%esi), %eax
- movl 12(%esi), %ecx
- movl %eax, 8(%esp)
- movl %ecx, 12(%esp)
- movl 16(%esi), %eax
- movl 20(%esi), %ecx
- movl %eax, 16(%esp)
- movl %ecx, 20(%esp)
- movl 24(%esi), %eax
- movl 28(%esi), %ecx
- movl %eax, 24(%esp)
- movl %ecx, 28(%esp)
- movl 32(%esi), %eax
- movl 36(%esi), %ecx
- movl %eax, 32(%esp)
- movl %ecx, 36(%esp)
- movl 40(%esi), %eax
- movl 44(%esi), %ecx
- movl %eax, 40(%esp)
- movl %ecx, 44(%esp)
- movl 48(%esi), %eax
- movl 52(%esi), %ecx
- movl %eax, 48(%esp)
- movl %ecx, 52(%esp)
- movl 56(%esi), %eax
- movl 60(%esi), %ecx
- movl %eax, 56(%esp)
- movl %ecx, 60(%esp)
- jmp .L001shortcut
-.L_sha1_block_asm_host_order_end:
- .size sha1_block_asm_host_order,.L_sha1_block_asm_host_order_end-sha1_block_asm_host_order
-.ident "desasm.pl"