summaryrefslogtreecommitdiffstats
path: root/secure/lib/libcrypto/i386/sha1-586.s
diff options
context:
space:
mode:
Diffstat (limited to 'secure/lib/libcrypto/i386/sha1-586.s')
-rw-r--r--secure/lib/libcrypto/i386/sha1-586.s1785
1 files changed, 681 insertions, 1104 deletions
diff --git a/secure/lib/libcrypto/i386/sha1-586.s b/secure/lib/libcrypto/i386/sha1-586.s
index b13c665..4a35f9d 100644
--- a/secure/lib/libcrypto/i386/sha1-586.s
+++ b/secure/lib/libcrypto/i386/sha1-586.s
@@ -1,9 +1,9 @@
# $FreeBSD$
- # Dont even think of reading this code
- # It was automatically generated by sha1-586.pl
- # Which is a perl program used to generate the x86 assember for
- # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris
- # eric <eay@cryptsoft.com>
+
+
+
+
+
.file "sha1-586.s"
.version "01.01"
@@ -27,1851 +27,1428 @@ sha1_block_asm_data_order:
movl 16(%ebp), %edi
movl 8(%ebp), %ebx
movl %ecx, 68(%esp)
- # First we need to setup the X array
+
.L000start:
- # First, load the words onto the stack in network byte order
+
movl (%esi), %eax
movl 4(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, (%esp)
movl %ecx, 4(%esp)
movl 8(%esi), %eax
movl 12(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, 8(%esp)
movl %ecx, 12(%esp)
movl 16(%esi), %eax
movl 20(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, 16(%esp)
movl %ecx, 20(%esp)
movl 24(%esi), %eax
movl 28(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, 24(%esp)
movl %ecx, 28(%esp)
movl 32(%esi), %eax
movl 36(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, 32(%esp)
movl %ecx, 36(%esp)
movl 40(%esi), %eax
movl 44(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, 40(%esp)
movl %ecx, 44(%esp)
movl 48(%esi), %eax
movl 52(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, 48(%esp)
movl %ecx, 52(%esp)
movl 56(%esi), %eax
movl 60(%esi), %ecx
.byte 15
-.byte 200 # bswapl %eax
+.byte 200
.byte 15
-.byte 201 # bswapl %ecx
+.byte 201
movl %eax, 56(%esp)
movl %ecx, 60(%esp)
- # We now have the X array on the stack
- # starting at sp-4
+
+
movl %esi, 132(%esp)
.L001shortcut:
- # Start processing
+
movl (%ebp), %eax
movl 4(%ebp), %ecx
- # 00_15 0
- movl %ebx, %esi
+
movl %eax, %ebp
- xorl %edx, %esi
+ movl %ebx, %esi
roll $5, %ebp
+ xorl %edx, %esi
andl %ecx, %esi
+ rorl $2, %ecx
addl %edi, %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
movl (%esp), %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
xorl %edx, %esi
leal 1518500249(%ebp,%edi,1),%ebp
- movl %ecx, %edi
addl %ebp, %esi
- xorl %ebx, %edi
+
movl %esi, %ebp
- andl %eax, %edi
+ movl %ecx, %edi
roll $5, %ebp
+ xorl %ebx, %edi
+ andl %eax, %edi
+ rorl $2, %eax
addl %edx, %ebp
movl 4(%esp), %edx
-.byte 209
-.byte 200 # rorl $1 %eax
xorl %ebx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
leal 1518500249(%ebp,%edx,1),%ebp
addl %ebp, %edi
- # 00_15 2
- movl %eax, %edx
+
movl %edi, %ebp
- xorl %ecx, %edx
+ movl %eax, %edx
roll $5, %ebp
+ xorl %ecx, %edx
andl %esi, %edx
+ rorl $2, %esi
addl %ebx, %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
movl 8(%esp), %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
xorl %ecx, %edx
leal 1518500249(%ebp,%ebx,1),%ebp
- movl %esi, %ebx
addl %ebp, %edx
- xorl %eax, %ebx
+
movl %edx, %ebp
- andl %edi, %ebx
+ movl %esi, %ebx
roll $5, %ebp
+ xorl %eax, %ebx
+ andl %edi, %ebx
+ rorl $2, %edi
addl %ecx, %ebp
movl 12(%esp), %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
xorl %eax, %ebx
-.byte 209
-.byte 207 # rorl $1 %edi
leal 1518500249(%ebp,%ecx,1),%ebp
addl %ebp, %ebx
- # 00_15 4
- movl %edi, %ecx
+
movl %ebx, %ebp
- xorl %esi, %ecx
+ movl %edi, %ecx
roll $5, %ebp
+ xorl %esi, %ecx
andl %edx, %ecx
+ rorl $2, %edx
addl %eax, %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
movl 16(%esp), %eax
-.byte 209
-.byte 202 # rorl $1 %edx
xorl %esi, %ecx
leal 1518500249(%ebp,%eax,1),%ebp
- movl %edx, %eax
addl %ebp, %ecx
- xorl %edi, %eax
+
movl %ecx, %ebp
- andl %ebx, %eax
+ movl %edx, %eax
roll $5, %ebp
+ xorl %edi, %eax
+ andl %ebx, %eax
+ rorl $2, %ebx
addl %esi, %ebp
movl 20(%esp), %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
xorl %edi, %eax
-.byte 209
-.byte 203 # rorl $1 %ebx
leal 1518500249(%ebp,%esi,1),%ebp
addl %ebp, %eax
- # 00_15 6
- movl %ebx, %esi
+
movl %eax, %ebp
- xorl %edx, %esi
+ movl %ebx, %esi
roll $5, %ebp
+ xorl %edx, %esi
andl %ecx, %esi
+ rorl $2, %ecx
addl %edi, %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
movl 24(%esp), %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
xorl %edx, %esi
leal 1518500249(%ebp,%edi,1),%ebp
- movl %ecx, %edi
addl %ebp, %esi
- xorl %ebx, %edi
+
movl %esi, %ebp
- andl %eax, %edi
+ movl %ecx, %edi
roll $5, %ebp
+ xorl %ebx, %edi
+ andl %eax, %edi
+ rorl $2, %eax
addl %edx, %ebp
movl 28(%esp), %edx
-.byte 209
-.byte 200 # rorl $1 %eax
xorl %ebx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
leal 1518500249(%ebp,%edx,1),%ebp
addl %ebp, %edi
- # 00_15 8
- movl %eax, %edx
+
movl %edi, %ebp
- xorl %ecx, %edx
+ movl %eax, %edx
roll $5, %ebp
+ xorl %ecx, %edx
andl %esi, %edx
+ rorl $2, %esi
addl %ebx, %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
movl 32(%esp), %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
xorl %ecx, %edx
leal 1518500249(%ebp,%ebx,1),%ebp
- movl %esi, %ebx
addl %ebp, %edx
- xorl %eax, %ebx
+
movl %edx, %ebp
- andl %edi, %ebx
+ movl %esi, %ebx
roll $5, %ebp
+ xorl %eax, %ebx
+ andl %edi, %ebx
+ rorl $2, %edi
addl %ecx, %ebp
movl 36(%esp), %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
xorl %eax, %ebx
-.byte 209
-.byte 207 # rorl $1 %edi
leal 1518500249(%ebp,%ecx,1),%ebp
addl %ebp, %ebx
- # 00_15 10
- movl %edi, %ecx
+
movl %ebx, %ebp
- xorl %esi, %ecx
+ movl %edi, %ecx
roll $5, %ebp
+ xorl %esi, %ecx
andl %edx, %ecx
+ rorl $2, %edx
addl %eax, %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
movl 40(%esp), %eax
-.byte 209
-.byte 202 # rorl $1 %edx
xorl %esi, %ecx
leal 1518500249(%ebp,%eax,1),%ebp
- movl %edx, %eax
addl %ebp, %ecx
- xorl %edi, %eax
+
movl %ecx, %ebp
- andl %ebx, %eax
+ movl %edx, %eax
roll $5, %ebp
+ xorl %edi, %eax
+ andl %ebx, %eax
+ rorl $2, %ebx
addl %esi, %ebp
movl 44(%esp), %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
xorl %edi, %eax
-.byte 209
-.byte 203 # rorl $1 %ebx
leal 1518500249(%ebp,%esi,1),%ebp
addl %ebp, %eax
- # 00_15 12
- movl %ebx, %esi
+
movl %eax, %ebp
- xorl %edx, %esi
+ movl %ebx, %esi
roll $5, %ebp
+ xorl %edx, %esi
andl %ecx, %esi
+ rorl $2, %ecx
addl %edi, %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
movl 48(%esp), %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
xorl %edx, %esi
leal 1518500249(%ebp,%edi,1),%ebp
- movl %ecx, %edi
addl %ebp, %esi
- xorl %ebx, %edi
+
movl %esi, %ebp
- andl %eax, %edi
+ movl %ecx, %edi
roll $5, %ebp
+ xorl %ebx, %edi
+ andl %eax, %edi
+ rorl $2, %eax
addl %edx, %ebp
movl 52(%esp), %edx
-.byte 209
-.byte 200 # rorl $1 %eax
xorl %ebx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
leal 1518500249(%ebp,%edx,1),%ebp
addl %ebp, %edi
- # 00_15 14
- movl %eax, %edx
+
movl %edi, %ebp
- xorl %ecx, %edx
+ movl %eax, %edx
roll $5, %ebp
+ xorl %ecx, %edx
andl %esi, %edx
+ rorl $2, %esi
addl %ebx, %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
movl 56(%esp), %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
xorl %ecx, %edx
leal 1518500249(%ebp,%ebx,1),%ebp
- movl %esi, %ebx
addl %ebp, %edx
- xorl %eax, %ebx
+
movl %edx, %ebp
- andl %edi, %ebx
+ movl %esi, %ebx
roll $5, %ebp
+ xorl %eax, %ebx
+ andl %edi, %ebx
+ rorl $2, %edi
addl %ecx, %ebp
movl 60(%esp), %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
xorl %eax, %ebx
-.byte 209
-.byte 207 # rorl $1 %edi
leal 1518500249(%ebp,%ecx,1),%ebp
addl %ebp, %ebx
- # 16_19 16
- nop
- movl (%esp), %ebp
+
movl 8(%esp), %ecx
- xorl %ebp, %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 52(%esp), %ebp
- xorl %ebp, %ecx
movl %edi, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl (%esp), %ecx
xorl %esi, %ebp
- movl %ecx, (%esp)
+ xorl 32(%esp), %ecx
andl %edx, %ebp
- leal 1518500249(%ecx,%eax,1),%ecx
+ xorl 52(%esp), %ecx
+ rorl $2, %edx
xorl %esi, %ebp
+.byte 209
+.byte 193
+ movl %ecx, (%esp)
+ leal 1518500249(%ecx,%eax,1),%ecx
movl %ebx, %eax
addl %ebp, %ecx
roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
addl %eax, %ecx
- movl 4(%esp), %eax
- movl 12(%esp), %ebp
- xorl %ebp, %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 56(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
+
+ movl 12(%esp), %eax
movl %edx, %ebp
+ xorl 4(%esp), %eax
xorl %edi, %ebp
- movl %eax, 4(%esp)
+ xorl 36(%esp), %eax
andl %ebx, %ebp
- leal 1518500249(%eax,%esi,1),%eax
+ xorl 56(%esp), %eax
+ rorl $2, %ebx
xorl %edi, %ebp
+.byte 209
+.byte 192
+ movl %eax, 4(%esp)
+ leal 1518500249(%eax,%esi,1),%eax
movl %ecx, %esi
+ addl %ebp, %eax
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %esi, %eax
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %eax
- # 16_19 18
- movl 8(%esp), %ebp
+
movl 16(%esp), %esi
- xorl %ebp, %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl 60(%esp), %ebp
- xorl %ebp, %esi
movl %ebx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 8(%esp), %esi
xorl %edx, %ebp
- movl %esi, 8(%esp)
+ xorl 40(%esp), %esi
andl %ecx, %ebp
- leal 1518500249(%esi,%edi,1),%esi
+ xorl 60(%esp), %esi
+ rorl $2, %ecx
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 8(%esp)
+ leal 1518500249(%esi,%edi,1),%esi
movl %eax, %edi
addl %ebp, %esi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
addl %edi, %esi
- movl 12(%esp), %edi
- movl 20(%esp), %ebp
- xorl %ebp, %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl (%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
+
+ movl 20(%esp), %edi
movl %ecx, %ebp
+ xorl 12(%esp), %edi
xorl %ebx, %ebp
- movl %edi, 12(%esp)
+ xorl 44(%esp), %edi
andl %eax, %ebp
- leal 1518500249(%edi,%edx,1),%edi
+ xorl (%esp), %edi
+ rorl $2, %eax
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 12(%esp)
+ leal 1518500249(%edi,%edx,1),%edi
movl %esi, %edx
+ addl %ebp, %edi
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
addl %edx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edi
- # 20_39 20
+
movl 16(%esp), %edx
- movl 24(%esp), %ebp
- xorl %ebp, %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 4(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 24(%esp), %edx
+ rorl $2, %esi
+ xorl 48(%esp), %edx
xorl %eax, %ebp
- movl %edx, 16(%esp)
+ xorl 4(%esp), %edx
xorl %ecx, %ebp
+.byte 209
+.byte 194
+ movl %edx, 16(%esp)
leal 1859775393(%edx,%ebx,1),%edx
movl %edi, %ebx
roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ addl %ebp, %edx
addl %ebx, %edx
- # 20_39 21
+
movl 20(%esp), %ebx
- movl 28(%esp), %ebp
- xorl %ebp, %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 8(%esp), %ebp
- xorl %ebp, %ebx
movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
+ xorl 28(%esp), %ebx
+ rorl $2, %edi
+ xorl 52(%esp), %ebx
xorl %esi, %ebp
- movl %ebx, 20(%esp)
+ xorl 8(%esp), %ebx
xorl %eax, %ebp
+.byte 209
+.byte 195
+ movl %ebx, 20(%esp)
leal 1859775393(%ebx,%ecx,1),%ebx
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
+ addl %ebp, %ebx
addl %ecx, %ebx
- # 20_39 22
+
movl 24(%esp), %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 12(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 32(%esp), %ecx
+ rorl $2, %edx
+ xorl 56(%esp), %ecx
xorl %edi, %ebp
- movl %ecx, 24(%esp)
+ xorl 12(%esp), %ecx
xorl %esi, %ebp
+.byte 209
+.byte 193
+ movl %ecx, 24(%esp)
leal 1859775393(%ecx,%eax,1),%ecx
movl %ebx, %eax
roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ addl %ebp, %ecx
addl %eax, %ecx
- # 20_39 23
+
movl 28(%esp), %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 16(%esp), %ebp
- xorl %ebp, %eax
movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
+ xorl 36(%esp), %eax
+ rorl $2, %ebx
+ xorl 60(%esp), %eax
xorl %edx, %ebp
- movl %eax, 28(%esp)
+ xorl 16(%esp), %eax
xorl %edi, %ebp
+.byte 209
+.byte 192
+ movl %eax, 28(%esp)
leal 1859775393(%eax,%esi,1),%eax
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
+ addl %ebp, %eax
addl %esi, %eax
- # 20_39 24
+
movl 32(%esp), %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 20(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 40(%esp), %esi
+ rorl $2, %ecx
+ xorl (%esp), %esi
xorl %ebx, %ebp
- movl %esi, 32(%esp)
+ xorl 20(%esp), %esi
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 32(%esp)
leal 1859775393(%esi,%edi,1),%esi
movl %eax, %edi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ addl %ebp, %esi
addl %edi, %esi
- # 20_39 25
+
movl 36(%esp), %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- xorl %ebp, %edi
movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
+ xorl 44(%esp), %edi
+ rorl $2, %eax
+ xorl 4(%esp), %edi
xorl %ecx, %ebp
- movl %edi, 36(%esp)
+ xorl 24(%esp), %edi
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 36(%esp)
leal 1859775393(%edi,%edx,1),%edi
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
+ addl %ebp, %edi
addl %edx, %edi
- # 20_39 26
+
movl 40(%esp), %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 8(%esp), %ebp
- xorl %ebp, %edx
- movl 28(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 48(%esp), %edx
+ rorl $2, %esi
+ xorl 8(%esp), %edx
xorl %eax, %ebp
- movl %edx, 40(%esp)
+ xorl 28(%esp), %edx
xorl %ecx, %ebp
+.byte 209
+.byte 194
+ movl %edx, 40(%esp)
leal 1859775393(%edx,%ebx,1),%edx
movl %edi, %ebx
roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ addl %ebp, %edx
addl %ebx, %edx
- # 20_39 27
+
movl 44(%esp), %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 12(%esp), %ebp
- xorl %ebp, %ebx
- movl 32(%esp), %ebp
- xorl %ebp, %ebx
movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
+ xorl 52(%esp), %ebx
+ rorl $2, %edi
+ xorl 12(%esp), %ebx
xorl %esi, %ebp
- movl %ebx, 44(%esp)
+ xorl 32(%esp), %ebx
xorl %eax, %ebp
+.byte 209
+.byte 195
+ movl %ebx, 44(%esp)
leal 1859775393(%ebx,%ecx,1),%ebx
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
+ addl %ebp, %ebx
addl %ecx, %ebx
- # 20_39 28
+
movl 48(%esp), %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 16(%esp), %ebp
- xorl %ebp, %ecx
- movl 36(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 56(%esp), %ecx
+ rorl $2, %edx
+ xorl 16(%esp), %ecx
xorl %edi, %ebp
- movl %ecx, 48(%esp)
+ xorl 36(%esp), %ecx
xorl %esi, %ebp
+.byte 209
+.byte 193
+ movl %ecx, 48(%esp)
leal 1859775393(%ecx,%eax,1),%ecx
movl %ebx, %eax
roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ addl %ebp, %ecx
addl %eax, %ecx
- # 20_39 29
+
movl 52(%esp), %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 20(%esp), %ebp
- xorl %ebp, %eax
- movl 40(%esp), %ebp
- xorl %ebp, %eax
movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
+ xorl 60(%esp), %eax
+ rorl $2, %ebx
+ xorl 20(%esp), %eax
xorl %edx, %ebp
- movl %eax, 52(%esp)
+ xorl 40(%esp), %eax
xorl %edi, %ebp
+.byte 209
+.byte 192
+ movl %eax, 52(%esp)
leal 1859775393(%eax,%esi,1),%eax
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
+ addl %ebp, %eax
addl %esi, %eax
- # 20_39 30
+
movl 56(%esp), %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- xorl %ebp, %esi
- movl 44(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl (%esp), %esi
+ rorl $2, %ecx
+ xorl 24(%esp), %esi
xorl %ebx, %ebp
- movl %esi, 56(%esp)
+ xorl 44(%esp), %esi
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 56(%esp)
leal 1859775393(%esi,%edi,1),%esi
movl %eax, %edi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ addl %ebp, %esi
addl %edi, %esi
- # 20_39 31
+
movl 60(%esp), %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- xorl %ebp, %edi
- movl 48(%esp), %ebp
- xorl %ebp, %edi
movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
+ xorl 4(%esp), %edi
+ rorl $2, %eax
+ xorl 28(%esp), %edi
xorl %ecx, %ebp
- movl %edi, 60(%esp)
+ xorl 48(%esp), %edi
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 60(%esp)
leal 1859775393(%edi,%edx,1),%edi
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
+ addl %ebp, %edi
addl %edx, %edi
- # 20_39 32
+
movl (%esp), %edx
- movl 8(%esp), %ebp
- xorl %ebp, %edx
- movl 32(%esp), %ebp
- xorl %ebp, %edx
- movl 52(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 8(%esp), %edx
+ rorl $2, %esi
+ xorl 32(%esp), %edx
xorl %eax, %ebp
- movl %edx, (%esp)
+ xorl 52(%esp), %edx
xorl %ecx, %ebp
+.byte 209
+.byte 194
+ movl %edx, (%esp)
leal 1859775393(%edx,%ebx,1),%edx
movl %edi, %ebx
roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ addl %ebp, %edx
addl %ebx, %edx
- # 20_39 33
+
movl 4(%esp), %ebx
- movl 12(%esp), %ebp
- xorl %ebp, %ebx
- movl 36(%esp), %ebp
- xorl %ebp, %ebx
- movl 56(%esp), %ebp
- xorl %ebp, %ebx
movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
+ xorl 12(%esp), %ebx
+ rorl $2, %edi
+ xorl 36(%esp), %ebx
xorl %esi, %ebp
- movl %ebx, 4(%esp)
+ xorl 56(%esp), %ebx
xorl %eax, %ebp
+.byte 209
+.byte 195
+ movl %ebx, 4(%esp)
leal 1859775393(%ebx,%ecx,1),%ebx
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
+ addl %ebp, %ebx
addl %ecx, %ebx
- # 20_39 34
+
movl 8(%esp), %ecx
- movl 16(%esp), %ebp
- xorl %ebp, %ecx
- movl 40(%esp), %ebp
- xorl %ebp, %ecx
- movl 60(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 16(%esp), %ecx
+ rorl $2, %edx
+ xorl 40(%esp), %ecx
xorl %edi, %ebp
- movl %ecx, 8(%esp)
+ xorl 60(%esp), %ecx
xorl %esi, %ebp
+.byte 209
+.byte 193
+ movl %ecx, 8(%esp)
leal 1859775393(%ecx,%eax,1),%ecx
movl %ebx, %eax
roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ addl %ebp, %ecx
addl %eax, %ecx
- # 20_39 35
+
movl 12(%esp), %eax
- movl 20(%esp), %ebp
- xorl %ebp, %eax
- movl 44(%esp), %ebp
- xorl %ebp, %eax
- movl (%esp), %ebp
- xorl %ebp, %eax
movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
+ xorl 20(%esp), %eax
+ rorl $2, %ebx
+ xorl 44(%esp), %eax
xorl %edx, %ebp
- movl %eax, 12(%esp)
+ xorl (%esp), %eax
xorl %edi, %ebp
+.byte 209
+.byte 192
+ movl %eax, 12(%esp)
leal 1859775393(%eax,%esi,1),%eax
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
+ addl %ebp, %eax
addl %esi, %eax
- # 20_39 36
+
movl 16(%esp), %esi
- movl 24(%esp), %ebp
- xorl %ebp, %esi
- movl 48(%esp), %ebp
- xorl %ebp, %esi
- movl 4(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 24(%esp), %esi
+ rorl $2, %ecx
+ xorl 48(%esp), %esi
xorl %ebx, %ebp
- movl %esi, 16(%esp)
+ xorl 4(%esp), %esi
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 16(%esp)
leal 1859775393(%esi,%edi,1),%esi
movl %eax, %edi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ addl %ebp, %esi
addl %edi, %esi
- # 20_39 37
+
movl 20(%esp), %edi
- movl 28(%esp), %ebp
- xorl %ebp, %edi
- movl 52(%esp), %ebp
- xorl %ebp, %edi
- movl 8(%esp), %ebp
- xorl %ebp, %edi
movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
+ xorl 28(%esp), %edi
+ rorl $2, %eax
+ xorl 52(%esp), %edi
xorl %ecx, %ebp
- movl %edi, 20(%esp)
+ xorl 8(%esp), %edi
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 20(%esp)
leal 1859775393(%edi,%edx,1),%edi
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
+ addl %ebp, %edi
addl %edx, %edi
- # 20_39 38
+
movl 24(%esp), %edx
- movl 32(%esp), %ebp
- xorl %ebp, %edx
- movl 56(%esp), %ebp
- xorl %ebp, %edx
- movl 12(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 32(%esp), %edx
+ rorl $2, %esi
+ xorl 56(%esp), %edx
xorl %eax, %ebp
- movl %edx, 24(%esp)
+ xorl 12(%esp), %edx
xorl %ecx, %ebp
+.byte 209
+.byte 194
+ movl %edx, 24(%esp)
leal 1859775393(%edx,%ebx,1),%edx
movl %edi, %ebx
roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ addl %ebp, %edx
addl %ebx, %edx
- # 20_39 39
+
movl 28(%esp), %ebx
- movl 36(%esp), %ebp
- xorl %ebp, %ebx
- movl 60(%esp), %ebp
- xorl %ebp, %ebx
- movl 16(%esp), %ebp
- xorl %ebp, %ebx
movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
+ xorl 36(%esp), %ebx
+ rorl $2, %edi
+ xorl 60(%esp), %ebx
xorl %esi, %ebp
- movl %ebx, 28(%esp)
+ xorl 16(%esp), %ebx
xorl %eax, %ebp
+.byte 209
+.byte 195
+ movl %ebx, 28(%esp)
leal 1859775393(%ebx,%ecx,1),%ebx
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
+ addl %ebp, %ebx
addl %ecx, %ebx
- # 40_59 40
+
movl 32(%esp), %ecx
- movl 40(%esp), %ebp
- xorl %ebp, %ecx
- movl (%esp), %ebp
- xorl %ebp, %ecx
- movl 20(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 40(%esp), %ecx
orl %edi, %ebp
- movl %ecx, 32(%esp)
+ xorl (%esp), %ecx
andl %esi, %ebp
+ xorl 20(%esp), %ecx
+.byte 209
+.byte 193
+ movl %ecx, 32(%esp)
leal 2400959708(%ecx,%eax,1),%ecx
movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ rorl $2, %edx
andl %edi, %eax
orl %eax, %ebp
movl %ebx, %eax
roll $5, %eax
addl %eax, %ebp
- movl 36(%esp), %eax
addl %ebp, %ecx
- movl 44(%esp), %ebp
- xorl %ebp, %eax
- movl 4(%esp), %ebp
- xorl %ebp, %eax
- movl 24(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
+
+ movl 36(%esp), %eax
movl %ebx, %ebp
- movl %eax, 36(%esp)
+ xorl 44(%esp), %eax
orl %edx, %ebp
+ xorl 4(%esp), %eax
+ andl %edi, %ebp
+ xorl 24(%esp), %eax
+.byte 209
+.byte 192
+ movl %eax, 36(%esp)
leal 2400959708(%eax,%esi,1),%eax
movl %ebx, %esi
- andl %edi, %ebp
+ rorl $2, %ebx
andl %edx, %esi
orl %esi, %ebp
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %ebp, %eax
- # 40_59 41
- # 40_59 42
+
movl 40(%esp), %esi
- movl 48(%esp), %ebp
- xorl %ebp, %esi
- movl 8(%esp), %ebp
- xorl %ebp, %esi
- movl 28(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 48(%esp), %esi
orl %ebx, %ebp
- movl %esi, 40(%esp)
+ xorl 8(%esp), %esi
andl %edx, %ebp
+ xorl 28(%esp), %esi
+.byte 209
+.byte 198
+ movl %esi, 40(%esp)
leal 2400959708(%esi,%edi,1),%esi
movl %ecx, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ rorl $2, %ecx
andl %ebx, %edi
orl %edi, %ebp
movl %eax, %edi
roll $5, %edi
addl %edi, %ebp
- movl 44(%esp), %edi
addl %ebp, %esi
- movl 52(%esp), %ebp
- xorl %ebp, %edi
- movl 12(%esp), %ebp
- xorl %ebp, %edi
- movl 32(%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
+
+ movl 44(%esp), %edi
movl %eax, %ebp
- movl %edi, 44(%esp)
+ xorl 52(%esp), %edi
orl %ecx, %ebp
+ xorl 12(%esp), %edi
+ andl %ebx, %ebp
+ xorl 32(%esp), %edi
+.byte 209
+.byte 199
+ movl %edi, 44(%esp)
leal 2400959708(%edi,%edx,1),%edi
movl %eax, %edx
- andl %ebx, %ebp
+ rorl $2, %eax
andl %ecx, %edx
orl %edx, %ebp
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
addl %edx, %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
addl %ebp, %edi
- # 40_59 43
- # 40_59 44
+
movl 48(%esp), %edx
- movl 56(%esp), %ebp
- xorl %ebp, %edx
- movl 16(%esp), %ebp
- xorl %ebp, %edx
- movl 36(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 56(%esp), %edx
orl %eax, %ebp
- movl %edx, 48(%esp)
+ xorl 16(%esp), %edx
andl %ecx, %ebp
+ xorl 36(%esp), %edx
+.byte 209
+.byte 194
+ movl %edx, 48(%esp)
leal 2400959708(%edx,%ebx,1),%edx
movl %esi, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ rorl $2, %esi
andl %eax, %ebx
orl %ebx, %ebp
movl %edi, %ebx
roll $5, %ebx
addl %ebx, %ebp
- movl 52(%esp), %ebx
addl %ebp, %edx
- movl 60(%esp), %ebp
- xorl %ebp, %ebx
- movl 20(%esp), %ebp
- xorl %ebp, %ebx
- movl 40(%esp), %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ebp, %ebx
-.byte 209
-.byte 195 # roll $1 %ebx
+
+ movl 52(%esp), %ebx
movl %edi, %ebp
- movl %ebx, 52(%esp)
+ xorl 60(%esp), %ebx
orl %esi, %ebp
+ xorl 20(%esp), %ebx
+ andl %eax, %ebp
+ xorl 40(%esp), %ebx
+.byte 209
+.byte 195
+ movl %ebx, 52(%esp)
leal 2400959708(%ebx,%ecx,1),%ebx
movl %edi, %ecx
- andl %eax, %ebp
+ rorl $2, %edi
andl %esi, %ecx
orl %ecx, %ebp
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
addl %ecx, %ebp
-.byte 209
-.byte 207 # rorl $1 %edi
addl %ebp, %ebx
- # 40_59 45
- # 40_59 46
+
movl 56(%esp), %ecx
- movl (%esp), %ebp
- xorl %ebp, %ecx
- movl 24(%esp), %ebp
- xorl %ebp, %ecx
- movl 44(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl (%esp), %ecx
orl %edi, %ebp
- movl %ecx, 56(%esp)
+ xorl 24(%esp), %ecx
andl %esi, %ebp
+ xorl 44(%esp), %ecx
+.byte 209
+.byte 193
+ movl %ecx, 56(%esp)
leal 2400959708(%ecx,%eax,1),%ecx
movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ rorl $2, %edx
andl %edi, %eax
orl %eax, %ebp
movl %ebx, %eax
roll $5, %eax
addl %eax, %ebp
- movl 60(%esp), %eax
addl %ebp, %ecx
- movl 4(%esp), %ebp
- xorl %ebp, %eax
- movl 28(%esp), %ebp
- xorl %ebp, %eax
- movl 48(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
+
+ movl 60(%esp), %eax
movl %ebx, %ebp
- movl %eax, 60(%esp)
+ xorl 4(%esp), %eax
orl %edx, %ebp
+ xorl 28(%esp), %eax
+ andl %edi, %ebp
+ xorl 48(%esp), %eax
+.byte 209
+.byte 192
+ movl %eax, 60(%esp)
leal 2400959708(%eax,%esi,1),%eax
movl %ebx, %esi
- andl %edi, %ebp
+ rorl $2, %ebx
andl %edx, %esi
orl %esi, %ebp
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %ebp, %eax
- # 40_59 47
- # 40_59 48
+
movl (%esp), %esi
- movl 8(%esp), %ebp
- xorl %ebp, %esi
- movl 32(%esp), %ebp
- xorl %ebp, %esi
- movl 52(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 8(%esp), %esi
orl %ebx, %ebp
- movl %esi, (%esp)
+ xorl 32(%esp), %esi
andl %edx, %ebp
+ xorl 52(%esp), %esi
+.byte 209
+.byte 198
+ movl %esi, (%esp)
leal 2400959708(%esi,%edi,1),%esi
movl %ecx, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ rorl $2, %ecx
andl %ebx, %edi
orl %edi, %ebp
movl %eax, %edi
roll $5, %edi
addl %edi, %ebp
- movl 4(%esp), %edi
addl %ebp, %esi
- movl 12(%esp), %ebp
- xorl %ebp, %edi
- movl 36(%esp), %ebp
- xorl %ebp, %edi
- movl 56(%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
+
+ movl 4(%esp), %edi
movl %eax, %ebp
- movl %edi, 4(%esp)
+ xorl 12(%esp), %edi
orl %ecx, %ebp
+ xorl 36(%esp), %edi
+ andl %ebx, %ebp
+ xorl 56(%esp), %edi
+.byte 209
+.byte 199
+ movl %edi, 4(%esp)
leal 2400959708(%edi,%edx,1),%edi
movl %eax, %edx
- andl %ebx, %ebp
+ rorl $2, %eax
andl %ecx, %edx
orl %edx, %ebp
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
addl %edx, %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
addl %ebp, %edi
- # 40_59 49
- # 40_59 50
+
movl 8(%esp), %edx
- movl 16(%esp), %ebp
- xorl %ebp, %edx
- movl 40(%esp), %ebp
- xorl %ebp, %edx
- movl 60(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 16(%esp), %edx
orl %eax, %ebp
- movl %edx, 8(%esp)
+ xorl 40(%esp), %edx
andl %ecx, %ebp
+ xorl 60(%esp), %edx
+.byte 209
+.byte 194
+ movl %edx, 8(%esp)
leal 2400959708(%edx,%ebx,1),%edx
movl %esi, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ rorl $2, %esi
andl %eax, %ebx
orl %ebx, %ebp
movl %edi, %ebx
roll $5, %ebx
addl %ebx, %ebp
- movl 12(%esp), %ebx
addl %ebp, %edx
- movl 20(%esp), %ebp
- xorl %ebp, %ebx
- movl 44(%esp), %ebp
- xorl %ebp, %ebx
- movl (%esp), %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ebp, %ebx
-.byte 209
-.byte 195 # roll $1 %ebx
+
+ movl 12(%esp), %ebx
movl %edi, %ebp
- movl %ebx, 12(%esp)
+ xorl 20(%esp), %ebx
orl %esi, %ebp
+ xorl 44(%esp), %ebx
+ andl %eax, %ebp
+ xorl (%esp), %ebx
+.byte 209
+.byte 195
+ movl %ebx, 12(%esp)
leal 2400959708(%ebx,%ecx,1),%ebx
movl %edi, %ecx
- andl %eax, %ebp
+ rorl $2, %edi
andl %esi, %ecx
orl %ecx, %ebp
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
addl %ecx, %ebp
-.byte 209
-.byte 207 # rorl $1 %edi
addl %ebp, %ebx
- # 40_59 51
- # 40_59 52
+
movl 16(%esp), %ecx
- movl 24(%esp), %ebp
- xorl %ebp, %ecx
- movl 48(%esp), %ebp
- xorl %ebp, %ecx
- movl 4(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 24(%esp), %ecx
orl %edi, %ebp
- movl %ecx, 16(%esp)
+ xorl 48(%esp), %ecx
andl %esi, %ebp
+ xorl 4(%esp), %ecx
+.byte 209
+.byte 193
+ movl %ecx, 16(%esp)
leal 2400959708(%ecx,%eax,1),%ecx
movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ rorl $2, %edx
andl %edi, %eax
orl %eax, %ebp
movl %ebx, %eax
roll $5, %eax
addl %eax, %ebp
- movl 20(%esp), %eax
addl %ebp, %ecx
- movl 28(%esp), %ebp
- xorl %ebp, %eax
- movl 52(%esp), %ebp
- xorl %ebp, %eax
- movl 8(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
+
+ movl 20(%esp), %eax
movl %ebx, %ebp
- movl %eax, 20(%esp)
+ xorl 28(%esp), %eax
orl %edx, %ebp
+ xorl 52(%esp), %eax
+ andl %edi, %ebp
+ xorl 8(%esp), %eax
+.byte 209
+.byte 192
+ movl %eax, 20(%esp)
leal 2400959708(%eax,%esi,1),%eax
movl %ebx, %esi
- andl %edi, %ebp
+ rorl $2, %ebx
andl %edx, %esi
orl %esi, %ebp
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %ebp, %eax
- # 40_59 53
- # 40_59 54
+
movl 24(%esp), %esi
- movl 32(%esp), %ebp
- xorl %ebp, %esi
- movl 56(%esp), %ebp
- xorl %ebp, %esi
- movl 12(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 32(%esp), %esi
orl %ebx, %ebp
- movl %esi, 24(%esp)
+ xorl 56(%esp), %esi
andl %edx, %ebp
+ xorl 12(%esp), %esi
+.byte 209
+.byte 198
+ movl %esi, 24(%esp)
leal 2400959708(%esi,%edi,1),%esi
movl %ecx, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ rorl $2, %ecx
andl %ebx, %edi
orl %edi, %ebp
movl %eax, %edi
roll $5, %edi
addl %edi, %ebp
- movl 28(%esp), %edi
addl %ebp, %esi
- movl 36(%esp), %ebp
- xorl %ebp, %edi
- movl 60(%esp), %ebp
- xorl %ebp, %edi
- movl 16(%esp), %ebp
-.byte 209
-.byte 201 # rorl $1 %ecx
- xorl %ebp, %edi
-.byte 209
-.byte 199 # roll $1 %edi
+
+ movl 28(%esp), %edi
movl %eax, %ebp
- movl %edi, 28(%esp)
+ xorl 36(%esp), %edi
orl %ecx, %ebp
+ xorl 60(%esp), %edi
+ andl %ebx, %ebp
+ xorl 16(%esp), %edi
+.byte 209
+.byte 199
+ movl %edi, 28(%esp)
leal 2400959708(%edi,%edx,1),%edi
movl %eax, %edx
- andl %ebx, %ebp
+ rorl $2, %eax
andl %ecx, %edx
orl %edx, %ebp
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
addl %edx, %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
addl %ebp, %edi
- # 40_59 55
- # 40_59 56
+
movl 32(%esp), %edx
- movl 40(%esp), %ebp
- xorl %ebp, %edx
- movl (%esp), %ebp
- xorl %ebp, %edx
- movl 20(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 40(%esp), %edx
orl %eax, %ebp
- movl %edx, 32(%esp)
+ xorl (%esp), %edx
andl %ecx, %ebp
+ xorl 20(%esp), %edx
+.byte 209
+.byte 194
+ movl %edx, 32(%esp)
leal 2400959708(%edx,%ebx,1),%edx
movl %esi, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ rorl $2, %esi
andl %eax, %ebx
orl %ebx, %ebp
movl %edi, %ebx
roll $5, %ebx
addl %ebx, %ebp
- movl 36(%esp), %ebx
addl %ebp, %edx
- movl 44(%esp), %ebp
- xorl %ebp, %ebx
- movl 4(%esp), %ebp
- xorl %ebp, %ebx
- movl 24(%esp), %ebp
-.byte 209
-.byte 206 # rorl $1 %esi
- xorl %ebp, %ebx
-.byte 209
-.byte 195 # roll $1 %ebx
+
+ movl 36(%esp), %ebx
movl %edi, %ebp
- movl %ebx, 36(%esp)
+ xorl 44(%esp), %ebx
orl %esi, %ebp
+ xorl 4(%esp), %ebx
+ andl %eax, %ebp
+ xorl 24(%esp), %ebx
+.byte 209
+.byte 195
+ movl %ebx, 36(%esp)
leal 2400959708(%ebx,%ecx,1),%ebx
movl %edi, %ecx
- andl %eax, %ebp
+ rorl $2, %edi
andl %esi, %ecx
orl %ecx, %ebp
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
addl %ecx, %ebp
-.byte 209
-.byte 207 # rorl $1 %edi
addl %ebp, %ebx
- # 40_59 57
- # 40_59 58
+
movl 40(%esp), %ecx
- movl 48(%esp), %ebp
- xorl %ebp, %ecx
- movl 8(%esp), %ebp
- xorl %ebp, %ecx
- movl 28(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 48(%esp), %ecx
orl %edi, %ebp
- movl %ecx, 40(%esp)
+ xorl 8(%esp), %ecx
andl %esi, %ebp
+ xorl 28(%esp), %ecx
+.byte 209
+.byte 193
+ movl %ecx, 40(%esp)
leal 2400959708(%ecx,%eax,1),%ecx
movl %edx, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ rorl $2, %edx
andl %edi, %eax
orl %eax, %ebp
movl %ebx, %eax
roll $5, %eax
addl %eax, %ebp
- movl 44(%esp), %eax
addl %ebp, %ecx
- movl 52(%esp), %ebp
- xorl %ebp, %eax
- movl 12(%esp), %ebp
- xorl %ebp, %eax
- movl 32(%esp), %ebp
-.byte 209
-.byte 202 # rorl $1 %edx
- xorl %ebp, %eax
-.byte 209
-.byte 192 # roll $1 %eax
+
+ movl 44(%esp), %eax
movl %ebx, %ebp
- movl %eax, 44(%esp)
+ xorl 52(%esp), %eax
orl %edx, %ebp
+ xorl 12(%esp), %eax
+ andl %edi, %ebp
+ xorl 32(%esp), %eax
+.byte 209
+.byte 192
+ movl %eax, 44(%esp)
leal 2400959708(%eax,%esi,1),%eax
movl %ebx, %esi
- andl %edi, %ebp
+ rorl $2, %ebx
andl %edx, %esi
orl %esi, %ebp
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %esi, %ebp
-.byte 209
-.byte 203 # rorl $1 %ebx
addl %ebp, %eax
- # 40_59 59
- # 20_39 60
+
movl 48(%esp), %esi
- movl 56(%esp), %ebp
- xorl %ebp, %esi
- movl 16(%esp), %ebp
- xorl %ebp, %esi
- movl 36(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 56(%esp), %esi
+ rorl $2, %ecx
+ xorl 16(%esp), %esi
xorl %ebx, %ebp
- movl %esi, 48(%esp)
+ xorl 36(%esp), %esi
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 48(%esp)
leal 3395469782(%esi,%edi,1),%esi
movl %eax, %edi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ addl %ebp, %esi
addl %edi, %esi
- # 20_39 61
+
movl 52(%esp), %edi
- movl 60(%esp), %ebp
- xorl %ebp, %edi
- movl 20(%esp), %ebp
- xorl %ebp, %edi
- movl 40(%esp), %ebp
- xorl %ebp, %edi
movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
+ xorl 60(%esp), %edi
+ rorl $2, %eax
+ xorl 20(%esp), %edi
xorl %ecx, %ebp
- movl %edi, 52(%esp)
+ xorl 40(%esp), %edi
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 52(%esp)
leal 3395469782(%edi,%edx,1),%edi
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
+ addl %ebp, %edi
addl %edx, %edi
- # 20_39 62
+
movl 56(%esp), %edx
- movl (%esp), %ebp
- xorl %ebp, %edx
- movl 24(%esp), %ebp
- xorl %ebp, %edx
- movl 44(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl (%esp), %edx
+ rorl $2, %esi
+ xorl 24(%esp), %edx
xorl %eax, %ebp
- movl %edx, 56(%esp)
+ xorl 44(%esp), %edx
xorl %ecx, %ebp
+.byte 209
+.byte 194
+ movl %edx, 56(%esp)
leal 3395469782(%edx,%ebx,1),%edx
movl %edi, %ebx
roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ addl %ebp, %edx
addl %ebx, %edx
- # 20_39 63
+
movl 60(%esp), %ebx
- movl 4(%esp), %ebp
- xorl %ebp, %ebx
- movl 28(%esp), %ebp
- xorl %ebp, %ebx
- movl 48(%esp), %ebp
- xorl %ebp, %ebx
movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
+ xorl 4(%esp), %ebx
+ rorl $2, %edi
+ xorl 28(%esp), %ebx
xorl %esi, %ebp
- movl %ebx, 60(%esp)
+ xorl 48(%esp), %ebx
xorl %eax, %ebp
+.byte 209
+.byte 195
+ movl %ebx, 60(%esp)
leal 3395469782(%ebx,%ecx,1),%ebx
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
+ addl %ebp, %ebx
addl %ecx, %ebx
- # 20_39 64
+
movl (%esp), %ecx
- movl 8(%esp), %ebp
- xorl %ebp, %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 52(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 8(%esp), %ecx
+ rorl $2, %edx
+ xorl 32(%esp), %ecx
xorl %edi, %ebp
- movl %ecx, (%esp)
+ xorl 52(%esp), %ecx
xorl %esi, %ebp
+.byte 209
+.byte 193
+ movl %ecx, (%esp)
leal 3395469782(%ecx,%eax,1),%ecx
movl %ebx, %eax
roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ addl %ebp, %ecx
addl %eax, %ecx
- # 20_39 65
+
movl 4(%esp), %eax
- movl 12(%esp), %ebp
- xorl %ebp, %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 56(%esp), %ebp
- xorl %ebp, %eax
movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
+ xorl 12(%esp), %eax
+ rorl $2, %ebx
+ xorl 36(%esp), %eax
xorl %edx, %ebp
- movl %eax, 4(%esp)
+ xorl 56(%esp), %eax
xorl %edi, %ebp
+.byte 209
+.byte 192
+ movl %eax, 4(%esp)
leal 3395469782(%eax,%esi,1),%eax
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
+ addl %ebp, %eax
addl %esi, %eax
- # 20_39 66
+
movl 8(%esp), %esi
- movl 16(%esp), %ebp
- xorl %ebp, %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl 60(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 16(%esp), %esi
+ rorl $2, %ecx
+ xorl 40(%esp), %esi
xorl %ebx, %ebp
- movl %esi, 8(%esp)
+ xorl 60(%esp), %esi
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 8(%esp)
leal 3395469782(%esi,%edi,1),%esi
movl %eax, %edi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ addl %ebp, %esi
addl %edi, %esi
- # 20_39 67
+
movl 12(%esp), %edi
- movl 20(%esp), %ebp
- xorl %ebp, %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl (%esp), %ebp
- xorl %ebp, %edi
movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
+ xorl 20(%esp), %edi
+ rorl $2, %eax
+ xorl 44(%esp), %edi
xorl %ecx, %ebp
- movl %edi, 12(%esp)
+ xorl (%esp), %edi
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 12(%esp)
leal 3395469782(%edi,%edx,1),%edi
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
+ addl %ebp, %edi
addl %edx, %edi
- # 20_39 68
+
movl 16(%esp), %edx
- movl 24(%esp), %ebp
- xorl %ebp, %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 4(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 24(%esp), %edx
+ rorl $2, %esi
+ xorl 48(%esp), %edx
xorl %eax, %ebp
- movl %edx, 16(%esp)
+ xorl 4(%esp), %edx
xorl %ecx, %ebp
+.byte 209
+.byte 194
+ movl %edx, 16(%esp)
leal 3395469782(%edx,%ebx,1),%edx
movl %edi, %ebx
roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ addl %ebp, %edx
addl %ebx, %edx
- # 20_39 69
+
movl 20(%esp), %ebx
- movl 28(%esp), %ebp
- xorl %ebp, %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 8(%esp), %ebp
- xorl %ebp, %ebx
movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
+ xorl 28(%esp), %ebx
+ rorl $2, %edi
+ xorl 52(%esp), %ebx
xorl %esi, %ebp
- movl %ebx, 20(%esp)
+ xorl 8(%esp), %ebx
xorl %eax, %ebp
+.byte 209
+.byte 195
+ movl %ebx, 20(%esp)
leal 3395469782(%ebx,%ecx,1),%ebx
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
+ addl %ebp, %ebx
addl %ecx, %ebx
- # 20_39 70
+
movl 24(%esp), %ecx
- movl 32(%esp), %ebp
- xorl %ebp, %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 12(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 32(%esp), %ecx
+ rorl $2, %edx
+ xorl 56(%esp), %ecx
xorl %edi, %ebp
- movl %ecx, 24(%esp)
+ xorl 12(%esp), %ecx
xorl %esi, %ebp
+.byte 209
+.byte 193
+ movl %ecx, 24(%esp)
leal 3395469782(%ecx,%eax,1),%ecx
movl %ebx, %eax
roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ addl %ebp, %ecx
addl %eax, %ecx
- # 20_39 71
+
movl 28(%esp), %eax
- movl 36(%esp), %ebp
- xorl %ebp, %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 16(%esp), %ebp
- xorl %ebp, %eax
movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
+ xorl 36(%esp), %eax
+ rorl $2, %ebx
+ xorl 60(%esp), %eax
xorl %edx, %ebp
- movl %eax, 28(%esp)
+ xorl 16(%esp), %eax
xorl %edi, %ebp
+.byte 209
+.byte 192
+ movl %eax, 28(%esp)
leal 3395469782(%eax,%esi,1),%eax
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
+ addl %ebp, %eax
addl %esi, %eax
- # 20_39 72
+
movl 32(%esp), %esi
- movl 40(%esp), %ebp
- xorl %ebp, %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 20(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl 40(%esp), %esi
+ rorl $2, %ecx
+ xorl (%esp), %esi
xorl %ebx, %ebp
- movl %esi, 32(%esp)
+ xorl 20(%esp), %esi
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 32(%esp)
leal 3395469782(%esi,%edi,1),%esi
movl %eax, %edi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ addl %ebp, %esi
addl %edi, %esi
- # 20_39 73
+
movl 36(%esp), %edi
- movl 44(%esp), %ebp
- xorl %ebp, %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 24(%esp), %ebp
- xorl %ebp, %edi
movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
+ xorl 44(%esp), %edi
+ rorl $2, %eax
+ xorl 4(%esp), %edi
xorl %ecx, %ebp
- movl %edi, 36(%esp)
+ xorl 24(%esp), %edi
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 36(%esp)
leal 3395469782(%edi,%edx,1),%edi
movl %esi, %edx
roll $5, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
- addl %ebp, %edx
-.byte 209
-.byte 200 # rorl $1 %eax
+ addl %ebp, %edi
addl %edx, %edi
- # 20_39 74
+
movl 40(%esp), %edx
- movl 48(%esp), %ebp
- xorl %ebp, %edx
- movl 8(%esp), %ebp
- xorl %ebp, %edx
- movl 28(%esp), %ebp
- xorl %ebp, %edx
movl %esi, %ebp
-.byte 209
-.byte 194 # roll $1 %edx
+ xorl 48(%esp), %edx
+ rorl $2, %esi
+ xorl 8(%esp), %edx
xorl %eax, %ebp
- movl %edx, 40(%esp)
+ xorl 28(%esp), %edx
xorl %ecx, %ebp
+.byte 209
+.byte 194
+ movl %edx, 40(%esp)
leal 3395469782(%edx,%ebx,1),%edx
movl %edi, %ebx
roll $5, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
- addl %ebp, %ebx
-.byte 209
-.byte 206 # rorl $1 %esi
+ addl %ebp, %edx
addl %ebx, %edx
- # 20_39 75
+
movl 44(%esp), %ebx
- movl 52(%esp), %ebp
- xorl %ebp, %ebx
- movl 12(%esp), %ebp
- xorl %ebp, %ebx
- movl 32(%esp), %ebp
- xorl %ebp, %ebx
movl %edi, %ebp
-.byte 209
-.byte 195 # roll $1 %ebx
+ xorl 52(%esp), %ebx
+ rorl $2, %edi
+ xorl 12(%esp), %ebx
xorl %esi, %ebp
- movl %ebx, 44(%esp)
+ xorl 32(%esp), %ebx
xorl %eax, %ebp
+.byte 209
+.byte 195
+ movl %ebx, 44(%esp)
leal 3395469782(%ebx,%ecx,1),%ebx
movl %edx, %ecx
roll $5, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
- addl %ebp, %ecx
-.byte 209
-.byte 207 # rorl $1 %edi
+ addl %ebp, %ebx
addl %ecx, %ebx
- # 20_39 76
+
movl 48(%esp), %ecx
- movl 56(%esp), %ebp
- xorl %ebp, %ecx
- movl 16(%esp), %ebp
- xorl %ebp, %ecx
- movl 36(%esp), %ebp
- xorl %ebp, %ecx
movl %edx, %ebp
-.byte 209
-.byte 193 # roll $1 %ecx
+ xorl 56(%esp), %ecx
+ rorl $2, %edx
+ xorl 16(%esp), %ecx
xorl %edi, %ebp
- movl %ecx, 48(%esp)
+ xorl 36(%esp), %ecx
xorl %esi, %ebp
+.byte 209
+.byte 193
+ movl %ecx, 48(%esp)
leal 3395469782(%ecx,%eax,1),%ecx
movl %ebx, %eax
roll $5, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
- addl %ebp, %eax
-.byte 209
-.byte 202 # rorl $1 %edx
+ addl %ebp, %ecx
addl %eax, %ecx
- # 20_39 77
+
movl 52(%esp), %eax
- movl 60(%esp), %ebp
- xorl %ebp, %eax
- movl 20(%esp), %ebp
- xorl %ebp, %eax
- movl 40(%esp), %ebp
- xorl %ebp, %eax
movl %ebx, %ebp
-.byte 209
-.byte 192 # roll $1 %eax
+ xorl 60(%esp), %eax
+ rorl $2, %ebx
+ xorl 20(%esp), %eax
xorl %edx, %ebp
- movl %eax, 52(%esp)
+ xorl 40(%esp), %eax
xorl %edi, %ebp
+.byte 209
+.byte 192
+ movl %eax, 52(%esp)
leal 3395469782(%eax,%esi,1),%eax
movl %ecx, %esi
roll $5, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
- addl %ebp, %esi
-.byte 209
-.byte 203 # rorl $1 %ebx
+ addl %ebp, %eax
addl %esi, %eax
- # 20_39 78
+
movl 56(%esp), %esi
- movl (%esp), %ebp
- xorl %ebp, %esi
- movl 24(%esp), %ebp
- xorl %ebp, %esi
- movl 44(%esp), %ebp
- xorl %ebp, %esi
movl %ecx, %ebp
-.byte 209
-.byte 198 # roll $1 %esi
+ xorl (%esp), %esi
+ rorl $2, %ecx
+ xorl 24(%esp), %esi
xorl %ebx, %ebp
- movl %esi, 56(%esp)
+ xorl 44(%esp), %esi
xorl %edx, %ebp
+.byte 209
+.byte 198
+ movl %esi, 56(%esp)
leal 3395469782(%esi,%edi,1),%esi
movl %eax, %edi
roll $5, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
- addl %ebp, %edi
-.byte 209
-.byte 201 # rorl $1 %ecx
+ addl %ebp, %esi
addl %edi, %esi
- # 20_39 79
+
movl 60(%esp), %edi
- movl 4(%esp), %ebp
- xorl %ebp, %edi
- movl 28(%esp), %ebp
- xorl %ebp, %edi
- movl 48(%esp), %ebp
- xorl %ebp, %edi
movl %eax, %ebp
-.byte 209
-.byte 199 # roll $1 %edi
+ xorl 4(%esp), %edi
+ rorl $2, %eax
+ xorl 28(%esp), %edi
xorl %ecx, %ebp
- movl %edi, 60(%esp)
+ xorl 48(%esp), %edi
xorl %ebx, %ebp
+.byte 209
+.byte 199
+ movl %edi, 60(%esp)
leal 3395469782(%edi,%edx,1),%edi
movl %esi, %edx
roll $5, %edx
- addl %ebp, %edx
- movl 128(%esp), %ebp
-.byte 209
-.byte 200 # rorl $1 %eax
+ addl %ebp, %edi
addl %edx, %edi
-.byte 209
-.byte 200 # rorl $1 %eax
- # End processing
+
+ movl 128(%esp), %ebp
movl 12(%ebp), %edx
addl %ecx, %edx
movl 4(%ebp), %ecx
@@ -1921,7 +1498,7 @@ sha1_block_asm_host_order:
movl 16(%ebp), %edi
movl 8(%ebp), %ebx
movl %ecx, 68(%esp)
- # First we need to setup the X array
+
movl (%esi), %eax
movl 4(%esi), %ecx
movl %eax, (%esp)
OpenPOWER on IntegriCloud