diff options
Diffstat (limited to 'secure/lib/libcrypto/i386/sha1-586.s')
-rw-r--r-- | secure/lib/libcrypto/i386/sha1-586.s | 1785 |
1 files changed, 681 insertions, 1104 deletions
diff --git a/secure/lib/libcrypto/i386/sha1-586.s b/secure/lib/libcrypto/i386/sha1-586.s index b13c665..4a35f9d 100644 --- a/secure/lib/libcrypto/i386/sha1-586.s +++ b/secure/lib/libcrypto/i386/sha1-586.s @@ -1,9 +1,9 @@ # $FreeBSD$ - # Dont even think of reading this code - # It was automatically generated by sha1-586.pl - # Which is a perl program used to generate the x86 assember for - # any of elf, a.out, BSDI, Win32, gaswin (for GNU as on Win32) or Solaris - # eric <eay@cryptsoft.com> + + + + + .file "sha1-586.s" .version "01.01" @@ -27,1851 +27,1428 @@ sha1_block_asm_data_order: movl 16(%ebp), %edi movl 8(%ebp), %ebx movl %ecx, 68(%esp) - # First we need to setup the X array + .L000start: - # First, load the words onto the stack in network byte order + movl (%esi), %eax movl 4(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, (%esp) movl %ecx, 4(%esp) movl 8(%esi), %eax movl 12(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, 8(%esp) movl %ecx, 12(%esp) movl 16(%esi), %eax movl 20(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, 16(%esp) movl %ecx, 20(%esp) movl 24(%esi), %eax movl 28(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, 24(%esp) movl %ecx, 28(%esp) movl 32(%esi), %eax movl 36(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, 32(%esp) movl %ecx, 36(%esp) movl 40(%esi), %eax movl 44(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, 40(%esp) movl %ecx, 44(%esp) movl 48(%esi), %eax movl 52(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, 48(%esp) movl %ecx, 52(%esp) movl 56(%esi), %eax movl 60(%esi), %ecx .byte 15 -.byte 200 # bswapl %eax +.byte 200 .byte 15 -.byte 201 # bswapl %ecx +.byte 201 movl %eax, 56(%esp) movl %ecx, 60(%esp) - # We now have the X array on the stack - # starting at sp-4 + + movl %esi, 132(%esp) .L001shortcut: - # Start processing + movl (%ebp), %eax movl 4(%ebp), %ecx - # 00_15 0 - movl %ebx, %esi + movl %eax, %ebp - xorl %edx, %esi + movl %ebx, %esi roll $5, %ebp + xorl %edx, %esi andl %ecx, %esi + rorl $2, %ecx addl %edi, %ebp -.byte 209 -.byte 201 # rorl $1 %ecx movl (%esp), %edi -.byte 209 -.byte 201 # rorl $1 %ecx xorl %edx, %esi leal 1518500249(%ebp,%edi,1),%ebp - movl %ecx, %edi addl %ebp, %esi - xorl %ebx, %edi + movl %esi, %ebp - andl %eax, %edi + movl %ecx, %edi roll $5, %ebp + xorl %ebx, %edi + andl %eax, %edi + rorl $2, %eax addl %edx, %ebp movl 4(%esp), %edx -.byte 209 -.byte 200 # rorl $1 %eax xorl %ebx, %edi -.byte 209 -.byte 200 # rorl $1 %eax leal 1518500249(%ebp,%edx,1),%ebp addl %ebp, %edi - # 00_15 2 - movl %eax, %edx + movl %edi, %ebp - xorl %ecx, %edx + movl %eax, %edx roll $5, %ebp + xorl %ecx, %edx andl %esi, %edx + rorl $2, %esi addl %ebx, %ebp -.byte 209 -.byte 206 # rorl $1 %esi movl 8(%esp), %ebx -.byte 209 -.byte 206 # rorl $1 %esi xorl %ecx, %edx leal 1518500249(%ebp,%ebx,1),%ebp - movl %esi, %ebx addl %ebp, %edx - xorl %eax, %ebx + movl %edx, %ebp - andl %edi, %ebx + movl %esi, %ebx roll $5, %ebp + xorl %eax, %ebx + andl %edi, %ebx + rorl $2, %edi addl %ecx, %ebp movl 12(%esp), %ecx -.byte 209 -.byte 207 # rorl $1 %edi xorl %eax, %ebx -.byte 209 -.byte 207 # rorl $1 %edi leal 1518500249(%ebp,%ecx,1),%ebp addl %ebp, %ebx - # 00_15 4 - movl %edi, %ecx + movl %ebx, %ebp - xorl %esi, %ecx + movl %edi, %ecx roll $5, %ebp + xorl %esi, %ecx andl %edx, %ecx + rorl $2, %edx addl %eax, %ebp -.byte 209 -.byte 202 # rorl $1 %edx movl 16(%esp), %eax -.byte 209 -.byte 202 # rorl $1 %edx xorl %esi, %ecx leal 1518500249(%ebp,%eax,1),%ebp - movl %edx, %eax addl %ebp, %ecx - xorl %edi, %eax + movl %ecx, %ebp - andl %ebx, %eax + movl %edx, %eax roll $5, %ebp + xorl %edi, %eax + andl %ebx, %eax + rorl $2, %ebx addl %esi, %ebp movl 20(%esp), %esi -.byte 209 -.byte 203 # rorl $1 %ebx xorl %edi, %eax -.byte 209 -.byte 203 # rorl $1 %ebx leal 1518500249(%ebp,%esi,1),%ebp addl %ebp, %eax - # 00_15 6 - movl %ebx, %esi + movl %eax, %ebp - xorl %edx, %esi + movl %ebx, %esi roll $5, %ebp + xorl %edx, %esi andl %ecx, %esi + rorl $2, %ecx addl %edi, %ebp -.byte 209 -.byte 201 # rorl $1 %ecx movl 24(%esp), %edi -.byte 209 -.byte 201 # rorl $1 %ecx xorl %edx, %esi leal 1518500249(%ebp,%edi,1),%ebp - movl %ecx, %edi addl %ebp, %esi - xorl %ebx, %edi + movl %esi, %ebp - andl %eax, %edi + movl %ecx, %edi roll $5, %ebp + xorl %ebx, %edi + andl %eax, %edi + rorl $2, %eax addl %edx, %ebp movl 28(%esp), %edx -.byte 209 -.byte 200 # rorl $1 %eax xorl %ebx, %edi -.byte 209 -.byte 200 # rorl $1 %eax leal 1518500249(%ebp,%edx,1),%ebp addl %ebp, %edi - # 00_15 8 - movl %eax, %edx + movl %edi, %ebp - xorl %ecx, %edx + movl %eax, %edx roll $5, %ebp + xorl %ecx, %edx andl %esi, %edx + rorl $2, %esi addl %ebx, %ebp -.byte 209 -.byte 206 # rorl $1 %esi movl 32(%esp), %ebx -.byte 209 -.byte 206 # rorl $1 %esi xorl %ecx, %edx leal 1518500249(%ebp,%ebx,1),%ebp - movl %esi, %ebx addl %ebp, %edx - xorl %eax, %ebx + movl %edx, %ebp - andl %edi, %ebx + movl %esi, %ebx roll $5, %ebp + xorl %eax, %ebx + andl %edi, %ebx + rorl $2, %edi addl %ecx, %ebp movl 36(%esp), %ecx -.byte 209 -.byte 207 # rorl $1 %edi xorl %eax, %ebx -.byte 209 -.byte 207 # rorl $1 %edi leal 1518500249(%ebp,%ecx,1),%ebp addl %ebp, %ebx - # 00_15 10 - movl %edi, %ecx + movl %ebx, %ebp - xorl %esi, %ecx + movl %edi, %ecx roll $5, %ebp + xorl %esi, %ecx andl %edx, %ecx + rorl $2, %edx addl %eax, %ebp -.byte 209 -.byte 202 # rorl $1 %edx movl 40(%esp), %eax -.byte 209 -.byte 202 # rorl $1 %edx xorl %esi, %ecx leal 1518500249(%ebp,%eax,1),%ebp - movl %edx, %eax addl %ebp, %ecx - xorl %edi, %eax + movl %ecx, %ebp - andl %ebx, %eax + movl %edx, %eax roll $5, %ebp + xorl %edi, %eax + andl %ebx, %eax + rorl $2, %ebx addl %esi, %ebp movl 44(%esp), %esi -.byte 209 -.byte 203 # rorl $1 %ebx xorl %edi, %eax -.byte 209 -.byte 203 # rorl $1 %ebx leal 1518500249(%ebp,%esi,1),%ebp addl %ebp, %eax - # 00_15 12 - movl %ebx, %esi + movl %eax, %ebp - xorl %edx, %esi + movl %ebx, %esi roll $5, %ebp + xorl %edx, %esi andl %ecx, %esi + rorl $2, %ecx addl %edi, %ebp -.byte 209 -.byte 201 # rorl $1 %ecx movl 48(%esp), %edi -.byte 209 -.byte 201 # rorl $1 %ecx xorl %edx, %esi leal 1518500249(%ebp,%edi,1),%ebp - movl %ecx, %edi addl %ebp, %esi - xorl %ebx, %edi + movl %esi, %ebp - andl %eax, %edi + movl %ecx, %edi roll $5, %ebp + xorl %ebx, %edi + andl %eax, %edi + rorl $2, %eax addl %edx, %ebp movl 52(%esp), %edx -.byte 209 -.byte 200 # rorl $1 %eax xorl %ebx, %edi -.byte 209 -.byte 200 # rorl $1 %eax leal 1518500249(%ebp,%edx,1),%ebp addl %ebp, %edi - # 00_15 14 - movl %eax, %edx + movl %edi, %ebp - xorl %ecx, %edx + movl %eax, %edx roll $5, %ebp + xorl %ecx, %edx andl %esi, %edx + rorl $2, %esi addl %ebx, %ebp -.byte 209 -.byte 206 # rorl $1 %esi movl 56(%esp), %ebx -.byte 209 -.byte 206 # rorl $1 %esi xorl %ecx, %edx leal 1518500249(%ebp,%ebx,1),%ebp - movl %esi, %ebx addl %ebp, %edx - xorl %eax, %ebx + movl %edx, %ebp - andl %edi, %ebx + movl %esi, %ebx roll $5, %ebp + xorl %eax, %ebx + andl %edi, %ebx + rorl $2, %edi addl %ecx, %ebp movl 60(%esp), %ecx -.byte 209 -.byte 207 # rorl $1 %edi xorl %eax, %ebx -.byte 209 -.byte 207 # rorl $1 %edi leal 1518500249(%ebp,%ecx,1),%ebp addl %ebp, %ebx - # 16_19 16 - nop - movl (%esp), %ebp + movl 8(%esp), %ecx - xorl %ebp, %ecx - movl 32(%esp), %ebp - xorl %ebp, %ecx - movl 52(%esp), %ebp - xorl %ebp, %ecx movl %edi, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl (%esp), %ecx xorl %esi, %ebp - movl %ecx, (%esp) + xorl 32(%esp), %ecx andl %edx, %ebp - leal 1518500249(%ecx,%eax,1),%ecx + xorl 52(%esp), %ecx + rorl $2, %edx xorl %esi, %ebp +.byte 209 +.byte 193 + movl %ecx, (%esp) + leal 1518500249(%ecx,%eax,1),%ecx movl %ebx, %eax addl %ebp, %ecx roll $5, %eax -.byte 209 -.byte 202 # rorl $1 %edx addl %eax, %ecx - movl 4(%esp), %eax - movl 12(%esp), %ebp - xorl %ebp, %eax - movl 36(%esp), %ebp - xorl %ebp, %eax - movl 56(%esp), %ebp -.byte 209 -.byte 202 # rorl $1 %edx - xorl %ebp, %eax -.byte 209 -.byte 192 # roll $1 %eax + + movl 12(%esp), %eax movl %edx, %ebp + xorl 4(%esp), %eax xorl %edi, %ebp - movl %eax, 4(%esp) + xorl 36(%esp), %eax andl %ebx, %ebp - leal 1518500249(%eax,%esi,1),%eax + xorl 56(%esp), %eax + rorl $2, %ebx xorl %edi, %ebp +.byte 209 +.byte 192 + movl %eax, 4(%esp) + leal 1518500249(%eax,%esi,1),%eax movl %ecx, %esi + addl %ebp, %eax roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx addl %esi, %eax -.byte 209 -.byte 203 # rorl $1 %ebx - addl %ebp, %eax - # 16_19 18 - movl 8(%esp), %ebp + movl 16(%esp), %esi - xorl %ebp, %esi - movl 40(%esp), %ebp - xorl %ebp, %esi - movl 60(%esp), %ebp - xorl %ebp, %esi movl %ebx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 8(%esp), %esi xorl %edx, %ebp - movl %esi, 8(%esp) + xorl 40(%esp), %esi andl %ecx, %ebp - leal 1518500249(%esi,%edi,1),%esi + xorl 60(%esp), %esi + rorl $2, %ecx xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 8(%esp) + leal 1518500249(%esi,%edi,1),%esi movl %eax, %edi addl %ebp, %esi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx addl %edi, %esi - movl 12(%esp), %edi - movl 20(%esp), %ebp - xorl %ebp, %edi - movl 44(%esp), %ebp - xorl %ebp, %edi - movl (%esp), %ebp -.byte 209 -.byte 201 # rorl $1 %ecx - xorl %ebp, %edi -.byte 209 -.byte 199 # roll $1 %edi + + movl 20(%esp), %edi movl %ecx, %ebp + xorl 12(%esp), %edi xorl %ebx, %ebp - movl %edi, 12(%esp) + xorl 44(%esp), %edi andl %eax, %ebp - leal 1518500249(%edi,%edx,1),%edi + xorl (%esp), %edi + rorl $2, %eax xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 12(%esp) + leal 1518500249(%edi,%edx,1),%edi movl %esi, %edx + addl %ebp, %edi roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax addl %edx, %edi -.byte 209 -.byte 200 # rorl $1 %eax - addl %ebp, %edi - # 20_39 20 + movl 16(%esp), %edx - movl 24(%esp), %ebp - xorl %ebp, %edx - movl 48(%esp), %ebp - xorl %ebp, %edx - movl 4(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 24(%esp), %edx + rorl $2, %esi + xorl 48(%esp), %edx xorl %eax, %ebp - movl %edx, 16(%esp) + xorl 4(%esp), %edx xorl %ecx, %ebp +.byte 209 +.byte 194 + movl %edx, 16(%esp) leal 1859775393(%edx,%ebx,1),%edx movl %edi, %ebx roll $5, %ebx -.byte 209 -.byte 206 # rorl $1 %esi - addl %ebp, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + addl %ebp, %edx addl %ebx, %edx - # 20_39 21 + movl 20(%esp), %ebx - movl 28(%esp), %ebp - xorl %ebp, %ebx - movl 52(%esp), %ebp - xorl %ebp, %ebx - movl 8(%esp), %ebp - xorl %ebp, %ebx movl %edi, %ebp -.byte 209 -.byte 195 # roll $1 %ebx + xorl 28(%esp), %ebx + rorl $2, %edi + xorl 52(%esp), %ebx xorl %esi, %ebp - movl %ebx, 20(%esp) + xorl 8(%esp), %ebx xorl %eax, %ebp +.byte 209 +.byte 195 + movl %ebx, 20(%esp) leal 1859775393(%ebx,%ecx,1),%ebx movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi - addl %ebp, %ecx -.byte 209 -.byte 207 # rorl $1 %edi + addl %ebp, %ebx addl %ecx, %ebx - # 20_39 22 + movl 24(%esp), %ecx - movl 32(%esp), %ebp - xorl %ebp, %ecx - movl 56(%esp), %ebp - xorl %ebp, %ecx - movl 12(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 32(%esp), %ecx + rorl $2, %edx + xorl 56(%esp), %ecx xorl %edi, %ebp - movl %ecx, 24(%esp) + xorl 12(%esp), %ecx xorl %esi, %ebp +.byte 209 +.byte 193 + movl %ecx, 24(%esp) leal 1859775393(%ecx,%eax,1),%ecx movl %ebx, %eax roll $5, %eax -.byte 209 -.byte 202 # rorl $1 %edx - addl %ebp, %eax -.byte 209 -.byte 202 # rorl $1 %edx + addl %ebp, %ecx addl %eax, %ecx - # 20_39 23 + movl 28(%esp), %eax - movl 36(%esp), %ebp - xorl %ebp, %eax - movl 60(%esp), %ebp - xorl %ebp, %eax - movl 16(%esp), %ebp - xorl %ebp, %eax movl %ebx, %ebp -.byte 209 -.byte 192 # roll $1 %eax + xorl 36(%esp), %eax + rorl $2, %ebx + xorl 60(%esp), %eax xorl %edx, %ebp - movl %eax, 28(%esp) + xorl 16(%esp), %eax xorl %edi, %ebp +.byte 209 +.byte 192 + movl %eax, 28(%esp) leal 1859775393(%eax,%esi,1),%eax movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx - addl %ebp, %esi -.byte 209 -.byte 203 # rorl $1 %ebx + addl %ebp, %eax addl %esi, %eax - # 20_39 24 + movl 32(%esp), %esi - movl 40(%esp), %ebp - xorl %ebp, %esi - movl (%esp), %ebp - xorl %ebp, %esi - movl 20(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 40(%esp), %esi + rorl $2, %ecx + xorl (%esp), %esi xorl %ebx, %ebp - movl %esi, 32(%esp) + xorl 20(%esp), %esi xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 32(%esp) leal 1859775393(%esi,%edi,1),%esi movl %eax, %edi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx - addl %ebp, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + addl %ebp, %esi addl %edi, %esi - # 20_39 25 + movl 36(%esp), %edi - movl 44(%esp), %ebp - xorl %ebp, %edi - movl 4(%esp), %ebp - xorl %ebp, %edi - movl 24(%esp), %ebp - xorl %ebp, %edi movl %eax, %ebp -.byte 209 -.byte 199 # roll $1 %edi + xorl 44(%esp), %edi + rorl $2, %eax + xorl 4(%esp), %edi xorl %ecx, %ebp - movl %edi, 36(%esp) + xorl 24(%esp), %edi xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 36(%esp) leal 1859775393(%edi,%edx,1),%edi movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax - addl %ebp, %edx -.byte 209 -.byte 200 # rorl $1 %eax + addl %ebp, %edi addl %edx, %edi - # 20_39 26 + movl 40(%esp), %edx - movl 48(%esp), %ebp - xorl %ebp, %edx - movl 8(%esp), %ebp - xorl %ebp, %edx - movl 28(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 48(%esp), %edx + rorl $2, %esi + xorl 8(%esp), %edx xorl %eax, %ebp - movl %edx, 40(%esp) + xorl 28(%esp), %edx xorl %ecx, %ebp +.byte 209 +.byte 194 + movl %edx, 40(%esp) leal 1859775393(%edx,%ebx,1),%edx movl %edi, %ebx roll $5, %ebx -.byte 209 -.byte 206 # rorl $1 %esi - addl %ebp, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + addl %ebp, %edx addl %ebx, %edx - # 20_39 27 + movl 44(%esp), %ebx - movl 52(%esp), %ebp - xorl %ebp, %ebx - movl 12(%esp), %ebp - xorl %ebp, %ebx - movl 32(%esp), %ebp - xorl %ebp, %ebx movl %edi, %ebp -.byte 209 -.byte 195 # roll $1 %ebx + xorl 52(%esp), %ebx + rorl $2, %edi + xorl 12(%esp), %ebx xorl %esi, %ebp - movl %ebx, 44(%esp) + xorl 32(%esp), %ebx xorl %eax, %ebp +.byte 209 +.byte 195 + movl %ebx, 44(%esp) leal 1859775393(%ebx,%ecx,1),%ebx movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi - addl %ebp, %ecx -.byte 209 -.byte 207 # rorl $1 %edi + addl %ebp, %ebx addl %ecx, %ebx - # 20_39 28 + movl 48(%esp), %ecx - movl 56(%esp), %ebp - xorl %ebp, %ecx - movl 16(%esp), %ebp - xorl %ebp, %ecx - movl 36(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 56(%esp), %ecx + rorl $2, %edx + xorl 16(%esp), %ecx xorl %edi, %ebp - movl %ecx, 48(%esp) + xorl 36(%esp), %ecx xorl %esi, %ebp +.byte 209 +.byte 193 + movl %ecx, 48(%esp) leal 1859775393(%ecx,%eax,1),%ecx movl %ebx, %eax roll $5, %eax -.byte 209 -.byte 202 # rorl $1 %edx - addl %ebp, %eax -.byte 209 -.byte 202 # rorl $1 %edx + addl %ebp, %ecx addl %eax, %ecx - # 20_39 29 + movl 52(%esp), %eax - movl 60(%esp), %ebp - xorl %ebp, %eax - movl 20(%esp), %ebp - xorl %ebp, %eax - movl 40(%esp), %ebp - xorl %ebp, %eax movl %ebx, %ebp -.byte 209 -.byte 192 # roll $1 %eax + xorl 60(%esp), %eax + rorl $2, %ebx + xorl 20(%esp), %eax xorl %edx, %ebp - movl %eax, 52(%esp) + xorl 40(%esp), %eax xorl %edi, %ebp +.byte 209 +.byte 192 + movl %eax, 52(%esp) leal 1859775393(%eax,%esi,1),%eax movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx - addl %ebp, %esi -.byte 209 -.byte 203 # rorl $1 %ebx + addl %ebp, %eax addl %esi, %eax - # 20_39 30 + movl 56(%esp), %esi - movl (%esp), %ebp - xorl %ebp, %esi - movl 24(%esp), %ebp - xorl %ebp, %esi - movl 44(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl (%esp), %esi + rorl $2, %ecx + xorl 24(%esp), %esi xorl %ebx, %ebp - movl %esi, 56(%esp) + xorl 44(%esp), %esi xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 56(%esp) leal 1859775393(%esi,%edi,1),%esi movl %eax, %edi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx - addl %ebp, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + addl %ebp, %esi addl %edi, %esi - # 20_39 31 + movl 60(%esp), %edi - movl 4(%esp), %ebp - xorl %ebp, %edi - movl 28(%esp), %ebp - xorl %ebp, %edi - movl 48(%esp), %ebp - xorl %ebp, %edi movl %eax, %ebp -.byte 209 -.byte 199 # roll $1 %edi + xorl 4(%esp), %edi + rorl $2, %eax + xorl 28(%esp), %edi xorl %ecx, %ebp - movl %edi, 60(%esp) + xorl 48(%esp), %edi xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 60(%esp) leal 1859775393(%edi,%edx,1),%edi movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax - addl %ebp, %edx -.byte 209 -.byte 200 # rorl $1 %eax + addl %ebp, %edi addl %edx, %edi - # 20_39 32 + movl (%esp), %edx - movl 8(%esp), %ebp - xorl %ebp, %edx - movl 32(%esp), %ebp - xorl %ebp, %edx - movl 52(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 8(%esp), %edx + rorl $2, %esi + xorl 32(%esp), %edx xorl %eax, %ebp - movl %edx, (%esp) + xorl 52(%esp), %edx xorl %ecx, %ebp +.byte 209 +.byte 194 + movl %edx, (%esp) leal 1859775393(%edx,%ebx,1),%edx movl %edi, %ebx roll $5, %ebx -.byte 209 -.byte 206 # rorl $1 %esi - addl %ebp, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + addl %ebp, %edx addl %ebx, %edx - # 20_39 33 + movl 4(%esp), %ebx - movl 12(%esp), %ebp - xorl %ebp, %ebx - movl 36(%esp), %ebp - xorl %ebp, %ebx - movl 56(%esp), %ebp - xorl %ebp, %ebx movl %edi, %ebp -.byte 209 -.byte 195 # roll $1 %ebx + xorl 12(%esp), %ebx + rorl $2, %edi + xorl 36(%esp), %ebx xorl %esi, %ebp - movl %ebx, 4(%esp) + xorl 56(%esp), %ebx xorl %eax, %ebp +.byte 209 +.byte 195 + movl %ebx, 4(%esp) leal 1859775393(%ebx,%ecx,1),%ebx movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi - addl %ebp, %ecx -.byte 209 -.byte 207 # rorl $1 %edi + addl %ebp, %ebx addl %ecx, %ebx - # 20_39 34 + movl 8(%esp), %ecx - movl 16(%esp), %ebp - xorl %ebp, %ecx - movl 40(%esp), %ebp - xorl %ebp, %ecx - movl 60(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 16(%esp), %ecx + rorl $2, %edx + xorl 40(%esp), %ecx xorl %edi, %ebp - movl %ecx, 8(%esp) + xorl 60(%esp), %ecx xorl %esi, %ebp +.byte 209 +.byte 193 + movl %ecx, 8(%esp) leal 1859775393(%ecx,%eax,1),%ecx movl %ebx, %eax roll $5, %eax -.byte 209 -.byte 202 # rorl $1 %edx - addl %ebp, %eax -.byte 209 -.byte 202 # rorl $1 %edx + addl %ebp, %ecx addl %eax, %ecx - # 20_39 35 + movl 12(%esp), %eax - movl 20(%esp), %ebp - xorl %ebp, %eax - movl 44(%esp), %ebp - xorl %ebp, %eax - movl (%esp), %ebp - xorl %ebp, %eax movl %ebx, %ebp -.byte 209 -.byte 192 # roll $1 %eax + xorl 20(%esp), %eax + rorl $2, %ebx + xorl 44(%esp), %eax xorl %edx, %ebp - movl %eax, 12(%esp) + xorl (%esp), %eax xorl %edi, %ebp +.byte 209 +.byte 192 + movl %eax, 12(%esp) leal 1859775393(%eax,%esi,1),%eax movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx - addl %ebp, %esi -.byte 209 -.byte 203 # rorl $1 %ebx + addl %ebp, %eax addl %esi, %eax - # 20_39 36 + movl 16(%esp), %esi - movl 24(%esp), %ebp - xorl %ebp, %esi - movl 48(%esp), %ebp - xorl %ebp, %esi - movl 4(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 24(%esp), %esi + rorl $2, %ecx + xorl 48(%esp), %esi xorl %ebx, %ebp - movl %esi, 16(%esp) + xorl 4(%esp), %esi xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 16(%esp) leal 1859775393(%esi,%edi,1),%esi movl %eax, %edi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx - addl %ebp, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + addl %ebp, %esi addl %edi, %esi - # 20_39 37 + movl 20(%esp), %edi - movl 28(%esp), %ebp - xorl %ebp, %edi - movl 52(%esp), %ebp - xorl %ebp, %edi - movl 8(%esp), %ebp - xorl %ebp, %edi movl %eax, %ebp -.byte 209 -.byte 199 # roll $1 %edi + xorl 28(%esp), %edi + rorl $2, %eax + xorl 52(%esp), %edi xorl %ecx, %ebp - movl %edi, 20(%esp) + xorl 8(%esp), %edi xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 20(%esp) leal 1859775393(%edi,%edx,1),%edi movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax - addl %ebp, %edx -.byte 209 -.byte 200 # rorl $1 %eax + addl %ebp, %edi addl %edx, %edi - # 20_39 38 + movl 24(%esp), %edx - movl 32(%esp), %ebp - xorl %ebp, %edx - movl 56(%esp), %ebp - xorl %ebp, %edx - movl 12(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 32(%esp), %edx + rorl $2, %esi + xorl 56(%esp), %edx xorl %eax, %ebp - movl %edx, 24(%esp) + xorl 12(%esp), %edx xorl %ecx, %ebp +.byte 209 +.byte 194 + movl %edx, 24(%esp) leal 1859775393(%edx,%ebx,1),%edx movl %edi, %ebx roll $5, %ebx -.byte 209 -.byte 206 # rorl $1 %esi - addl %ebp, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + addl %ebp, %edx addl %ebx, %edx - # 20_39 39 + movl 28(%esp), %ebx - movl 36(%esp), %ebp - xorl %ebp, %ebx - movl 60(%esp), %ebp - xorl %ebp, %ebx - movl 16(%esp), %ebp - xorl %ebp, %ebx movl %edi, %ebp -.byte 209 -.byte 195 # roll $1 %ebx + xorl 36(%esp), %ebx + rorl $2, %edi + xorl 60(%esp), %ebx xorl %esi, %ebp - movl %ebx, 28(%esp) + xorl 16(%esp), %ebx xorl %eax, %ebp +.byte 209 +.byte 195 + movl %ebx, 28(%esp) leal 1859775393(%ebx,%ecx,1),%ebx movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi - addl %ebp, %ecx -.byte 209 -.byte 207 # rorl $1 %edi + addl %ebp, %ebx addl %ecx, %ebx - # 40_59 40 + movl 32(%esp), %ecx - movl 40(%esp), %ebp - xorl %ebp, %ecx - movl (%esp), %ebp - xorl %ebp, %ecx - movl 20(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 40(%esp), %ecx orl %edi, %ebp - movl %ecx, 32(%esp) + xorl (%esp), %ecx andl %esi, %ebp + xorl 20(%esp), %ecx +.byte 209 +.byte 193 + movl %ecx, 32(%esp) leal 2400959708(%ecx,%eax,1),%ecx movl %edx, %eax -.byte 209 -.byte 202 # rorl $1 %edx + rorl $2, %edx andl %edi, %eax orl %eax, %ebp movl %ebx, %eax roll $5, %eax addl %eax, %ebp - movl 36(%esp), %eax addl %ebp, %ecx - movl 44(%esp), %ebp - xorl %ebp, %eax - movl 4(%esp), %ebp - xorl %ebp, %eax - movl 24(%esp), %ebp -.byte 209 -.byte 202 # rorl $1 %edx - xorl %ebp, %eax -.byte 209 -.byte 192 # roll $1 %eax + + movl 36(%esp), %eax movl %ebx, %ebp - movl %eax, 36(%esp) + xorl 44(%esp), %eax orl %edx, %ebp + xorl 4(%esp), %eax + andl %edi, %ebp + xorl 24(%esp), %eax +.byte 209 +.byte 192 + movl %eax, 36(%esp) leal 2400959708(%eax,%esi,1),%eax movl %ebx, %esi - andl %edi, %ebp + rorl $2, %ebx andl %edx, %esi orl %esi, %ebp movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx addl %esi, %ebp -.byte 209 -.byte 203 # rorl $1 %ebx addl %ebp, %eax - # 40_59 41 - # 40_59 42 + movl 40(%esp), %esi - movl 48(%esp), %ebp - xorl %ebp, %esi - movl 8(%esp), %ebp - xorl %ebp, %esi - movl 28(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 48(%esp), %esi orl %ebx, %ebp - movl %esi, 40(%esp) + xorl 8(%esp), %esi andl %edx, %ebp + xorl 28(%esp), %esi +.byte 209 +.byte 198 + movl %esi, 40(%esp) leal 2400959708(%esi,%edi,1),%esi movl %ecx, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + rorl $2, %ecx andl %ebx, %edi orl %edi, %ebp movl %eax, %edi roll $5, %edi addl %edi, %ebp - movl 44(%esp), %edi addl %ebp, %esi - movl 52(%esp), %ebp - xorl %ebp, %edi - movl 12(%esp), %ebp - xorl %ebp, %edi - movl 32(%esp), %ebp -.byte 209 -.byte 201 # rorl $1 %ecx - xorl %ebp, %edi -.byte 209 -.byte 199 # roll $1 %edi + + movl 44(%esp), %edi movl %eax, %ebp - movl %edi, 44(%esp) + xorl 52(%esp), %edi orl %ecx, %ebp + xorl 12(%esp), %edi + andl %ebx, %ebp + xorl 32(%esp), %edi +.byte 209 +.byte 199 + movl %edi, 44(%esp) leal 2400959708(%edi,%edx,1),%edi movl %eax, %edx - andl %ebx, %ebp + rorl $2, %eax andl %ecx, %edx orl %edx, %ebp movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax addl %edx, %ebp -.byte 209 -.byte 200 # rorl $1 %eax addl %ebp, %edi - # 40_59 43 - # 40_59 44 + movl 48(%esp), %edx - movl 56(%esp), %ebp - xorl %ebp, %edx - movl 16(%esp), %ebp - xorl %ebp, %edx - movl 36(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 56(%esp), %edx orl %eax, %ebp - movl %edx, 48(%esp) + xorl 16(%esp), %edx andl %ecx, %ebp + xorl 36(%esp), %edx +.byte 209 +.byte 194 + movl %edx, 48(%esp) leal 2400959708(%edx,%ebx,1),%edx movl %esi, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + rorl $2, %esi andl %eax, %ebx orl %ebx, %ebp movl %edi, %ebx roll $5, %ebx addl %ebx, %ebp - movl 52(%esp), %ebx addl %ebp, %edx - movl 60(%esp), %ebp - xorl %ebp, %ebx - movl 20(%esp), %ebp - xorl %ebp, %ebx - movl 40(%esp), %ebp -.byte 209 -.byte 206 # rorl $1 %esi - xorl %ebp, %ebx -.byte 209 -.byte 195 # roll $1 %ebx + + movl 52(%esp), %ebx movl %edi, %ebp - movl %ebx, 52(%esp) + xorl 60(%esp), %ebx orl %esi, %ebp + xorl 20(%esp), %ebx + andl %eax, %ebp + xorl 40(%esp), %ebx +.byte 209 +.byte 195 + movl %ebx, 52(%esp) leal 2400959708(%ebx,%ecx,1),%ebx movl %edi, %ecx - andl %eax, %ebp + rorl $2, %edi andl %esi, %ecx orl %ecx, %ebp movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi addl %ecx, %ebp -.byte 209 -.byte 207 # rorl $1 %edi addl %ebp, %ebx - # 40_59 45 - # 40_59 46 + movl 56(%esp), %ecx - movl (%esp), %ebp - xorl %ebp, %ecx - movl 24(%esp), %ebp - xorl %ebp, %ecx - movl 44(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl (%esp), %ecx orl %edi, %ebp - movl %ecx, 56(%esp) + xorl 24(%esp), %ecx andl %esi, %ebp + xorl 44(%esp), %ecx +.byte 209 +.byte 193 + movl %ecx, 56(%esp) leal 2400959708(%ecx,%eax,1),%ecx movl %edx, %eax -.byte 209 -.byte 202 # rorl $1 %edx + rorl $2, %edx andl %edi, %eax orl %eax, %ebp movl %ebx, %eax roll $5, %eax addl %eax, %ebp - movl 60(%esp), %eax addl %ebp, %ecx - movl 4(%esp), %ebp - xorl %ebp, %eax - movl 28(%esp), %ebp - xorl %ebp, %eax - movl 48(%esp), %ebp -.byte 209 -.byte 202 # rorl $1 %edx - xorl %ebp, %eax -.byte 209 -.byte 192 # roll $1 %eax + + movl 60(%esp), %eax movl %ebx, %ebp - movl %eax, 60(%esp) + xorl 4(%esp), %eax orl %edx, %ebp + xorl 28(%esp), %eax + andl %edi, %ebp + xorl 48(%esp), %eax +.byte 209 +.byte 192 + movl %eax, 60(%esp) leal 2400959708(%eax,%esi,1),%eax movl %ebx, %esi - andl %edi, %ebp + rorl $2, %ebx andl %edx, %esi orl %esi, %ebp movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx addl %esi, %ebp -.byte 209 -.byte 203 # rorl $1 %ebx addl %ebp, %eax - # 40_59 47 - # 40_59 48 + movl (%esp), %esi - movl 8(%esp), %ebp - xorl %ebp, %esi - movl 32(%esp), %ebp - xorl %ebp, %esi - movl 52(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 8(%esp), %esi orl %ebx, %ebp - movl %esi, (%esp) + xorl 32(%esp), %esi andl %edx, %ebp + xorl 52(%esp), %esi +.byte 209 +.byte 198 + movl %esi, (%esp) leal 2400959708(%esi,%edi,1),%esi movl %ecx, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + rorl $2, %ecx andl %ebx, %edi orl %edi, %ebp movl %eax, %edi roll $5, %edi addl %edi, %ebp - movl 4(%esp), %edi addl %ebp, %esi - movl 12(%esp), %ebp - xorl %ebp, %edi - movl 36(%esp), %ebp - xorl %ebp, %edi - movl 56(%esp), %ebp -.byte 209 -.byte 201 # rorl $1 %ecx - xorl %ebp, %edi -.byte 209 -.byte 199 # roll $1 %edi + + movl 4(%esp), %edi movl %eax, %ebp - movl %edi, 4(%esp) + xorl 12(%esp), %edi orl %ecx, %ebp + xorl 36(%esp), %edi + andl %ebx, %ebp + xorl 56(%esp), %edi +.byte 209 +.byte 199 + movl %edi, 4(%esp) leal 2400959708(%edi,%edx,1),%edi movl %eax, %edx - andl %ebx, %ebp + rorl $2, %eax andl %ecx, %edx orl %edx, %ebp movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax addl %edx, %ebp -.byte 209 -.byte 200 # rorl $1 %eax addl %ebp, %edi - # 40_59 49 - # 40_59 50 + movl 8(%esp), %edx - movl 16(%esp), %ebp - xorl %ebp, %edx - movl 40(%esp), %ebp - xorl %ebp, %edx - movl 60(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 16(%esp), %edx orl %eax, %ebp - movl %edx, 8(%esp) + xorl 40(%esp), %edx andl %ecx, %ebp + xorl 60(%esp), %edx +.byte 209 +.byte 194 + movl %edx, 8(%esp) leal 2400959708(%edx,%ebx,1),%edx movl %esi, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + rorl $2, %esi andl %eax, %ebx orl %ebx, %ebp movl %edi, %ebx roll $5, %ebx addl %ebx, %ebp - movl 12(%esp), %ebx addl %ebp, %edx - movl 20(%esp), %ebp - xorl %ebp, %ebx - movl 44(%esp), %ebp - xorl %ebp, %ebx - movl (%esp), %ebp -.byte 209 -.byte 206 # rorl $1 %esi - xorl %ebp, %ebx -.byte 209 -.byte 195 # roll $1 %ebx + + movl 12(%esp), %ebx movl %edi, %ebp - movl %ebx, 12(%esp) + xorl 20(%esp), %ebx orl %esi, %ebp + xorl 44(%esp), %ebx + andl %eax, %ebp + xorl (%esp), %ebx +.byte 209 +.byte 195 + movl %ebx, 12(%esp) leal 2400959708(%ebx,%ecx,1),%ebx movl %edi, %ecx - andl %eax, %ebp + rorl $2, %edi andl %esi, %ecx orl %ecx, %ebp movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi addl %ecx, %ebp -.byte 209 -.byte 207 # rorl $1 %edi addl %ebp, %ebx - # 40_59 51 - # 40_59 52 + movl 16(%esp), %ecx - movl 24(%esp), %ebp - xorl %ebp, %ecx - movl 48(%esp), %ebp - xorl %ebp, %ecx - movl 4(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 24(%esp), %ecx orl %edi, %ebp - movl %ecx, 16(%esp) + xorl 48(%esp), %ecx andl %esi, %ebp + xorl 4(%esp), %ecx +.byte 209 +.byte 193 + movl %ecx, 16(%esp) leal 2400959708(%ecx,%eax,1),%ecx movl %edx, %eax -.byte 209 -.byte 202 # rorl $1 %edx + rorl $2, %edx andl %edi, %eax orl %eax, %ebp movl %ebx, %eax roll $5, %eax addl %eax, %ebp - movl 20(%esp), %eax addl %ebp, %ecx - movl 28(%esp), %ebp - xorl %ebp, %eax - movl 52(%esp), %ebp - xorl %ebp, %eax - movl 8(%esp), %ebp -.byte 209 -.byte 202 # rorl $1 %edx - xorl %ebp, %eax -.byte 209 -.byte 192 # roll $1 %eax + + movl 20(%esp), %eax movl %ebx, %ebp - movl %eax, 20(%esp) + xorl 28(%esp), %eax orl %edx, %ebp + xorl 52(%esp), %eax + andl %edi, %ebp + xorl 8(%esp), %eax +.byte 209 +.byte 192 + movl %eax, 20(%esp) leal 2400959708(%eax,%esi,1),%eax movl %ebx, %esi - andl %edi, %ebp + rorl $2, %ebx andl %edx, %esi orl %esi, %ebp movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx addl %esi, %ebp -.byte 209 -.byte 203 # rorl $1 %ebx addl %ebp, %eax - # 40_59 53 - # 40_59 54 + movl 24(%esp), %esi - movl 32(%esp), %ebp - xorl %ebp, %esi - movl 56(%esp), %ebp - xorl %ebp, %esi - movl 12(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 32(%esp), %esi orl %ebx, %ebp - movl %esi, 24(%esp) + xorl 56(%esp), %esi andl %edx, %ebp + xorl 12(%esp), %esi +.byte 209 +.byte 198 + movl %esi, 24(%esp) leal 2400959708(%esi,%edi,1),%esi movl %ecx, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + rorl $2, %ecx andl %ebx, %edi orl %edi, %ebp movl %eax, %edi roll $5, %edi addl %edi, %ebp - movl 28(%esp), %edi addl %ebp, %esi - movl 36(%esp), %ebp - xorl %ebp, %edi - movl 60(%esp), %ebp - xorl %ebp, %edi - movl 16(%esp), %ebp -.byte 209 -.byte 201 # rorl $1 %ecx - xorl %ebp, %edi -.byte 209 -.byte 199 # roll $1 %edi + + movl 28(%esp), %edi movl %eax, %ebp - movl %edi, 28(%esp) + xorl 36(%esp), %edi orl %ecx, %ebp + xorl 60(%esp), %edi + andl %ebx, %ebp + xorl 16(%esp), %edi +.byte 209 +.byte 199 + movl %edi, 28(%esp) leal 2400959708(%edi,%edx,1),%edi movl %eax, %edx - andl %ebx, %ebp + rorl $2, %eax andl %ecx, %edx orl %edx, %ebp movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax addl %edx, %ebp -.byte 209 -.byte 200 # rorl $1 %eax addl %ebp, %edi - # 40_59 55 - # 40_59 56 + movl 32(%esp), %edx - movl 40(%esp), %ebp - xorl %ebp, %edx - movl (%esp), %ebp - xorl %ebp, %edx - movl 20(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 40(%esp), %edx orl %eax, %ebp - movl %edx, 32(%esp) + xorl (%esp), %edx andl %ecx, %ebp + xorl 20(%esp), %edx +.byte 209 +.byte 194 + movl %edx, 32(%esp) leal 2400959708(%edx,%ebx,1),%edx movl %esi, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + rorl $2, %esi andl %eax, %ebx orl %ebx, %ebp movl %edi, %ebx roll $5, %ebx addl %ebx, %ebp - movl 36(%esp), %ebx addl %ebp, %edx - movl 44(%esp), %ebp - xorl %ebp, %ebx - movl 4(%esp), %ebp - xorl %ebp, %ebx - movl 24(%esp), %ebp -.byte 209 -.byte 206 # rorl $1 %esi - xorl %ebp, %ebx -.byte 209 -.byte 195 # roll $1 %ebx + + movl 36(%esp), %ebx movl %edi, %ebp - movl %ebx, 36(%esp) + xorl 44(%esp), %ebx orl %esi, %ebp + xorl 4(%esp), %ebx + andl %eax, %ebp + xorl 24(%esp), %ebx +.byte 209 +.byte 195 + movl %ebx, 36(%esp) leal 2400959708(%ebx,%ecx,1),%ebx movl %edi, %ecx - andl %eax, %ebp + rorl $2, %edi andl %esi, %ecx orl %ecx, %ebp movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi addl %ecx, %ebp -.byte 209 -.byte 207 # rorl $1 %edi addl %ebp, %ebx - # 40_59 57 - # 40_59 58 + movl 40(%esp), %ecx - movl 48(%esp), %ebp - xorl %ebp, %ecx - movl 8(%esp), %ebp - xorl %ebp, %ecx - movl 28(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 48(%esp), %ecx orl %edi, %ebp - movl %ecx, 40(%esp) + xorl 8(%esp), %ecx andl %esi, %ebp + xorl 28(%esp), %ecx +.byte 209 +.byte 193 + movl %ecx, 40(%esp) leal 2400959708(%ecx,%eax,1),%ecx movl %edx, %eax -.byte 209 -.byte 202 # rorl $1 %edx + rorl $2, %edx andl %edi, %eax orl %eax, %ebp movl %ebx, %eax roll $5, %eax addl %eax, %ebp - movl 44(%esp), %eax addl %ebp, %ecx - movl 52(%esp), %ebp - xorl %ebp, %eax - movl 12(%esp), %ebp - xorl %ebp, %eax - movl 32(%esp), %ebp -.byte 209 -.byte 202 # rorl $1 %edx - xorl %ebp, %eax -.byte 209 -.byte 192 # roll $1 %eax + + movl 44(%esp), %eax movl %ebx, %ebp - movl %eax, 44(%esp) + xorl 52(%esp), %eax orl %edx, %ebp + xorl 12(%esp), %eax + andl %edi, %ebp + xorl 32(%esp), %eax +.byte 209 +.byte 192 + movl %eax, 44(%esp) leal 2400959708(%eax,%esi,1),%eax movl %ebx, %esi - andl %edi, %ebp + rorl $2, %ebx andl %edx, %esi orl %esi, %ebp movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx addl %esi, %ebp -.byte 209 -.byte 203 # rorl $1 %ebx addl %ebp, %eax - # 40_59 59 - # 20_39 60 + movl 48(%esp), %esi - movl 56(%esp), %ebp - xorl %ebp, %esi - movl 16(%esp), %ebp - xorl %ebp, %esi - movl 36(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 56(%esp), %esi + rorl $2, %ecx + xorl 16(%esp), %esi xorl %ebx, %ebp - movl %esi, 48(%esp) + xorl 36(%esp), %esi xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 48(%esp) leal 3395469782(%esi,%edi,1),%esi movl %eax, %edi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx - addl %ebp, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + addl %ebp, %esi addl %edi, %esi - # 20_39 61 + movl 52(%esp), %edi - movl 60(%esp), %ebp - xorl %ebp, %edi - movl 20(%esp), %ebp - xorl %ebp, %edi - movl 40(%esp), %ebp - xorl %ebp, %edi movl %eax, %ebp -.byte 209 -.byte 199 # roll $1 %edi + xorl 60(%esp), %edi + rorl $2, %eax + xorl 20(%esp), %edi xorl %ecx, %ebp - movl %edi, 52(%esp) + xorl 40(%esp), %edi xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 52(%esp) leal 3395469782(%edi,%edx,1),%edi movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax - addl %ebp, %edx -.byte 209 -.byte 200 # rorl $1 %eax + addl %ebp, %edi addl %edx, %edi - # 20_39 62 + movl 56(%esp), %edx - movl (%esp), %ebp - xorl %ebp, %edx - movl 24(%esp), %ebp - xorl %ebp, %edx - movl 44(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl (%esp), %edx + rorl $2, %esi + xorl 24(%esp), %edx xorl %eax, %ebp - movl %edx, 56(%esp) + xorl 44(%esp), %edx xorl %ecx, %ebp +.byte 209 +.byte 194 + movl %edx, 56(%esp) leal 3395469782(%edx,%ebx,1),%edx movl %edi, %ebx roll $5, %ebx -.byte 209 -.byte 206 # rorl $1 %esi - addl %ebp, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + addl %ebp, %edx addl %ebx, %edx - # 20_39 63 + movl 60(%esp), %ebx - movl 4(%esp), %ebp - xorl %ebp, %ebx - movl 28(%esp), %ebp - xorl %ebp, %ebx - movl 48(%esp), %ebp - xorl %ebp, %ebx movl %edi, %ebp -.byte 209 -.byte 195 # roll $1 %ebx + xorl 4(%esp), %ebx + rorl $2, %edi + xorl 28(%esp), %ebx xorl %esi, %ebp - movl %ebx, 60(%esp) + xorl 48(%esp), %ebx xorl %eax, %ebp +.byte 209 +.byte 195 + movl %ebx, 60(%esp) leal 3395469782(%ebx,%ecx,1),%ebx movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi - addl %ebp, %ecx -.byte 209 -.byte 207 # rorl $1 %edi + addl %ebp, %ebx addl %ecx, %ebx - # 20_39 64 + movl (%esp), %ecx - movl 8(%esp), %ebp - xorl %ebp, %ecx - movl 32(%esp), %ebp - xorl %ebp, %ecx - movl 52(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 8(%esp), %ecx + rorl $2, %edx + xorl 32(%esp), %ecx xorl %edi, %ebp - movl %ecx, (%esp) + xorl 52(%esp), %ecx xorl %esi, %ebp +.byte 209 +.byte 193 + movl %ecx, (%esp) leal 3395469782(%ecx,%eax,1),%ecx movl %ebx, %eax roll $5, %eax -.byte 209 -.byte 202 # rorl $1 %edx - addl %ebp, %eax -.byte 209 -.byte 202 # rorl $1 %edx + addl %ebp, %ecx addl %eax, %ecx - # 20_39 65 + movl 4(%esp), %eax - movl 12(%esp), %ebp - xorl %ebp, %eax - movl 36(%esp), %ebp - xorl %ebp, %eax - movl 56(%esp), %ebp - xorl %ebp, %eax movl %ebx, %ebp -.byte 209 -.byte 192 # roll $1 %eax + xorl 12(%esp), %eax + rorl $2, %ebx + xorl 36(%esp), %eax xorl %edx, %ebp - movl %eax, 4(%esp) + xorl 56(%esp), %eax xorl %edi, %ebp +.byte 209 +.byte 192 + movl %eax, 4(%esp) leal 3395469782(%eax,%esi,1),%eax movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx - addl %ebp, %esi -.byte 209 -.byte 203 # rorl $1 %ebx + addl %ebp, %eax addl %esi, %eax - # 20_39 66 + movl 8(%esp), %esi - movl 16(%esp), %ebp - xorl %ebp, %esi - movl 40(%esp), %ebp - xorl %ebp, %esi - movl 60(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 16(%esp), %esi + rorl $2, %ecx + xorl 40(%esp), %esi xorl %ebx, %ebp - movl %esi, 8(%esp) + xorl 60(%esp), %esi xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 8(%esp) leal 3395469782(%esi,%edi,1),%esi movl %eax, %edi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx - addl %ebp, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + addl %ebp, %esi addl %edi, %esi - # 20_39 67 + movl 12(%esp), %edi - movl 20(%esp), %ebp - xorl %ebp, %edi - movl 44(%esp), %ebp - xorl %ebp, %edi - movl (%esp), %ebp - xorl %ebp, %edi movl %eax, %ebp -.byte 209 -.byte 199 # roll $1 %edi + xorl 20(%esp), %edi + rorl $2, %eax + xorl 44(%esp), %edi xorl %ecx, %ebp - movl %edi, 12(%esp) + xorl (%esp), %edi xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 12(%esp) leal 3395469782(%edi,%edx,1),%edi movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax - addl %ebp, %edx -.byte 209 -.byte 200 # rorl $1 %eax + addl %ebp, %edi addl %edx, %edi - # 20_39 68 + movl 16(%esp), %edx - movl 24(%esp), %ebp - xorl %ebp, %edx - movl 48(%esp), %ebp - xorl %ebp, %edx - movl 4(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 24(%esp), %edx + rorl $2, %esi + xorl 48(%esp), %edx xorl %eax, %ebp - movl %edx, 16(%esp) + xorl 4(%esp), %edx xorl %ecx, %ebp +.byte 209 +.byte 194 + movl %edx, 16(%esp) leal 3395469782(%edx,%ebx,1),%edx movl %edi, %ebx roll $5, %ebx -.byte 209 -.byte 206 # rorl $1 %esi - addl %ebp, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + addl %ebp, %edx addl %ebx, %edx - # 20_39 69 + movl 20(%esp), %ebx - movl 28(%esp), %ebp - xorl %ebp, %ebx - movl 52(%esp), %ebp - xorl %ebp, %ebx - movl 8(%esp), %ebp - xorl %ebp, %ebx movl %edi, %ebp -.byte 209 -.byte 195 # roll $1 %ebx + xorl 28(%esp), %ebx + rorl $2, %edi + xorl 52(%esp), %ebx xorl %esi, %ebp - movl %ebx, 20(%esp) + xorl 8(%esp), %ebx xorl %eax, %ebp +.byte 209 +.byte 195 + movl %ebx, 20(%esp) leal 3395469782(%ebx,%ecx,1),%ebx movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi - addl %ebp, %ecx -.byte 209 -.byte 207 # rorl $1 %edi + addl %ebp, %ebx addl %ecx, %ebx - # 20_39 70 + movl 24(%esp), %ecx - movl 32(%esp), %ebp - xorl %ebp, %ecx - movl 56(%esp), %ebp - xorl %ebp, %ecx - movl 12(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 32(%esp), %ecx + rorl $2, %edx + xorl 56(%esp), %ecx xorl %edi, %ebp - movl %ecx, 24(%esp) + xorl 12(%esp), %ecx xorl %esi, %ebp +.byte 209 +.byte 193 + movl %ecx, 24(%esp) leal 3395469782(%ecx,%eax,1),%ecx movl %ebx, %eax roll $5, %eax -.byte 209 -.byte 202 # rorl $1 %edx - addl %ebp, %eax -.byte 209 -.byte 202 # rorl $1 %edx + addl %ebp, %ecx addl %eax, %ecx - # 20_39 71 + movl 28(%esp), %eax - movl 36(%esp), %ebp - xorl %ebp, %eax - movl 60(%esp), %ebp - xorl %ebp, %eax - movl 16(%esp), %ebp - xorl %ebp, %eax movl %ebx, %ebp -.byte 209 -.byte 192 # roll $1 %eax + xorl 36(%esp), %eax + rorl $2, %ebx + xorl 60(%esp), %eax xorl %edx, %ebp - movl %eax, 28(%esp) + xorl 16(%esp), %eax xorl %edi, %ebp +.byte 209 +.byte 192 + movl %eax, 28(%esp) leal 3395469782(%eax,%esi,1),%eax movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx - addl %ebp, %esi -.byte 209 -.byte 203 # rorl $1 %ebx + addl %ebp, %eax addl %esi, %eax - # 20_39 72 + movl 32(%esp), %esi - movl 40(%esp), %ebp - xorl %ebp, %esi - movl (%esp), %ebp - xorl %ebp, %esi - movl 20(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl 40(%esp), %esi + rorl $2, %ecx + xorl (%esp), %esi xorl %ebx, %ebp - movl %esi, 32(%esp) + xorl 20(%esp), %esi xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 32(%esp) leal 3395469782(%esi,%edi,1),%esi movl %eax, %edi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx - addl %ebp, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + addl %ebp, %esi addl %edi, %esi - # 20_39 73 + movl 36(%esp), %edi - movl 44(%esp), %ebp - xorl %ebp, %edi - movl 4(%esp), %ebp - xorl %ebp, %edi - movl 24(%esp), %ebp - xorl %ebp, %edi movl %eax, %ebp -.byte 209 -.byte 199 # roll $1 %edi + xorl 44(%esp), %edi + rorl $2, %eax + xorl 4(%esp), %edi xorl %ecx, %ebp - movl %edi, 36(%esp) + xorl 24(%esp), %edi xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 36(%esp) leal 3395469782(%edi,%edx,1),%edi movl %esi, %edx roll $5, %edx -.byte 209 -.byte 200 # rorl $1 %eax - addl %ebp, %edx -.byte 209 -.byte 200 # rorl $1 %eax + addl %ebp, %edi addl %edx, %edi - # 20_39 74 + movl 40(%esp), %edx - movl 48(%esp), %ebp - xorl %ebp, %edx - movl 8(%esp), %ebp - xorl %ebp, %edx - movl 28(%esp), %ebp - xorl %ebp, %edx movl %esi, %ebp -.byte 209 -.byte 194 # roll $1 %edx + xorl 48(%esp), %edx + rorl $2, %esi + xorl 8(%esp), %edx xorl %eax, %ebp - movl %edx, 40(%esp) + xorl 28(%esp), %edx xorl %ecx, %ebp +.byte 209 +.byte 194 + movl %edx, 40(%esp) leal 3395469782(%edx,%ebx,1),%edx movl %edi, %ebx roll $5, %ebx -.byte 209 -.byte 206 # rorl $1 %esi - addl %ebp, %ebx -.byte 209 -.byte 206 # rorl $1 %esi + addl %ebp, %edx addl %ebx, %edx - # 20_39 75 + movl 44(%esp), %ebx - movl 52(%esp), %ebp - xorl %ebp, %ebx - movl 12(%esp), %ebp - xorl %ebp, %ebx - movl 32(%esp), %ebp - xorl %ebp, %ebx movl %edi, %ebp -.byte 209 -.byte 195 # roll $1 %ebx + xorl 52(%esp), %ebx + rorl $2, %edi + xorl 12(%esp), %ebx xorl %esi, %ebp - movl %ebx, 44(%esp) + xorl 32(%esp), %ebx xorl %eax, %ebp +.byte 209 +.byte 195 + movl %ebx, 44(%esp) leal 3395469782(%ebx,%ecx,1),%ebx movl %edx, %ecx roll $5, %ecx -.byte 209 -.byte 207 # rorl $1 %edi - addl %ebp, %ecx -.byte 209 -.byte 207 # rorl $1 %edi + addl %ebp, %ebx addl %ecx, %ebx - # 20_39 76 + movl 48(%esp), %ecx - movl 56(%esp), %ebp - xorl %ebp, %ecx - movl 16(%esp), %ebp - xorl %ebp, %ecx - movl 36(%esp), %ebp - xorl %ebp, %ecx movl %edx, %ebp -.byte 209 -.byte 193 # roll $1 %ecx + xorl 56(%esp), %ecx + rorl $2, %edx + xorl 16(%esp), %ecx xorl %edi, %ebp - movl %ecx, 48(%esp) + xorl 36(%esp), %ecx xorl %esi, %ebp +.byte 209 +.byte 193 + movl %ecx, 48(%esp) leal 3395469782(%ecx,%eax,1),%ecx movl %ebx, %eax roll $5, %eax -.byte 209 -.byte 202 # rorl $1 %edx - addl %ebp, %eax -.byte 209 -.byte 202 # rorl $1 %edx + addl %ebp, %ecx addl %eax, %ecx - # 20_39 77 + movl 52(%esp), %eax - movl 60(%esp), %ebp - xorl %ebp, %eax - movl 20(%esp), %ebp - xorl %ebp, %eax - movl 40(%esp), %ebp - xorl %ebp, %eax movl %ebx, %ebp -.byte 209 -.byte 192 # roll $1 %eax + xorl 60(%esp), %eax + rorl $2, %ebx + xorl 20(%esp), %eax xorl %edx, %ebp - movl %eax, 52(%esp) + xorl 40(%esp), %eax xorl %edi, %ebp +.byte 209 +.byte 192 + movl %eax, 52(%esp) leal 3395469782(%eax,%esi,1),%eax movl %ecx, %esi roll $5, %esi -.byte 209 -.byte 203 # rorl $1 %ebx - addl %ebp, %esi -.byte 209 -.byte 203 # rorl $1 %ebx + addl %ebp, %eax addl %esi, %eax - # 20_39 78 + movl 56(%esp), %esi - movl (%esp), %ebp - xorl %ebp, %esi - movl 24(%esp), %ebp - xorl %ebp, %esi - movl 44(%esp), %ebp - xorl %ebp, %esi movl %ecx, %ebp -.byte 209 -.byte 198 # roll $1 %esi + xorl (%esp), %esi + rorl $2, %ecx + xorl 24(%esp), %esi xorl %ebx, %ebp - movl %esi, 56(%esp) + xorl 44(%esp), %esi xorl %edx, %ebp +.byte 209 +.byte 198 + movl %esi, 56(%esp) leal 3395469782(%esi,%edi,1),%esi movl %eax, %edi roll $5, %edi -.byte 209 -.byte 201 # rorl $1 %ecx - addl %ebp, %edi -.byte 209 -.byte 201 # rorl $1 %ecx + addl %ebp, %esi addl %edi, %esi - # 20_39 79 + movl 60(%esp), %edi - movl 4(%esp), %ebp - xorl %ebp, %edi - movl 28(%esp), %ebp - xorl %ebp, %edi - movl 48(%esp), %ebp - xorl %ebp, %edi movl %eax, %ebp -.byte 209 -.byte 199 # roll $1 %edi + xorl 4(%esp), %edi + rorl $2, %eax + xorl 28(%esp), %edi xorl %ecx, %ebp - movl %edi, 60(%esp) + xorl 48(%esp), %edi xorl %ebx, %ebp +.byte 209 +.byte 199 + movl %edi, 60(%esp) leal 3395469782(%edi,%edx,1),%edi movl %esi, %edx roll $5, %edx - addl %ebp, %edx - movl 128(%esp), %ebp -.byte 209 -.byte 200 # rorl $1 %eax + addl %ebp, %edi addl %edx, %edi -.byte 209 -.byte 200 # rorl $1 %eax - # End processing + + movl 128(%esp), %ebp movl 12(%ebp), %edx addl %ecx, %edx movl 4(%ebp), %ecx @@ -1921,7 +1498,7 @@ sha1_block_asm_host_order: movl 16(%ebp), %edi movl 8(%ebp), %ebx movl %ecx, 68(%esp) - # First we need to setup the X array + movl (%esi), %eax movl 4(%esi), %ecx movl %eax, (%esp) |