freebsd-dev/secure/lib/libcrypto/i386/co-586.s
Mark Murray f3d90904b0 Re-add the hand-optimised assembler versions of some of the ciphers
to the build.

Should have done this ages ago:	markm
Reminded above to do this:	peter
2004-03-23 08:32:29 +00:00

1271 lines
18 KiB
ArmAsm

# $FreeBSD$
.file "/usr/src/secure/lib/libcrypto/../../../crypto/openssl/crypto/bn/asm/co-586.s"
.version "01.01"
gcc2_compiled.:
.text
.align 16
.globl bn_mul_comba8
.type bn_mul_comba8,@function
bn_mul_comba8:
pushl %esi
movl 12(%esp), %esi
pushl %edi
movl 20(%esp), %edi
pushl %ebp
pushl %ebx
xorl %ebx, %ebx
movl (%esi), %eax
xorl %ecx, %ecx
movl (%edi), %edx
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
movl (%edi), %edx
adcl $0, %ebp
movl %ebx, (%eax)
movl 4(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %ecx
movl (%esi), %eax
adcl %edx, %ebp
movl 4(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esp), %eax
adcl %edx, %ebp
movl (%edi), %edx
adcl $0, %ebx
movl %ecx, 4(%eax)
movl 8(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %ebp
movl 4(%esi), %eax
adcl %edx, %ebx
movl 4(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl (%esi), %eax
adcl %edx, %ebx
movl 8(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esp), %eax
adcl %edx, %ebx
movl (%edi), %edx
adcl $0, %ecx
movl %ebp, 8(%eax)
movl 12(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 8(%esi), %eax
adcl %edx, %ecx
movl 4(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 4(%esi), %eax
adcl %edx, %ecx
movl 8(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl (%esi), %eax
adcl %edx, %ecx
movl 12(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
movl (%edi), %edx
adcl $0, %ebp
movl %ebx, 12(%eax)
movl 16(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %ecx
movl 12(%esi), %eax
adcl %edx, %ebp
movl 4(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 8(%esi), %eax
adcl %edx, %ebp
movl 8(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 4(%esi), %eax
adcl %edx, %ebp
movl 12(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl (%esi), %eax
adcl %edx, %ebp
movl 16(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esp), %eax
adcl %edx, %ebp
movl (%edi), %edx
adcl $0, %ebx
movl %ecx, 16(%eax)
movl 20(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %ebp
movl 16(%esi), %eax
adcl %edx, %ebx
movl 4(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 12(%esi), %eax
adcl %edx, %ebx
movl 8(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 8(%esi), %eax
adcl %edx, %ebx
movl 12(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 4(%esi), %eax
adcl %edx, %ebx
movl 16(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl (%esi), %eax
adcl %edx, %ebx
movl 20(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esp), %eax
adcl %edx, %ebx
movl (%edi), %edx
adcl $0, %ecx
movl %ebp, 20(%eax)
movl 24(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esi), %eax
adcl %edx, %ecx
movl 4(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 16(%esi), %eax
adcl %edx, %ecx
movl 8(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 12(%esi), %eax
adcl %edx, %ecx
movl 12(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 8(%esi), %eax
adcl %edx, %ecx
movl 16(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 4(%esi), %eax
adcl %edx, %ecx
movl 20(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl (%esi), %eax
adcl %edx, %ecx
movl 24(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
movl (%edi), %edx
adcl $0, %ebp
movl %ebx, 24(%eax)
movl 28(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %ecx
movl 24(%esi), %eax
adcl %edx, %ebp
movl 4(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esi), %eax
adcl %edx, %ebp
movl 8(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 16(%esi), %eax
adcl %edx, %ebp
movl 12(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 12(%esi), %eax
adcl %edx, %ebp
movl 16(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 8(%esi), %eax
adcl %edx, %ebp
movl 20(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 4(%esi), %eax
adcl %edx, %ebp
movl 24(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl (%esi), %eax
adcl %edx, %ebp
movl 28(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esp), %eax
adcl %edx, %ebp
movl 4(%edi), %edx
adcl $0, %ebx
movl %ecx, 28(%eax)
movl 28(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %ebp
movl 24(%esi), %eax
adcl %edx, %ebx
movl 8(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esi), %eax
adcl %edx, %ebx
movl 12(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 16(%esi), %eax
adcl %edx, %ebx
movl 16(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 12(%esi), %eax
adcl %edx, %ebx
movl 20(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 8(%esi), %eax
adcl %edx, %ebx
movl 24(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 4(%esi), %eax
adcl %edx, %ebx
movl 28(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esp), %eax
adcl %edx, %ebx
movl 8(%edi), %edx
adcl $0, %ecx
movl %ebp, 32(%eax)
movl 28(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 24(%esi), %eax
adcl %edx, %ecx
movl 12(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esi), %eax
adcl %edx, %ecx
movl 16(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 16(%esi), %eax
adcl %edx, %ecx
movl 20(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 12(%esi), %eax
adcl %edx, %ecx
movl 24(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 8(%esi), %eax
adcl %edx, %ecx
movl 28(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
movl 12(%edi), %edx
adcl $0, %ebp
movl %ebx, 36(%eax)
movl 28(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %ecx
movl 24(%esi), %eax
adcl %edx, %ebp
movl 16(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esi), %eax
adcl %edx, %ebp
movl 20(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 16(%esi), %eax
adcl %edx, %ebp
movl 24(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 12(%esi), %eax
adcl %edx, %ebp
movl 28(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esp), %eax
adcl %edx, %ebp
movl 16(%edi), %edx
adcl $0, %ebx
movl %ecx, 40(%eax)
movl 28(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %ebp
movl 24(%esi), %eax
adcl %edx, %ebx
movl 20(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esi), %eax
adcl %edx, %ebx
movl 24(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 16(%esi), %eax
adcl %edx, %ebx
movl 28(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esp), %eax
adcl %edx, %ebx
movl 20(%edi), %edx
adcl $0, %ecx
movl %ebp, 44(%eax)
movl 28(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 24(%esi), %eax
adcl %edx, %ecx
movl 24(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esi), %eax
adcl %edx, %ecx
movl 28(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
movl 24(%edi), %edx
adcl $0, %ebp
movl %ebx, 48(%eax)
movl 28(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %ecx
movl 24(%esi), %eax
adcl %edx, %ebp
movl 28(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esp), %eax
adcl %edx, %ebp
movl 28(%edi), %edx
adcl $0, %ebx
movl %ecx, 52(%eax)
movl 28(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esp), %eax
adcl %edx, %ebx
adcl $0, %ecx
movl %ebp, 56(%eax)
movl %ebx, 60(%eax)
popl %ebx
popl %ebp
popl %edi
popl %esi
ret
.L_bn_mul_comba8_end:
.size bn_mul_comba8,.L_bn_mul_comba8_end-bn_mul_comba8
.ident "desasm.pl"
.text
.align 16
.globl bn_mul_comba4
.type bn_mul_comba4,@function
bn_mul_comba4:
pushl %esi
movl 12(%esp), %esi
pushl %edi
movl 20(%esp), %edi
pushl %ebp
pushl %ebx
xorl %ebx, %ebx
movl (%esi), %eax
xorl %ecx, %ecx
movl (%edi), %edx
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
movl (%edi), %edx
adcl $0, %ebp
movl %ebx, (%eax)
movl 4(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %ecx
movl (%esi), %eax
adcl %edx, %ebp
movl 4(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esp), %eax
adcl %edx, %ebp
movl (%edi), %edx
adcl $0, %ebx
movl %ecx, 4(%eax)
movl 8(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %ebp
movl 4(%esi), %eax
adcl %edx, %ebx
movl 4(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl (%esi), %eax
adcl %edx, %ebx
movl 8(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esp), %eax
adcl %edx, %ebx
movl (%edi), %edx
adcl $0, %ecx
movl %ebp, 8(%eax)
movl 12(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 8(%esi), %eax
adcl %edx, %ecx
movl 4(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 4(%esi), %eax
adcl %edx, %ecx
movl 8(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl (%esi), %eax
adcl %edx, %ecx
movl 12(%edi), %edx
adcl $0, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
movl 4(%edi), %edx
adcl $0, %ebp
movl %ebx, 12(%eax)
movl 12(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %ecx
movl 8(%esi), %eax
adcl %edx, %ebp
movl 8(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 4(%esi), %eax
adcl %edx, %ebp
movl 12(%edi), %edx
adcl $0, %ebx
mull %edx
addl %eax, %ecx
movl 20(%esp), %eax
adcl %edx, %ebp
movl 8(%edi), %edx
adcl $0, %ebx
movl %ecx, 16(%eax)
movl 12(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %ebp
movl 8(%esi), %eax
adcl %edx, %ebx
movl 12(%edi), %edx
adcl $0, %ecx
mull %edx
addl %eax, %ebp
movl 20(%esp), %eax
adcl %edx, %ebx
movl 12(%edi), %edx
adcl $0, %ecx
movl %ebp, 20(%eax)
movl 12(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %ebx
movl 20(%esp), %eax
adcl %edx, %ecx
adcl $0, %ebp
movl %ebx, 24(%eax)
movl %ecx, 28(%eax)
popl %ebx
popl %ebp
popl %edi
popl %esi
ret
.L_bn_mul_comba4_end:
.size bn_mul_comba4,.L_bn_mul_comba4_end-bn_mul_comba4
.ident "desasm.pl"
.text
.align 16
.globl bn_sqr_comba8
.type bn_sqr_comba8,@function
bn_sqr_comba8:
pushl %esi
pushl %edi
pushl %ebp
pushl %ebx
movl 20(%esp), %edi
movl 24(%esp), %esi
xorl %ebx, %ebx
xorl %ecx, %ecx
movl (%esi), %eax
xorl %ebp, %ebp
mull %eax
addl %eax, %ebx
adcl %edx, %ecx
movl (%esi), %edx
adcl $0, %ebp
movl %ebx, (%edi)
movl 4(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 8(%esi), %eax
adcl $0, %ebx
movl %ecx, 4(%edi)
movl (%esi), %edx
xorl %ecx, %ecx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 4(%esi), %eax
adcl $0, %ecx
mull %eax
addl %eax, %ebp
adcl %edx, %ebx
movl (%esi), %edx
adcl $0, %ecx
movl %ebp, 8(%edi)
movl 12(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 8(%esi), %eax
adcl $0, %ebp
movl 4(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 16(%esi), %eax
adcl $0, %ebp
movl %ebx, 12(%edi)
movl (%esi), %edx
xorl %ebx, %ebx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 12(%esi), %eax
adcl $0, %ebx
movl 4(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 8(%esi), %eax
adcl $0, %ebx
mull %eax
addl %eax, %ecx
adcl %edx, %ebp
movl (%esi), %edx
adcl $0, %ebx
movl %ecx, 16(%edi)
movl 20(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 16(%esi), %eax
adcl $0, %ecx
movl 4(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 12(%esi), %eax
adcl $0, %ecx
movl 8(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 24(%esi), %eax
adcl $0, %ecx
movl %ebp, 20(%edi)
movl (%esi), %edx
xorl %ebp, %ebp
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 20(%esi), %eax
adcl $0, %ebp
movl 4(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 16(%esi), %eax
adcl $0, %ebp
movl 8(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 12(%esi), %eax
adcl $0, %ebp
mull %eax
addl %eax, %ebx
adcl %edx, %ecx
movl (%esi), %edx
adcl $0, %ebp
movl %ebx, 24(%edi)
movl 28(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 24(%esi), %eax
adcl $0, %ebx
movl 4(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 20(%esi), %eax
adcl $0, %ebx
movl 8(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 16(%esi), %eax
adcl $0, %ebx
movl 12(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 28(%esi), %eax
adcl $0, %ebx
movl %ecx, 28(%edi)
movl 4(%esi), %edx
xorl %ecx, %ecx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 24(%esi), %eax
adcl $0, %ecx
movl 8(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 20(%esi), %eax
adcl $0, %ecx
movl 12(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 16(%esi), %eax
adcl $0, %ecx
mull %eax
addl %eax, %ebp
adcl %edx, %ebx
movl 8(%esi), %edx
adcl $0, %ecx
movl %ebp, 32(%edi)
movl 28(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 24(%esi), %eax
adcl $0, %ebp
movl 12(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 20(%esi), %eax
adcl $0, %ebp
movl 16(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 28(%esi), %eax
adcl $0, %ebp
movl %ebx, 36(%edi)
movl 12(%esi), %edx
xorl %ebx, %ebx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 24(%esi), %eax
adcl $0, %ebx
movl 16(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 20(%esi), %eax
adcl $0, %ebx
mull %eax
addl %eax, %ecx
adcl %edx, %ebp
movl 16(%esi), %edx
adcl $0, %ebx
movl %ecx, 40(%edi)
movl 28(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 24(%esi), %eax
adcl $0, %ecx
movl 20(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 28(%esi), %eax
adcl $0, %ecx
movl %ebp, 44(%edi)
movl 20(%esi), %edx
xorl %ebp, %ebp
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 24(%esi), %eax
adcl $0, %ebp
mull %eax
addl %eax, %ebx
adcl %edx, %ecx
movl 24(%esi), %edx
adcl $0, %ebp
movl %ebx, 48(%edi)
movl 28(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 28(%esi), %eax
adcl $0, %ebx
movl %ecx, 52(%edi)
xorl %ecx, %ecx
mull %eax
addl %eax, %ebp
adcl %edx, %ebx
adcl $0, %ecx
movl %ebp, 56(%edi)
movl %ebx, 60(%edi)
popl %ebx
popl %ebp
popl %edi
popl %esi
ret
.L_bn_sqr_comba8_end:
.size bn_sqr_comba8,.L_bn_sqr_comba8_end-bn_sqr_comba8
.ident "desasm.pl"
.text
.align 16
.globl bn_sqr_comba4
.type bn_sqr_comba4,@function
bn_sqr_comba4:
pushl %esi
pushl %edi
pushl %ebp
pushl %ebx
movl 20(%esp), %edi
movl 24(%esp), %esi
xorl %ebx, %ebx
xorl %ecx, %ecx
movl (%esi), %eax
xorl %ebp, %ebp
mull %eax
addl %eax, %ebx
adcl %edx, %ecx
movl (%esi), %edx
adcl $0, %ebp
movl %ebx, (%edi)
movl 4(%esi), %eax
xorl %ebx, %ebx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 8(%esi), %eax
adcl $0, %ebx
movl %ecx, 4(%edi)
movl (%esi), %edx
xorl %ecx, %ecx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 4(%esi), %eax
adcl $0, %ecx
mull %eax
addl %eax, %ebp
adcl %edx, %ebx
movl (%esi), %edx
adcl $0, %ecx
movl %ebp, 8(%edi)
movl 12(%esi), %eax
xorl %ebp, %ebp
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 8(%esi), %eax
adcl $0, %ebp
movl 4(%esi), %edx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebp
addl %eax, %ebx
adcl %edx, %ecx
movl 12(%esi), %eax
adcl $0, %ebp
movl %ebx, 12(%edi)
movl 4(%esi), %edx
xorl %ebx, %ebx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ebx
addl %eax, %ecx
adcl %edx, %ebp
movl 8(%esi), %eax
adcl $0, %ebx
mull %eax
addl %eax, %ecx
adcl %edx, %ebp
movl 8(%esi), %edx
adcl $0, %ebx
movl %ecx, 16(%edi)
movl 12(%esi), %eax
xorl %ecx, %ecx
mull %edx
addl %eax, %eax
adcl %edx, %edx
adcl $0, %ecx
addl %eax, %ebp
adcl %edx, %ebx
movl 12(%esi), %eax
adcl $0, %ecx
movl %ebp, 20(%edi)
xorl %ebp, %ebp
mull %eax
addl %eax, %ebx
adcl %edx, %ecx
adcl $0, %ebp
movl %ebx, 24(%edi)
movl %ecx, 28(%edi)
popl %ebx
popl %ebp
popl %edi
popl %esi
ret
.L_bn_sqr_comba4_end:
.size bn_sqr_comba4,.L_bn_sqr_comba4_end-bn_sqr_comba4
.ident "desasm.pl"