freebsd-dev/secure/lib/libcrypto/amd64/x86_64-mont.S
Jung-uk Kim 1f13597d10 Merge OpenSSL 1.0.1c.
Approved by:	benl (maintainer)
2012-07-12 19:30:53 +00:00

1376 lines
21 KiB
ArmAsm

# $FreeBSD$
.text
.globl bn_mul_mont
.type bn_mul_mont,@function
.align 16
bn_mul_mont:
testl $3,%r9d
jnz .Lmul_enter
cmpl $8,%r9d
jb .Lmul_enter
cmpq %rsi,%rdx
jne .Lmul4x_enter
jmp .Lsqr4x_enter
.align 16
.Lmul_enter:
pushq %rbx
pushq %rbp
pushq %r12
pushq %r13
pushq %r14
pushq %r15
movl %r9d,%r9d
leaq 2(%r9),%r10
movq %rsp,%r11
negq %r10
leaq (%rsp,%r10,8),%rsp
andq $-1024,%rsp
movq %r11,8(%rsp,%r9,8)
.Lmul_body:
movq %rdx,%r12
movq (%r8),%r8
movq (%r12),%rbx
movq (%rsi),%rax
xorq %r14,%r14
xorq %r15,%r15
movq %r8,%rbp
mulq %rbx
movq %rax,%r10
movq (%rcx),%rax
imulq %r10,%rbp
movq %rdx,%r11
mulq %rbp
addq %rax,%r10
movq 8(%rsi),%rax
adcq $0,%rdx
movq %rdx,%r13
leaq 1(%r15),%r15
jmp .L1st_enter
.align 16
.L1st:
addq %rax,%r13
movq (%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r11,%r13
movq %r10,%r11
adcq $0,%rdx
movq %r13,-16(%rsp,%r15,8)
movq %rdx,%r13
.L1st_enter:
mulq %rbx
addq %rax,%r11
movq (%rcx,%r15,8),%rax
adcq $0,%rdx
leaq 1(%r15),%r15
movq %rdx,%r10
mulq %rbp
cmpq %r9,%r15
jne .L1st
addq %rax,%r13
movq (%rsi),%rax
adcq $0,%rdx
addq %r11,%r13
adcq $0,%rdx
movq %r13,-16(%rsp,%r15,8)
movq %rdx,%r13
movq %r10,%r11
xorq %rdx,%rdx
addq %r11,%r13
adcq $0,%rdx
movq %r13,-8(%rsp,%r9,8)
movq %rdx,(%rsp,%r9,8)
leaq 1(%r14),%r14
jmp .Louter
.align 16
.Louter:
movq (%r12,%r14,8),%rbx
xorq %r15,%r15
movq %r8,%rbp
movq (%rsp),%r10
mulq %rbx
addq %rax,%r10
movq (%rcx),%rax
adcq $0,%rdx
imulq %r10,%rbp
movq %rdx,%r11
mulq %rbp
addq %rax,%r10
movq 8(%rsi),%rax
adcq $0,%rdx
movq 8(%rsp),%r10
movq %rdx,%r13
leaq 1(%r15),%r15
jmp .Linner_enter
.align 16
.Linner:
addq %rax,%r13
movq (%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r10,%r13
movq (%rsp,%r15,8),%r10
adcq $0,%rdx
movq %r13,-16(%rsp,%r15,8)
movq %rdx,%r13
.Linner_enter:
mulq %rbx
addq %rax,%r11
movq (%rcx,%r15,8),%rax
adcq $0,%rdx
addq %r11,%r10
movq %rdx,%r11
adcq $0,%r11
leaq 1(%r15),%r15
mulq %rbp
cmpq %r9,%r15
jne .Linner
addq %rax,%r13
movq (%rsi),%rax
adcq $0,%rdx
addq %r10,%r13
movq (%rsp,%r15,8),%r10
adcq $0,%rdx
movq %r13,-16(%rsp,%r15,8)
movq %rdx,%r13
xorq %rdx,%rdx
addq %r11,%r13
adcq $0,%rdx
addq %r10,%r13
adcq $0,%rdx
movq %r13,-8(%rsp,%r9,8)
movq %rdx,(%rsp,%r9,8)
leaq 1(%r14),%r14
cmpq %r9,%r14
jl .Louter
xorq %r14,%r14
movq (%rsp),%rax
leaq (%rsp),%rsi
movq %r9,%r15
jmp .Lsub
.align 16
.Lsub: sbbq (%rcx,%r14,8),%rax
movq %rax,(%rdi,%r14,8)
movq 8(%rsi,%r14,8),%rax
leaq 1(%r14),%r14
decq %r15
jnz .Lsub
sbbq $0,%rax
xorq %r14,%r14
andq %rax,%rsi
notq %rax
movq %rdi,%rcx
andq %rax,%rcx
movq %r9,%r15
orq %rcx,%rsi
.align 16
.Lcopy:
movq (%rsi,%r14,8),%rax
movq %r14,(%rsp,%r14,8)
movq %rax,(%rdi,%r14,8)
leaq 1(%r14),%r14
subq $1,%r15
jnz .Lcopy
movq 8(%rsp,%r9,8),%rsi
movq $1,%rax
movq (%rsi),%r15
movq 8(%rsi),%r14
movq 16(%rsi),%r13
movq 24(%rsi),%r12
movq 32(%rsi),%rbp
movq 40(%rsi),%rbx
leaq 48(%rsi),%rsp
.Lmul_epilogue:
.byte 0xf3,0xc3
.size bn_mul_mont,.-bn_mul_mont
.type bn_mul4x_mont,@function
.align 16
bn_mul4x_mont:
.Lmul4x_enter:
pushq %rbx
pushq %rbp
pushq %r12
pushq %r13
pushq %r14
pushq %r15
movl %r9d,%r9d
leaq 4(%r9),%r10
movq %rsp,%r11
negq %r10
leaq (%rsp,%r10,8),%rsp
andq $-1024,%rsp
movq %r11,8(%rsp,%r9,8)
.Lmul4x_body:
movq %rdi,16(%rsp,%r9,8)
movq %rdx,%r12
movq (%r8),%r8
movq (%r12),%rbx
movq (%rsi),%rax
xorq %r14,%r14
xorq %r15,%r15
movq %r8,%rbp
mulq %rbx
movq %rax,%r10
movq (%rcx),%rax
imulq %r10,%rbp
movq %rdx,%r11
mulq %rbp
addq %rax,%r10
movq 8(%rsi),%rax
adcq $0,%rdx
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq 8(%rcx),%rax
adcq $0,%rdx
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq 16(%rsi),%rax
adcq $0,%rdx
addq %r11,%rdi
leaq 4(%r15),%r15
adcq $0,%rdx
movq %rdi,(%rsp)
movq %rdx,%r13
jmp .L1st4x
.align 16
.L1st4x:
mulq %rbx
addq %rax,%r10
movq -16(%rcx,%r15,8),%rax
adcq $0,%rdx
movq %rdx,%r11
mulq %rbp
addq %rax,%r13
movq -8(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r10,%r13
adcq $0,%rdx
movq %r13,-24(%rsp,%r15,8)
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq -8(%rcx,%r15,8),%rax
adcq $0,%rdx
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq (%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r11,%rdi
adcq $0,%rdx
movq %rdi,-16(%rsp,%r15,8)
movq %rdx,%r13
mulq %rbx
addq %rax,%r10
movq (%rcx,%r15,8),%rax
adcq $0,%rdx
movq %rdx,%r11
mulq %rbp
addq %rax,%r13
movq 8(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r10,%r13
adcq $0,%rdx
movq %r13,-8(%rsp,%r15,8)
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq 8(%rcx,%r15,8),%rax
adcq $0,%rdx
leaq 4(%r15),%r15
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq -16(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r11,%rdi
adcq $0,%rdx
movq %rdi,-32(%rsp,%r15,8)
movq %rdx,%r13
cmpq %r9,%r15
jl .L1st4x
mulq %rbx
addq %rax,%r10
movq -16(%rcx,%r15,8),%rax
adcq $0,%rdx
movq %rdx,%r11
mulq %rbp
addq %rax,%r13
movq -8(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r10,%r13
adcq $0,%rdx
movq %r13,-24(%rsp,%r15,8)
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq -8(%rcx,%r15,8),%rax
adcq $0,%rdx
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq (%rsi),%rax
adcq $0,%rdx
addq %r11,%rdi
adcq $0,%rdx
movq %rdi,-16(%rsp,%r15,8)
movq %rdx,%r13
xorq %rdi,%rdi
addq %r10,%r13
adcq $0,%rdi
movq %r13,-8(%rsp,%r15,8)
movq %rdi,(%rsp,%r15,8)
leaq 1(%r14),%r14
.align 4
.Louter4x:
movq (%r12,%r14,8),%rbx
xorq %r15,%r15
movq (%rsp),%r10
movq %r8,%rbp
mulq %rbx
addq %rax,%r10
movq (%rcx),%rax
adcq $0,%rdx
imulq %r10,%rbp
movq %rdx,%r11
mulq %rbp
addq %rax,%r10
movq 8(%rsi),%rax
adcq $0,%rdx
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq 8(%rcx),%rax
adcq $0,%rdx
addq 8(%rsp),%r11
adcq $0,%rdx
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq 16(%rsi),%rax
adcq $0,%rdx
addq %r11,%rdi
leaq 4(%r15),%r15
adcq $0,%rdx
movq %rdi,(%rsp)
movq %rdx,%r13
jmp .Linner4x
.align 16
.Linner4x:
mulq %rbx
addq %rax,%r10
movq -16(%rcx,%r15,8),%rax
adcq $0,%rdx
addq -16(%rsp,%r15,8),%r10
adcq $0,%rdx
movq %rdx,%r11
mulq %rbp
addq %rax,%r13
movq -8(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r10,%r13
adcq $0,%rdx
movq %r13,-24(%rsp,%r15,8)
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq -8(%rcx,%r15,8),%rax
adcq $0,%rdx
addq -8(%rsp,%r15,8),%r11
adcq $0,%rdx
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq (%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r11,%rdi
adcq $0,%rdx
movq %rdi,-16(%rsp,%r15,8)
movq %rdx,%r13
mulq %rbx
addq %rax,%r10
movq (%rcx,%r15,8),%rax
adcq $0,%rdx
addq (%rsp,%r15,8),%r10
adcq $0,%rdx
movq %rdx,%r11
mulq %rbp
addq %rax,%r13
movq 8(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r10,%r13
adcq $0,%rdx
movq %r13,-8(%rsp,%r15,8)
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq 8(%rcx,%r15,8),%rax
adcq $0,%rdx
addq 8(%rsp,%r15,8),%r11
adcq $0,%rdx
leaq 4(%r15),%r15
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq -16(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r11,%rdi
adcq $0,%rdx
movq %rdi,-32(%rsp,%r15,8)
movq %rdx,%r13
cmpq %r9,%r15
jl .Linner4x
mulq %rbx
addq %rax,%r10
movq -16(%rcx,%r15,8),%rax
adcq $0,%rdx
addq -16(%rsp,%r15,8),%r10
adcq $0,%rdx
movq %rdx,%r11
mulq %rbp
addq %rax,%r13
movq -8(%rsi,%r15,8),%rax
adcq $0,%rdx
addq %r10,%r13
adcq $0,%rdx
movq %r13,-24(%rsp,%r15,8)
movq %rdx,%rdi
mulq %rbx
addq %rax,%r11
movq -8(%rcx,%r15,8),%rax
adcq $0,%rdx
addq -8(%rsp,%r15,8),%r11
adcq $0,%rdx
leaq 1(%r14),%r14
movq %rdx,%r10
mulq %rbp
addq %rax,%rdi
movq (%rsi),%rax
adcq $0,%rdx
addq %r11,%rdi
adcq $0,%rdx
movq %rdi,-16(%rsp,%r15,8)
movq %rdx,%r13
xorq %rdi,%rdi
addq %r10,%r13
adcq $0,%rdi
addq (%rsp,%r9,8),%r13
adcq $0,%rdi
movq %r13,-8(%rsp,%r15,8)
movq %rdi,(%rsp,%r15,8)
cmpq %r9,%r14
jl .Louter4x
movq 16(%rsp,%r9,8),%rdi
movq 0(%rsp),%rax
pxor %xmm0,%xmm0
movq 8(%rsp),%rdx
shrq $2,%r9
leaq (%rsp),%rsi
xorq %r14,%r14
subq 0(%rcx),%rax
movq 16(%rsi),%rbx
movq 24(%rsi),%rbp
sbbq 8(%rcx),%rdx
leaq -1(%r9),%r15
jmp .Lsub4x
.align 16
.Lsub4x:
movq %rax,0(%rdi,%r14,8)
movq %rdx,8(%rdi,%r14,8)
sbbq 16(%rcx,%r14,8),%rbx
movq 32(%rsi,%r14,8),%rax
movq 40(%rsi,%r14,8),%rdx
sbbq 24(%rcx,%r14,8),%rbp
movq %rbx,16(%rdi,%r14,8)
movq %rbp,24(%rdi,%r14,8)
sbbq 32(%rcx,%r14,8),%rax
movq 48(%rsi,%r14,8),%rbx
movq 56(%rsi,%r14,8),%rbp
sbbq 40(%rcx,%r14,8),%rdx
leaq 4(%r14),%r14
decq %r15
jnz .Lsub4x
movq %rax,0(%rdi,%r14,8)
movq 32(%rsi,%r14,8),%rax
sbbq 16(%rcx,%r14,8),%rbx
movq %rdx,8(%rdi,%r14,8)
sbbq 24(%rcx,%r14,8),%rbp
movq %rbx,16(%rdi,%r14,8)
sbbq $0,%rax
movq %rbp,24(%rdi,%r14,8)
xorq %r14,%r14
andq %rax,%rsi
notq %rax
movq %rdi,%rcx
andq %rax,%rcx
leaq -1(%r9),%r15
orq %rcx,%rsi
movdqu (%rsi),%xmm1
movdqa %xmm0,(%rsp)
movdqu %xmm1,(%rdi)
jmp .Lcopy4x
.align 16
.Lcopy4x:
movdqu 16(%rsi,%r14,1),%xmm2
movdqu 32(%rsi,%r14,1),%xmm1
movdqa %xmm0,16(%rsp,%r14,1)
movdqu %xmm2,16(%rdi,%r14,1)
movdqa %xmm0,32(%rsp,%r14,1)
movdqu %xmm1,32(%rdi,%r14,1)
leaq 32(%r14),%r14
decq %r15
jnz .Lcopy4x
shlq $2,%r9
movdqu 16(%rsi,%r14,1),%xmm2
movdqa %xmm0,16(%rsp,%r14,1)
movdqu %xmm2,16(%rdi,%r14,1)
movq 8(%rsp,%r9,8),%rsi
movq $1,%rax
movq (%rsi),%r15
movq 8(%rsi),%r14
movq 16(%rsi),%r13
movq 24(%rsi),%r12
movq 32(%rsi),%rbp
movq 40(%rsi),%rbx
leaq 48(%rsi),%rsp
.Lmul4x_epilogue:
.byte 0xf3,0xc3
.size bn_mul4x_mont,.-bn_mul4x_mont
.type bn_sqr4x_mont,@function
.align 16
bn_sqr4x_mont:
.Lsqr4x_enter:
pushq %rbx
pushq %rbp
pushq %r12
pushq %r13
pushq %r14
pushq %r15
shll $3,%r9d
xorq %r10,%r10
movq %rsp,%r11
subq %r9,%r10
movq (%r8),%r8
leaq -72(%rsp,%r10,2),%rsp
andq $-1024,%rsp
movq %rdi,32(%rsp)
movq %rcx,40(%rsp)
movq %r8,48(%rsp)
movq %r11,56(%rsp)
.Lsqr4x_body:
leaq 32(%r10),%rbp
leaq (%rsi,%r9,1),%rsi
movq %r9,%rcx
movq -32(%rsi,%rbp,1),%r14
leaq 64(%rsp,%r9,2),%rdi
movq -24(%rsi,%rbp,1),%rax
leaq -32(%rdi,%rbp,1),%rdi
movq -16(%rsi,%rbp,1),%rbx
movq %rax,%r15
mulq %r14
movq %rax,%r10
movq %rbx,%rax
movq %rdx,%r11
movq %r10,-24(%rdi,%rbp,1)
xorq %r10,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
movq %r11,-16(%rdi,%rbp,1)
leaq -16(%rbp),%rcx
movq 8(%rsi,%rcx,1),%rbx
mulq %r15
movq %rax,%r12
movq %rbx,%rax
movq %rdx,%r13
xorq %r11,%r11
addq %r12,%r10
leaq 16(%rcx),%rcx
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,-8(%rdi,%rcx,1)
jmp .Lsqr4x_1st
.align 16
.Lsqr4x_1st:
movq (%rsi,%rcx,1),%rbx
xorq %r12,%r12
mulq %r15
addq %rax,%r13
movq %rbx,%rax
adcq %rdx,%r12
xorq %r10,%r10
addq %r13,%r11
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
movq %r11,(%rdi,%rcx,1)
movq 8(%rsi,%rcx,1),%rbx
xorq %r13,%r13
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq %rdx,%r13
xorq %r11,%r11
addq %r12,%r10
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,8(%rdi,%rcx,1)
movq 16(%rsi,%rcx,1),%rbx
xorq %r12,%r12
mulq %r15
addq %rax,%r13
movq %rbx,%rax
adcq %rdx,%r12
xorq %r10,%r10
addq %r13,%r11
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
movq %r11,16(%rdi,%rcx,1)
movq 24(%rsi,%rcx,1),%rbx
xorq %r13,%r13
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq %rdx,%r13
xorq %r11,%r11
addq %r12,%r10
leaq 32(%rcx),%rcx
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,-8(%rdi,%rcx,1)
cmpq $0,%rcx
jne .Lsqr4x_1st
xorq %r12,%r12
addq %r11,%r13
adcq $0,%r12
mulq %r15
addq %rax,%r13
adcq %rdx,%r12
movq %r13,(%rdi)
leaq 16(%rbp),%rbp
movq %r12,8(%rdi)
jmp .Lsqr4x_outer
.align 16
.Lsqr4x_outer:
movq -32(%rsi,%rbp,1),%r14
leaq 64(%rsp,%r9,2),%rdi
movq -24(%rsi,%rbp,1),%rax
leaq -32(%rdi,%rbp,1),%rdi
movq -16(%rsi,%rbp,1),%rbx
movq %rax,%r15
movq -24(%rdi,%rbp,1),%r10
xorq %r11,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,-24(%rdi,%rbp,1)
xorq %r10,%r10
addq -16(%rdi,%rbp,1),%r11
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
movq %r11,-16(%rdi,%rbp,1)
leaq -16(%rbp),%rcx
xorq %r12,%r12
movq 8(%rsi,%rcx,1),%rbx
xorq %r13,%r13
addq 8(%rdi,%rcx,1),%r12
adcq $0,%r13
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq %rdx,%r13
xorq %r11,%r11
addq %r12,%r10
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,8(%rdi,%rcx,1)
leaq 16(%rcx),%rcx
jmp .Lsqr4x_inner
.align 16
.Lsqr4x_inner:
movq (%rsi,%rcx,1),%rbx
xorq %r12,%r12
addq (%rdi,%rcx,1),%r13
adcq $0,%r12
mulq %r15
addq %rax,%r13
movq %rbx,%rax
adcq %rdx,%r12
xorq %r10,%r10
addq %r13,%r11
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
movq %r11,(%rdi,%rcx,1)
movq 8(%rsi,%rcx,1),%rbx
xorq %r13,%r13
addq 8(%rdi,%rcx,1),%r12
adcq $0,%r13
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq %rdx,%r13
xorq %r11,%r11
addq %r12,%r10
leaq 16(%rcx),%rcx
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,-8(%rdi,%rcx,1)
cmpq $0,%rcx
jne .Lsqr4x_inner
xorq %r12,%r12
addq %r11,%r13
adcq $0,%r12
mulq %r15
addq %rax,%r13
adcq %rdx,%r12
movq %r13,(%rdi)
movq %r12,8(%rdi)
addq $16,%rbp
jnz .Lsqr4x_outer
movq -32(%rsi),%r14
leaq 64(%rsp,%r9,2),%rdi
movq -24(%rsi),%rax
leaq -32(%rdi,%rbp,1),%rdi
movq -16(%rsi),%rbx
movq %rax,%r15
xorq %r11,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,-24(%rdi)
xorq %r10,%r10
addq %r13,%r11
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
movq %r11,-16(%rdi)
movq -8(%rsi),%rbx
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq $0,%rdx
xorq %r11,%r11
addq %r12,%r10
movq %rdx,%r13
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %rbx,%rax
adcq %rdx,%r11
movq %r10,-8(%rdi)
xorq %r12,%r12
addq %r11,%r13
adcq $0,%r12
mulq %r15
addq %rax,%r13
movq -16(%rsi),%rax
adcq %rdx,%r12
movq %r13,(%rdi)
movq %r12,8(%rdi)
mulq %rbx
addq $16,%rbp
xorq %r14,%r14
subq %r9,%rbp
xorq %r15,%r15
addq %r12,%rax
adcq $0,%rdx
movq %rax,8(%rdi)
movq %rdx,16(%rdi)
movq %r15,24(%rdi)
movq -16(%rsi,%rbp,1),%rax
leaq 64(%rsp,%r9,2),%rdi
xorq %r10,%r10
movq -24(%rdi,%rbp,2),%r11
leaq (%r14,%r10,2),%r12
shrq $63,%r10
leaq (%rcx,%r11,2),%r13
shrq $63,%r11
orq %r10,%r13
movq -16(%rdi,%rbp,2),%r10
movq %r11,%r14
mulq %rax
negq %r15
movq -8(%rdi,%rbp,2),%r11
adcq %rax,%r12
movq -8(%rsi,%rbp,1),%rax
movq %r12,-32(%rdi,%rbp,2)
adcq %rdx,%r13
leaq (%r14,%r10,2),%rbx
movq %r13,-24(%rdi,%rbp,2)
sbbq %r15,%r15
shrq $63,%r10
leaq (%rcx,%r11,2),%r8
shrq $63,%r11
orq %r10,%r8
movq 0(%rdi,%rbp,2),%r10
movq %r11,%r14
mulq %rax
negq %r15
movq 8(%rdi,%rbp,2),%r11
adcq %rax,%rbx
movq 0(%rsi,%rbp,1),%rax
movq %rbx,-16(%rdi,%rbp,2)
adcq %rdx,%r8
leaq 16(%rbp),%rbp
movq %r8,-40(%rdi,%rbp,2)
sbbq %r15,%r15
jmp .Lsqr4x_shift_n_add
.align 16
.Lsqr4x_shift_n_add:
leaq (%r14,%r10,2),%r12
shrq $63,%r10
leaq (%rcx,%r11,2),%r13
shrq $63,%r11
orq %r10,%r13
movq -16(%rdi,%rbp,2),%r10
movq %r11,%r14
mulq %rax
negq %r15
movq -8(%rdi,%rbp,2),%r11
adcq %rax,%r12
movq -8(%rsi,%rbp,1),%rax
movq %r12,-32(%rdi,%rbp,2)
adcq %rdx,%r13
leaq (%r14,%r10,2),%rbx
movq %r13,-24(%rdi,%rbp,2)
sbbq %r15,%r15
shrq $63,%r10
leaq (%rcx,%r11,2),%r8
shrq $63,%r11
orq %r10,%r8
movq 0(%rdi,%rbp,2),%r10
movq %r11,%r14
mulq %rax
negq %r15
movq 8(%rdi,%rbp,2),%r11
adcq %rax,%rbx
movq 0(%rsi,%rbp,1),%rax
movq %rbx,-16(%rdi,%rbp,2)
adcq %rdx,%r8
leaq (%r14,%r10,2),%r12
movq %r8,-8(%rdi,%rbp,2)
sbbq %r15,%r15
shrq $63,%r10
leaq (%rcx,%r11,2),%r13
shrq $63,%r11
orq %r10,%r13
movq 16(%rdi,%rbp,2),%r10
movq %r11,%r14
mulq %rax
negq %r15
movq 24(%rdi,%rbp,2),%r11
adcq %rax,%r12
movq 8(%rsi,%rbp,1),%rax
movq %r12,0(%rdi,%rbp,2)
adcq %rdx,%r13
leaq (%r14,%r10,2),%rbx
movq %r13,8(%rdi,%rbp,2)
sbbq %r15,%r15
shrq $63,%r10
leaq (%rcx,%r11,2),%r8
shrq $63,%r11
orq %r10,%r8
movq 32(%rdi,%rbp,2),%r10
movq %r11,%r14
mulq %rax
negq %r15
movq 40(%rdi,%rbp,2),%r11
adcq %rax,%rbx
movq 16(%rsi,%rbp,1),%rax
movq %rbx,16(%rdi,%rbp,2)
adcq %rdx,%r8
movq %r8,24(%rdi,%rbp,2)
sbbq %r15,%r15
addq $32,%rbp
jnz .Lsqr4x_shift_n_add
leaq (%r14,%r10,2),%r12
shrq $63,%r10
leaq (%rcx,%r11,2),%r13
shrq $63,%r11
orq %r10,%r13
movq -16(%rdi),%r10
movq %r11,%r14
mulq %rax
negq %r15
movq -8(%rdi),%r11
adcq %rax,%r12
movq -8(%rsi),%rax
movq %r12,-32(%rdi)
adcq %rdx,%r13
leaq (%r14,%r10,2),%rbx
movq %r13,-24(%rdi)
sbbq %r15,%r15
shrq $63,%r10
leaq (%rcx,%r11,2),%r8
shrq $63,%r11
orq %r10,%r8
mulq %rax
negq %r15
adcq %rax,%rbx
adcq %rdx,%r8
movq %rbx,-16(%rdi)
movq %r8,-8(%rdi)
movq 40(%rsp),%rsi
movq 48(%rsp),%r8
xorq %rcx,%rcx
movq %r9,0(%rsp)
subq %r9,%rcx
movq 64(%rsp),%r10
movq %r8,%r14
leaq 64(%rsp,%r9,2),%rax
leaq 64(%rsp,%r9,1),%rdi
movq %rax,8(%rsp)
leaq (%rsi,%r9,1),%rsi
xorq %rbp,%rbp
movq 0(%rsi,%rcx,1),%rax
movq 8(%rsi,%rcx,1),%r9
imulq %r10,%r14
movq %rax,%rbx
jmp .Lsqr4x_mont_outer
.align 16
.Lsqr4x_mont_outer:
xorq %r11,%r11
mulq %r14
addq %rax,%r10
movq %r9,%rax
adcq %rdx,%r11
movq %r8,%r15
xorq %r10,%r10
addq 8(%rdi,%rcx,1),%r11
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
imulq %r11,%r15
movq 16(%rsi,%rcx,1),%rbx
xorq %r13,%r13
addq %r11,%r12
adcq $0,%r13
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq %rdx,%r13
movq %r12,8(%rdi,%rcx,1)
xorq %r11,%r11
addq 16(%rdi,%rcx,1),%r10
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %r9,%rax
adcq %rdx,%r11
movq 24(%rsi,%rcx,1),%r9
xorq %r12,%r12
addq %r10,%r13
adcq $0,%r12
mulq %r15
addq %rax,%r13
movq %r9,%rax
adcq %rdx,%r12
movq %r13,16(%rdi,%rcx,1)
xorq %r10,%r10
addq 24(%rdi,%rcx,1),%r11
leaq 32(%rcx),%rcx
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
jmp .Lsqr4x_mont_inner
.align 16
.Lsqr4x_mont_inner:
movq (%rsi,%rcx,1),%rbx
xorq %r13,%r13
addq %r11,%r12
adcq $0,%r13
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq %rdx,%r13
movq %r12,-8(%rdi,%rcx,1)
xorq %r11,%r11
addq (%rdi,%rcx,1),%r10
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %r9,%rax
adcq %rdx,%r11
movq 8(%rsi,%rcx,1),%r9
xorq %r12,%r12
addq %r10,%r13
adcq $0,%r12
mulq %r15
addq %rax,%r13
movq %r9,%rax
adcq %rdx,%r12
movq %r13,(%rdi,%rcx,1)
xorq %r10,%r10
addq 8(%rdi,%rcx,1),%r11
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
movq 16(%rsi,%rcx,1),%rbx
xorq %r13,%r13
addq %r11,%r12
adcq $0,%r13
mulq %r15
addq %rax,%r12
movq %rbx,%rax
adcq %rdx,%r13
movq %r12,8(%rdi,%rcx,1)
xorq %r11,%r11
addq 16(%rdi,%rcx,1),%r10
adcq $0,%r11
mulq %r14
addq %rax,%r10
movq %r9,%rax
adcq %rdx,%r11
movq 24(%rsi,%rcx,1),%r9
xorq %r12,%r12
addq %r10,%r13
adcq $0,%r12
mulq %r15
addq %rax,%r13
movq %r9,%rax
adcq %rdx,%r12
movq %r13,16(%rdi,%rcx,1)
xorq %r10,%r10
addq 24(%rdi,%rcx,1),%r11
leaq 32(%rcx),%rcx
adcq $0,%r10
mulq %r14
addq %rax,%r11
movq %rbx,%rax
adcq %rdx,%r10
cmpq $0,%rcx
jne .Lsqr4x_mont_inner
subq 0(%rsp),%rcx
movq %r8,%r14
xorq %r13,%r13
addq %r11,%r12
adcq $0,%r13
mulq %r15
addq %rax,%r12
movq %r9,%rax
adcq %rdx,%r13
movq %r12,-8(%rdi)
xorq %r11,%r11
addq (%rdi),%r10
adcq $0,%r11
movq 0(%rsi,%rcx,1),%rbx
addq %rbp,%r10
adcq $0,%r11
imulq 16(%rdi,%rcx,1),%r14
xorq %r12,%r12
movq 8(%rsi,%rcx,1),%r9
addq %r10,%r13
movq 16(%rdi,%rcx,1),%r10
adcq $0,%r12
mulq %r15
addq %rax,%r13
movq %rbx,%rax
adcq %rdx,%r12
movq %r13,(%rdi)
xorq %rbp,%rbp
addq 8(%rdi),%r12
adcq %rbp,%rbp
addq %r11,%r12
leaq 16(%rdi),%rdi
adcq $0,%rbp
movq %r12,-8(%rdi)
cmpq 8(%rsp),%rdi
jb .Lsqr4x_mont_outer
movq 0(%rsp),%r9
movq %rbp,(%rdi)
movq 64(%rsp,%r9,1),%rax
leaq 64(%rsp,%r9,1),%rbx
movq 40(%rsp),%rsi
shrq $5,%r9
movq 8(%rbx),%rdx
xorq %rbp,%rbp
movq 32(%rsp),%rdi
subq 0(%rsi),%rax
movq 16(%rbx),%r10
movq 24(%rbx),%r11
sbbq 8(%rsi),%rdx
leaq -1(%r9),%rcx
jmp .Lsqr4x_sub
.align 16
.Lsqr4x_sub:
movq %rax,0(%rdi,%rbp,8)
movq %rdx,8(%rdi,%rbp,8)
sbbq 16(%rsi,%rbp,8),%r10
movq 32(%rbx,%rbp,8),%rax
movq 40(%rbx,%rbp,8),%rdx
sbbq 24(%rsi,%rbp,8),%r11
movq %r10,16(%rdi,%rbp,8)
movq %r11,24(%rdi,%rbp,8)
sbbq 32(%rsi,%rbp,8),%rax
movq 48(%rbx,%rbp,8),%r10
movq 56(%rbx,%rbp,8),%r11
sbbq 40(%rsi,%rbp,8),%rdx
leaq 4(%rbp),%rbp
decq %rcx
jnz .Lsqr4x_sub
movq %rax,0(%rdi,%rbp,8)
movq 32(%rbx,%rbp,8),%rax
sbbq 16(%rsi,%rbp,8),%r10
movq %rdx,8(%rdi,%rbp,8)
sbbq 24(%rsi,%rbp,8),%r11
movq %r10,16(%rdi,%rbp,8)
sbbq $0,%rax
movq %r11,24(%rdi,%rbp,8)
xorq %rbp,%rbp
andq %rax,%rbx
notq %rax
movq %rdi,%rsi
andq %rax,%rsi
leaq -1(%r9),%rcx
orq %rsi,%rbx
pxor %xmm0,%xmm0
leaq 64(%rsp,%r9,8),%rsi
movdqu (%rbx),%xmm1
leaq (%rsi,%r9,8),%rsi
movdqa %xmm0,64(%rsp)
movdqa %xmm0,(%rsi)
movdqu %xmm1,(%rdi)
jmp .Lsqr4x_copy
.align 16
.Lsqr4x_copy:
movdqu 16(%rbx,%rbp,1),%xmm2
movdqu 32(%rbx,%rbp,1),%xmm1
movdqa %xmm0,80(%rsp,%rbp,1)
movdqa %xmm0,96(%rsp,%rbp,1)
movdqa %xmm0,16(%rsi,%rbp,1)
movdqa %xmm0,32(%rsi,%rbp,1)
movdqu %xmm2,16(%rdi,%rbp,1)
movdqu %xmm1,32(%rdi,%rbp,1)
leaq 32(%rbp),%rbp
decq %rcx
jnz .Lsqr4x_copy
movdqu 16(%rbx,%rbp,1),%xmm2
movdqa %xmm0,80(%rsp,%rbp,1)
movdqa %xmm0,16(%rsi,%rbp,1)
movdqu %xmm2,16(%rdi,%rbp,1)
movq 56(%rsp),%rsi
movq $1,%rax
movq 0(%rsi),%r15
movq 8(%rsi),%r14
movq 16(%rsi),%r13
movq 24(%rsi),%r12
movq 32(%rsi),%rbp
movq 40(%rsi),%rbx
leaq 48(%rsi),%rsp
.Lsqr4x_epilogue:
.byte 0xf3,0xc3
.size bn_sqr4x_mont,.-bn_sqr4x_mont
.byte 77,111,110,116,103,111,109,101,114,121,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
.align 16