You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1408 lines
22 KiB
1408 lines
22 KiB
.text |
|
|
|
.globl bn_mul_mont |
|
.type bn_mul_mont,@function |
|
.align 16 |
|
bn_mul_mont: |
|
testl $3,%r9d |
|
jnz .Lmul_enter |
|
cmpl $8,%r9d |
|
jb .Lmul_enter |
|
cmpq %rsi,%rdx |
|
jne .Lmul4x_enter |
|
jmp .Lsqr4x_enter |
|
|
|
.align 16 |
|
.Lmul_enter: |
|
pushq %rbx |
|
pushq %rbp |
|
pushq %r12 |
|
pushq %r13 |
|
pushq %r14 |
|
pushq %r15 |
|
|
|
movl %r9d,%r9d |
|
leaq 2(%r9),%r10 |
|
movq %rsp,%r11 |
|
negq %r10 |
|
leaq (%rsp,%r10,8),%rsp |
|
andq $-1024,%rsp |
|
|
|
movq %r11,8(%rsp,%r9,8) |
|
.Lmul_body: |
|
|
|
|
|
|
|
|
|
|
|
|
|
subq %rsp,%r11 |
|
andq $-4096,%r11 |
|
.Lmul_page_walk: |
|
movq (%rsp,%r11,1),%r10 |
|
subq $4096,%r11 |
|
.byte 0x66,0x2e |
|
jnc .Lmul_page_walk |
|
|
|
movq %rdx,%r12 |
|
movq (%r8),%r8 |
|
movq (%r12),%rbx |
|
movq (%rsi),%rax |
|
|
|
xorq %r14,%r14 |
|
xorq %r15,%r15 |
|
|
|
movq %r8,%rbp |
|
mulq %rbx |
|
movq %rax,%r10 |
|
movq (%rcx),%rax |
|
|
|
imulq %r10,%rbp |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r10 |
|
movq 8(%rsi),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%r13 |
|
|
|
leaq 1(%r15),%r15 |
|
jmp .L1st_enter |
|
|
|
.align 16 |
|
.L1st: |
|
addq %rax,%r13 |
|
movq (%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r11,%r13 |
|
movq %r10,%r11 |
|
adcq $0,%rdx |
|
movq %r13,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
|
|
.L1st_enter: |
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq (%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
leaq 1(%r15),%r15 |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
cmpq %r9,%r15 |
|
jne .L1st |
|
|
|
addq %rax,%r13 |
|
movq (%rsi),%rax |
|
adcq $0,%rdx |
|
addq %r11,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
movq %r10,%r11 |
|
|
|
xorq %rdx,%rdx |
|
addq %r11,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-8(%rsp,%r9,8) |
|
movq %rdx,(%rsp,%r9,8) |
|
|
|
leaq 1(%r14),%r14 |
|
jmp .Louter |
|
.align 16 |
|
.Louter: |
|
movq (%r12,%r14,8),%rbx |
|
xorq %r15,%r15 |
|
movq %r8,%rbp |
|
movq (%rsp),%r10 |
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq (%rcx),%rax |
|
adcq $0,%rdx |
|
|
|
imulq %r10,%rbp |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r10 |
|
movq 8(%rsi),%rax |
|
adcq $0,%rdx |
|
movq 8(%rsp),%r10 |
|
movq %rdx,%r13 |
|
|
|
leaq 1(%r15),%r15 |
|
jmp .Linner_enter |
|
|
|
.align 16 |
|
.Linner: |
|
addq %rax,%r13 |
|
movq (%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
movq (%rsp,%r15,8),%r10 |
|
adcq $0,%rdx |
|
movq %r13,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
|
|
.Linner_enter: |
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq (%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r11,%r10 |
|
movq %rdx,%r11 |
|
adcq $0,%r11 |
|
leaq 1(%r15),%r15 |
|
|
|
mulq %rbp |
|
cmpq %r9,%r15 |
|
jne .Linner |
|
|
|
addq %rax,%r13 |
|
movq (%rsi),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
movq (%rsp,%r15,8),%r10 |
|
adcq $0,%rdx |
|
movq %r13,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
|
|
xorq %rdx,%rdx |
|
addq %r11,%r13 |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-8(%rsp,%r9,8) |
|
movq %rdx,(%rsp,%r9,8) |
|
|
|
leaq 1(%r14),%r14 |
|
cmpq %r9,%r14 |
|
jl .Louter |
|
|
|
xorq %r14,%r14 |
|
movq (%rsp),%rax |
|
leaq (%rsp),%rsi |
|
movq %r9,%r15 |
|
jmp .Lsub |
|
.align 16 |
|
.Lsub: sbbq (%rcx,%r14,8),%rax |
|
movq %rax,(%rdi,%r14,8) |
|
movq 8(%rsi,%r14,8),%rax |
|
leaq 1(%r14),%r14 |
|
decq %r15 |
|
jnz .Lsub |
|
|
|
sbbq $0,%rax |
|
xorq %r14,%r14 |
|
andq %rax,%rsi |
|
notq %rax |
|
movq %rdi,%rcx |
|
andq %rax,%rcx |
|
movq %r9,%r15 |
|
orq %rcx,%rsi |
|
.align 16 |
|
.Lcopy: |
|
movq (%rsi,%r14,8),%rax |
|
movq %r14,(%rsp,%r14,8) |
|
movq %rax,(%rdi,%r14,8) |
|
leaq 1(%r14),%r14 |
|
subq $1,%r15 |
|
jnz .Lcopy |
|
|
|
movq 8(%rsp,%r9,8),%rsi |
|
movq $1,%rax |
|
movq (%rsi),%r15 |
|
movq 8(%rsi),%r14 |
|
movq 16(%rsi),%r13 |
|
movq 24(%rsi),%r12 |
|
movq 32(%rsi),%rbp |
|
movq 40(%rsi),%rbx |
|
leaq 48(%rsi),%rsp |
|
.Lmul_epilogue: |
|
.byte 0xf3,0xc3 |
|
.size bn_mul_mont,.-bn_mul_mont |
|
.type bn_mul4x_mont,@function |
|
.align 16 |
|
bn_mul4x_mont: |
|
.Lmul4x_enter: |
|
pushq %rbx |
|
pushq %rbp |
|
pushq %r12 |
|
pushq %r13 |
|
pushq %r14 |
|
pushq %r15 |
|
|
|
movl %r9d,%r9d |
|
leaq 4(%r9),%r10 |
|
movq %rsp,%r11 |
|
negq %r10 |
|
leaq (%rsp,%r10,8),%rsp |
|
andq $-1024,%rsp |
|
|
|
movq %r11,8(%rsp,%r9,8) |
|
.Lmul4x_body: |
|
subq %rsp,%r11 |
|
andq $-4096,%r11 |
|
.Lmul4x_page_walk: |
|
movq (%rsp,%r11,1),%r10 |
|
subq $4096,%r11 |
|
.byte 0x2e |
|
jnc .Lmul4x_page_walk |
|
|
|
movq %rdi,16(%rsp,%r9,8) |
|
movq %rdx,%r12 |
|
movq (%r8),%r8 |
|
movq (%r12),%rbx |
|
movq (%rsi),%rax |
|
|
|
xorq %r14,%r14 |
|
xorq %r15,%r15 |
|
|
|
movq %r8,%rbp |
|
mulq %rbx |
|
movq %rax,%r10 |
|
movq (%rcx),%rax |
|
|
|
imulq %r10,%rbp |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r10 |
|
movq 8(%rsi),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq 8(%rcx),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq 16(%rsi),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
leaq 4(%r15),%r15 |
|
adcq $0,%rdx |
|
movq %rdi,(%rsp) |
|
movq %rdx,%r13 |
|
jmp .L1st4x |
|
.align 16 |
|
.L1st4x: |
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq -16(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r13 |
|
movq -8(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-24(%rsp,%r15,8) |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq -8(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq (%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
adcq $0,%rdx |
|
movq %rdi,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
|
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq (%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r13 |
|
movq 8(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-8(%rsp,%r15,8) |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq 8(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
leaq 4(%r15),%r15 |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq -16(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
adcq $0,%rdx |
|
movq %rdi,-32(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
cmpq %r9,%r15 |
|
jl .L1st4x |
|
|
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq -16(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r13 |
|
movq -8(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-24(%rsp,%r15,8) |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq -8(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq (%rsi),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
adcq $0,%rdx |
|
movq %rdi,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
|
|
xorq %rdi,%rdi |
|
addq %r10,%r13 |
|
adcq $0,%rdi |
|
movq %r13,-8(%rsp,%r15,8) |
|
movq %rdi,(%rsp,%r15,8) |
|
|
|
leaq 1(%r14),%r14 |
|
.align 4 |
|
.Louter4x: |
|
movq (%r12,%r14,8),%rbx |
|
xorq %r15,%r15 |
|
movq (%rsp),%r10 |
|
movq %r8,%rbp |
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq (%rcx),%rax |
|
adcq $0,%rdx |
|
|
|
imulq %r10,%rbp |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r10 |
|
movq 8(%rsi),%rax |
|
adcq $0,%rdx |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq 8(%rcx),%rax |
|
adcq $0,%rdx |
|
addq 8(%rsp),%r11 |
|
adcq $0,%rdx |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq 16(%rsi),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
leaq 4(%r15),%r15 |
|
adcq $0,%rdx |
|
movq %rdi,(%rsp) |
|
movq %rdx,%r13 |
|
jmp .Linner4x |
|
.align 16 |
|
.Linner4x: |
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq -16(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq -16(%rsp,%r15,8),%r10 |
|
adcq $0,%rdx |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r13 |
|
movq -8(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-24(%rsp,%r15,8) |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq -8(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq -8(%rsp,%r15,8),%r11 |
|
adcq $0,%rdx |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq (%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
adcq $0,%rdx |
|
movq %rdi,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
|
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq (%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq (%rsp,%r15,8),%r10 |
|
adcq $0,%rdx |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r13 |
|
movq 8(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-8(%rsp,%r15,8) |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq 8(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq 8(%rsp,%r15,8),%r11 |
|
adcq $0,%rdx |
|
leaq 4(%r15),%r15 |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq -16(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
adcq $0,%rdx |
|
movq %rdi,-32(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
cmpq %r9,%r15 |
|
jl .Linner4x |
|
|
|
mulq %rbx |
|
addq %rax,%r10 |
|
movq -16(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq -16(%rsp,%r15,8),%r10 |
|
adcq $0,%rdx |
|
movq %rdx,%r11 |
|
|
|
mulq %rbp |
|
addq %rax,%r13 |
|
movq -8(%rsi,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq %r10,%r13 |
|
adcq $0,%rdx |
|
movq %r13,-24(%rsp,%r15,8) |
|
movq %rdx,%rdi |
|
|
|
mulq %rbx |
|
addq %rax,%r11 |
|
movq -8(%rcx,%r15,8),%rax |
|
adcq $0,%rdx |
|
addq -8(%rsp,%r15,8),%r11 |
|
adcq $0,%rdx |
|
leaq 1(%r14),%r14 |
|
movq %rdx,%r10 |
|
|
|
mulq %rbp |
|
addq %rax,%rdi |
|
movq (%rsi),%rax |
|
adcq $0,%rdx |
|
addq %r11,%rdi |
|
adcq $0,%rdx |
|
movq %rdi,-16(%rsp,%r15,8) |
|
movq %rdx,%r13 |
|
|
|
xorq %rdi,%rdi |
|
addq %r10,%r13 |
|
adcq $0,%rdi |
|
addq (%rsp,%r9,8),%r13 |
|
adcq $0,%rdi |
|
movq %r13,-8(%rsp,%r15,8) |
|
movq %rdi,(%rsp,%r15,8) |
|
|
|
cmpq %r9,%r14 |
|
jl .Louter4x |
|
movq 16(%rsp,%r9,8),%rdi |
|
movq 0(%rsp),%rax |
|
pxor %xmm0,%xmm0 |
|
movq 8(%rsp),%rdx |
|
shrq $2,%r9 |
|
leaq (%rsp),%rsi |
|
xorq %r14,%r14 |
|
|
|
subq 0(%rcx),%rax |
|
movq 16(%rsi),%rbx |
|
movq 24(%rsi),%rbp |
|
sbbq 8(%rcx),%rdx |
|
leaq -1(%r9),%r15 |
|
jmp .Lsub4x |
|
.align 16 |
|
.Lsub4x: |
|
movq %rax,0(%rdi,%r14,8) |
|
movq %rdx,8(%rdi,%r14,8) |
|
sbbq 16(%rcx,%r14,8),%rbx |
|
movq 32(%rsi,%r14,8),%rax |
|
movq 40(%rsi,%r14,8),%rdx |
|
sbbq 24(%rcx,%r14,8),%rbp |
|
movq %rbx,16(%rdi,%r14,8) |
|
movq %rbp,24(%rdi,%r14,8) |
|
sbbq 32(%rcx,%r14,8),%rax |
|
movq 48(%rsi,%r14,8),%rbx |
|
movq 56(%rsi,%r14,8),%rbp |
|
sbbq 40(%rcx,%r14,8),%rdx |
|
leaq 4(%r14),%r14 |
|
decq %r15 |
|
jnz .Lsub4x |
|
|
|
movq %rax,0(%rdi,%r14,8) |
|
movq 32(%rsi,%r14,8),%rax |
|
sbbq 16(%rcx,%r14,8),%rbx |
|
movq %rdx,8(%rdi,%r14,8) |
|
sbbq 24(%rcx,%r14,8),%rbp |
|
movq %rbx,16(%rdi,%r14,8) |
|
|
|
sbbq $0,%rax |
|
movq %rbp,24(%rdi,%r14,8) |
|
xorq %r14,%r14 |
|
andq %rax,%rsi |
|
notq %rax |
|
movq %rdi,%rcx |
|
andq %rax,%rcx |
|
leaq -1(%r9),%r15 |
|
orq %rcx,%rsi |
|
|
|
movdqu (%rsi),%xmm1 |
|
movdqa %xmm0,(%rsp) |
|
movdqu %xmm1,(%rdi) |
|
jmp .Lcopy4x |
|
.align 16 |
|
.Lcopy4x: |
|
movdqu 16(%rsi,%r14,1),%xmm2 |
|
movdqu 32(%rsi,%r14,1),%xmm1 |
|
movdqa %xmm0,16(%rsp,%r14,1) |
|
movdqu %xmm2,16(%rdi,%r14,1) |
|
movdqa %xmm0,32(%rsp,%r14,1) |
|
movdqu %xmm1,32(%rdi,%r14,1) |
|
leaq 32(%r14),%r14 |
|
decq %r15 |
|
jnz .Lcopy4x |
|
|
|
shlq $2,%r9 |
|
movdqu 16(%rsi,%r14,1),%xmm2 |
|
movdqa %xmm0,16(%rsp,%r14,1) |
|
movdqu %xmm2,16(%rdi,%r14,1) |
|
movq 8(%rsp,%r9,8),%rsi |
|
movq $1,%rax |
|
movq (%rsi),%r15 |
|
movq 8(%rsi),%r14 |
|
movq 16(%rsi),%r13 |
|
movq 24(%rsi),%r12 |
|
movq 32(%rsi),%rbp |
|
movq 40(%rsi),%rbx |
|
leaq 48(%rsi),%rsp |
|
.Lmul4x_epilogue: |
|
.byte 0xf3,0xc3 |
|
.size bn_mul4x_mont,.-bn_mul4x_mont |
|
.type bn_sqr4x_mont,@function |
|
.align 16 |
|
bn_sqr4x_mont: |
|
.Lsqr4x_enter: |
|
movq %rsp,%rax |
|
pushq %rbx |
|
pushq %rbp |
|
pushq %r12 |
|
pushq %r13 |
|
pushq %r14 |
|
pushq %r15 |
|
|
|
shll $3,%r9d |
|
movq %rsp,%r11 |
|
negq %r9 |
|
movq (%r8),%r8 |
|
leaq -72(%rsp,%r9,2),%rsp |
|
andq $-1024,%rsp |
|
|
|
subq %rsp,%r11 |
|
andq $-4096,%r11 |
|
.Lsqr4x_page_walk: |
|
movq (%rsp,%r11,1),%r10 |
|
subq $4096,%r11 |
|
.byte 0x2e |
|
jnc .Lsqr4x_page_walk |
|
|
|
movq %r9,%r10 |
|
negq %r9 |
|
leaq -48(%rax),%r11 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
movq %rdi,32(%rsp) |
|
movq %rcx,40(%rsp) |
|
movq %r8,48(%rsp) |
|
movq %r11,56(%rsp) |
|
.Lsqr4x_body: |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
leaq 32(%r10),%rbp |
|
leaq (%rsi,%r9,1),%rsi |
|
|
|
movq %r9,%rcx |
|
|
|
|
|
movq -32(%rsi,%rbp,1),%r14 |
|
leaq 64(%rsp,%r9,2),%rdi |
|
movq -24(%rsi,%rbp,1),%rax |
|
leaq -32(%rdi,%rbp,1),%rdi |
|
movq -16(%rsi,%rbp,1),%rbx |
|
movq %rax,%r15 |
|
|
|
mulq %r14 |
|
movq %rax,%r10 |
|
movq %rbx,%rax |
|
movq %rdx,%r11 |
|
movq %r10,-24(%rdi,%rbp,1) |
|
|
|
xorq %r10,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
movq %r11,-16(%rdi,%rbp,1) |
|
|
|
leaq -16(%rbp),%rcx |
|
|
|
|
|
movq 8(%rsi,%rcx,1),%rbx |
|
mulq %r15 |
|
movq %rax,%r12 |
|
movq %rbx,%rax |
|
movq %rdx,%r13 |
|
|
|
xorq %r11,%r11 |
|
addq %r12,%r10 |
|
leaq 16(%rcx),%rcx |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,-8(%rdi,%rcx,1) |
|
jmp .Lsqr4x_1st |
|
|
|
.align 16 |
|
.Lsqr4x_1st: |
|
movq (%rsi,%rcx,1),%rbx |
|
xorq %r12,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq %rbx,%rax |
|
adcq %rdx,%r12 |
|
|
|
xorq %r10,%r10 |
|
addq %r13,%r11 |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
movq %r11,(%rdi,%rcx,1) |
|
|
|
|
|
movq 8(%rsi,%rcx,1),%rbx |
|
xorq %r13,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq %rdx,%r13 |
|
|
|
xorq %r11,%r11 |
|
addq %r12,%r10 |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,8(%rdi,%rcx,1) |
|
|
|
movq 16(%rsi,%rcx,1),%rbx |
|
xorq %r12,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq %rbx,%rax |
|
adcq %rdx,%r12 |
|
|
|
xorq %r10,%r10 |
|
addq %r13,%r11 |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
movq %r11,16(%rdi,%rcx,1) |
|
|
|
|
|
movq 24(%rsi,%rcx,1),%rbx |
|
xorq %r13,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq %rdx,%r13 |
|
|
|
xorq %r11,%r11 |
|
addq %r12,%r10 |
|
leaq 32(%rcx),%rcx |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,-8(%rdi,%rcx,1) |
|
|
|
cmpq $0,%rcx |
|
jne .Lsqr4x_1st |
|
|
|
xorq %r12,%r12 |
|
addq %r11,%r13 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
adcq %rdx,%r12 |
|
|
|
movq %r13,(%rdi) |
|
leaq 16(%rbp),%rbp |
|
movq %r12,8(%rdi) |
|
jmp .Lsqr4x_outer |
|
|
|
.align 16 |
|
.Lsqr4x_outer: |
|
movq -32(%rsi,%rbp,1),%r14 |
|
leaq 64(%rsp,%r9,2),%rdi |
|
movq -24(%rsi,%rbp,1),%rax |
|
leaq -32(%rdi,%rbp,1),%rdi |
|
movq -16(%rsi,%rbp,1),%rbx |
|
movq %rax,%r15 |
|
|
|
movq -24(%rdi,%rbp,1),%r10 |
|
xorq %r11,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,-24(%rdi,%rbp,1) |
|
|
|
xorq %r10,%r10 |
|
addq -16(%rdi,%rbp,1),%r11 |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
movq %r11,-16(%rdi,%rbp,1) |
|
|
|
leaq -16(%rbp),%rcx |
|
xorq %r12,%r12 |
|
|
|
|
|
movq 8(%rsi,%rcx,1),%rbx |
|
xorq %r13,%r13 |
|
addq 8(%rdi,%rcx,1),%r12 |
|
adcq $0,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq %rdx,%r13 |
|
|
|
xorq %r11,%r11 |
|
addq %r12,%r10 |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,8(%rdi,%rcx,1) |
|
|
|
leaq 16(%rcx),%rcx |
|
jmp .Lsqr4x_inner |
|
|
|
.align 16 |
|
.Lsqr4x_inner: |
|
movq (%rsi,%rcx,1),%rbx |
|
xorq %r12,%r12 |
|
addq (%rdi,%rcx,1),%r13 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq %rbx,%rax |
|
adcq %rdx,%r12 |
|
|
|
xorq %r10,%r10 |
|
addq %r13,%r11 |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
movq %r11,(%rdi,%rcx,1) |
|
|
|
movq 8(%rsi,%rcx,1),%rbx |
|
xorq %r13,%r13 |
|
addq 8(%rdi,%rcx,1),%r12 |
|
adcq $0,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq %rdx,%r13 |
|
|
|
xorq %r11,%r11 |
|
addq %r12,%r10 |
|
leaq 16(%rcx),%rcx |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,-8(%rdi,%rcx,1) |
|
|
|
cmpq $0,%rcx |
|
jne .Lsqr4x_inner |
|
|
|
xorq %r12,%r12 |
|
addq %r11,%r13 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
adcq %rdx,%r12 |
|
|
|
movq %r13,(%rdi) |
|
movq %r12,8(%rdi) |
|
|
|
addq $16,%rbp |
|
jnz .Lsqr4x_outer |
|
|
|
|
|
movq -32(%rsi),%r14 |
|
leaq 64(%rsp,%r9,2),%rdi |
|
movq -24(%rsi),%rax |
|
leaq -32(%rdi,%rbp,1),%rdi |
|
movq -16(%rsi),%rbx |
|
movq %rax,%r15 |
|
|
|
xorq %r11,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,-24(%rdi) |
|
|
|
xorq %r10,%r10 |
|
addq %r13,%r11 |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
movq %r11,-16(%rdi) |
|
|
|
movq -8(%rsi),%rbx |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq $0,%rdx |
|
|
|
xorq %r11,%r11 |
|
addq %r12,%r10 |
|
movq %rdx,%r13 |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %rbx,%rax |
|
adcq %rdx,%r11 |
|
movq %r10,-8(%rdi) |
|
|
|
xorq %r12,%r12 |
|
addq %r11,%r13 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq -16(%rsi),%rax |
|
adcq %rdx,%r12 |
|
|
|
movq %r13,(%rdi) |
|
movq %r12,8(%rdi) |
|
|
|
mulq %rbx |
|
addq $16,%rbp |
|
xorq %r14,%r14 |
|
subq %r9,%rbp |
|
xorq %r15,%r15 |
|
|
|
addq %r12,%rax |
|
adcq $0,%rdx |
|
movq %rax,8(%rdi) |
|
movq %rdx,16(%rdi) |
|
movq %r15,24(%rdi) |
|
|
|
movq -16(%rsi,%rbp,1),%rax |
|
leaq 64(%rsp,%r9,2),%rdi |
|
xorq %r10,%r10 |
|
movq -24(%rdi,%rbp,2),%r11 |
|
|
|
leaq (%r14,%r10,2),%r12 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r13 |
|
shrq $63,%r11 |
|
orq %r10,%r13 |
|
movq -16(%rdi,%rbp,2),%r10 |
|
movq %r11,%r14 |
|
mulq %rax |
|
negq %r15 |
|
movq -8(%rdi,%rbp,2),%r11 |
|
adcq %rax,%r12 |
|
movq -8(%rsi,%rbp,1),%rax |
|
movq %r12,-32(%rdi,%rbp,2) |
|
adcq %rdx,%r13 |
|
|
|
leaq (%r14,%r10,2),%rbx |
|
movq %r13,-24(%rdi,%rbp,2) |
|
sbbq %r15,%r15 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r8 |
|
shrq $63,%r11 |
|
orq %r10,%r8 |
|
movq 0(%rdi,%rbp,2),%r10 |
|
movq %r11,%r14 |
|
mulq %rax |
|
negq %r15 |
|
movq 8(%rdi,%rbp,2),%r11 |
|
adcq %rax,%rbx |
|
movq 0(%rsi,%rbp,1),%rax |
|
movq %rbx,-16(%rdi,%rbp,2) |
|
adcq %rdx,%r8 |
|
leaq 16(%rbp),%rbp |
|
movq %r8,-40(%rdi,%rbp,2) |
|
sbbq %r15,%r15 |
|
jmp .Lsqr4x_shift_n_add |
|
|
|
.align 16 |
|
.Lsqr4x_shift_n_add: |
|
leaq (%r14,%r10,2),%r12 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r13 |
|
shrq $63,%r11 |
|
orq %r10,%r13 |
|
movq -16(%rdi,%rbp,2),%r10 |
|
movq %r11,%r14 |
|
mulq %rax |
|
negq %r15 |
|
movq -8(%rdi,%rbp,2),%r11 |
|
adcq %rax,%r12 |
|
movq -8(%rsi,%rbp,1),%rax |
|
movq %r12,-32(%rdi,%rbp,2) |
|
adcq %rdx,%r13 |
|
|
|
leaq (%r14,%r10,2),%rbx |
|
movq %r13,-24(%rdi,%rbp,2) |
|
sbbq %r15,%r15 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r8 |
|
shrq $63,%r11 |
|
orq %r10,%r8 |
|
movq 0(%rdi,%rbp,2),%r10 |
|
movq %r11,%r14 |
|
mulq %rax |
|
negq %r15 |
|
movq 8(%rdi,%rbp,2),%r11 |
|
adcq %rax,%rbx |
|
movq 0(%rsi,%rbp,1),%rax |
|
movq %rbx,-16(%rdi,%rbp,2) |
|
adcq %rdx,%r8 |
|
|
|
leaq (%r14,%r10,2),%r12 |
|
movq %r8,-8(%rdi,%rbp,2) |
|
sbbq %r15,%r15 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r13 |
|
shrq $63,%r11 |
|
orq %r10,%r13 |
|
movq 16(%rdi,%rbp,2),%r10 |
|
movq %r11,%r14 |
|
mulq %rax |
|
negq %r15 |
|
movq 24(%rdi,%rbp,2),%r11 |
|
adcq %rax,%r12 |
|
movq 8(%rsi,%rbp,1),%rax |
|
movq %r12,0(%rdi,%rbp,2) |
|
adcq %rdx,%r13 |
|
|
|
leaq (%r14,%r10,2),%rbx |
|
movq %r13,8(%rdi,%rbp,2) |
|
sbbq %r15,%r15 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r8 |
|
shrq $63,%r11 |
|
orq %r10,%r8 |
|
movq 32(%rdi,%rbp,2),%r10 |
|
movq %r11,%r14 |
|
mulq %rax |
|
negq %r15 |
|
movq 40(%rdi,%rbp,2),%r11 |
|
adcq %rax,%rbx |
|
movq 16(%rsi,%rbp,1),%rax |
|
movq %rbx,16(%rdi,%rbp,2) |
|
adcq %rdx,%r8 |
|
movq %r8,24(%rdi,%rbp,2) |
|
sbbq %r15,%r15 |
|
addq $32,%rbp |
|
jnz .Lsqr4x_shift_n_add |
|
|
|
leaq (%r14,%r10,2),%r12 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r13 |
|
shrq $63,%r11 |
|
orq %r10,%r13 |
|
movq -16(%rdi),%r10 |
|
movq %r11,%r14 |
|
mulq %rax |
|
negq %r15 |
|
movq -8(%rdi),%r11 |
|
adcq %rax,%r12 |
|
movq -8(%rsi),%rax |
|
movq %r12,-32(%rdi) |
|
adcq %rdx,%r13 |
|
|
|
leaq (%r14,%r10,2),%rbx |
|
movq %r13,-24(%rdi) |
|
sbbq %r15,%r15 |
|
shrq $63,%r10 |
|
leaq (%rcx,%r11,2),%r8 |
|
shrq $63,%r11 |
|
orq %r10,%r8 |
|
mulq %rax |
|
negq %r15 |
|
adcq %rax,%rbx |
|
adcq %rdx,%r8 |
|
movq %rbx,-16(%rdi) |
|
movq %r8,-8(%rdi) |
|
movq 40(%rsp),%rsi |
|
movq 48(%rsp),%r8 |
|
xorq %rcx,%rcx |
|
movq %r9,0(%rsp) |
|
subq %r9,%rcx |
|
movq 64(%rsp),%r10 |
|
movq %r8,%r14 |
|
leaq 64(%rsp,%r9,2),%rax |
|
leaq 64(%rsp,%r9,1),%rdi |
|
movq %rax,8(%rsp) |
|
leaq (%rsi,%r9,1),%rsi |
|
xorq %rbp,%rbp |
|
|
|
movq 0(%rsi,%rcx,1),%rax |
|
movq 8(%rsi,%rcx,1),%r9 |
|
imulq %r10,%r14 |
|
movq %rax,%rbx |
|
jmp .Lsqr4x_mont_outer |
|
|
|
.align 16 |
|
.Lsqr4x_mont_outer: |
|
xorq %r11,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %r9,%rax |
|
adcq %rdx,%r11 |
|
movq %r8,%r15 |
|
|
|
xorq %r10,%r10 |
|
addq 8(%rdi,%rcx,1),%r11 |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
|
|
imulq %r11,%r15 |
|
|
|
movq 16(%rsi,%rcx,1),%rbx |
|
xorq %r13,%r13 |
|
addq %r11,%r12 |
|
adcq $0,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq %rdx,%r13 |
|
movq %r12,8(%rdi,%rcx,1) |
|
|
|
xorq %r11,%r11 |
|
addq 16(%rdi,%rcx,1),%r10 |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %r9,%rax |
|
adcq %rdx,%r11 |
|
|
|
movq 24(%rsi,%rcx,1),%r9 |
|
xorq %r12,%r12 |
|
addq %r10,%r13 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq %r9,%rax |
|
adcq %rdx,%r12 |
|
movq %r13,16(%rdi,%rcx,1) |
|
|
|
xorq %r10,%r10 |
|
addq 24(%rdi,%rcx,1),%r11 |
|
leaq 32(%rcx),%rcx |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
jmp .Lsqr4x_mont_inner |
|
|
|
.align 16 |
|
.Lsqr4x_mont_inner: |
|
movq (%rsi,%rcx,1),%rbx |
|
xorq %r13,%r13 |
|
addq %r11,%r12 |
|
adcq $0,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq %rdx,%r13 |
|
movq %r12,-8(%rdi,%rcx,1) |
|
|
|
xorq %r11,%r11 |
|
addq (%rdi,%rcx,1),%r10 |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %r9,%rax |
|
adcq %rdx,%r11 |
|
|
|
movq 8(%rsi,%rcx,1),%r9 |
|
xorq %r12,%r12 |
|
addq %r10,%r13 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq %r9,%rax |
|
adcq %rdx,%r12 |
|
movq %r13,(%rdi,%rcx,1) |
|
|
|
xorq %r10,%r10 |
|
addq 8(%rdi,%rcx,1),%r11 |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
|
|
|
|
movq 16(%rsi,%rcx,1),%rbx |
|
xorq %r13,%r13 |
|
addq %r11,%r12 |
|
adcq $0,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %rbx,%rax |
|
adcq %rdx,%r13 |
|
movq %r12,8(%rdi,%rcx,1) |
|
|
|
xorq %r11,%r11 |
|
addq 16(%rdi,%rcx,1),%r10 |
|
adcq $0,%r11 |
|
mulq %r14 |
|
addq %rax,%r10 |
|
movq %r9,%rax |
|
adcq %rdx,%r11 |
|
|
|
movq 24(%rsi,%rcx,1),%r9 |
|
xorq %r12,%r12 |
|
addq %r10,%r13 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq %r9,%rax |
|
adcq %rdx,%r12 |
|
movq %r13,16(%rdi,%rcx,1) |
|
|
|
xorq %r10,%r10 |
|
addq 24(%rdi,%rcx,1),%r11 |
|
leaq 32(%rcx),%rcx |
|
adcq $0,%r10 |
|
mulq %r14 |
|
addq %rax,%r11 |
|
movq %rbx,%rax |
|
adcq %rdx,%r10 |
|
cmpq $0,%rcx |
|
jne .Lsqr4x_mont_inner |
|
|
|
subq 0(%rsp),%rcx |
|
movq %r8,%r14 |
|
|
|
xorq %r13,%r13 |
|
addq %r11,%r12 |
|
adcq $0,%r13 |
|
mulq %r15 |
|
addq %rax,%r12 |
|
movq %r9,%rax |
|
adcq %rdx,%r13 |
|
movq %r12,-8(%rdi) |
|
|
|
xorq %r11,%r11 |
|
addq (%rdi),%r10 |
|
adcq $0,%r11 |
|
movq 0(%rsi,%rcx,1),%rbx |
|
addq %rbp,%r10 |
|
adcq $0,%r11 |
|
|
|
imulq 16(%rdi,%rcx,1),%r14 |
|
xorq %r12,%r12 |
|
movq 8(%rsi,%rcx,1),%r9 |
|
addq %r10,%r13 |
|
movq 16(%rdi,%rcx,1),%r10 |
|
adcq $0,%r12 |
|
mulq %r15 |
|
addq %rax,%r13 |
|
movq %rbx,%rax |
|
adcq %rdx,%r12 |
|
movq %r13,(%rdi) |
|
|
|
xorq %rbp,%rbp |
|
addq 8(%rdi),%r12 |
|
adcq %rbp,%rbp |
|
addq %r11,%r12 |
|
leaq 16(%rdi),%rdi |
|
adcq $0,%rbp |
|
movq %r12,-8(%rdi) |
|
cmpq 8(%rsp),%rdi |
|
jb .Lsqr4x_mont_outer |
|
|
|
movq 0(%rsp),%r9 |
|
movq %rbp,(%rdi) |
|
movq 64(%rsp,%r9,1),%rax |
|
leaq 64(%rsp,%r9,1),%rbx |
|
movq 40(%rsp),%rsi |
|
shrq $5,%r9 |
|
movq 8(%rbx),%rdx |
|
xorq %rbp,%rbp |
|
|
|
movq 32(%rsp),%rdi |
|
subq 0(%rsi),%rax |
|
movq 16(%rbx),%r10 |
|
movq 24(%rbx),%r11 |
|
sbbq 8(%rsi),%rdx |
|
leaq -1(%r9),%rcx |
|
jmp .Lsqr4x_sub |
|
.align 16 |
|
.Lsqr4x_sub: |
|
movq %rax,0(%rdi,%rbp,8) |
|
movq %rdx,8(%rdi,%rbp,8) |
|
sbbq 16(%rsi,%rbp,8),%r10 |
|
movq 32(%rbx,%rbp,8),%rax |
|
movq 40(%rbx,%rbp,8),%rdx |
|
sbbq 24(%rsi,%rbp,8),%r11 |
|
movq %r10,16(%rdi,%rbp,8) |
|
movq %r11,24(%rdi,%rbp,8) |
|
sbbq 32(%rsi,%rbp,8),%rax |
|
movq 48(%rbx,%rbp,8),%r10 |
|
movq 56(%rbx,%rbp,8),%r11 |
|
sbbq 40(%rsi,%rbp,8),%rdx |
|
leaq 4(%rbp),%rbp |
|
decq %rcx |
|
jnz .Lsqr4x_sub |
|
|
|
movq %rax,0(%rdi,%rbp,8) |
|
movq 32(%rbx,%rbp,8),%rax |
|
sbbq 16(%rsi,%rbp,8),%r10 |
|
movq %rdx,8(%rdi,%rbp,8) |
|
sbbq 24(%rsi,%rbp,8),%r11 |
|
movq %r10,16(%rdi,%rbp,8) |
|
|
|
sbbq $0,%rax |
|
movq %r11,24(%rdi,%rbp,8) |
|
xorq %rbp,%rbp |
|
andq %rax,%rbx |
|
notq %rax |
|
movq %rdi,%rsi |
|
andq %rax,%rsi |
|
leaq -1(%r9),%rcx |
|
orq %rsi,%rbx |
|
|
|
pxor %xmm0,%xmm0 |
|
leaq 64(%rsp,%r9,8),%rsi |
|
movdqu (%rbx),%xmm1 |
|
leaq (%rsi,%r9,8),%rsi |
|
movdqa %xmm0,64(%rsp) |
|
movdqa %xmm0,(%rsi) |
|
movdqu %xmm1,(%rdi) |
|
jmp .Lsqr4x_copy |
|
.align 16 |
|
.Lsqr4x_copy: |
|
movdqu 16(%rbx,%rbp,1),%xmm2 |
|
movdqu 32(%rbx,%rbp,1),%xmm1 |
|
movdqa %xmm0,80(%rsp,%rbp,1) |
|
movdqa %xmm0,96(%rsp,%rbp,1) |
|
movdqa %xmm0,16(%rsi,%rbp,1) |
|
movdqa %xmm0,32(%rsi,%rbp,1) |
|
movdqu %xmm2,16(%rdi,%rbp,1) |
|
movdqu %xmm1,32(%rdi,%rbp,1) |
|
leaq 32(%rbp),%rbp |
|
decq %rcx |
|
jnz .Lsqr4x_copy |
|
|
|
movdqu 16(%rbx,%rbp,1),%xmm2 |
|
movdqa %xmm0,80(%rsp,%rbp,1) |
|
movdqa %xmm0,16(%rsi,%rbp,1) |
|
movdqu %xmm2,16(%rdi,%rbp,1) |
|
movq 56(%rsp),%rsi |
|
movq $1,%rax |
|
movq 0(%rsi),%r15 |
|
movq 8(%rsi),%r14 |
|
movq 16(%rsi),%r13 |
|
movq 24(%rsi),%r12 |
|
movq 32(%rsi),%rbp |
|
movq 40(%rsi),%rbx |
|
leaq 48(%rsi),%rsp |
|
.Lsqr4x_epilogue: |
|
.byte 0xf3,0xc3 |
|
.size bn_sqr4x_mont,.-bn_sqr4x_mont |
|
.byte 77,111,110,116,103,111,109,101,114,121,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |
|
.align 16
|
|
|