-rw-r--r-- 2370 lib25519-20221222/crypto_mGnP/ed25519/amd64-maax/fe25519_nsquare.S raw
// linker define fe25519_nsquare
// linker use mask63
/* Assembly for feedback field squaring. */
#define mask63 CRYPTO_SHARED_NAMESPACE(mask63)
.p2align 5
.globl _CRYPTO_SHARED_NAMESPACE(fe25519_nsquare)
.globl CRYPTO_SHARED_NAMESPACE(fe25519_nsquare)
_CRYPTO_SHARED_NAMESPACE(fe25519_nsquare):
CRYPTO_SHARED_NAMESPACE(fe25519_nsquare):
movq %rsp,%r11
andq $-32,%rsp
subq $64,%rsp
movq %r11,0(%rsp)
movq %r12,8(%rsp)
movq %r13,16(%rsp)
movq %r14,24(%rsp)
movq %r15,32(%rsp)
movq %rbx,40(%rsp)
movq %rbp,48(%rsp)
movq %rdi,56(%rsp)
movq 0(%rsi),%rbx
movq 8(%rsi),%rbp
movq 16(%rsi),%rax
movq 24(%rsi),%rsi
movq %rdx,%rdi
.L:
xorq %r13,%r13
movq %rbx,%rdx
mulx %rbp,%r9,%r10
mulx %rax,%rcx,%r11
adcx %rcx,%r10
mulx %rsi,%rcx,%r12
adcx %rcx,%r11
adcx %r13,%r12
movq %rbp,%rdx
xorq %r14,%r14
mulx %rax,%rcx,%rdx
adcx %rcx,%r11
adox %rdx,%r12
movq %rbp,%rdx
mulx %rsi,%rcx,%rdx
adcx %rcx,%r12
adox %rdx,%r13
adcx %r14,%r13
xorq %r15,%r15
movq %rax,%rdx
mulx %rsi,%rcx,%r14
adcx %rcx,%r13
adcx %r15,%r14
shld $1,%r14,%r15
shld $1,%r13,%r14
shld $1,%r12,%r13
shld $1,%r11,%r12
shld $1,%r10,%r11
shld $1,%r9,%r10
shlq $1,%r9
xorq %rdx,%rdx
movq %rbx,%rdx
mulx %rdx,%r8,%rdx
adcx %rdx,%r9
movq %rbp,%rdx
mulx %rdx,%rcx,%rdx
adcx %rcx,%r10
adcx %rdx,%r11
movq %rax,%rdx
mulx %rdx,%rcx,%rdx
adcx %rcx,%r12
adcx %rdx,%r13
movq %rsi,%rdx
mulx %rdx,%rcx,%rdx
adcx %rcx,%r14
adcx %rdx,%r15
xorq %rbx,%rbx
movq $38,%rdx
mulx %r12,%r12,%rbp
adcx %r8,%r12
adox %r9,%rbp
mulx %r13,%rcx,%rax
adcx %rcx,%rbp
adox %r10,%rax
mulx %r14,%rcx,%rsi
adcx %rcx,%rax
adox %r11,%rsi
mulx %r15,%rcx,%r15
adcx %rcx,%rsi
adox %rbx,%r15
adcx %rbx,%r15
shld $1,%rsi,%r15
andq mask63(%rip),%rsi
imul $19,%r15,%rbx
addq %r12,%rbx
adcq $0,%rbp
adcq $0,%rax
adcq $0,%rsi
subq $1,%rdi
cmpq $0,%rdi
jne .L
movq 56(%rsp),%rdi
movq %rbx,0(%rdi)
movq %rbp,8(%rdi)
movq %rax,16(%rdi)
movq %rsi,24(%rdi)
movq 0(%rsp),%r11
movq 8(%rsp),%r12
movq 16(%rsp),%r13
movq 24(%rsp),%r14
movq 32(%rsp),%r15
movq 40(%rsp),%rbx
movq 48(%rsp),%rbp
movq %r11,%rsp
ret