-rw-r--r-- 2948 lib25519-20221222/crypto_mGnP/ed25519/amd64-maa4/fe25519_mul.S raw
// linker define fe25519_mul // linker use mask63 /* Assembly for field multiplication. */ #define mask63 CRYPTO_SHARED_NAMESPACE(mask63) .p2align 5 .globl _CRYPTO_SHARED_NAMESPACE(fe25519_mul) .globl CRYPTO_SHARED_NAMESPACE(fe25519_mul) _CRYPTO_SHARED_NAMESPACE(fe25519_mul): CRYPTO_SHARED_NAMESPACE(fe25519_mul): movq %rsp,%r11 andq $-32,%rsp subq $40,%rsp movq %r11,0(%rsp) movq %r12,8(%rsp) movq %r13,16(%rsp) movq %r14,24(%rsp) movq %r15,32(%rsp) movq %rdx,%rcx movq 8(%rsi),%rax mulq 24(%rcx) movq %rax,%r8 movq $0,%r9 movq %rdx,%r10 movq $0,%r11 movq 16(%rsi),%rax mulq 16(%rcx) addq %rax,%r8 adcq $0,%r9 addq %rdx,%r10 adcq $0,%r11 movq 24(%rsi),%rax mulq 8(%rcx) addq %rax,%r8 adcq $0,%r9 addq %rdx,%r10 adcq $0,%r11 movq 16(%rsi),%rax mulq 24(%rcx) addq %rax,%r10 adcq $0,%r11 movq %rdx,%r12 movq $0,%r13 movq 24(%rsi),%rax mulq 16(%rcx) addq %rax,%r10 adcq $0,%r11 addq %rdx,%r12 adcq $0,%r13 movq $38,%rax mulq %r10 imul $38,%r11,%r11 movq %rax,%r10 addq %rdx,%r11 movq 24(%rsi),%rax mulq 24(%rcx) addq %rax,%r12 adcq $0,%r13 movq $38,%rax mulq %rdx movq %rax,%r14 movq %rdx,%r15 movq $38,%rax mulq %r12 imul $38,%r13,%r13 movq %rax,%r12 addq %rdx,%r13 movq 0(%rsi),%rax mulq 24(%rcx) addq %rax,%r14 adcq $0,%r15 addq %rdx,%r8 adcq $0,%r9 movq 8(%rsi),%rax mulq 16(%rcx) addq %rax,%r14 adcq $0,%r15 addq %rdx,%r8 adcq $0,%r9 movq 16(%rsi),%rax mulq 8(%rcx) addq %rax,%r14 adcq $0,%r15 addq %rdx,%r8 adcq $0,%r9 movq 24(%rsi),%rax mulq 0(%rcx) addq %rax,%r14 adcq $0,%r15 addq %rdx,%r8 adcq $0,%r9 movq $38,%rax mulq %r8 imul $38,%r9,%r9 movq %rax,%r8 addq %rdx,%r9 movq 0(%rsi),%rax mulq 0(%rcx) addq %rax,%r8 adcq $0,%r9 addq %rdx,%r10 adcq $0,%r11 movq 0(%rsi),%rax mulq 8(%rcx) addq %rax,%r10 adcq $0,%r11 addq %rdx,%r12 adcq $0,%r13 movq 8(%rsi),%rax mulq 0(%rcx) addq %rax,%r10 adcq $0,%r11 addq %rdx,%r12 adcq $0,%r13 movq 0(%rsi),%rax mulq 16(%rcx) addq %rax,%r12 adcq $0,%r13 addq %rdx,%r14 adcq $0,%r15 movq 8(%rsi),%rax mulq 8(%rcx) addq %rax,%r12 adcq $0,%r13 addq %rdx,%r14 adcq $0,%r15 movq 16(%rsi),%rax mulq 0(%rcx) addq %rax,%r12 adcq $0,%r13 addq %rdx,%r14 adcq $0,%r15 addq %r9,%r10 adcq $0,%r11 addq %r11,%r12 adcq $0,%r13 addq %r13,%r14 adcq $0,%r15 shld $1,%r14,%r15 andq mask63(%rip),%r14 imul $19,%r15,%r15 addq %r15,%r8 adcq $0,%r10 adcq $0,%r12 adcq $0,%r14 movq %r8,0(%rdi) movq %r10,8(%rdi) movq %r12,16(%rdi) movq %r14,24(%rdi) movq 0(%rsp),%r11 movq 8(%rsp),%r12 movq 16(%rsp),%r13 movq 24(%rsp),%r14 movq 32(%rsp),%r15 movq %r11,%rsp ret