-rw-r--r-- 9035 lib25519-20230630/crypto_nP/montgomery25519/amd64-64/work_cswap.S raw
#include "crypto_asm_hidden.h" // linker define work_cswap # qhasm: int64 workp # qhasm: int64 swap # qhasm: input workp # qhasm: input swap # qhasm: int64 w0 # qhasm: int64 w1 # qhasm: int64 w2 # qhasm: int64 w3 # qhasm: int64 w4 # qhasm: int64 w5 # qhasm: int64 w6 # qhasm: int64 w7 # qhasm: int64 w8 # qhasm: int64 w9 # qhasm: int64 w10 # qhasm: int64 w11 # qhasm: int64 w12 # qhasm: int64 w13 # qhasm: int64 w14 # qhasm: int64 w15 # qhasm: int64 t # qhasm: int64 caller1 # qhasm: int64 caller2 # qhasm: int64 caller3 # qhasm: int64 caller4 # qhasm: int64 caller5 # qhasm: int64 caller6 # qhasm: int64 caller7 # qhasm: caller caller1 # qhasm: caller caller2 # qhasm: caller caller3 # qhasm: caller caller4 # qhasm: caller caller5 # qhasm: caller caller6 # qhasm: caller caller7 # qhasm: stack64 caller1_stack # qhasm: stack64 caller2_stack # qhasm: stack64 caller3_stack # qhasm: stack64 caller4_stack # qhasm: stack64 caller5_stack # qhasm: stack64 caller6_stack # qhasm: stack64 caller7_stack # qhasm: enter CRYPTO_SHARED_NAMESPACE(work_cswap) .text .p2align 5 ASM_HIDDEN _CRYPTO_SHARED_NAMESPACE(work_cswap) .globl _CRYPTO_SHARED_NAMESPACE(work_cswap) ASM_HIDDEN CRYPTO_SHARED_NAMESPACE(work_cswap) .globl CRYPTO_SHARED_NAMESPACE(work_cswap) _CRYPTO_SHARED_NAMESPACE(work_cswap): CRYPTO_SHARED_NAMESPACE(work_cswap): mov %rsp,%r11 and $31,%r11 add $0,%r11 sub %r11,%rsp # qhasm: =? swap - 1 # asm 1: cmp $1,<swap=int64#2 # asm 2: cmp $1,<swap=%rsi cmp $1,%rsi # qhasm: w0 = *(uint64 *)(workp + 0) # asm 1: movq 0(<workp=int64#1),>w0=int64#2 # asm 2: movq 0(<workp=%rdi),>w0=%rsi movq 0(%rdi),%rsi # qhasm: w8 = *(uint64 *)(workp + 64) # asm 1: movq 64(<workp=int64#1),>w8=int64#3 # asm 2: movq 64(<workp=%rdi),>w8=%rdx movq 64(%rdi),%rdx # qhasm: t = w0 # asm 1: mov <w0=int64#2,>t=int64#4 # asm 2: mov <w0=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w0 = w8 if = # asm 1: cmove <w8=int64#3,<w0=int64#2 # asm 2: cmove <w8=%rdx,<w0=%rsi cmove %rdx,%rsi # qhasm: w8 = t if = # asm 1: cmove <t=int64#4,<w8=int64#3 # asm 2: cmove <t=%rcx,<w8=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 0) = w0 # asm 1: movq <w0=int64#2,0(<workp=int64#1) # asm 2: movq <w0=%rsi,0(<workp=%rdi) movq %rsi,0(%rdi) # qhasm: *(uint64 *)(workp + 64) = w8 # asm 1: movq <w8=int64#3,64(<workp=int64#1) # asm 2: movq <w8=%rdx,64(<workp=%rdi) movq %rdx,64(%rdi) # qhasm: w1 = *(uint64 *)(workp + 8) # asm 1: movq 8(<workp=int64#1),>w1=int64#2 # asm 2: movq 8(<workp=%rdi),>w1=%rsi movq 8(%rdi),%rsi # qhasm: w9 = *(uint64 *)(workp + 72) # asm 1: movq 72(<workp=int64#1),>w9=int64#3 # asm 2: movq 72(<workp=%rdi),>w9=%rdx movq 72(%rdi),%rdx # qhasm: t = w1 # asm 1: mov <w1=int64#2,>t=int64#4 # asm 2: mov <w1=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w1 = w9 if = # asm 1: cmove <w9=int64#3,<w1=int64#2 # asm 2: cmove <w9=%rdx,<w1=%rsi cmove %rdx,%rsi # qhasm: w9 = t if = # asm 1: cmove <t=int64#4,<w9=int64#3 # asm 2: cmove <t=%rcx,<w9=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 8) = w1 # asm 1: movq <w1=int64#2,8(<workp=int64#1) # asm 2: movq <w1=%rsi,8(<workp=%rdi) movq %rsi,8(%rdi) # qhasm: *(uint64 *)(workp + 72) = w9 # asm 1: movq <w9=int64#3,72(<workp=int64#1) # asm 2: movq <w9=%rdx,72(<workp=%rdi) movq %rdx,72(%rdi) # qhasm: w2 = *(uint64 *)(workp + 16) # asm 1: movq 16(<workp=int64#1),>w2=int64#2 # asm 2: movq 16(<workp=%rdi),>w2=%rsi movq 16(%rdi),%rsi # qhasm: w10 = *(uint64 *)(workp + 80) # asm 1: movq 80(<workp=int64#1),>w10=int64#3 # asm 2: movq 80(<workp=%rdi),>w10=%rdx movq 80(%rdi),%rdx # qhasm: t = w2 # asm 1: mov <w2=int64#2,>t=int64#4 # asm 2: mov <w2=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w2 = w10 if = # asm 1: cmove <w10=int64#3,<w2=int64#2 # asm 2: cmove <w10=%rdx,<w2=%rsi cmove %rdx,%rsi # qhasm: w10 = t if = # asm 1: cmove <t=int64#4,<w10=int64#3 # asm 2: cmove <t=%rcx,<w10=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 16) = w2 # asm 1: movq <w2=int64#2,16(<workp=int64#1) # asm 2: movq <w2=%rsi,16(<workp=%rdi) movq %rsi,16(%rdi) # qhasm: *(uint64 *)(workp + 80) = w10 # asm 1: movq <w10=int64#3,80(<workp=int64#1) # asm 2: movq <w10=%rdx,80(<workp=%rdi) movq %rdx,80(%rdi) # qhasm: w3 = *(uint64 *)(workp + 24) # asm 1: movq 24(<workp=int64#1),>w3=int64#2 # asm 2: movq 24(<workp=%rdi),>w3=%rsi movq 24(%rdi),%rsi # qhasm: w11 = *(uint64 *)(workp + 88) # asm 1: movq 88(<workp=int64#1),>w11=int64#3 # asm 2: movq 88(<workp=%rdi),>w11=%rdx movq 88(%rdi),%rdx # qhasm: t = w3 # asm 1: mov <w3=int64#2,>t=int64#4 # asm 2: mov <w3=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w3 = w11 if = # asm 1: cmove <w11=int64#3,<w3=int64#2 # asm 2: cmove <w11=%rdx,<w3=%rsi cmove %rdx,%rsi # qhasm: w11 = t if = # asm 1: cmove <t=int64#4,<w11=int64#3 # asm 2: cmove <t=%rcx,<w11=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 24) = w3 # asm 1: movq <w3=int64#2,24(<workp=int64#1) # asm 2: movq <w3=%rsi,24(<workp=%rdi) movq %rsi,24(%rdi) # qhasm: *(uint64 *)(workp + 88) = w11 # asm 1: movq <w11=int64#3,88(<workp=int64#1) # asm 2: movq <w11=%rdx,88(<workp=%rdi) movq %rdx,88(%rdi) # qhasm: w4 = *(uint64 *)(workp + 32) # asm 1: movq 32(<workp=int64#1),>w4=int64#2 # asm 2: movq 32(<workp=%rdi),>w4=%rsi movq 32(%rdi),%rsi # qhasm: w12 = *(uint64 *)(workp + 96) # asm 1: movq 96(<workp=int64#1),>w12=int64#3 # asm 2: movq 96(<workp=%rdi),>w12=%rdx movq 96(%rdi),%rdx # qhasm: t = w4 # asm 1: mov <w4=int64#2,>t=int64#4 # asm 2: mov <w4=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w4 = w12 if = # asm 1: cmove <w12=int64#3,<w4=int64#2 # asm 2: cmove <w12=%rdx,<w4=%rsi cmove %rdx,%rsi # qhasm: w12 = t if = # asm 1: cmove <t=int64#4,<w12=int64#3 # asm 2: cmove <t=%rcx,<w12=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 32) = w4 # asm 1: movq <w4=int64#2,32(<workp=int64#1) # asm 2: movq <w4=%rsi,32(<workp=%rdi) movq %rsi,32(%rdi) # qhasm: *(uint64 *)(workp + 96) = w12 # asm 1: movq <w12=int64#3,96(<workp=int64#1) # asm 2: movq <w12=%rdx,96(<workp=%rdi) movq %rdx,96(%rdi) # qhasm: w5 = *(uint64 *)(workp + 40) # asm 1: movq 40(<workp=int64#1),>w5=int64#2 # asm 2: movq 40(<workp=%rdi),>w5=%rsi movq 40(%rdi),%rsi # qhasm: w13 = *(uint64 *)(workp + 104) # asm 1: movq 104(<workp=int64#1),>w13=int64#3 # asm 2: movq 104(<workp=%rdi),>w13=%rdx movq 104(%rdi),%rdx # qhasm: t = w5 # asm 1: mov <w5=int64#2,>t=int64#4 # asm 2: mov <w5=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w5 = w13 if = # asm 1: cmove <w13=int64#3,<w5=int64#2 # asm 2: cmove <w13=%rdx,<w5=%rsi cmove %rdx,%rsi # qhasm: w13 = t if = # asm 1: cmove <t=int64#4,<w13=int64#3 # asm 2: cmove <t=%rcx,<w13=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 40) = w5 # asm 1: movq <w5=int64#2,40(<workp=int64#1) # asm 2: movq <w5=%rsi,40(<workp=%rdi) movq %rsi,40(%rdi) # qhasm: *(uint64 *)(workp + 104) = w13 # asm 1: movq <w13=int64#3,104(<workp=int64#1) # asm 2: movq <w13=%rdx,104(<workp=%rdi) movq %rdx,104(%rdi) # qhasm: w6 = *(uint64 *)(workp + 48) # asm 1: movq 48(<workp=int64#1),>w6=int64#2 # asm 2: movq 48(<workp=%rdi),>w6=%rsi movq 48(%rdi),%rsi # qhasm: w14 = *(uint64 *)(workp + 112) # asm 1: movq 112(<workp=int64#1),>w14=int64#3 # asm 2: movq 112(<workp=%rdi),>w14=%rdx movq 112(%rdi),%rdx # qhasm: t = w6 # asm 1: mov <w6=int64#2,>t=int64#4 # asm 2: mov <w6=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w6 = w14 if = # asm 1: cmove <w14=int64#3,<w6=int64#2 # asm 2: cmove <w14=%rdx,<w6=%rsi cmove %rdx,%rsi # qhasm: w14 = t if = # asm 1: cmove <t=int64#4,<w14=int64#3 # asm 2: cmove <t=%rcx,<w14=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 48) = w6 # asm 1: movq <w6=int64#2,48(<workp=int64#1) # asm 2: movq <w6=%rsi,48(<workp=%rdi) movq %rsi,48(%rdi) # qhasm: *(uint64 *)(workp + 112) = w14 # asm 1: movq <w14=int64#3,112(<workp=int64#1) # asm 2: movq <w14=%rdx,112(<workp=%rdi) movq %rdx,112(%rdi) # qhasm: w7 = *(uint64 *)(workp + 56) # asm 1: movq 56(<workp=int64#1),>w7=int64#2 # asm 2: movq 56(<workp=%rdi),>w7=%rsi movq 56(%rdi),%rsi # qhasm: w15 = *(uint64 *)(workp + 120) # asm 1: movq 120(<workp=int64#1),>w15=int64#3 # asm 2: movq 120(<workp=%rdi),>w15=%rdx movq 120(%rdi),%rdx # qhasm: t = w7 # asm 1: mov <w7=int64#2,>t=int64#4 # asm 2: mov <w7=%rsi,>t=%rcx mov %rsi,%rcx # qhasm: w7 = w15 if = # asm 1: cmove <w15=int64#3,<w7=int64#2 # asm 2: cmove <w15=%rdx,<w7=%rsi cmove %rdx,%rsi # qhasm: w15 = t if = # asm 1: cmove <t=int64#4,<w15=int64#3 # asm 2: cmove <t=%rcx,<w15=%rdx cmove %rcx,%rdx # qhasm: *(uint64 *)(workp + 56) = w7 # asm 1: movq <w7=int64#2,56(<workp=int64#1) # asm 2: movq <w7=%rsi,56(<workp=%rdi) movq %rsi,56(%rdi) # qhasm: *(uint64 *)(workp + 120) = w15 # asm 1: movq <w15=int64#3,120(<workp=int64#1) # asm 2: movq <w15=%rdx,120(<workp=%rdi) movq %rdx,120(%rdi) # qhasm: leave add %r11,%rsp mov %rdi,%rax mov %rsi,%rdx ret