summaryrefslogtreecommitdiff
path: root/secure/lib/libcrypto/amd64/x25519-x86_64.S
diff options
context:
space:
mode:
Diffstat (limited to 'secure/lib/libcrypto/amd64/x25519-x86_64.S')
-rw-r--r--secure/lib/libcrypto/amd64/x25519-x86_64.S804
1 files changed, 0 insertions, 804 deletions
diff --git a/secure/lib/libcrypto/amd64/x25519-x86_64.S b/secure/lib/libcrypto/amd64/x25519-x86_64.S
deleted file mode 100644
index 7448e866aaf6d..0000000000000
--- a/secure/lib/libcrypto/amd64/x25519-x86_64.S
+++ /dev/null
@@ -1,804 +0,0 @@
-/* $FreeBSD$ */
-/* Do not modify. This file is auto-generated from x25519-x86_64.pl. */
-.text
-
-.globl x25519_fe51_mul
-.type x25519_fe51_mul,@function
-.align 32
-x25519_fe51_mul:
-.cfi_startproc
- pushq %rbp
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbp,-16
- pushq %rbx
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbx,-24
- pushq %r12
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r12,-32
- pushq %r13
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r13,-40
- pushq %r14
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r14,-48
- pushq %r15
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r15,-56
- leaq -40(%rsp),%rsp
-.cfi_adjust_cfa_offset 40
-.Lfe51_mul_body:
-
- movq 0(%rsi),%rax
- movq 0(%rdx),%r11
- movq 8(%rdx),%r12
- movq 16(%rdx),%r13
- movq 24(%rdx),%rbp
- movq 32(%rdx),%r14
-
- movq %rdi,32(%rsp)
- movq %rax,%rdi
- mulq %r11
- movq %r11,0(%rsp)
- movq %rax,%rbx
- movq %rdi,%rax
- movq %rdx,%rcx
- mulq %r12
- movq %r12,8(%rsp)
- movq %rax,%r8
- movq %rdi,%rax
- leaq (%r14,%r14,8),%r15
- movq %rdx,%r9
- mulq %r13
- movq %r13,16(%rsp)
- movq %rax,%r10
- movq %rdi,%rax
- leaq (%r14,%r15,2),%rdi
- movq %rdx,%r11
- mulq %rbp
- movq %rax,%r12
- movq 0(%rsi),%rax
- movq %rdx,%r13
- mulq %r14
- movq %rax,%r14
- movq 8(%rsi),%rax
- movq %rdx,%r15
-
- mulq %rdi
- addq %rax,%rbx
- movq 16(%rsi),%rax
- adcq %rdx,%rcx
- mulq %rdi
- addq %rax,%r8
- movq 24(%rsi),%rax
- adcq %rdx,%r9
- mulq %rdi
- addq %rax,%r10
- movq 32(%rsi),%rax
- adcq %rdx,%r11
- mulq %rdi
- imulq $19,%rbp,%rdi
- addq %rax,%r12
- movq 8(%rsi),%rax
- adcq %rdx,%r13
- mulq %rbp
- movq 16(%rsp),%rbp
- addq %rax,%r14
- movq 16(%rsi),%rax
- adcq %rdx,%r15
-
- mulq %rdi
- addq %rax,%rbx
- movq 24(%rsi),%rax
- adcq %rdx,%rcx
- mulq %rdi
- addq %rax,%r8
- movq 32(%rsi),%rax
- adcq %rdx,%r9
- mulq %rdi
- imulq $19,%rbp,%rdi
- addq %rax,%r10
- movq 8(%rsi),%rax
- adcq %rdx,%r11
- mulq %rbp
- addq %rax,%r12
- movq 16(%rsi),%rax
- adcq %rdx,%r13
- mulq %rbp
- movq 8(%rsp),%rbp
- addq %rax,%r14
- movq 24(%rsi),%rax
- adcq %rdx,%r15
-
- mulq %rdi
- addq %rax,%rbx
- movq 32(%rsi),%rax
- adcq %rdx,%rcx
- mulq %rdi
- addq %rax,%r8
- movq 8(%rsi),%rax
- adcq %rdx,%r9
- mulq %rbp
- imulq $19,%rbp,%rdi
- addq %rax,%r10
- movq 16(%rsi),%rax
- adcq %rdx,%r11
- mulq %rbp
- addq %rax,%r12
- movq 24(%rsi),%rax
- adcq %rdx,%r13
- mulq %rbp
- movq 0(%rsp),%rbp
- addq %rax,%r14
- movq 32(%rsi),%rax
- adcq %rdx,%r15
-
- mulq %rdi
- addq %rax,%rbx
- movq 8(%rsi),%rax
- adcq %rdx,%rcx
- mulq %rbp
- addq %rax,%r8
- movq 16(%rsi),%rax
- adcq %rdx,%r9
- mulq %rbp
- addq %rax,%r10
- movq 24(%rsi),%rax
- adcq %rdx,%r11
- mulq %rbp
- addq %rax,%r12
- movq 32(%rsi),%rax
- adcq %rdx,%r13
- mulq %rbp
- addq %rax,%r14
- adcq %rdx,%r15
-
- movq 32(%rsp),%rdi
- jmp .Lreduce51
-.Lfe51_mul_epilogue:
-.cfi_endproc
-.size x25519_fe51_mul,.-x25519_fe51_mul
-
-.globl x25519_fe51_sqr
-.type x25519_fe51_sqr,@function
-.align 32
-x25519_fe51_sqr:
-.cfi_startproc
- pushq %rbp
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbp,-16
- pushq %rbx
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbx,-24
- pushq %r12
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r12,-32
- pushq %r13
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r13,-40
- pushq %r14
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r14,-48
- pushq %r15
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r15,-56
- leaq -40(%rsp),%rsp
-.cfi_adjust_cfa_offset 40
-.Lfe51_sqr_body:
-
- movq 0(%rsi),%rax
- movq 16(%rsi),%r15
- movq 32(%rsi),%rbp
-
- movq %rdi,32(%rsp)
- leaq (%rax,%rax,1),%r14
- mulq %rax
- movq %rax,%rbx
- movq 8(%rsi),%rax
- movq %rdx,%rcx
- mulq %r14
- movq %rax,%r8
- movq %r15,%rax
- movq %r15,0(%rsp)
- movq %rdx,%r9
- mulq %r14
- movq %rax,%r10
- movq 24(%rsi),%rax
- movq %rdx,%r11
- imulq $19,%rbp,%rdi
- mulq %r14
- movq %rax,%r12
- movq %rbp,%rax
- movq %rdx,%r13
- mulq %r14
- movq %rax,%r14
- movq %rbp,%rax
- movq %rdx,%r15
-
- mulq %rdi
- addq %rax,%r12
- movq 8(%rsi),%rax
- adcq %rdx,%r13
-
- movq 24(%rsi),%rsi
- leaq (%rax,%rax,1),%rbp
- mulq %rax
- addq %rax,%r10
- movq 0(%rsp),%rax
- adcq %rdx,%r11
- mulq %rbp
- addq %rax,%r12
- movq %rbp,%rax
- adcq %rdx,%r13
- mulq %rsi
- addq %rax,%r14
- movq %rbp,%rax
- adcq %rdx,%r15
- imulq $19,%rsi,%rbp
- mulq %rdi
- addq %rax,%rbx
- leaq (%rsi,%rsi,1),%rax
- adcq %rdx,%rcx
-
- mulq %rdi
- addq %rax,%r10
- movq %rsi,%rax
- adcq %rdx,%r11
- mulq %rbp
- addq %rax,%r8
- movq 0(%rsp),%rax
- adcq %rdx,%r9
-
- leaq (%rax,%rax,1),%rsi
- mulq %rax
- addq %rax,%r14
- movq %rbp,%rax
- adcq %rdx,%r15
- mulq %rsi
- addq %rax,%rbx
- movq %rsi,%rax
- adcq %rdx,%rcx
- mulq %rdi
- addq %rax,%r8
- adcq %rdx,%r9
-
- movq 32(%rsp),%rdi
- jmp .Lreduce51
-
-.align 32
-.Lreduce51:
- movq $0x7ffffffffffff,%rbp
-
- movq %r10,%rdx
- shrq $51,%r10
- shlq $13,%r11
- andq %rbp,%rdx
- orq %r10,%r11
- addq %r11,%r12
- adcq $0,%r13
-
- movq %rbx,%rax
- shrq $51,%rbx
- shlq $13,%rcx
- andq %rbp,%rax
- orq %rbx,%rcx
- addq %rcx,%r8
- adcq $0,%r9
-
- movq %r12,%rbx
- shrq $51,%r12
- shlq $13,%r13
- andq %rbp,%rbx
- orq %r12,%r13
- addq %r13,%r14
- adcq $0,%r15
-
- movq %r8,%rcx
- shrq $51,%r8
- shlq $13,%r9
- andq %rbp,%rcx
- orq %r8,%r9
- addq %r9,%rdx
-
- movq %r14,%r10
- shrq $51,%r14
- shlq $13,%r15
- andq %rbp,%r10
- orq %r14,%r15
-
- leaq (%r15,%r15,8),%r14
- leaq (%r15,%r14,2),%r15
- addq %r15,%rax
-
- movq %rdx,%r8
- andq %rbp,%rdx
- shrq $51,%r8
- addq %r8,%rbx
-
- movq %rax,%r9
- andq %rbp,%rax
- shrq $51,%r9
- addq %r9,%rcx
-
- movq %rax,0(%rdi)
- movq %rcx,8(%rdi)
- movq %rdx,16(%rdi)
- movq %rbx,24(%rdi)
- movq %r10,32(%rdi)
-
- movq 40(%rsp),%r15
-.cfi_restore %r15
- movq 48(%rsp),%r14
-.cfi_restore %r14
- movq 56(%rsp),%r13
-.cfi_restore %r13
- movq 64(%rsp),%r12
-.cfi_restore %r12
- movq 72(%rsp),%rbx
-.cfi_restore %rbx
- movq 80(%rsp),%rbp
-.cfi_restore %rbp
- leaq 88(%rsp),%rsp
-.cfi_adjust_cfa_offset 88
-.Lfe51_sqr_epilogue:
- .byte 0xf3,0xc3
-.cfi_endproc
-.size x25519_fe51_sqr,.-x25519_fe51_sqr
-
-.globl x25519_fe51_mul121666
-.type x25519_fe51_mul121666,@function
-.align 32
-x25519_fe51_mul121666:
-.cfi_startproc
- pushq %rbp
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbp,-16
- pushq %rbx
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbx,-24
- pushq %r12
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r12,-32
- pushq %r13
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r13,-40
- pushq %r14
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r14,-48
- pushq %r15
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r15,-56
- leaq -40(%rsp),%rsp
-.cfi_adjust_cfa_offset 40
-.Lfe51_mul121666_body:
- movl $121666,%eax
-
- mulq 0(%rsi)
- movq %rax,%rbx
- movl $121666,%eax
- movq %rdx,%rcx
- mulq 8(%rsi)
- movq %rax,%r8
- movl $121666,%eax
- movq %rdx,%r9
- mulq 16(%rsi)
- movq %rax,%r10
- movl $121666,%eax
- movq %rdx,%r11
- mulq 24(%rsi)
- movq %rax,%r12
- movl $121666,%eax
- movq %rdx,%r13
- mulq 32(%rsi)
- movq %rax,%r14
- movq %rdx,%r15
-
- jmp .Lreduce51
-.Lfe51_mul121666_epilogue:
-.cfi_endproc
-.size x25519_fe51_mul121666,.-x25519_fe51_mul121666
-
-.globl x25519_fe64_eligible
-.type x25519_fe64_eligible,@function
-.align 32
-x25519_fe64_eligible:
-.cfi_startproc
- movl OPENSSL_ia32cap_P+8(%rip),%ecx
- xorl %eax,%eax
- andl $0x80100,%ecx
- cmpl $0x80100,%ecx
- cmovel %ecx,%eax
- .byte 0xf3,0xc3
-.cfi_endproc
-.size x25519_fe64_eligible,.-x25519_fe64_eligible
-
-.globl x25519_fe64_mul
-.type x25519_fe64_mul,@function
-.align 32
-x25519_fe64_mul:
-.cfi_startproc
- pushq %rbp
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbp,-16
- pushq %rbx
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbx,-24
- pushq %r12
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r12,-32
- pushq %r13
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r13,-40
- pushq %r14
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r14,-48
- pushq %r15
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r15,-56
- pushq %rdi
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rdi,-64
- leaq -16(%rsp),%rsp
-.cfi_adjust_cfa_offset 16
-.Lfe64_mul_body:
-
- movq %rdx,%rax
- movq 0(%rdx),%rbp
- movq 0(%rsi),%rdx
- movq 8(%rax),%rcx
- movq 16(%rax),%r14
- movq 24(%rax),%r15
-
- mulxq %rbp,%r8,%rax
- xorl %edi,%edi
- mulxq %rcx,%r9,%rbx
- adcxq %rax,%r9
- mulxq %r14,%r10,%rax
- adcxq %rbx,%r10
- mulxq %r15,%r11,%r12
- movq 8(%rsi),%rdx
- adcxq %rax,%r11
- movq %r14,(%rsp)
- adcxq %rdi,%r12
-
- mulxq %rbp,%rax,%rbx
- adoxq %rax,%r9
- adcxq %rbx,%r10
- mulxq %rcx,%rax,%rbx
- adoxq %rax,%r10
- adcxq %rbx,%r11
- mulxq %r14,%rax,%rbx
- adoxq %rax,%r11
- adcxq %rbx,%r12
- mulxq %r15,%rax,%r13
- movq 16(%rsi),%rdx
- adoxq %rax,%r12
- adcxq %rdi,%r13
- adoxq %rdi,%r13
-
- mulxq %rbp,%rax,%rbx
- adcxq %rax,%r10
- adoxq %rbx,%r11
- mulxq %rcx,%rax,%rbx
- adcxq %rax,%r11
- adoxq %rbx,%r12
- mulxq %r14,%rax,%rbx
- adcxq %rax,%r12
- adoxq %rbx,%r13
- mulxq %r15,%rax,%r14
- movq 24(%rsi),%rdx
- adcxq %rax,%r13
- adoxq %rdi,%r14
- adcxq %rdi,%r14
-
- mulxq %rbp,%rax,%rbx
- adoxq %rax,%r11
- adcxq %rbx,%r12
- mulxq %rcx,%rax,%rbx
- adoxq %rax,%r12
- adcxq %rbx,%r13
- mulxq (%rsp),%rax,%rbx
- adoxq %rax,%r13
- adcxq %rbx,%r14
- mulxq %r15,%rax,%r15
- movl $38,%edx
- adoxq %rax,%r14
- adcxq %rdi,%r15
- adoxq %rdi,%r15
-
- jmp .Lreduce64
-.Lfe64_mul_epilogue:
-.cfi_endproc
-.size x25519_fe64_mul,.-x25519_fe64_mul
-
-.globl x25519_fe64_sqr
-.type x25519_fe64_sqr,@function
-.align 32
-x25519_fe64_sqr:
-.cfi_startproc
- pushq %rbp
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbp,-16
- pushq %rbx
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rbx,-24
- pushq %r12
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r12,-32
- pushq %r13
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r13,-40
- pushq %r14
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r14,-48
- pushq %r15
-.cfi_adjust_cfa_offset 8
-.cfi_offset %r15,-56
- pushq %rdi
-.cfi_adjust_cfa_offset 8
-.cfi_offset %rdi,-64
- leaq -16(%rsp),%rsp
-.cfi_adjust_cfa_offset 16
-.Lfe64_sqr_body:
-
- movq 0(%rsi),%rdx
- movq 8(%rsi),%rcx
- movq 16(%rsi),%rbp
- movq 24(%rsi),%rsi
-
-
- mulxq %rdx,%r8,%r15
- mulxq %rcx,%r9,%rax
- xorl %edi,%edi
- mulxq %rbp,%r10,%rbx
- adcxq %rax,%r10
- mulxq %rsi,%r11,%r12
- movq %rcx,%rdx
- adcxq %rbx,%r11
- adcxq %rdi,%r12
-
-
- mulxq %rbp,%rax,%rbx
- adoxq %rax,%r11
- adcxq %rbx,%r12
- mulxq %rsi,%rax,%r13
- movq %rbp,%rdx
- adoxq %rax,%r12
- adcxq %rdi,%r13
-
-
- mulxq %rsi,%rax,%r14
- movq %rcx,%rdx
- adoxq %rax,%r13
- adcxq %rdi,%r14
- adoxq %rdi,%r14
-
- adcxq %r9,%r9
- adoxq %r15,%r9
- adcxq %r10,%r10
- mulxq %rdx,%rax,%rbx
- movq %rbp,%rdx
- adcxq %r11,%r11
- adoxq %rax,%r10
- adcxq %r12,%r12
- adoxq %rbx,%r11
- mulxq %rdx,%rax,%rbx
- movq %rsi,%rdx
- adcxq %r13,%r13
- adoxq %rax,%r12
- adcxq %r14,%r14
- adoxq %rbx,%r13
- mulxq %rdx,%rax,%r15
- movl $38,%edx
- adoxq %rax,%r14
- adcxq %rdi,%r15
- adoxq %rdi,%r15
- jmp .Lreduce64
-
-.align 32
-.Lreduce64:
- mulxq %r12,%rax,%rbx
- adcxq %rax,%r8
- adoxq %rbx,%r9
- mulxq %r13,%rax,%rbx
- adcxq %rax,%r9
- adoxq %rbx,%r10
- mulxq %r14,%rax,%rbx
- adcxq %rax,%r10
- adoxq %rbx,%r11
- mulxq %r15,%rax,%r12
- adcxq %rax,%r11
- adoxq %rdi,%r12
- adcxq %rdi,%r12
-
- movq 16(%rsp),%rdi
- imulq %rdx,%r12
-
- addq %r12,%r8
- adcq $0,%r9
- adcq $0,%r10
- adcq $0,%r11
-
- sbbq %rax,%rax
- andq $38,%rax
-
- addq %rax,%r8
- movq %r9,8(%rdi)
- movq %r10,16(%rdi)
- movq %r11,24(%rdi)
- movq %r8,0(%rdi)
-
- movq 24(%rsp),%r15
-.cfi_restore %r15
- movq 32(%rsp),%r14
-.cfi_restore %r14
- movq 40(%rsp),%r13
-.cfi_restore %r13
- movq 48(%rsp),%r12
-.cfi_restore %r12
- movq 56(%rsp),%rbx
-.cfi_restore %rbx
- movq 64(%rsp),%rbp
-.cfi_restore %rbp
- leaq 72(%rsp),%rsp
-.cfi_adjust_cfa_offset 88
-.Lfe64_sqr_epilogue:
- .byte 0xf3,0xc3
-.cfi_endproc
-.size x25519_fe64_sqr,.-x25519_fe64_sqr
-
-.globl x25519_fe64_mul121666
-.type x25519_fe64_mul121666,@function
-.align 32
-x25519_fe64_mul121666:
-.Lfe64_mul121666_body:
-.cfi_startproc
- movl $121666,%edx
- mulxq 0(%rsi),%r8,%rcx
- mulxq 8(%rsi),%r9,%rax
- addq %rcx,%r9
- mulxq 16(%rsi),%r10,%rcx
- adcq %rax,%r10
- mulxq 24(%rsi),%r11,%rax
- adcq %rcx,%r11
- adcq $0,%rax
-
- imulq $38,%rax,%rax
-
- addq %rax,%r8
- adcq $0,%r9
- adcq $0,%r10
- adcq $0,%r11
-
- sbbq %rax,%rax
- andq $38,%rax
-
- addq %rax,%r8
- movq %r9,8(%rdi)
- movq %r10,16(%rdi)
- movq %r11,24(%rdi)
- movq %r8,0(%rdi)
-
-.Lfe64_mul121666_epilogue:
- .byte 0xf3,0xc3
-.cfi_endproc
-.size x25519_fe64_mul121666,.-x25519_fe64_mul121666
-
-.globl x25519_fe64_add
-.type x25519_fe64_add,@function
-.align 32
-x25519_fe64_add:
-.Lfe64_add_body:
-.cfi_startproc
- movq 0(%rsi),%r8
- movq 8(%rsi),%r9
- movq 16(%rsi),%r10
- movq 24(%rsi),%r11
-
- addq 0(%rdx),%r8
- adcq 8(%rdx),%r9
- adcq 16(%rdx),%r10
- adcq 24(%rdx),%r11
-
- sbbq %rax,%rax
- andq $38,%rax
-
- addq %rax,%r8
- adcq $0,%r9
- adcq $0,%r10
- movq %r9,8(%rdi)
- adcq $0,%r11
- movq %r10,16(%rdi)
- sbbq %rax,%rax
- movq %r11,24(%rdi)
- andq $38,%rax
-
- addq %rax,%r8
- movq %r8,0(%rdi)
-
-.Lfe64_add_epilogue:
- .byte 0xf3,0xc3
-.cfi_endproc
-.size x25519_fe64_add,.-x25519_fe64_add
-
-.globl x25519_fe64_sub
-.type x25519_fe64_sub,@function
-.align 32
-x25519_fe64_sub:
-.Lfe64_sub_body:
-.cfi_startproc
- movq 0(%rsi),%r8
- movq 8(%rsi),%r9
- movq 16(%rsi),%r10
- movq 24(%rsi),%r11
-
- subq 0(%rdx),%r8
- sbbq 8(%rdx),%r9
- sbbq 16(%rdx),%r10
- sbbq 24(%rdx),%r11
-
- sbbq %rax,%rax
- andq $38,%rax
-
- subq %rax,%r8
- sbbq $0,%r9
- sbbq $0,%r10
- movq %r9,8(%rdi)
- sbbq $0,%r11
- movq %r10,16(%rdi)
- sbbq %rax,%rax
- movq %r11,24(%rdi)
- andq $38,%rax
-
- subq %rax,%r8
- movq %r8,0(%rdi)
-
-.Lfe64_sub_epilogue:
- .byte 0xf3,0xc3
-.cfi_endproc
-.size x25519_fe64_sub,.-x25519_fe64_sub
-
-.globl x25519_fe64_tobytes
-.type x25519_fe64_tobytes,@function
-.align 32
-x25519_fe64_tobytes:
-.Lfe64_to_body:
-.cfi_startproc
- movq 0(%rsi),%r8
- movq 8(%rsi),%r9
- movq 16(%rsi),%r10
- movq 24(%rsi),%r11
-
-
- leaq (%r11,%r11,1),%rax
- sarq $63,%r11
- shrq $1,%rax
- andq $19,%r11
- addq $19,%r11
-
- addq %r11,%r8
- adcq $0,%r9
- adcq $0,%r10
- adcq $0,%rax
-
- leaq (%rax,%rax,1),%r11
- sarq $63,%rax
- shrq $1,%r11
- notq %rax
- andq $19,%rax
-
- subq %rax,%r8
- sbbq $0,%r9
- sbbq $0,%r10
- sbbq $0,%r11
-
- movq %r8,0(%rdi)
- movq %r9,8(%rdi)
- movq %r10,16(%rdi)
- movq %r11,24(%rdi)
-
-.Lfe64_to_epilogue:
- .byte 0xf3,0xc3
-.cfi_endproc
-.size x25519_fe64_tobytes,.-x25519_fe64_tobytes
-.byte 88,50,53,53,49,57,32,112,114,105,109,105,116,105,118,101,115,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0