diff options
Diffstat (limited to 'secure/lib/libcrypto/amd64/x25519-x86_64.S')
-rw-r--r-- | secure/lib/libcrypto/amd64/x25519-x86_64.S | 804 |
1 files changed, 0 insertions, 804 deletions
diff --git a/secure/lib/libcrypto/amd64/x25519-x86_64.S b/secure/lib/libcrypto/amd64/x25519-x86_64.S deleted file mode 100644 index 7448e866aaf6d..0000000000000 --- a/secure/lib/libcrypto/amd64/x25519-x86_64.S +++ /dev/null @@ -1,804 +0,0 @@ -/* $FreeBSD$ */ -/* Do not modify. This file is auto-generated from x25519-x86_64.pl. */ -.text - -.globl x25519_fe51_mul -.type x25519_fe51_mul,@function -.align 32 -x25519_fe51_mul: -.cfi_startproc - pushq %rbp -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbp,-16 - pushq %rbx -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbx,-24 - pushq %r12 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r12,-32 - pushq %r13 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r13,-40 - pushq %r14 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r14,-48 - pushq %r15 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r15,-56 - leaq -40(%rsp),%rsp -.cfi_adjust_cfa_offset 40 -.Lfe51_mul_body: - - movq 0(%rsi),%rax - movq 0(%rdx),%r11 - movq 8(%rdx),%r12 - movq 16(%rdx),%r13 - movq 24(%rdx),%rbp - movq 32(%rdx),%r14 - - movq %rdi,32(%rsp) - movq %rax,%rdi - mulq %r11 - movq %r11,0(%rsp) - movq %rax,%rbx - movq %rdi,%rax - movq %rdx,%rcx - mulq %r12 - movq %r12,8(%rsp) - movq %rax,%r8 - movq %rdi,%rax - leaq (%r14,%r14,8),%r15 - movq %rdx,%r9 - mulq %r13 - movq %r13,16(%rsp) - movq %rax,%r10 - movq %rdi,%rax - leaq (%r14,%r15,2),%rdi - movq %rdx,%r11 - mulq %rbp - movq %rax,%r12 - movq 0(%rsi),%rax - movq %rdx,%r13 - mulq %r14 - movq %rax,%r14 - movq 8(%rsi),%rax - movq %rdx,%r15 - - mulq %rdi - addq %rax,%rbx - movq 16(%rsi),%rax - adcq %rdx,%rcx - mulq %rdi - addq %rax,%r8 - movq 24(%rsi),%rax - adcq %rdx,%r9 - mulq %rdi - addq %rax,%r10 - movq 32(%rsi),%rax - adcq %rdx,%r11 - mulq %rdi - imulq $19,%rbp,%rdi - addq %rax,%r12 - movq 8(%rsi),%rax - adcq %rdx,%r13 - mulq %rbp - movq 16(%rsp),%rbp - addq %rax,%r14 - movq 16(%rsi),%rax - adcq %rdx,%r15 - - mulq %rdi - addq %rax,%rbx - movq 24(%rsi),%rax - adcq %rdx,%rcx - mulq %rdi - addq %rax,%r8 - movq 32(%rsi),%rax - adcq %rdx,%r9 - mulq %rdi - imulq $19,%rbp,%rdi - addq %rax,%r10 - movq 8(%rsi),%rax - adcq %rdx,%r11 - mulq %rbp - addq %rax,%r12 - movq 16(%rsi),%rax - adcq %rdx,%r13 - mulq %rbp - movq 8(%rsp),%rbp - addq %rax,%r14 - movq 24(%rsi),%rax - adcq %rdx,%r15 - - mulq %rdi - addq %rax,%rbx - movq 32(%rsi),%rax - adcq %rdx,%rcx - mulq %rdi - addq %rax,%r8 - movq 8(%rsi),%rax - adcq %rdx,%r9 - mulq %rbp - imulq $19,%rbp,%rdi - addq %rax,%r10 - movq 16(%rsi),%rax - adcq %rdx,%r11 - mulq %rbp - addq %rax,%r12 - movq 24(%rsi),%rax - adcq %rdx,%r13 - mulq %rbp - movq 0(%rsp),%rbp - addq %rax,%r14 - movq 32(%rsi),%rax - adcq %rdx,%r15 - - mulq %rdi - addq %rax,%rbx - movq 8(%rsi),%rax - adcq %rdx,%rcx - mulq %rbp - addq %rax,%r8 - movq 16(%rsi),%rax - adcq %rdx,%r9 - mulq %rbp - addq %rax,%r10 - movq 24(%rsi),%rax - adcq %rdx,%r11 - mulq %rbp - addq %rax,%r12 - movq 32(%rsi),%rax - adcq %rdx,%r13 - mulq %rbp - addq %rax,%r14 - adcq %rdx,%r15 - - movq 32(%rsp),%rdi - jmp .Lreduce51 -.Lfe51_mul_epilogue: -.cfi_endproc -.size x25519_fe51_mul,.-x25519_fe51_mul - -.globl x25519_fe51_sqr -.type x25519_fe51_sqr,@function -.align 32 -x25519_fe51_sqr: -.cfi_startproc - pushq %rbp -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbp,-16 - pushq %rbx -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbx,-24 - pushq %r12 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r12,-32 - pushq %r13 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r13,-40 - pushq %r14 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r14,-48 - pushq %r15 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r15,-56 - leaq -40(%rsp),%rsp -.cfi_adjust_cfa_offset 40 -.Lfe51_sqr_body: - - movq 0(%rsi),%rax - movq 16(%rsi),%r15 - movq 32(%rsi),%rbp - - movq %rdi,32(%rsp) - leaq (%rax,%rax,1),%r14 - mulq %rax - movq %rax,%rbx - movq 8(%rsi),%rax - movq %rdx,%rcx - mulq %r14 - movq %rax,%r8 - movq %r15,%rax - movq %r15,0(%rsp) - movq %rdx,%r9 - mulq %r14 - movq %rax,%r10 - movq 24(%rsi),%rax - movq %rdx,%r11 - imulq $19,%rbp,%rdi - mulq %r14 - movq %rax,%r12 - movq %rbp,%rax - movq %rdx,%r13 - mulq %r14 - movq %rax,%r14 - movq %rbp,%rax - movq %rdx,%r15 - - mulq %rdi - addq %rax,%r12 - movq 8(%rsi),%rax - adcq %rdx,%r13 - - movq 24(%rsi),%rsi - leaq (%rax,%rax,1),%rbp - mulq %rax - addq %rax,%r10 - movq 0(%rsp),%rax - adcq %rdx,%r11 - mulq %rbp - addq %rax,%r12 - movq %rbp,%rax - adcq %rdx,%r13 - mulq %rsi - addq %rax,%r14 - movq %rbp,%rax - adcq %rdx,%r15 - imulq $19,%rsi,%rbp - mulq %rdi - addq %rax,%rbx - leaq (%rsi,%rsi,1),%rax - adcq %rdx,%rcx - - mulq %rdi - addq %rax,%r10 - movq %rsi,%rax - adcq %rdx,%r11 - mulq %rbp - addq %rax,%r8 - movq 0(%rsp),%rax - adcq %rdx,%r9 - - leaq (%rax,%rax,1),%rsi - mulq %rax - addq %rax,%r14 - movq %rbp,%rax - adcq %rdx,%r15 - mulq %rsi - addq %rax,%rbx - movq %rsi,%rax - adcq %rdx,%rcx - mulq %rdi - addq %rax,%r8 - adcq %rdx,%r9 - - movq 32(%rsp),%rdi - jmp .Lreduce51 - -.align 32 -.Lreduce51: - movq $0x7ffffffffffff,%rbp - - movq %r10,%rdx - shrq $51,%r10 - shlq $13,%r11 - andq %rbp,%rdx - orq %r10,%r11 - addq %r11,%r12 - adcq $0,%r13 - - movq %rbx,%rax - shrq $51,%rbx - shlq $13,%rcx - andq %rbp,%rax - orq %rbx,%rcx - addq %rcx,%r8 - adcq $0,%r9 - - movq %r12,%rbx - shrq $51,%r12 - shlq $13,%r13 - andq %rbp,%rbx - orq %r12,%r13 - addq %r13,%r14 - adcq $0,%r15 - - movq %r8,%rcx - shrq $51,%r8 - shlq $13,%r9 - andq %rbp,%rcx - orq %r8,%r9 - addq %r9,%rdx - - movq %r14,%r10 - shrq $51,%r14 - shlq $13,%r15 - andq %rbp,%r10 - orq %r14,%r15 - - leaq (%r15,%r15,8),%r14 - leaq (%r15,%r14,2),%r15 - addq %r15,%rax - - movq %rdx,%r8 - andq %rbp,%rdx - shrq $51,%r8 - addq %r8,%rbx - - movq %rax,%r9 - andq %rbp,%rax - shrq $51,%r9 - addq %r9,%rcx - - movq %rax,0(%rdi) - movq %rcx,8(%rdi) - movq %rdx,16(%rdi) - movq %rbx,24(%rdi) - movq %r10,32(%rdi) - - movq 40(%rsp),%r15 -.cfi_restore %r15 - movq 48(%rsp),%r14 -.cfi_restore %r14 - movq 56(%rsp),%r13 -.cfi_restore %r13 - movq 64(%rsp),%r12 -.cfi_restore %r12 - movq 72(%rsp),%rbx -.cfi_restore %rbx - movq 80(%rsp),%rbp -.cfi_restore %rbp - leaq 88(%rsp),%rsp -.cfi_adjust_cfa_offset 88 -.Lfe51_sqr_epilogue: - .byte 0xf3,0xc3 -.cfi_endproc -.size x25519_fe51_sqr,.-x25519_fe51_sqr - -.globl x25519_fe51_mul121666 -.type x25519_fe51_mul121666,@function -.align 32 -x25519_fe51_mul121666: -.cfi_startproc - pushq %rbp -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbp,-16 - pushq %rbx -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbx,-24 - pushq %r12 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r12,-32 - pushq %r13 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r13,-40 - pushq %r14 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r14,-48 - pushq %r15 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r15,-56 - leaq -40(%rsp),%rsp -.cfi_adjust_cfa_offset 40 -.Lfe51_mul121666_body: - movl $121666,%eax - - mulq 0(%rsi) - movq %rax,%rbx - movl $121666,%eax - movq %rdx,%rcx - mulq 8(%rsi) - movq %rax,%r8 - movl $121666,%eax - movq %rdx,%r9 - mulq 16(%rsi) - movq %rax,%r10 - movl $121666,%eax - movq %rdx,%r11 - mulq 24(%rsi) - movq %rax,%r12 - movl $121666,%eax - movq %rdx,%r13 - mulq 32(%rsi) - movq %rax,%r14 - movq %rdx,%r15 - - jmp .Lreduce51 -.Lfe51_mul121666_epilogue: -.cfi_endproc -.size x25519_fe51_mul121666,.-x25519_fe51_mul121666 - -.globl x25519_fe64_eligible -.type x25519_fe64_eligible,@function -.align 32 -x25519_fe64_eligible: -.cfi_startproc - movl OPENSSL_ia32cap_P+8(%rip),%ecx - xorl %eax,%eax - andl $0x80100,%ecx - cmpl $0x80100,%ecx - cmovel %ecx,%eax - .byte 0xf3,0xc3 -.cfi_endproc -.size x25519_fe64_eligible,.-x25519_fe64_eligible - -.globl x25519_fe64_mul -.type x25519_fe64_mul,@function -.align 32 -x25519_fe64_mul: -.cfi_startproc - pushq %rbp -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbp,-16 - pushq %rbx -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbx,-24 - pushq %r12 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r12,-32 - pushq %r13 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r13,-40 - pushq %r14 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r14,-48 - pushq %r15 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r15,-56 - pushq %rdi -.cfi_adjust_cfa_offset 8 -.cfi_offset %rdi,-64 - leaq -16(%rsp),%rsp -.cfi_adjust_cfa_offset 16 -.Lfe64_mul_body: - - movq %rdx,%rax - movq 0(%rdx),%rbp - movq 0(%rsi),%rdx - movq 8(%rax),%rcx - movq 16(%rax),%r14 - movq 24(%rax),%r15 - - mulxq %rbp,%r8,%rax - xorl %edi,%edi - mulxq %rcx,%r9,%rbx - adcxq %rax,%r9 - mulxq %r14,%r10,%rax - adcxq %rbx,%r10 - mulxq %r15,%r11,%r12 - movq 8(%rsi),%rdx - adcxq %rax,%r11 - movq %r14,(%rsp) - adcxq %rdi,%r12 - - mulxq %rbp,%rax,%rbx - adoxq %rax,%r9 - adcxq %rbx,%r10 - mulxq %rcx,%rax,%rbx - adoxq %rax,%r10 - adcxq %rbx,%r11 - mulxq %r14,%rax,%rbx - adoxq %rax,%r11 - adcxq %rbx,%r12 - mulxq %r15,%rax,%r13 - movq 16(%rsi),%rdx - adoxq %rax,%r12 - adcxq %rdi,%r13 - adoxq %rdi,%r13 - - mulxq %rbp,%rax,%rbx - adcxq %rax,%r10 - adoxq %rbx,%r11 - mulxq %rcx,%rax,%rbx - adcxq %rax,%r11 - adoxq %rbx,%r12 - mulxq %r14,%rax,%rbx - adcxq %rax,%r12 - adoxq %rbx,%r13 - mulxq %r15,%rax,%r14 - movq 24(%rsi),%rdx - adcxq %rax,%r13 - adoxq %rdi,%r14 - adcxq %rdi,%r14 - - mulxq %rbp,%rax,%rbx - adoxq %rax,%r11 - adcxq %rbx,%r12 - mulxq %rcx,%rax,%rbx - adoxq %rax,%r12 - adcxq %rbx,%r13 - mulxq (%rsp),%rax,%rbx - adoxq %rax,%r13 - adcxq %rbx,%r14 - mulxq %r15,%rax,%r15 - movl $38,%edx - adoxq %rax,%r14 - adcxq %rdi,%r15 - adoxq %rdi,%r15 - - jmp .Lreduce64 -.Lfe64_mul_epilogue: -.cfi_endproc -.size x25519_fe64_mul,.-x25519_fe64_mul - -.globl x25519_fe64_sqr -.type x25519_fe64_sqr,@function -.align 32 -x25519_fe64_sqr: -.cfi_startproc - pushq %rbp -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbp,-16 - pushq %rbx -.cfi_adjust_cfa_offset 8 -.cfi_offset %rbx,-24 - pushq %r12 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r12,-32 - pushq %r13 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r13,-40 - pushq %r14 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r14,-48 - pushq %r15 -.cfi_adjust_cfa_offset 8 -.cfi_offset %r15,-56 - pushq %rdi -.cfi_adjust_cfa_offset 8 -.cfi_offset %rdi,-64 - leaq -16(%rsp),%rsp -.cfi_adjust_cfa_offset 16 -.Lfe64_sqr_body: - - movq 0(%rsi),%rdx - movq 8(%rsi),%rcx - movq 16(%rsi),%rbp - movq 24(%rsi),%rsi - - - mulxq %rdx,%r8,%r15 - mulxq %rcx,%r9,%rax - xorl %edi,%edi - mulxq %rbp,%r10,%rbx - adcxq %rax,%r10 - mulxq %rsi,%r11,%r12 - movq %rcx,%rdx - adcxq %rbx,%r11 - adcxq %rdi,%r12 - - - mulxq %rbp,%rax,%rbx - adoxq %rax,%r11 - adcxq %rbx,%r12 - mulxq %rsi,%rax,%r13 - movq %rbp,%rdx - adoxq %rax,%r12 - adcxq %rdi,%r13 - - - mulxq %rsi,%rax,%r14 - movq %rcx,%rdx - adoxq %rax,%r13 - adcxq %rdi,%r14 - adoxq %rdi,%r14 - - adcxq %r9,%r9 - adoxq %r15,%r9 - adcxq %r10,%r10 - mulxq %rdx,%rax,%rbx - movq %rbp,%rdx - adcxq %r11,%r11 - adoxq %rax,%r10 - adcxq %r12,%r12 - adoxq %rbx,%r11 - mulxq %rdx,%rax,%rbx - movq %rsi,%rdx - adcxq %r13,%r13 - adoxq %rax,%r12 - adcxq %r14,%r14 - adoxq %rbx,%r13 - mulxq %rdx,%rax,%r15 - movl $38,%edx - adoxq %rax,%r14 - adcxq %rdi,%r15 - adoxq %rdi,%r15 - jmp .Lreduce64 - -.align 32 -.Lreduce64: - mulxq %r12,%rax,%rbx - adcxq %rax,%r8 - adoxq %rbx,%r9 - mulxq %r13,%rax,%rbx - adcxq %rax,%r9 - adoxq %rbx,%r10 - mulxq %r14,%rax,%rbx - adcxq %rax,%r10 - adoxq %rbx,%r11 - mulxq %r15,%rax,%r12 - adcxq %rax,%r11 - adoxq %rdi,%r12 - adcxq %rdi,%r12 - - movq 16(%rsp),%rdi - imulq %rdx,%r12 - - addq %r12,%r8 - adcq $0,%r9 - adcq $0,%r10 - adcq $0,%r11 - - sbbq %rax,%rax - andq $38,%rax - - addq %rax,%r8 - movq %r9,8(%rdi) - movq %r10,16(%rdi) - movq %r11,24(%rdi) - movq %r8,0(%rdi) - - movq 24(%rsp),%r15 -.cfi_restore %r15 - movq 32(%rsp),%r14 -.cfi_restore %r14 - movq 40(%rsp),%r13 -.cfi_restore %r13 - movq 48(%rsp),%r12 -.cfi_restore %r12 - movq 56(%rsp),%rbx -.cfi_restore %rbx - movq 64(%rsp),%rbp -.cfi_restore %rbp - leaq 72(%rsp),%rsp -.cfi_adjust_cfa_offset 88 -.Lfe64_sqr_epilogue: - .byte 0xf3,0xc3 -.cfi_endproc -.size x25519_fe64_sqr,.-x25519_fe64_sqr - -.globl x25519_fe64_mul121666 -.type x25519_fe64_mul121666,@function -.align 32 -x25519_fe64_mul121666: -.Lfe64_mul121666_body: -.cfi_startproc - movl $121666,%edx - mulxq 0(%rsi),%r8,%rcx - mulxq 8(%rsi),%r9,%rax - addq %rcx,%r9 - mulxq 16(%rsi),%r10,%rcx - adcq %rax,%r10 - mulxq 24(%rsi),%r11,%rax - adcq %rcx,%r11 - adcq $0,%rax - - imulq $38,%rax,%rax - - addq %rax,%r8 - adcq $0,%r9 - adcq $0,%r10 - adcq $0,%r11 - - sbbq %rax,%rax - andq $38,%rax - - addq %rax,%r8 - movq %r9,8(%rdi) - movq %r10,16(%rdi) - movq %r11,24(%rdi) - movq %r8,0(%rdi) - -.Lfe64_mul121666_epilogue: - .byte 0xf3,0xc3 -.cfi_endproc -.size x25519_fe64_mul121666,.-x25519_fe64_mul121666 - -.globl x25519_fe64_add -.type x25519_fe64_add,@function -.align 32 -x25519_fe64_add: -.Lfe64_add_body: -.cfi_startproc - movq 0(%rsi),%r8 - movq 8(%rsi),%r9 - movq 16(%rsi),%r10 - movq 24(%rsi),%r11 - - addq 0(%rdx),%r8 - adcq 8(%rdx),%r9 - adcq 16(%rdx),%r10 - adcq 24(%rdx),%r11 - - sbbq %rax,%rax - andq $38,%rax - - addq %rax,%r8 - adcq $0,%r9 - adcq $0,%r10 - movq %r9,8(%rdi) - adcq $0,%r11 - movq %r10,16(%rdi) - sbbq %rax,%rax - movq %r11,24(%rdi) - andq $38,%rax - - addq %rax,%r8 - movq %r8,0(%rdi) - -.Lfe64_add_epilogue: - .byte 0xf3,0xc3 -.cfi_endproc -.size x25519_fe64_add,.-x25519_fe64_add - -.globl x25519_fe64_sub -.type x25519_fe64_sub,@function -.align 32 -x25519_fe64_sub: -.Lfe64_sub_body: -.cfi_startproc - movq 0(%rsi),%r8 - movq 8(%rsi),%r9 - movq 16(%rsi),%r10 - movq 24(%rsi),%r11 - - subq 0(%rdx),%r8 - sbbq 8(%rdx),%r9 - sbbq 16(%rdx),%r10 - sbbq 24(%rdx),%r11 - - sbbq %rax,%rax - andq $38,%rax - - subq %rax,%r8 - sbbq $0,%r9 - sbbq $0,%r10 - movq %r9,8(%rdi) - sbbq $0,%r11 - movq %r10,16(%rdi) - sbbq %rax,%rax - movq %r11,24(%rdi) - andq $38,%rax - - subq %rax,%r8 - movq %r8,0(%rdi) - -.Lfe64_sub_epilogue: - .byte 0xf3,0xc3 -.cfi_endproc -.size x25519_fe64_sub,.-x25519_fe64_sub - -.globl x25519_fe64_tobytes -.type x25519_fe64_tobytes,@function -.align 32 -x25519_fe64_tobytes: -.Lfe64_to_body: -.cfi_startproc - movq 0(%rsi),%r8 - movq 8(%rsi),%r9 - movq 16(%rsi),%r10 - movq 24(%rsi),%r11 - - - leaq (%r11,%r11,1),%rax - sarq $63,%r11 - shrq $1,%rax - andq $19,%r11 - addq $19,%r11 - - addq %r11,%r8 - adcq $0,%r9 - adcq $0,%r10 - adcq $0,%rax - - leaq (%rax,%rax,1),%r11 - sarq $63,%rax - shrq $1,%r11 - notq %rax - andq $19,%rax - - subq %rax,%r8 - sbbq $0,%r9 - sbbq $0,%r10 - sbbq $0,%r11 - - movq %r8,0(%rdi) - movq %r9,8(%rdi) - movq %r10,16(%rdi) - movq %r11,24(%rdi) - -.Lfe64_to_epilogue: - .byte 0xf3,0xc3 -.cfi_endproc -.size x25519_fe64_tobytes,.-x25519_fe64_tobytes -.byte 88,50,53,53,49,57,32,112,114,105,109,105,116,105,118,101,115,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |