/* Do not modify. This file is auto-generated from x25519-x86_64.pl. */ .text .globl x25519_fe51_mul .type x25519_fe51_mul,@function .align 32 x25519_fe51_mul: .cfi_startproc pushq %rbp .cfi_adjust_cfa_offset 8 .cfi_offset %rbp,-16 pushq %rbx .cfi_adjust_cfa_offset 8 .cfi_offset %rbx,-24 pushq %r12 .cfi_adjust_cfa_offset 8 .cfi_offset %r12,-32 pushq %r13 .cfi_adjust_cfa_offset 8 .cfi_offset %r13,-40 pushq %r14 .cfi_adjust_cfa_offset 8 .cfi_offset %r14,-48 pushq %r15 .cfi_adjust_cfa_offset 8 .cfi_offset %r15,-56 leaq -40(%rsp),%rsp .cfi_adjust_cfa_offset 40 .Lfe51_mul_body: movq 0(%rsi),%rax movq 0(%rdx),%r11 movq 8(%rdx),%r12 movq 16(%rdx),%r13 movq 24(%rdx),%rbp movq 32(%rdx),%r14 movq %rdi,32(%rsp) movq %rax,%rdi mulq %r11 movq %r11,0(%rsp) movq %rax,%rbx movq %rdi,%rax movq %rdx,%rcx mulq %r12 movq %r12,8(%rsp) movq %rax,%r8 movq %rdi,%rax leaq (%r14,%r14,8),%r15 movq %rdx,%r9 mulq %r13 movq %r13,16(%rsp) movq %rax,%r10 movq %rdi,%rax leaq (%r14,%r15,2),%rdi movq %rdx,%r11 mulq %rbp movq %rax,%r12 movq 0(%rsi),%rax movq %rdx,%r13 mulq %r14 movq %rax,%r14 movq 8(%rsi),%rax movq %rdx,%r15 mulq %rdi addq %rax,%rbx movq 16(%rsi),%rax adcq %rdx,%rcx mulq %rdi addq %rax,%r8 movq 24(%rsi),%rax adcq %rdx,%r9 mulq %rdi addq %rax,%r10 movq 32(%rsi),%rax adcq %rdx,%r11 mulq %rdi imulq $19,%rbp,%rdi addq %rax,%r12 movq 8(%rsi),%rax adcq %rdx,%r13 mulq %rbp movq 16(%rsp),%rbp addq %rax,%r14 movq 16(%rsi),%rax adcq %rdx,%r15 mulq %rdi addq %rax,%rbx movq 24(%rsi),%rax adcq %rdx,%rcx mulq %rdi addq %rax,%r8 movq 32(%rsi),%rax adcq %rdx,%r9 mulq %rdi imulq $19,%rbp,%rdi addq %rax,%r10 movq 8(%rsi),%rax adcq %rdx,%r11 mulq %rbp addq %rax,%r12 movq 16(%rsi),%rax adcq %rdx,%r13 mulq %rbp movq 8(%rsp),%rbp addq %rax,%r14 movq 24(%rsi),%rax adcq %rdx,%r15 mulq %rdi addq %rax,%rbx movq 32(%rsi),%rax adcq %rdx,%rcx mulq %rdi addq %rax,%r8 movq 8(%rsi),%rax adcq %rdx,%r9 mulq %rbp imulq $19,%rbp,%rdi addq %rax,%r10 movq 16(%rsi),%rax adcq %rdx,%r11 mulq %rbp addq %rax,%r12 movq 24(%rsi),%rax adcq %rdx,%r13 mulq %rbp movq 0(%rsp),%rbp addq %rax,%r14 movq 32(%rsi),%rax adcq %rdx,%r15 mulq %rdi addq %rax,%rbx movq 8(%rsi),%rax adcq %rdx,%rcx mulq %rbp addq %rax,%r8 movq 16(%rsi),%rax adcq %rdx,%r9 mulq %rbp addq %rax,%r10 movq 24(%rsi),%rax adcq %rdx,%r11 mulq %rbp addq %rax,%r12 movq 32(%rsi),%rax adcq %rdx,%r13 mulq %rbp addq %rax,%r14 adcq %rdx,%r15 movq 32(%rsp),%rdi jmp .Lreduce51 .Lfe51_mul_epilogue: .cfi_endproc .size x25519_fe51_mul,.-x25519_fe51_mul .globl x25519_fe51_sqr .type x25519_fe51_sqr,@function .align 32 x25519_fe51_sqr: .cfi_startproc pushq %rbp .cfi_adjust_cfa_offset 8 .cfi_offset %rbp,-16 pushq %rbx .cfi_adjust_cfa_offset 8 .cfi_offset %rbx,-24 pushq %r12 .cfi_adjust_cfa_offset 8 .cfi_offset %r12,-32 pushq %r13 .cfi_adjust_cfa_offset 8 .cfi_offset %r13,-40 pushq %r14 .cfi_adjust_cfa_offset 8 .cfi_offset %r14,-48 pushq %r15 .cfi_adjust_cfa_offset 8 .cfi_offset %r15,-56 leaq -40(%rsp),%rsp .cfi_adjust_cfa_offset 40 .Lfe51_sqr_body: movq 0(%rsi),%rax movq 16(%rsi),%r15 movq 32(%rsi),%rbp movq %rdi,32(%rsp) leaq (%rax,%rax,1),%r14 mulq %rax movq %rax,%rbx movq 8(%rsi),%rax movq %rdx,%rcx mulq %r14 movq %rax,%r8 movq %r15,%rax movq %r15,0(%rsp) movq %rdx,%r9 mulq %r14 movq %rax,%r10 movq 24(%rsi),%rax movq %rdx,%r11 imulq $19,%rbp,%rdi mulq %r14 movq %rax,%r12 movq %rbp,%rax movq %rdx,%r13 mulq %r14 movq %rax,%r14 movq %rbp,%rax movq %rdx,%r15 mulq %rdi addq %rax,%r12 movq 8(%rsi),%rax adcq %rdx,%r13 movq 24(%rsi),%rsi leaq (%rax,%rax,1),%rbp mulq %rax addq %rax,%r10 movq 0(%rsp),%rax adcq %rdx,%r11 mulq %rbp addq %rax,%r12 movq %rbp,%rax adcq %rdx,%r13 mulq %rsi addq %rax,%r14 movq %rbp,%rax adcq %rdx,%r15 imulq $19,%rsi,%rbp mulq %rdi addq %rax,%rbx leaq (%rsi,%rsi,1),%rax adcq %rdx,%rcx mulq %rdi addq %rax,%r10 movq %rsi,%rax adcq %rdx,%r11 mulq %rbp addq %rax,%r8 movq 0(%rsp),%rax adcq %rdx,%r9 leaq (%rax,%rax,1),%rsi mulq %rax addq %rax,%r14 movq %rbp,%rax adcq %rdx,%r15 mulq %rsi addq %rax,%rbx movq %rsi,%rax adcq %rdx,%rcx mulq %rdi addq %rax,%r8 adcq %rdx,%r9 movq 32(%rsp),%rdi jmp .Lreduce51 .align 32 .Lreduce51: movq $0x7ffffffffffff,%rbp movq %r10,%rdx shrq $51,%r10 shlq $13,%r11 andq %rbp,%rdx orq %r10,%r11 addq %r11,%r12 adcq $0,%r13 movq %rbx,%rax shrq $51,%rbx shlq $13,%rcx andq %rbp,%rax orq %rbx,%rcx addq %rcx,%r8 adcq $0,%r9 movq %r12,%rbx shrq $51,%r12 shlq $13,%r13 andq %rbp,%rbx orq %r12,%r13 addq %r13,%r14 adcq $0,%r15 movq %r8,%rcx shrq $51,%r8 shlq $13,%r9 andq %rbp,%rcx orq %r8,%r9 addq %r9,%rdx movq %r14,%r10 shrq $51,%r14 shlq $13,%r15 andq %rbp,%r10 orq %r14,%r15 leaq (%r15,%r15,8),%r14 leaq (%r15,%r14,2),%r15 addq %r15,%rax movq %rdx,%r8 andq %rbp,%rdx shrq $51,%r8 addq %r8,%rbx movq %rax,%r9 andq %rbp,%rax shrq $51,%r9 addq %r9,%rcx movq %rax,0(%rdi) movq %rcx,8(%rdi) movq %rdx,16(%rdi) movq %rbx,24(%rdi) movq %r10,32(%rdi) movq 40(%rsp),%r15 .cfi_restore %r15 movq 48(%rsp),%r14 .cfi_restore %r14 movq 56(%rsp),%r13 .cfi_restore %r13 movq 64(%rsp),%r12 .cfi_restore %r12 movq 72(%rsp),%rbx .cfi_restore %rbx movq 80(%rsp),%rbp .cfi_restore %rbp leaq 88(%rsp),%rsp .cfi_adjust_cfa_offset 88 .Lfe51_sqr_epilogue: .byte 0xf3,0xc3 .cfi_endproc .size x25519_fe51_sqr,.-x25519_fe51_sqr .globl x25519_fe51_mul121666 .type x25519_fe51_mul121666,@function .align 32 x25519_fe51_mul121666: .cfi_startproc pushq %rbp .cfi_adjust_cfa_offset 8 .cfi_offset %rbp,-16 pushq %rbx .cfi_adjust_cfa_offset 8 .cfi_offset %rbx,-24 pushq %r12 .cfi_adjust_cfa_offset 8 .cfi_offset %r12,-32 pushq %r13 .cfi_adjust_cfa_offset 8 .cfi_offset %r13,-40 pushq %r14 .cfi_adjust_cfa_offset 8 .cfi_offset %r14,-48 pushq %r15 .cfi_adjust_cfa_offset 8 .cfi_offset %r15,-56 leaq -40(%rsp),%rsp .cfi_adjust_cfa_offset 40 .Lfe51_mul121666_body: movl $121666,%eax mulq 0(%rsi) movq %rax,%rbx movl $121666,%eax movq %rdx,%rcx mulq 8(%rsi) movq %rax,%r8 movl $121666,%eax movq %rdx,%r9 mulq 16(%rsi) movq %rax,%r10 movl $121666,%eax movq %rdx,%r11 mulq 24(%rsi) movq %rax,%r12 movl $121666,%eax movq %rdx,%r13 mulq 32(%rsi) movq %rax,%r14 movq %rdx,%r15 jmp .Lreduce51 .Lfe51_mul121666_epilogue: .cfi_endproc .size x25519_fe51_mul121666,.-x25519_fe51_mul121666 .globl x25519_fe64_eligible .type x25519_fe64_eligible,@function .align 32 x25519_fe64_eligible: .cfi_startproc movl OPENSSL_ia32cap_P+8(%rip),%ecx xorl %eax,%eax andl $0x80100,%ecx cmpl $0x80100,%ecx cmovel %ecx,%eax .byte 0xf3,0xc3 .cfi_endproc .size x25519_fe64_eligible,.-x25519_fe64_eligible .globl x25519_fe64_mul .type x25519_fe64_mul,@function .align 32 x25519_fe64_mul: .cfi_startproc pushq %rbp .cfi_adjust_cfa_offset 8 .cfi_offset %rbp,-16 pushq %rbx .cfi_adjust_cfa_offset 8 .cfi_offset %rbx,-24 pushq %r12 .cfi_adjust_cfa_offset 8 .cfi_offset %r12,-32 pushq %r13 .cfi_adjust_cfa_offset 8 .cfi_offset %r13,-40 pushq %r14 .cfi_adjust_cfa_offset 8 .cfi_offset %r14,-48 pushq %r15 .cfi_adjust_cfa_offset 8 .cfi_offset %r15,-56 pushq %rdi .cfi_adjust_cfa_offset 8 .cfi_offset %rdi,-64 leaq -16(%rsp),%rsp .cfi_adjust_cfa_offset 16 .Lfe64_mul_body: movq %rdx,%rax movq 0(%rdx),%rbp movq 0(%rsi),%rdx movq 8(%rax),%rcx movq 16(%rax),%r14 movq 24(%rax),%r15 mulxq %rbp,%r8,%rax xorl %edi,%edi mulxq %rcx,%r9,%rbx adcxq %rax,%r9 mulxq %r14,%r10,%rax adcxq %rbx,%r10 mulxq %r15,%r11,%r12 movq 8(%rsi),%rdx adcxq %rax,%r11 movq %r14,(%rsp) adcxq %rdi,%r12 mulxq %rbp,%rax,%rbx adoxq %rax,%r9 adcxq %rbx,%r10 mulxq %rcx,%rax,%rbx adoxq %rax,%r10 adcxq %rbx,%r11 mulxq %r14,%rax,%rbx adoxq %rax,%r11 adcxq %rbx,%r12 mulxq %r15,%rax,%r13 movq 16(%rsi),%rdx adoxq %rax,%r12 adcxq %rdi,%r13 adoxq %rdi,%r13 mulxq %rbp,%rax,%rbx adcxq %rax,%r10 adoxq %rbx,%r11 mulxq %rcx,%rax,%rbx adcxq %rax,%r11 adoxq %rbx,%r12 mulxq %r14,%rax,%rbx adcxq %rax,%r12 adoxq %rbx,%r13 mulxq %r15,%rax,%r14 movq 24(%rsi),%rdx adcxq %rax,%r13 adoxq %rdi,%r14 adcxq %rdi,%r14 mulxq %rbp,%rax,%rbx adoxq %rax,%r11 adcxq %rbx,%r12 mulxq %rcx,%rax,%rbx adoxq %rax,%r12 adcxq %rbx,%r13 mulxq (%rsp),%rax,%rbx adoxq %rax,%r13 adcxq %rbx,%r14 mulxq %r15,%rax,%r15 movl $38,%edx adoxq %rax,%r14 adcxq %rdi,%r15 adoxq %rdi,%r15 jmp .Lreduce64 .Lfe64_mul_epilogue: .cfi_endproc .size x25519_fe64_mul,.-x25519_fe64_mul .globl x25519_fe64_sqr .type x25519_fe64_sqr,@function .align 32 x25519_fe64_sqr: .cfi_startproc pushq %rbp .cfi_adjust_cfa_offset 8 .cfi_offset %rbp,-16 pushq %rbx .cfi_adjust_cfa_offset 8 .cfi_offset %rbx,-24 pushq %r12 .cfi_adjust_cfa_offset 8 .cfi_offset %r12,-32 pushq %r13 .cfi_adjust_cfa_offset 8 .cfi_offset %r13,-40 pushq %r14 .cfi_adjust_cfa_offset 8 .cfi_offset %r14,-48 pushq %r15 .cfi_adjust_cfa_offset 8 .cfi_offset %r15,-56 pushq %rdi .cfi_adjust_cfa_offset 8 .cfi_offset %rdi,-64 leaq -16(%rsp),%rsp .cfi_adjust_cfa_offset 16 .Lfe64_sqr_body: movq 0(%rsi),%rdx movq 8(%rsi),%rcx movq 16(%rsi),%rbp movq 24(%rsi),%rsi mulxq %rdx,%r8,%r15 mulxq %rcx,%r9,%rax xorl %edi,%edi mulxq %rbp,%r10,%rbx adcxq %rax,%r10 mulxq %rsi,%r11,%r12 movq %rcx,%rdx adcxq %rbx,%r11 adcxq %rdi,%r12 mulxq %rbp,%rax,%rbx adoxq %rax,%r11 adcxq %rbx,%r12 mulxq %rsi,%rax,%r13 movq %rbp,%rdx adoxq %rax,%r12 adcxq %rdi,%r13 mulxq %rsi,%rax,%r14 movq %rcx,%rdx adoxq %rax,%r13 adcxq %rdi,%r14 adoxq %rdi,%r14 adcxq %r9,%r9 adoxq %r15,%r9 adcxq %r10,%r10 mulxq %rdx,%rax,%rbx movq %rbp,%rdx adcxq %r11,%r11 adoxq %rax,%r10 adcxq %r12,%r12 adoxq %rbx,%r11 mulxq %rdx,%rax,%rbx movq %rsi,%rdx adcxq %r13,%r13 adoxq %rax,%r12 adcxq %r14,%r14 adoxq %rbx,%r13 mulxq %rdx,%rax,%r15 movl $38,%edx adoxq %rax,%r14 adcxq %rdi,%r15 adoxq %rdi,%r15 jmp .Lreduce64 .align 32 .Lreduce64: mulxq %r12,%rax,%rbx adcxq %rax,%r8 adoxq %rbx,%r9 mulxq %r13,%rax,%rbx adcxq %rax,%r9 adoxq %rbx,%r10 mulxq %r14,%rax,%rbx adcxq %rax,%r10 adoxq %rbx,%r11 mulxq %r15,%rax,%r12 adcxq %rax,%r11 adoxq %rdi,%r12 adcxq %rdi,%r12 movq 16(%rsp),%rdi imulq %rdx,%r12 addq %r12,%r8 adcq $0,%r9 adcq $0,%r10 adcq $0,%r11 sbbq %rax,%rax andq $38,%rax addq %rax,%r8 movq %r9,8(%rdi) movq %r10,16(%rdi) movq %r11,24(%rdi) movq %r8,0(%rdi) movq 24(%rsp),%r15 .cfi_restore %r15 movq 32(%rsp),%r14 .cfi_restore %r14 movq 40(%rsp),%r13 .cfi_restore %r13 movq 48(%rsp),%r12 .cfi_restore %r12 movq 56(%rsp),%rbx .cfi_restore %rbx movq 64(%rsp),%rbp .cfi_restore %rbp leaq 72(%rsp),%rsp .cfi_adjust_cfa_offset 88 .Lfe64_sqr_epilogue: .byte 0xf3,0xc3 .cfi_endproc .size x25519_fe64_sqr,.-x25519_fe64_sqr .globl x25519_fe64_mul121666 .type x25519_fe64_mul121666,@function .align 32 x25519_fe64_mul121666: .Lfe64_mul121666_body: .cfi_startproc movl $121666,%edx mulxq 0(%rsi),%r8,%rcx mulxq 8(%rsi),%r9,%rax addq %rcx,%r9 mulxq 16(%rsi),%r10,%rcx adcq %rax,%r10 mulxq 24(%rsi),%r11,%rax adcq %rcx,%r11 adcq $0,%rax imulq $38,%rax,%rax addq %rax,%r8 adcq $0,%r9 adcq $0,%r10 adcq $0,%r11 sbbq %rax,%rax andq $38,%rax addq %rax,%r8 movq %r9,8(%rdi) movq %r10,16(%rdi) movq %r11,24(%rdi) movq %r8,0(%rdi) .Lfe64_mul121666_epilogue: .byte 0xf3,0xc3 .cfi_endproc .size x25519_fe64_mul121666,.-x25519_fe64_mul121666 .globl x25519_fe64_add .type x25519_fe64_add,@function .align 32 x25519_fe64_add: .Lfe64_add_body: .cfi_startproc movq 0(%rsi),%r8 movq 8(%rsi),%r9 movq 16(%rsi),%r10 movq 24(%rsi),%r11 addq 0(%rdx),%r8 adcq 8(%rdx),%r9 adcq 16(%rdx),%r10 adcq 24(%rdx),%r11 sbbq %rax,%rax andq $38,%rax addq %rax,%r8 adcq $0,%r9 adcq $0,%r10 movq %r9,8(%rdi) adcq $0,%r11 movq %r10,16(%rdi) sbbq %rax,%rax movq %r11,24(%rdi) andq $38,%rax addq %rax,%r8 movq %r8,0(%rdi) .Lfe64_add_epilogue: .byte 0xf3,0xc3 .cfi_endproc .size x25519_fe64_add,.-x25519_fe64_add .globl x25519_fe64_sub .type x25519_fe64_sub,@function .align 32 x25519_fe64_sub: .Lfe64_sub_body: .cfi_startproc movq 0(%rsi),%r8 movq 8(%rsi),%r9 movq 16(%rsi),%r10 movq 24(%rsi),%r11 subq 0(%rdx),%r8 sbbq 8(%rdx),%r9 sbbq 16(%rdx),%r10 sbbq 24(%rdx),%r11 sbbq %rax,%rax andq $38,%rax subq %rax,%r8 sbbq $0,%r9 sbbq $0,%r10 movq %r9,8(%rdi) sbbq $0,%r11 movq %r10,16(%rdi) sbbq %rax,%rax movq %r11,24(%rdi) andq $38,%rax subq %rax,%r8 movq %r8,0(%rdi) .Lfe64_sub_epilogue: .byte 0xf3,0xc3 .cfi_endproc .size x25519_fe64_sub,.-x25519_fe64_sub .globl x25519_fe64_tobytes .type x25519_fe64_tobytes,@function .align 32 x25519_fe64_tobytes: .Lfe64_to_body: .cfi_startproc movq 0(%rsi),%r8 movq 8(%rsi),%r9 movq 16(%rsi),%r10 movq 24(%rsi),%r11 leaq (%r11,%r11,1),%rax sarq $63,%r11 shrq $1,%rax andq $19,%r11 addq $19,%r11 addq %r11,%r8 adcq $0,%r9 adcq $0,%r10 adcq $0,%rax leaq (%rax,%rax,1),%r11 sarq $63,%rax shrq $1,%r11 notq %rax andq $19,%rax subq %rax,%r8 sbbq $0,%r9 sbbq $0,%r10 sbbq $0,%r11 movq %r8,0(%rdi) movq %r9,8(%rdi) movq %r10,16(%rdi) movq %r11,24(%rdi) .Lfe64_to_epilogue: .byte 0xf3,0xc3 .cfi_endproc .size x25519_fe64_tobytes,.-x25519_fe64_tobytes .byte 88,50,53,53,49,57,32,112,114,105,109,105,116,105,118,101,115,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0