.text
.globl mul_mont_pasta
.hidden mul_mont_pasta
.type mul_mont_pasta,@function
.align 32
mul_mont_pasta:
.cfi_startproc
.byte 0xf3,0x0f,0x1e,0xfa
pushq %rbp
.cfi_adjust_cfa_offset 8
.cfi_offset %rbp,-16
pushq %rbx
.cfi_adjust_cfa_offset 8
.cfi_offset %rbx,-24
pushq %r12
.cfi_adjust_cfa_offset 8
.cfi_offset %r12,-32
pushq %r13
.cfi_adjust_cfa_offset 8
.cfi_offset %r13,-40
pushq %r14
.cfi_adjust_cfa_offset 8
.cfi_offset %r14,-48
pushq %r15
.cfi_adjust_cfa_offset 8
.cfi_offset %r15,-56
pushq %rdi
.cfi_adjust_cfa_offset 8
movq 0(%rdx),%rax
movq 0(%rsi),%r13
movq 8(%rsi),%r14
movq 16(%rsi),%r12
movq 24(%rsi),%rbp
movq %rdx,%rbx
movq %rax,%r15
mulq %r13
movq %rax,%r9
movq %r15,%rax
movq %rdx,%r10
call __mulq_mont_pasta
movq 8(%rsp),%r15
.cfi_restore %r15
movq 16(%rsp),%r14
.cfi_restore %r14
movq 24(%rsp),%r13
.cfi_restore %r13
movq 32(%rsp),%r12
.cfi_restore %r12
movq 40(%rsp),%rbx
.cfi_restore %rbx
movq 48(%rsp),%rbp
.cfi_restore %rbp
leaq 56(%rsp),%rsp
.cfi_adjust_cfa_offset -56
.byte 0xf3,0xc3
.cfi_endproc
.size mul_mont_pasta,.-mul_mont_pasta
.globl sqr_mont_pasta
.hidden sqr_mont_pasta
.type sqr_mont_pasta,@function
.align 32
sqr_mont_pasta:
.cfi_startproc
.byte 0xf3,0x0f,0x1e,0xfa
pushq %rbp
.cfi_adjust_cfa_offset 8
.cfi_offset %rbp,-16
pushq %rbx
.cfi_adjust_cfa_offset 8
.cfi_offset %rbx,-24
pushq %r12
.cfi_adjust_cfa_offset 8
.cfi_offset %r12,-32
pushq %r13
.cfi_adjust_cfa_offset 8
.cfi_offset %r13,-40
pushq %r14
.cfi_adjust_cfa_offset 8
.cfi_offset %r14,-48
pushq %r15
.cfi_adjust_cfa_offset 8
.cfi_offset %r15,-56
pushq %rdi
.cfi_adjust_cfa_offset 8
movq 0(%rsi),%rax
movq %rcx,%r8
movq 8(%rsi),%r14
movq %rdx,%rcx
movq 16(%rsi),%r12
leaq (%rsi),%rbx
movq 24(%rsi),%rbp
movq %rax,%r15
mulq %rax
movq %rax,%r9
movq %r15,%rax
movq %rdx,%r10
call __mulq_mont_pasta
movq 8(%rsp),%r15
.cfi_restore %r15
movq 16(%rsp),%r14
.cfi_restore %r14
movq 24(%rsp),%r13
.cfi_restore %r13
movq 32(%rsp),%r12
.cfi_restore %r12
movq 40(%rsp),%rbx
.cfi_restore %rbx
movq 48(%rsp),%rbp
.cfi_restore %rbp
leaq 56(%rsp),%rsp
.cfi_adjust_cfa_offset -56
.byte 0xf3,0xc3
.cfi_endproc
.size sqr_mont_pasta,.-sqr_mont_pasta
.type __mulq_mont_pasta,@function
.align 32
__mulq_mont_pasta:
.cfi_startproc
.byte 0xf3,0x0f,0x1e,0xfa
mulq %r14
addq %rax,%r10
movq %r15,%rax
adcq $0,%rdx
movq %rdx,%r11
mulq %r12
addq %rax,%r11
movq %r15,%rax
adcq $0,%rdx
movq %rdx,%r12
mulq %rbp
addq %rax,%r12
movq 8(%rbx),%rax
adcq $0,%rdx
xorq %r14,%r14
movq %rdx,%r13
movq %r9,%rdi
imulq %r8,%r9
movq %rax,%r15
mulq 0(%rsi)
addq %rax,%r10
movq %r15,%rax
adcq $0,%rdx
movq %rdx,%rbp
mulq 8(%rsi)
addq %rax,%r11
movq %r15,%rax
adcq $0,%rdx
addq %rbp,%r11
adcq $0,%rdx
movq %rdx,%rbp
mulq 16(%rsi)
addq %rax,%r12
movq %r15,%rax
adcq $0,%rdx
addq %rbp,%r12
adcq $0,%rdx
movq %rdx,%rbp
mulq 24(%rsi)
addq %rax,%r13
movq %r9,%rax
adcq $0,%rdx
addq %rbp,%r13
adcq $0,%rdx
movq %rdx,%r14
mulq 0(%rcx)
addq %rax,%rdi
movq %r9,%rax
adcq %rdx,%rdi
mulq 8(%rcx)
addq %rax,%r10
movq %r9,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %rdi,%r10
adcq %rdx,%r11
adcq $0,%rbp
mulq 24(%rcx)
addq %rax,%r12
movq 16(%rbx),%rax
adcq $0,%rdx
addq %rbp,%r12
adcq $0,%rdx
addq %rdx,%r13
adcq $0,%r14
movq %r10,%rdi
imulq %r8,%r10
movq %rax,%r9
mulq 0(%rsi)
addq %rax,%r11
movq %r9,%rax
adcq $0,%rdx
movq %rdx,%rbp
mulq 8(%rsi)
addq %rax,%r12
movq %r9,%rax
adcq $0,%rdx
addq %rbp,%r12
adcq $0,%rdx
movq %rdx,%rbp
mulq 16(%rsi)
addq %rax,%r13
movq %r9,%rax
adcq $0,%rdx
addq %rbp,%r13
adcq $0,%rdx
movq %rdx,%rbp
mulq 24(%rsi)
addq %rax,%r14
movq %r10,%rax
adcq $0,%rdx
addq %rbp,%r14
adcq $0,%rdx
movq %rdx,%r15
mulq 0(%rcx)
addq %rax,%rdi
movq %r10,%rax
adcq %rdx,%rdi
mulq 8(%rcx)
addq %rax,%r11
movq %r10,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %rdi,%r11
adcq %rdx,%r12
adcq $0,%rbp
mulq 24(%rcx)
addq %rax,%r13
movq 24(%rbx),%rax
adcq $0,%rdx
addq %rbp,%r13
adcq $0,%rdx
addq %rdx,%r14
adcq $0,%r15
movq %r11,%rdi
imulq %r8,%r11
movq %rax,%r10
mulq 0(%rsi)
addq %rax,%r12
movq %r10,%rax
adcq $0,%rdx
movq %rdx,%rbp
mulq 8(%rsi)
addq %rax,%r13
movq %r10,%rax
adcq $0,%rdx
addq %rbp,%r13
adcq $0,%rdx
movq %rdx,%rbp
mulq 16(%rsi)
addq %rax,%r14
movq %r10,%rax
adcq $0,%rdx
addq %rbp,%r14
adcq $0,%rdx
movq %rdx,%rbp
mulq 24(%rsi)
addq %rax,%r15
movq %r11,%rax
adcq $0,%rdx
addq %rbp,%r15
adcq $0,%rdx
movq %rdx,%r9
mulq 0(%rcx)
addq %rax,%rdi
movq %r11,%rax
adcq %rdx,%rdi
mulq 8(%rcx)
addq %rax,%r12
movq %r11,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %rdi,%r12
adcq %rdx,%r13
adcq $0,%rbp
mulq 24(%rcx)
addq %rax,%r14
movq %r12,%rax
adcq $0,%rdx
addq %rbp,%r14
adcq $0,%rdx
addq %rdx,%r15
adcq $0,%r9
imulq %r8,%rax
movq 8(%rsp),%rsi
xorq %r10,%r10
movq %rax,%r11
mulq 0(%rcx)
addq %rax,%r12
movq %r11,%rax
adcq %rdx,%r12
mulq 8(%rcx)
addq %rax,%r13
movq %r11,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %r12,%r13
adcq %rdx,%r14
adcq $0,%rbp
mulq 24(%rcx)
movq %r14,%rbx
addq %rbp,%r15
adcq $0,%rdx
addq %rax,%r15
movq %r13,%rax
adcq $0,%rdx
addq %rdx,%r9
adcq $0,%r10
movq %r15,%r12
subq 0(%rcx),%r13
sbbq 8(%rcx),%r14
sbbq 16(%rcx),%r15
movq %r9,%rbp
sbbq 24(%rcx),%r9
sbbq $0,%r10
cmovcq %rax,%r13
cmovcq %rbx,%r14
cmovcq %r12,%r15
movq %r13,0(%rsi)
cmovcq %rbp,%r9
movq %r14,8(%rsi)
movq %r15,16(%rsi)
movq %r9,24(%rsi)
.byte 0xf3,0xc3
.cfi_endproc
.size __mulq_mont_pasta,.-__mulq_mont_pasta
.globl from_mont_pasta
.hidden from_mont_pasta
.type from_mont_pasta,@function
.align 32
from_mont_pasta:
.cfi_startproc
.byte 0xf3,0x0f,0x1e,0xfa
pushq %rbp
.cfi_adjust_cfa_offset 8
.cfi_offset %rbp,-16
pushq %rbx
.cfi_adjust_cfa_offset 8
.cfi_offset %rbx,-24
pushq %r12
.cfi_adjust_cfa_offset 8
.cfi_offset %r12,-32
pushq %r13
.cfi_adjust_cfa_offset 8
.cfi_offset %r13,-40
pushq %r14
.cfi_adjust_cfa_offset 8
.cfi_offset %r14,-48
pushq %r15
.cfi_adjust_cfa_offset 8
.cfi_offset %r15,-56
subq $8,%rsp
.cfi_adjust_cfa_offset 8
movq %rdx,%rbx
call __mulq_by_1_mont_pasta
movq %r14,%r10
movq %r15,%r11
movq %r9,%r12
subq 0(%rbx),%r13
sbbq 8(%rbx),%r14
sbbq 16(%rbx),%r15
sbbq 24(%rbx),%r9
cmovncq %r13,%rax
cmovncq %r14,%r10
cmovncq %r15,%r11
movq %rax,0(%rdi)
cmovncq %r9,%r12
movq %r10,8(%rdi)
movq %r11,16(%rdi)
movq %r12,24(%rdi)
movq 8(%rsp),%r15
.cfi_restore %r15
movq 16(%rsp),%r14
.cfi_restore %r14
movq 24(%rsp),%r13
.cfi_restore %r13
movq 32(%rsp),%r12
.cfi_restore %r12
movq 40(%rsp),%rbx
.cfi_restore %rbx
movq 48(%rsp),%rbp
.cfi_restore %rbp
leaq 56(%rsp),%rsp
.cfi_adjust_cfa_offset -56
.byte 0xf3,0xc3
.cfi_endproc
.size from_mont_pasta,.-from_mont_pasta
.globl redc_mont_pasta
.hidden redc_mont_pasta
.type redc_mont_pasta,@function
.align 32
redc_mont_pasta:
.cfi_startproc
.byte 0xf3,0x0f,0x1e,0xfa
pushq %rbp
.cfi_adjust_cfa_offset 8
.cfi_offset %rbp,-16
pushq %rbx
.cfi_adjust_cfa_offset 8
.cfi_offset %rbx,-24
pushq %r12
.cfi_adjust_cfa_offset 8
.cfi_offset %r12,-32
pushq %r13
.cfi_adjust_cfa_offset 8
.cfi_offset %r13,-40
pushq %r14
.cfi_adjust_cfa_offset 8
.cfi_offset %r14,-48
pushq %r15
.cfi_adjust_cfa_offset 8
.cfi_offset %r15,-56
subq $8,%rsp
.cfi_adjust_cfa_offset 8
movq %rdx,%rbx
call __mulq_by_1_mont_pasta
addq 32(%rsi),%r13
adcq 40(%rsi),%r14
movq %r13,%rax
adcq 48(%rsi),%r15
movq %r14,%r10
adcq 56(%rsi),%r9
sbbq %rsi,%rsi
movq %r15,%r11
subq 0(%rbx),%r13
sbbq 8(%rbx),%r14
sbbq 16(%rbx),%r15
movq %r9,%r12
sbbq 24(%rbx),%r9
sbbq $0,%rsi
cmovncq %r13,%rax
cmovncq %r14,%r10
cmovncq %r15,%r11
movq %rax,0(%rdi)
cmovncq %r9,%r12
movq %r10,8(%rdi)
movq %r11,16(%rdi)
movq %r12,24(%rdi)
movq 8(%rsp),%r15
.cfi_restore %r15
movq 16(%rsp),%r14
.cfi_restore %r14
movq 24(%rsp),%r13
.cfi_restore %r13
movq 32(%rsp),%r12
.cfi_restore %r12
movq 40(%rsp),%rbx
.cfi_restore %rbx
movq 48(%rsp),%rbp
.cfi_restore %rbp
leaq 56(%rsp),%rsp
.cfi_adjust_cfa_offset -56
.byte 0xf3,0xc3
.cfi_endproc
.size redc_mont_pasta,.-redc_mont_pasta
.type __mulq_by_1_mont_pasta,@function
.align 32
__mulq_by_1_mont_pasta:
.cfi_startproc
.byte 0xf3,0x0f,0x1e,0xfa
movq 0(%rsi),%rax
movq 8(%rsi),%r10
movq 16(%rsi),%r11
movq 24(%rsi),%r12
movq %rax,%r13
imulq %rcx,%rax
movq %rax,%r9
mulq 0(%rbx)
addq %rax,%r13
movq %r9,%rax
adcq %rdx,%r13
mulq 8(%rbx)
addq %rax,%r10
movq %r9,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %r13,%r10
adcq %rdx,%r11
adcq $0,%rbp
movq %r10,%r14
imulq %rcx,%r10
mulq 24(%rbx)
addq %rax,%r12
movq %r10,%rax
adcq $0,%rdx
addq %rbp,%r12
adcq $0,%rdx
movq %rdx,%r13
mulq 0(%rbx)
addq %rax,%r14
movq %r10,%rax
adcq %rdx,%r14
mulq 8(%rbx)
addq %rax,%r11
movq %r10,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %r14,%r11
adcq %rdx,%r12
adcq $0,%rbp
movq %r11,%r15
imulq %rcx,%r11
mulq 24(%rbx)
addq %rax,%r13
movq %r11,%rax
adcq $0,%rdx
addq %rbp,%r13
adcq $0,%rdx
movq %rdx,%r14
mulq 0(%rbx)
addq %rax,%r15
movq %r11,%rax
adcq %rdx,%r15
mulq 8(%rbx)
addq %rax,%r12
movq %r11,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %r15,%r12
adcq %rdx,%r13
adcq $0,%rbp
movq %r12,%r9
imulq %rcx,%r12
mulq 24(%rbx)
addq %rax,%r14
movq %r12,%rax
adcq $0,%rdx
addq %rbp,%r14
adcq $0,%rdx
movq %rdx,%r15
mulq 0(%rbx)
addq %rax,%r9
movq %r12,%rax
adcq %rdx,%r9
mulq 8(%rbx)
addq %rax,%r13
movq %r12,%rax
adcq $0,%rdx
xorq %rbp,%rbp
addq %r9,%r13
adcq %rdx,%r14
adcq $0,%rbp
mulq 24(%rbx)
addq %rax,%r15
movq %r13,%rax
adcq $0,%rdx
addq %rbp,%r15
adcq $0,%rdx
movq %rdx,%r9
.byte 0xf3,0xc3
.cfi_endproc
.size __mulq_by_1_mont_pasta,.-__mulq_by_1_mont_pasta
.globl sqr_n_mul_mont_pasta
.hidden sqr_n_mul_mont_pasta
.type sqr_n_mul_mont_pasta,@function
.align 32
sqr_n_mul_mont_pasta:
.cfi_startproc
.byte 0xf3,0x0f,0x1e,0xfa
pushq %rbp
.cfi_adjust_cfa_offset 8
.cfi_offset %rbp,-16
pushq %rbx
.cfi_adjust_cfa_offset 8
.cfi_offset %rbx,-24
pushq %r12
.cfi_adjust_cfa_offset 8
.cfi_offset %r12,-32
pushq %r13
.cfi_adjust_cfa_offset 8
.cfi_offset %r13,-40
pushq %r14
.cfi_adjust_cfa_offset 8
.cfi_offset %r14,-48
pushq %r15
.cfi_adjust_cfa_offset 8
.cfi_offset %r15,-56
subq $40,%rsp
.cfi_adjust_cfa_offset 40
movq %rcx,16(%rsp)
movq %rdi,24(%rsp)
movq 0(%rsi),%rax
movq 8(%rsi),%rbp
movq 16(%rsi),%rdi
movq 24(%rsi),%rsi
movq %rdx,%rbx
movq %r9,(%rsp)
movq %rax,%rcx
jmp .Loop_sqrq_pasta
.align 32
.Loop_sqrq_pasta:
movq %rbx,8(%rsp)
mulq %rax
movq %rax,%r9
movq %rcx,%rax
movq %rdx,%r10
mulq %rbp
addq %rax,%r10
movq %rcx,%rax
adcq $0,%rdx
movq %rdx,%r11
mulq %rdi
addq %rax,%r11
movq %rcx,%rax
adcq $0,%rdx
movq %rdx,%r12
mulq %rsi
addq %rax,%r12
movq %rbp,%rax
adcq $0,%rdx
movq %rdx,%r13
movq %r9,%r15
imulq (%rsp),%r9
mulq %rcx
addq %rax,%r10
movq %rbp,%rax
adcq $0,%rdx
movq %rdx,%rbx
mulq %rbp
addq %rax,%r11
movq %rbp,%rax
adcq $0,%rdx
addq %rbx,%r11
adcq $0,%rdx
movq %rdx,%rbx
mulq %rdi
addq %rax,%r12
movq %rbp,%rax
adcq $0,%rdx
addq %rbx,%r12
adcq $0,%rdx
movq %rdx,%rbx
mulq %rsi
addq %rax,%r13
movq %r9,%rax
adcq $0,%rdx
addq %rbx,%r13
adcq $0,%rdx
movq %rdx,%r14
mulq 0(%r8)
addq %rax,%r15
movq %r9,%rax
adcq %rdx,%r15
mulq 8(%r8)
addq %rax,%r10
movq %r9,%rax
adcq $0,%rdx
xorq %rbx,%rbx
addq %r15,%r10
adcq %rdx,%r11
adcq $0,%rbx
mulq 24(%r8)
addq %rax,%r12
movq %rdi,%rax
adcq $0,%rdx
addq %rbx,%r12
adcq %rdx,%r13
adcq $0,%r14
movq %r10,%r9
imulq (%rsp),%r10
mulq %rcx
addq %rax,%r11
movq %rdi,%rax
adcq $0,%rdx
movq %rdx,%rbx
mulq %rbp
addq %rax,%r12
movq %rdi,%rax
adcq $0,%rdx
addq %rbx,%r12
adcq $0,%rdx
movq %rdx,%rbx
mulq %rdi
addq %rax,%r13
movq %rdi,%rax
adcq $0,%rdx
addq %rbx,%r13
adcq $0,%rdx
movq %rdx,%rbx
mulq %rsi
addq %rax,%r14
movq %r10,%rax
adcq $0,%rdx
addq %rbx,%r14
adcq $0,%rdx
movq %rdx,%r15
mulq 0(%r8)
addq %rax,%r9
movq %r10,%rax
adcq %rdx,%r9
mulq 8(%r8)
addq %rax,%r11
movq %r10,%rax
adcq $0,%rdx
xorq %rbx,%rbx
addq %r9,%r11
adcq %rdx,%r12
adcq $0,%rbx
mulq 24(%r8)
addq %rax,%r13
movq %rsi,%rax
adcq $0,%rdx
addq %rbx,%r13
adcq %rdx,%r14
adcq $0,%r15
movq %r11,%r10
imulq (%rsp),%r11
mulq %rcx
addq %rax,%r12
movq %rsi,%rax
adcq $0,%rdx
movq %rdx,%rbx
mulq %rbp
addq %rax,%r13
movq %rsi,%rax
adcq $0,%rdx
addq %rbx,%r13
adcq $0,%rdx
movq %rdx,%rbx
mulq %rdi
addq %rax,%r14
movq %rsi,%rax
adcq $0,%rdx
addq %rbx,%r14
adcq $0,%rdx
movq %rdx,%rbx
mulq %rsi
addq %rax,%r15
movq %r11,%rax
adcq $0,%rdx
addq %rbx,%r15
adcq $0,%rdx
movq %rdx,%r9
mulq 0(%r8)
addq %rax,%r10
movq %r11,%rax
adcq %rdx,%r10
mulq 8(%r8)
addq %rax,%r12
movq %r11,%rax
adcq $0,%rdx
xorq %rbx,%rbx
addq %r10,%r12
adcq %rdx,%r13
adcq $0,%rbx
mulq 24(%r8)
addq %rax,%r14
movq %r12,%rax
adcq $0,%rdx
addq %rbx,%r14
adcq %rdx,%r15
adcq $0,%r9
imulq (%rsp),%rax
movq %r13,%rcx
movq %r14,%rbp
movq 8(%rsp),%rbx
movq %r9,%rsi
movq 16(%rsp),%r11
movq %rax,%r10
mulq 0(%r8)
addq %rax,%r12
movq %r10,%rax
adcq %rdx,%r12
mulq 8(%r8)
addq %rax,%rcx
movq %r10,%rax
adcq $0,%rdx
xorq %rdi,%rdi
addq %r12,%rcx
adcq %rdx,%rbp
adcq $0,%rdi
mulq 24(%r8)
addq %rax,%rdi
movq %rcx,%rax
adcq $0,%rdx
addq %r15,%rdi
adcq %rdx,%rsi
subq $1,%rbx
jnz .Loop_sqrq_pasta
mulq 0(%r11)
movq %rax,%r12
movq %rcx,%rax
movq %rdx,%r13
mulq 8(%r11)
addq %rax,%r13
movq %rcx,%rax
adcq $0,%rdx
movq %rdx,%r14
mulq 16(%r11)
addq %rax,%r14
movq %rcx,%rax
adcq $0,%rdx
movq %rdx,%r15
mulq 24(%r11)
addq %rax,%r15
movq %rbp,%rax
adcq $0,%rdx
movq %rdx,%r9
movq %r12,%rcx
imulq (%rsp),%r12
mulq 0(%r11)
addq %rax,%r13
movq %rbp,%rax
adcq $0,%rdx
movq %rdx,%rbx
mulq 8(%r11)
addq %rax,%r14
movq %rbp,%rax
adcq $0,%rdx
addq %rbx,%r14
adcq $0,%rdx
movq %rdx,%rbx
mulq 16(%r11)
addq %rax,%r15
movq %rbp,%rax
adcq $0,%rdx
addq %rbx,%r15
adcq $0,%rdx
movq %rdx,%rbx
mulq 24(%r11)
addq %rax,%r9
movq %r12,%rax
adcq $0,%rdx
addq %rbx,%r9
adcq $0,%rdx
movq %rdx,%r10
mulq 0(%r8)
addq %rax,%rcx
movq %r12,%rax
adcq %rdx,%rcx
mulq 8(%r8)
addq %rax,%r13
movq %r12,%rax
adcq $0,%rdx
xorq %rbx,%rbx
addq %rcx,%r13
adcq %rdx,%r14
adcq $0,%rbx
mulq 24(%r8)
addq %rax,%r15
movq %rdi,%rax
adcq $0,%rdx
addq %rbx,%r15
adcq %rdx,%r9
adcq $0,%r10
movq %r13,%r12
imulq (%rsp),%r13
mulq 0(%r11)
addq %rax,%r14
movq %rdi,%rax
adcq $0,%rdx
movq %rdx,%rbx
mulq 8(%r11)
addq %rax,%r15
movq %rdi,%rax
adcq $0,%rdx
addq %rbx,%r15
adcq $0,%rdx
movq %rdx,%rbx
mulq 16(%r11)
addq %rax,%r9
movq %rdi,%rax
adcq $0,%rdx
addq %rbx,%r9
adcq $0,%rdx
movq %rdx,%rbx
mulq 24(%r11)
addq %rax,%r10
movq %r13,%rax
adcq $0,%rdx
addq %rbx,%r10
adcq $0,%rdx
movq %rdx,%rcx
mulq 0(%r8)
addq %rax,%r12
movq %r13,%rax
adcq %rdx,%r12
mulq 8(%r8)
addq %rax,%r14
movq %r13,%rax
adcq $0,%rdx
xorq %rbx,%rbx
addq %r12,%r14
adcq %rdx,%r15
adcq $0,%rbx
mulq 24(%r8)
addq %rax,%r9
movq %rsi,%rax
adcq $0,%rdx
addq %rbx,%r9
adcq %rdx,%r10
adcq $0,%rcx
movq %r14,%r13
imulq (%rsp),%r14
mulq 0(%r11)
addq %rax,%r15
movq %rsi,%rax
adcq $0,%rdx
movq %rdx,%rbx
mulq 8(%r11)
addq %rax,%r9
movq %rsi,%rax
adcq $0,%rdx
addq %rbx,%r9
adcq $0,%rdx
movq %rdx,%rbx
mulq 16(%r11)
addq %rax,%r10
movq %rsi,%rax
adcq $0,%rdx
addq %rbx,%r10
adcq $0,%rdx
movq %rdx,%rbx
mulq 24(%r11)
addq %rax,%rcx
movq %r14,%rax
adcq $0,%rdx
addq %rbx,%rcx
adcq $0,%rdx
movq %rdx,%r12
mulq 0(%r8)
addq %rax,%r13
movq %r14,%rax
adcq %rdx,%r13
mulq 8(%r8)
addq %rax,%r15
movq %r14,%rax
adcq $0,%rdx
xorq %rbx,%rbx
addq %r13,%r15
adcq %rdx,%r9
adcq $0,%rbx
mulq 24(%r8)
addq %rax,%r10
movq %r15,%rax
adcq $0,%rdx
addq %rbx,%r10
adcq %rdx,%rcx
adcq $0,%r12
imulq (%rsp),%rax
movq 24(%rsp),%rdi
movq %rax,%r14
mulq 0(%r8)
addq %rax,%r15
movq %r14,%rax
adcq %rdx,%r15
mulq 8(%r8)
addq %rax,%r9
movq %r14,%rax
adcq $0,%rdx
xorq %rbx,%rbx
addq %r15,%r9
adcq %rdx,%r10
adcq $0,%rbx
mulq 24(%r8)
movq %r9,%r13
addq %rbx,%rcx
adcq $0,%rdx
movq %r10,%r14
addq %rax,%rcx
adcq %rdx,%r12
movq %rcx,%r15
subq 0(%r8),%r9
sbbq 8(%r8),%r10
sbbq 16(%r8),%rcx
movq %r12,%rbx
sbbq 24(%r8),%r12
cmovcq %r13,%r9
cmovcq %r14,%r10
cmovcq %r15,%rcx
movq %r9,0(%rdi)
cmovcq %rbx,%r12
movq %r10,8(%rdi)
movq %rcx,16(%rdi)
movq %r12,24(%rdi)
movq 40(%rsp),%r15
.cfi_restore %r15
movq 48(%rsp),%r14
.cfi_restore %r14
movq 56(%rsp),%r13
.cfi_restore %r13
movq 64(%rsp),%r12
.cfi_restore %r12
movq 72(%rsp),%rbx
.cfi_restore %rbx
movq 80(%rsp),%rbp
.cfi_restore %rbp
leaq 88(%rsp),%rsp
.cfi_adjust_cfa_offset -88
.byte 0xf3,0xc3
.cfi_endproc
.size sqr_n_mul_mont_pasta,.-sqr_n_mul_mont_pasta
.section .note.gnu.property,"a",@note
.long 4,2f-1f,5
.byte 0x47,0x4E,0x55,0
1: .long 0xc0000002,4,3
.align 8
2: