207 lines
3.7 KiB
NASM
207 lines
3.7 KiB
NASM
dnl mpn_mod_1_3
|
|
|
|
dnl Copyright 2009 Jason Moxham
|
|
|
|
dnl This file is part of the MPIR Library.
|
|
|
|
dnl The MPIR Library is free software; you can redistribute it and/or modify
|
|
dnl it under the terms of the GNU Lesser General Public License as published
|
|
dnl by the Free Software Foundation; either version 2.1 of the License, or (at
|
|
dnl your option) any later version.
|
|
|
|
dnl The MPIR Library is distributed in the hope that it will be useful, but
|
|
dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
|
|
dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
|
|
dnl License for more details.
|
|
|
|
dnl You should have received a copy of the GNU Lesser General Public License
|
|
dnl along with the MPIR Library; see the file COPYING.LIB. If not, write
|
|
dnl to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
|
|
dnl Boston, MA 02110-1301, USA.
|
|
|
|
include(`../config.m4')
|
|
|
|
C (rdi,2)= not fully reduced remainder of (rsi,rdx) / divisor , and top limb <d
|
|
C where (rcx,4) contains B^i % divisor
|
|
|
|
ASM_START()
|
|
PROLOGUE(mpn_mod_1_3)
|
|
C // require rdx >=5
|
|
push %r12
|
|
push %r13
|
|
push %r14
|
|
push %r15
|
|
mov -8(%rsi,%rdx,8),%r15
|
|
mov -16(%rsi,%rdx,8),%r14
|
|
mov -32(%rsi,%rdx,8),%rax
|
|
mov -40(%rsi,%rdx,8),%r12
|
|
mov (%rcx),%r8
|
|
mov 8(%rcx),%r9
|
|
mov 16(%rcx),%r10
|
|
mov 24(%rcx),%r11
|
|
mov %rdx,%rcx
|
|
sub $8,%rcx
|
|
jc skiplp
|
|
ALIGN(16)
|
|
C // r15 r14 -8() -16()=rax -24()=r12
|
|
lp:
|
|
mul %r8
|
|
add %rax,%r12
|
|
mov 40(%rsi,%rcx,8),%rax
|
|
mov $0,%r13
|
|
adc %rdx,%r13
|
|
mul %r9
|
|
add %rax,%r12
|
|
nop
|
|
adc %rdx,%r13
|
|
mov %r10,%rax
|
|
mul %r14
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
mov %r12,%r14
|
|
mov %r11,%rax
|
|
mul %r15
|
|
add %rax,%r14
|
|
mov 0(%rsi,%rcx,8),%r12
|
|
mov %r13,%r15
|
|
mov 8(%rsi,%rcx,8),%rax
|
|
adc %rdx,%r15
|
|
sub $3,%rcx
|
|
jnc lp
|
|
skiplp:
|
|
C // we have loaded up the next two limbs
|
|
C // but because they are out of order we can have to do 3 limbs min
|
|
cmp $-2,%rcx
|
|
jl case1
|
|
je case2
|
|
case3:
|
|
C //two more limbs is 4 limbs
|
|
C // r15 r14 40() 8+24()=rax 0+24()=r12
|
|
mul %r8
|
|
add %rax,%r12
|
|
mov 40(%rsi,%rcx,8),%rax
|
|
mov $0,%r13
|
|
adc %rdx,%r13
|
|
mul %r9
|
|
add %rax,%r12
|
|
nop
|
|
adc %rdx,%r13
|
|
mov %r10,%rax
|
|
mul %r14
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
mov %r12,%r14
|
|
mov %r11,%rax
|
|
mul %r15
|
|
add %rax,%r14
|
|
mov 8(%rsi,%rcx,8),%r12
|
|
mov %r13,%r15
|
|
mov 16(%rsi,%rcx,8),%rax
|
|
adc %rdx,%r15
|
|
C // r15 r14 rax r12
|
|
mov $0,%r13
|
|
mul %r8
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
mov %r9,%rax
|
|
mul %r14
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
mov %r10,%rax
|
|
mul %r15
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
C // r13 r12
|
|
mov %r8,%rax
|
|
mul %r13
|
|
add %rax,%r12
|
|
adc $0,%rdx
|
|
mov %r12,(%rdi)
|
|
mov %rdx,8(%rdi)
|
|
pop %r15
|
|
pop %r14
|
|
pop %r13
|
|
pop %r12
|
|
ret
|
|
ALIGN(16)
|
|
case2:
|
|
C //two more limbs is 4 limbs
|
|
C // r15 r14 40() 8+24()=rax 0+24()=r12
|
|
mul %r8
|
|
add %rax,%r12
|
|
mov 40(%rsi,%rcx,8),%rax
|
|
mov $0,%r13
|
|
adc %rdx,%r13
|
|
mul %r9
|
|
add %rax,%r12
|
|
nop
|
|
adc %rdx,%r13
|
|
mov %r10,%rax
|
|
mul %r14
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
mov %r12,%r14
|
|
mov %r11,%rax
|
|
mul %r15
|
|
add %rax,%r14
|
|
mov 16(%rsi,%rcx,8),%r12
|
|
mov %r13,%r15
|
|
adc %rdx,%r15
|
|
C // r15 r14 r12
|
|
mov $0,%r13
|
|
mov %r8,%rax
|
|
mul %r14
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
mov %r9,%rax
|
|
mul %r15
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
C // r13 r12
|
|
mov %r8,%rax
|
|
mul %r13
|
|
add %rax,%r12
|
|
adc $0,%rdx
|
|
mov %r12,(%rdi)
|
|
mov %rdx,8(%rdi)
|
|
pop %r15
|
|
pop %r14
|
|
pop %r13
|
|
pop %r12
|
|
ret
|
|
ALIGN(16)
|
|
case1:
|
|
C // one more is 3 limbs
|
|
C // r15 r14 40() 8+24()=rax 0+24()=r12
|
|
mul %r8
|
|
add %rax,%r12
|
|
mov 40(%rsi,%rcx,8),%rax
|
|
mov $0,%r13
|
|
adc %rdx,%r13
|
|
mul %r9
|
|
add %rax,%r12
|
|
nop
|
|
adc %rdx,%r13
|
|
mov %r10,%rax
|
|
mul %r14
|
|
add %rax,%r12
|
|
adc %rdx,%r13
|
|
mov %r12,%r14
|
|
mov %r11,%rax
|
|
mul %r15
|
|
add %rax,%r14
|
|
mov %r13,%r15
|
|
adc %rdx,%r15
|
|
mov %r8,%rax
|
|
mul %r15
|
|
add %rax,%r14
|
|
adc $0,%rdx
|
|
mov %r14,(%rdi)
|
|
mov %rdx,8(%rdi)
|
|
pop %r15
|
|
pop %r14
|
|
pop %r13
|
|
pop %r12
|
|
ret
|
|
EPILOGUE()
|