mpir/mpn/x86_64/k8/karasub.asm
2011-07-04 04:34:15 +00:00

321 lines
6.4 KiB
NASM

dnl mpn_karasub
dnl Copyright 2011 The Code Cavern
dnl This file is part of the MPIR Library.
dnl The MPIR Library is free software; you can redistribute it and/or modify
dnl it under the terms of the GNU Lesser General Public License as published
dnl by the Free Software Foundation; either version 2.1 of the License, or (at
dnl your option) any later version.
dnl The MPIR Library is distributed in the hope that it will be useful, but
dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
dnl License for more details.
dnl You should have received a copy of the GNU Lesser General Public License
dnl along with the MPIR Library; see the file COPYING.LIB. If not, write
dnl to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
dnl Boston, MA 02110-1301, USA.
include(`../config.m4')
ASM_START()
PROLOGUE(mpn_karasub)
push %rbx
push %rbp
push %r12
push %r13
push %r14
push %r15
#rp is rdi
#tp is rsi
#n is rdx and put it on the stack
push %rdx
shr $1,%rdx
#n2 is rdx
lea (%rdx,%rdx,1),%rcx
# 2*n2 is rcx
# L is rdi
# H is rbp
# tp is rsi
lea (%rdi,%rcx,8),%rbp
xor %rax,%rax
xor %rbx,%rbx
# rax rbx are the carrys
lea -24(%rdi,%rdx,8),%rdi
lea -24(%rsi,%rdx,8),%rsi
lea -24(%rbp,%rdx,8),%rbp
mov $0,%rcx
neg %rdx
add $3,%rdx
add $3,%rcx
# assume n>8 or some such
.align 16
lp: bt $2,%rbx
mov (%rdi,%rcx,8),%r8
adc (%rbp,%rdx,8),%r8
mov %r8,%r12
mov 8(%rdi,%rcx,8),%r9
adc 8(%rbp,%rdx,8),%r9
mov 16(%rdi,%rcx,8),%r10
adc 16(%rbp,%rdx,8),%r10
mov 24(%rdi,%rcx,8),%r11
adc 24(%rbp,%rdx,8),%r11
rcl $1,%rbx
bt $1,%rax
mov %r11,%r15
adc (%rdi,%rdx,8),%r8
mov %r9,%r13
adc 8(%rdi,%rdx,8),%r9
mov %r10,%r14
adc 16(%rdi,%rdx,8),%r10
adc 24(%rdi,%rdx,8),%r11
rcl $1,%rax
bt $2,%rbx
adc (%rbp,%rcx,8),%r12
adc 8(%rbp,%rcx,8),%r13
adc 16(%rbp,%rcx,8),%r14
adc 24(%rbp,%rcx,8),%r15
rcl $1,%rbx
bt $1,%rax
sbb (%rsi,%rdx,8),%r8
sbb 8(%rsi,%rdx,8),%r9
sbb 16(%rsi,%rdx,8),%r10
sbb 24(%rsi,%rdx,8),%r11
mov %r10,16(%rdi,%rcx,8)
mov %r11,24(%rdi,%rcx,8)
rcl $1,%rax
bt $2,%rbx
mov %r8,(%rdi,%rcx,8)
mov %r9,8(%rdi,%rcx,8)
sbb (%rsi,%rcx,8),%r12
sbb 8(%rsi,%rcx,8),%r13
sbb 16(%rsi,%rcx,8),%r14
sbb 24(%rsi,%rcx,8),%r15
rcl $1,%rbx
add $4,%rcx
mov %r12,(%rbp,%rdx,8)
mov %r13,8(%rbp,%rdx,8)
mov %r14,16(%rbp,%rdx,8)
mov %r15,24(%rbp,%rdx,8)
add $4,%rdx
jnc lp
cmp $2,%rdx
jg case0
jz case1
jp case2
case3: bt $2,%rbx
mov (%rdi,%rcx,8),%r8
adc (%rbp,%rdx,8),%r8
mov %r8,%r12
mov 8(%rdi,%rcx,8),%r9
adc 8(%rbp,%rdx,8),%r9
mov 16(%rdi,%rcx,8),%r10
adc 16(%rbp,%rdx,8),%r10
#mov 24(%rdi,%rcx,8),%r11
#adc 24(%rbp,%rdx,8),%r11
rcl $1,%rbx
bt $1,%rax
#mov %r11,%r15
adc (%rdi,%rdx,8),%r8
mov %r9,%r13
adc 8(%rdi,%rdx,8),%r9
mov %r10,%r14
adc 16(%rdi,%rdx,8),%r10
#adc 24(%rdi,%rdx,8),%r11
rcl $1,%rax
bt $2,%rbx
adc (%rbp,%rcx,8),%r12
adc 8(%rbp,%rcx,8),%r13
adc 16(%rbp,%rcx,8),%r14
#adc 24(%rbp,%rcx,8),%r15
rcl $1,%rbx
bt $1,%rax
sbb (%rsi,%rdx,8),%r8
sbb 8(%rsi,%rdx,8),%r9
sbb 16(%rsi,%rdx,8),%r10
#sbb 24(%rsi,%rdx,8),%r11
mov %r10,16(%rdi,%rcx,8)
#mov %r11,24(%rdi,%rcx,8)
rcl $1,%rax
bt $2,%rbx
mov %r8,(%rdi,%rcx,8)
mov %r9,8(%rdi,%rcx,8)
sbb (%rsi,%rcx,8),%r12
sbb 8(%rsi,%rcx,8),%r13
sbb 16(%rsi,%rcx,8),%r14
#sbb 24(%rsi,%rcx,8),%r15
rcl $1,%rbx
add $3,%rcx
mov %r12,(%rbp,%rdx,8)
mov %r13,8(%rbp,%rdx,8)
mov %r14,16(%rbp,%rdx,8)
#mov %r15,24(%rbp,%rdx,8)
add $3,%rdx
jmp fin
case2: bt $2,%rbx
mov (%rdi,%rcx,8),%r8
adc (%rbp,%rdx,8),%r8
mov %r8,%r12
mov 8(%rdi,%rcx,8),%r9
adc 8(%rbp,%rdx,8),%r9
#mov 16(%rdi,%rcx,8),%r10
#adc 16(%rbp,%rdx,8),%r10
#mov 24(%rdi,%rcx,8),%r11
#adc 24(%rbp,%rdx,8),%r11
rcl $1,%rbx
bt $1,%rax
#mov %r11,%r15
adc (%rdi,%rdx,8),%r8
mov %r9,%r13
adc 8(%rdi,%rdx,8),%r9
#mov %r10,%r14
#adc 16(%rdi,%rdx,8),%r10
#adc 24(%rdi,%rdx,8),%r11
rcl $1,%rax
bt $2,%rbx
adc (%rbp,%rcx,8),%r12
adc 8(%rbp,%rcx,8),%r13
#adc 16(%rbp,%rcx,8),%r14
#adc 24(%rbp,%rcx,8),%r15
rcl $1,%rbx
bt $1,%rax
sbb (%rsi,%rdx,8),%r8
sbb 8(%rsi,%rdx,8),%r9
#sbb 16(%rsi,%rdx,8),%r10
#sbb 24(%rsi,%rdx,8),%r11
#mov %r10,16(%rdi,%rcx,8)
#mov %r11,24(%rdi,%rcx,8)
rcl $1,%rax
bt $2,%rbx
mov %r8,(%rdi,%rcx,8)
mov %r9,8(%rdi,%rcx,8)
sbb (%rsi,%rcx,8),%r12
sbb 8(%rsi,%rcx,8),%r13
#sbb 16(%rsi,%rcx,8),%r14
#sbb 24(%rsi,%rcx,8),%r15
rcl $1,%rbx
add $2,%rcx
mov %r12,(%rbp,%rdx,8)
mov %r13,8(%rbp,%rdx,8)
#mov %r14,16(%rbp,%rdx,8)
#mov %r15,24(%rbp,%rdx,8)
add $2,%rdx
jmp fin
case1: bt $2,%rbx
mov (%rdi,%rcx,8),%r8
adc (%rbp,%rdx,8),%r8
mov %r8,%r12
#mov 8(%rdi,%rcx,8),%r9
#adc 8(%rbp,%rdx,8),%r9
#mov 16(%rdi,%rcx,8),%r10
#adc 16(%rbp,%rdx,8),%r10
#mov 24(%rdi,%rcx,8),%r11
#adc 24(%rbp,%rdx,8),%r11
rcl $1,%rbx
bt $1,%rax
#mov %r11,%r15
adc (%rdi,%rdx,8),%r8
#mov %r9,%r13
#adc 8(%rdi,%rdx,8),%r9
#mov %r10,%r14
#adc 16(%rdi,%rdx,8),%r10
#adc 24(%rdi,%rdx,8),%r11
rcl $1,%rax
bt $2,%rbx
adc (%rbp,%rcx,8),%r12
#adc 8(%rbp,%rcx,8),%r13
#adc 16(%rbp,%rcx,8),%r14
#adc 24(%rbp,%rcx,8),%r15
rcl $1,%rbx
bt $1,%rax
sbb (%rsi,%rdx,8),%r8
#sbb 8(%rsi,%rdx,8),%r9
#sbb 16(%rsi,%rdx,8),%r10
#sbb 24(%rsi,%rdx,8),%r11
#mov %r10,16(%rdi,%rcx,8)
#mov %r11,24(%rdi,%rcx,8)
rcl $1,%rax
bt $2,%rbx
mov %r8,(%rdi,%rcx,8)
#mov %r9,8(%rdi,%rcx,8)
sbb (%rsi,%rcx,8),%r12
#sbb 8(%rsi,%rcx,8),%r13
#sbb 16(%rsi,%rcx,8),%r14
#sbb 24(%rsi,%rcx,8),%r15
rcl $1,%rbx
add $1,%rcx
mov %r12,(%rbp,%rdx,8)
#mov %r13,8(%rbp,%rdx,8)
#mov %r14,16(%rbp,%rdx,8)
#mov %r15,24(%rbp,%rdx,8)
add $1,%rdx
jmp fin
case0:
fin: # if odd the do next two
pop %r8
bt $0,%r8
jnc notodd
xor %r10,%r10
mov (%rbp,%rcx,8),%r8
mov 8(%rbp,%rcx,8),%r9
sub (%rsi,%rcx,8),%r8
sbb 8(%rsi,%rcx,8),%r9
rcl $1,%r10
add %r8,(%rbp,%rdx,8)
adc %r9,8(%rbp,%rdx,8)
mov %rdx,%rsi
l7: adc $0,16(%rbp,%rdx,8)
inc %rdx
jc l7
mov %rsi,%rdx
bt $0,%r10
l8: sbb $0,16(%rbp,%rdx,8)
inc %rdx
jc l8
mov %rsi,%rdx
# add in all carrys
# should we do the borrows last as it may be possible to underflow
# could use popcount
notodd: mov %rcx,%rsi
bt $0,%rax
l1: sbb $0,(%rdi,%rcx,8)
inc %rcx
jc l1
mov %rsi,%rcx
bt $1,%rax
l2: adc $0,(%rdi,%rcx,8)
inc %rcx
jc l2
mov %rsi,%rcx
bt $2,%rbx
l3: adc $0,(%rdi,%rcx,8)
inc %rcx
jc l3
mov %rdx,%rsi
bt $0,%rbx
l4: sbb $0,(%rbp,%rdx,8)
inc %rdx
jc l4
mov %rsi,%rdx
bt $1,%rbx
l5: adc $0,(%rbp,%rdx,8)
inc %rdx
jc l5
mov %rsi,%rdx
bt $2,%rbx
l6: adc $0,(%rbp,%rdx,8)
inc %rdx
jc l6
pop %r15
pop %r14
pop %r13
pop %r12
pop %rbp
pop %rbx
ret
EPILOGUE()