mpir/mpn/x86_64/divrem_euclidean_qr_2.as
2009-05-27 14:28:30 +00:00

169 lines
3.1 KiB
ActionScript

; mpn_divrem_euclidean_qr_2
; Copyright 2009 Jason Moxham
; This file is part of the MPIR Library.
; The MPIR Library is free software; you can redistribute it and/or modify
; it under the terms of the GNU Lesser General Public License as published
; by the Free Software Foundation; either version 2.1 of the License, or (at
; your option) any later version.
; The MPIR Library is distributed in the hope that it will be useful, but
; WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
; or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
; License for more details.
; You should have received a copy of the GNU Lesser General Public License
; along with the MPIR Library; see the file COPYING.LIB. If not, write
; to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
; Boston, MA 02110-1301, USA.
%include 'yasm_mac.inc'
GLOBAL_FUNC mpn_divrem_euclidean_qr_2
; rax MULTEMP
; rbx qn
; rcx dp
; rdx MULTEMP
; rdi qp
; rsi xp
; rbp i
; r8 l
; r9 h
; r10 t[0],xh
; r11
; r12 t1[0]
; r13 t1[1]
; r14 qf
; r15 q,mask,temp
push rbx
push rbp
push r12
push r13
push r14
push r15
mov rbx, rdx
dec rbx
mov r9, 0
mov rdx, [rcx+8]
not rdx
mov rax, -1
div qword [rcx+8]
mov rbp, rax
mov r8, [rsi+rbx*8]
dec rbx
mov r10, [rsi+rbx*8]
cmp r8, [rcx+8]
jae L_j2
mov rdx, r8
mov r9, r8
mov r10, [rsi+rbx*8]
mov r8, r10
mov r14, 0
jmp L_j1
L_j2:
mov r14, 1
mov rdx, r8
sub rdx, [rcx+8]
sub r10, [rcx]
sbb rdx, 0
jnc L_j3
dec r14
add r10, [rcx]
adc rdx, [rcx+8]
L_j3:
mov r9, rdx
mov r8, r10
L_j1:
dec rbx
mov rax, r9
js L_skiplp
align 16
L_lp:
cmp rax, [rcx+8]
je L_j4
mov r9, rax
bt r8, 63
adc rax, 0
mov r15, r8
mov r13, 0
sar r15, 63
and r15, [rcx+8]
add r15, r8
mul rbp
add rax, r15
mov rax, [rcx]
adc rdx, r9
mov r10, rdx
mul rdx
mov r12, [rcx]
add r12, rax
adc r13, rdx
mov rax, [rcx+8]
not r10
mul r10
sub r9, [rcx+8]
add rax, r8
adc rdx, r9
mov r9, rdx
and r9, [rcx+8]
; swapped r9,rax from here
add rax, r9
mov r15, rdx
sub r15, r10
and rdx, [rcx]
sub r12, rdx
sbb r13, 0
mov r8, [rsi+rbx*8]
sub r8, r12
sbb rax, r13
mov r10, [rcx]
mov r11, [rcx+8]
sbb r9, r9
and r10, r9
and r11, r9
add r15, r9
add r8, r10
adc rax, r11
adc r9, 0
and r10, r9
and r11, r9
add r15, r9
add r8, r10
adc rax, r11
L_j6:
mov [rdi+rbx*8], r15
dec rbx
jns L_lp
L_skiplp:
mov [rsi+8], rax
mov [rsi], r8
mov rax, r14
pop r15
pop r14
pop r13
pop r12
pop rbp
pop rbx
ret
L_j4:
mov r10, [rsi+rbx*8]
mov r15, -1
mov rdx, r8
add r10, [rcx]
adc rdx, rax
sbb rax, rax
sub rdx, [rcx]
adc rax, 0
jz L_j5
dec r15
add r10, [rcx]
adc rdx, [rcx+8]
L_j5:
mov rax, rdx
mov r8, r10
jmp L_j6
end