mpir/mpn/x86_64w/divebyfobm1.asm

109 lines
2.8 KiB
NASM

; Copyright 2008 Jason Moxham and Brian Gladman
;
; This file is part of the MPIR Library.
; The MPIR Library is free software; you can redistribute it and/or modify
; it under the terms of the GNU Lesser General Public License as published
; by the Free Software Foundation; either version 2.1 of the License, or (at
; your option) any later veSRCon.
; The MPIR Library is distributed in the hope that it will be useful, but
; WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
; or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
; License for more details.
; You should have received a copy of the GNU Lesser General Public License
; along with the MPIR Library; see the file COPYING.LIB. If not, write
; to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
; Boston, MA 02110-1301, USA.
;
; rcx[r8] = rdx[r8] / r9 where [rsp+40] must be set to (B - 1) / r9 on entry
;
; mp_limb_t mpn_divexact_fobm1(mp_ptr, mp_ptr, mp_size_t, mp_limb_t, mp_limb_t)
; rax rdi rsi rdx rcx r8
; rax rcx rdx r8 r9 [rsp+40]
%include "yasm_mac.inc"
%define CARRY_OUT
BITS 64
LEAF_PROC mpn_divexact_fobm1
mov rax, r8
mov r8, [rsp+40]
%ifdef CARRY_OUT
mov [rsp+32], r9
%endif
lea r10, [rdx+rax*8-24]
lea r11, [rcx+rax*8-24]
mov ecx, 3
%ifdef CARRY_IN
mov r9, [rsp+48] ; r9 is the carry in
imul r9, r8 ; this is needed if we have non-zero carry in
%else
mov r9, 0
%endif
sub rcx, rax
jnc .2
xalign 16
.1:
mov rax, [r10+rcx*8]
mul r8
sub r9, rax
mov [r11+rcx*8], r9
sbb r9, rdx
mov rax, [r10+rcx*8+8]
mul r8
sub r9, rax
mov [r11+rcx*8+8], r9
sbb r9, rdx
mov rax, [r10+rcx*8+16]
mul r8
sub r9, rax
mov [r11+rcx*8+16], r9
sbb r9, rdx
mov rax, [r10+rcx*8+24]
mul r8
sub r9, rax
mov [r11+rcx*8+24], r9
sbb r9, rdx
add rcx, 4
jnc .1
.2:
test rcx, 2
jnz .3
mov rax, [r10+rcx*8]
mul r8
sub r9, rax
mov [r11+rcx*8], r9
sbb r9, rdx
mov rax, [r10+rcx*8+8]
mul r8
sub r9, rax
mov [r11+rcx*8+8], r9
sbb r9, rdx
add rcx, 2
.3:
test rcx, 1
jnz .4
mov rax, [r10+rcx*8]
mul r8
sub r9, rax
mov [r11+rcx*8], r9
sbb r9, rdx
.4:
%ifdef CARRY_OUT
imul r9, [rsp+32]
%endif
mov rax, r9
neg rax
ret
end