mpir/mpn/x86_64w/core2/copyi.asm
gladman 1485e3dc1b 1. Add Jason's new assembler code to the Windows builds
2. Tidy up assembler to prepare for Windows nehalem build
2009-12-02 16:24:00 +00:00

222 lines
4.2 KiB
NASM

; Copyright 2009 Jason Moxham
;
; Windows Conversion Copyright 2008 Brian Gladman
;
; This file is part of the MPIR Library.
; The MPIR Library is free software; you can redistribute it and/or modify
; it under the terms of the GNU Lesser General Public License as published
; by the Free Software Foundation; either version 2.1 of the License, or (at
; your option) any later version.
; The MPIR Library is distributed in the hope that it will be useful, but
; WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
; or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
; License for more details.
; You should have received a copy of the GNU Lesser General Public License
; along with the MPIR Library; see the file COPYING.LIB. If not, write
; to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
; Boston, MA 02110-1301, USA.
;
; void mpn_copyi(mp_ptr, mp_ptr, mp_size_t)
; rdi rsi rdx
; rcx rdx r8d
%define SMALL_LOOP 1
%include "..\yasm_mac.inc"
CPU Core2
BITS 64
LEAF_PROC mpn_copyi
movsxd r8, r8d
cmp r8, 0
jz .9
%if SMALL_LOOP <> 0
cmp r8, 10
jge .2
xalign 16
.1:
mov rax, [rdx]
mov [rcx], rax
lea rdx, [rdx+8]
lea rcx, [rcx+8]
sub r8, 1
jnz .1
ret
%endif
.2:
mov rax, rcx
sub rax, rdx
test rax, 0xF
jz .17
test rcx, 0xF
jz .10
mov r9, 5
sub r9, r8
lea rdx, [rdx+r8*8-40]
lea rcx, [rcx+r8*8-40]
movapd xmm1, [rdx+r9*8]
movq [rcx+r9*8], xmm1
add rcx, 8
%if SMALL_LOOP = 0
cmp r8, 1
jz .9
%endif
cmp r9, 0
jge .4
xalign 16
.3:
movapd xmm0, [rdx+r9*8+16]
add r9, 4
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8-32], xmm1
movapd xmm1, [rdx+r9*8+0]
shufpd xmm0, xmm1, 1
movapd [rcx+r9*8-16], xmm0
jnc .3
.4:
cmp r9, 2
ja .8
jz .7
jp .6
xalign 16
.5:
movapd xmm0, [rdx+r9*8+16]
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8], xmm1
movapd xmm1, [rdx+r9*8+32]
shufpd xmm0, xmm1, 1
movapd [rcx+r9*8+16], xmm0
ret
xalign 16
.6:
movapd xmm0, [rdx+r9*8+16]
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8], xmm1
movhpd [rcx+r9*8+16], xmm0
ret
xalign 16
.7:
movapd xmm0, [rdx+r9*8+16]
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8], xmm1
ret
xalign 16
.8:
movhpd [rcx+r9*8], xmm1
.9:
ret
.10:
mov r9, 4
sub r9, r8
lea rdx, [rdx+r8*8-32]
lea rcx, [rcx+r8*8-32]
movapd xmm1, [rdx+r9*8-8]
sub rdx, 8
cmp r9, 0
jge .12
xalign 16
.11:
movapd xmm0, [rdx+r9*8+16]
add r9, 4
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8-32], xmm1
movapd xmm1, [rdx+r9*8+0]
shufpd xmm0, xmm1, 1
movapd [rcx+r9*8-16], xmm0
jnc .11
.12:
cmp r9, 2
ja .16
jz .15
jp .14
xalign 16
.13:
movapd xmm0, [rdx+r9*8+16]
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8], xmm1
movapd xmm1, [rdx+r9*8+32]
shufpd xmm0, xmm1, 1
movapd [rcx+r9*8+16], xmm0
ret
xalign 16
.14:
movapd xmm0, [rdx+r9*8+16]
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8], xmm1
movhpd [rcx+r9*8+16], xmm0
ret
xalign 16
.15:
movapd xmm0, [rdx+r9*8+16]
shufpd xmm1, xmm0, 1
movapd [rcx+r9*8], xmm1
ret
xalign 16
.16:movhpd [rcx+r9*8], xmm1
ret
xalign 16
.17:
mov r9, 3
sub r9, r8
test rcx, 0xF
lea rdx, [rdx+r8*8-24]
lea rcx, [rcx+r8*8-24]
jz .18
mov rax, [rdx+r9*8]
mov [rcx+r9*8], rax
add r9, 1
.18:
cmp r9, 0
jge .20
xalign 16
.19:
add r9, 4
movapd xmm0, [rdx+r9*8-32]
movapd [rcx+r9*8-32], xmm0
movapd xmm1, [rdx+r9*8-16]
movapd [rcx+r9*8-16], xmm1
jnc .19
.20:
cmp r9, 2
ja .22
je .24
jp .23
.21:
movapd xmm0, [rdx+r9*8]
movapd [rcx+r9*8], xmm0
mov rax, [rdx+r9*8+16]
mov [rcx+r9*8+16], rax
.22:
ret
xalign 16
.23:
movapd xmm0, [rdx+r9*8]
movapd [rcx+r9*8], xmm0
ret
xalign 16
.24:
mov rax, [rdx+r9*8]
mov [rcx+r9*8], rax
ret
end