dnl IA-64 mpn_rsh1add_n/mpn_rsh1sub_n -- rp[] = (up[] +- vp[]) >> 1. dnl Copyright 2003, 2004, 2005 Free Software Foundation, Inc. dnl This file is part of the GNU MP Library. dnl The GNU MP Library is free software; you can redistribute it and/or modify dnl it under the terms of the GNU Lesser General Public License as published dnl by the Free Software Foundation; either version 3 of the License, or (at dnl your option) any later version. dnl The GNU MP Library is distributed in the hope that it will be useful, but dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public dnl License for more details. dnl You should have received a copy of the GNU Lesser General Public License dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. include(`../config.m4') C cycles/limb C Itanium: 2.5 C Itanium 2: 1.5 C TODO C * Rewrite function entry code using aorslsh1_n.asm style. C * Micro-optimize feed-in and wind-down code. C INPUT PARAMETERS define(`rp',`r32') define(`up',`r33') define(`vp',`r34') define(`n',`r35') define(OPERATION_rsh1add_n,1) ifdef(`OPERATION_rsh1add_n',` define(ADDSUB, add) define(PRED, ltu) define(INCR, 1) define(LIM, -1) define(func, mpn_rsh1add_n) ') ifdef(`OPERATION_rsh1sub_n',` define(ADDSUB, sub) define(PRED, gtu) define(INCR, -1) define(LIM, 0) define(func, mpn_rsh1sub_n) ') C Some useful aliases for registers we use define(`u0',`r14') define(`u1',`r15') define(`u2',`r16') define(`u3',`r17') define(`v0',`r18') define(`v1',`r19') define(`v2',`r20') define(`v3',`r21') define(`w0',`r22') define(`w1',`r23') define(`w2',`r24') define(`w3',`r25') define(`x0',`r26') define(`x1',`r9') define(`x2',`r30') define(`x3',`r31') ASM_START() PROLOGUE(func) .prologue .save ar.lc, r2 .body ifdef(`HAVE_ABI_32',` addp4 rp = 0, rp C M I addp4 up = 0, up C M I addp4 vp = 0, vp C M I zxt4 n = n C I ;; ') {.mmi; ld8 r11 = [vp], 8 C M01 ld8 r10 = [up], 8 C M01 mov.i r2 = ar.lc C I0 }{.mmi; and r14 = 3, n C M I cmp.lt p15, p0 = 4, n C M I add n = -4, n C M I ;; }{.mmi; cmp.eq p6, p0 = 1, r14 C M I cmp.eq p7, p0 = 2, r14 C M I cmp.eq p8, p0 = 3, r14 C M I }{.bbb (p6) br.dptk .Lb01 C B (p7) br.dptk .Lb10 C B (p8) br.dptk .Lb11 C B } .Lb00: ld8 v0 = [vp], 8 C M01 ld8 u0 = [up], 8 C M01 shr.u n = n, 2 C I0 ;; ld8 v1 = [vp], 8 C M01 ld8 u1 = [up], 8 C M01 ADDSUB w3 = r10, r11 C M I ;; ld8 v2 = [vp], 8 C M01 ld8 u2 = [up], 8 C M01 (p15) br.dpnt .grt4 C B ;; cmp.PRED p7, p0 = w3, r10 C M I and r8 = 1, w3 C M I ADDSUB w0 = u0, v0 C M I ;; cmp.PRED p8, p0 = w0, u0 C M I ADDSUB w1 = u1, v1 C M I ;; cmp.PRED p9, p0 = w1, u1 C M I (p7) cmp.eq.or p8, p0 = LIM, w0 C M I (p7) add w0 = INCR, w0 C M I ;; shrp x3 = w0, w3, 1 C I0 ADDSUB w2 = u2, v2 C M I (p8) cmp.eq.or p9, p0 = LIM, w1 C M I (p8) add w1 = INCR, w1 C M I br .Lcj4 C B .grt4: ld8 v3 = [vp], 8 C M01 cmp.PRED p7, p0 = w3, r10 C M I ld8 u3 = [up], 8 C M01 and r8 = 1, w3 C M I ;; ADDSUB w0 = u0, v0 C M I ld8 v0 = [vp], 8 C M01 add n = -1, n ;; cmp.PRED p8, p0 = w0, u0 C M I ld8 u0 = [up], 8 C M01 ADDSUB w1 = u1, v1 C M I ;; ld8 v1 = [vp], 8 C M01 mov.i ar.lc = n C I0 cmp.PRED p9, p0 = w1, u1 C M I ld8 u1 = [up], 8 C M01 (p7) cmp.eq.or p8, p0 = LIM, w0 C M I (p7) add w0 = INCR, w0 C M I ;; ADDSUB w2 = u2, v2 C M I ld8 v2 = [vp], 8 C M01 shrp x3 = w0, w3, 1 C I0 (p8) cmp.eq.or p9, p0 = LIM, w1 C M I (p8) add w1 = INCR, w1 C M I br .LL00 C B .Lb01: ADDSUB w2 = r10, r11 C M I shr.u n = n, 2 C I0 (p15) br.dpnt .grt1 C B ;; cmp.PRED p6, p7 = w2, r10 C M I shr.u x2 = w2, 1 C I0 and r8 = 1, w2 C M I ;; (p6) dep x2 = -1, x2, 63, 1 C I0 br .Lcj1 C B .grt1: ld8 v3 = [vp], 8 C M01 ld8 u3 = [up], 8 C M01 ;; ld8 v0 = [vp], 8 C M01 ld8 u0 = [up], 8 C M01 mov.i ar.lc = n C FIXME swap with next I0 ;; ld8 v1 = [vp], 8 C M01 ld8 u1 = [up], 8 C M01 ;; ld8 v2 = [vp], 8 C M01 ld8 u2 = [up], 8 C M01 cmp.PRED p6, p0 = w2, r10 C M I and r8 = 1, w2 C M I ADDSUB w3 = u3, v3 C M I br.cloop.dptk .grt5 C B ;; cmp.PRED p7, p0 = w3, u3 C M I ;; ADDSUB w0 = u0, v0 C M I (p6) cmp.eq.or p7, p0 = LIM, w3 C M I (p6) add w3 = INCR, w3 C M I ;; cmp.PRED p8, p0 = w0, u0 C M I shrp x2 = w3, w2, 1 C I0 ADDSUB w1 = u1, v1 C M I ;; cmp.PRED p9, p0 = w1, u1 C M I (p7) cmp.eq.or p8, p0 = LIM, w0 C M I (p7) add w0 = INCR, w0 C M I br .Lcj5 C B .grt5: ld8 v3 = [vp], 8 C M01 cmp.PRED p7, p0 = w3, u3 C M I ld8 u3 = [up], 8 C M01 ;; ADDSUB w0 = u0, v0 C M I ld8 v0 = [vp], 8 C M01 (p6) cmp.eq.or p7, p0 = LIM, w3 C M I (p6) add w3 = INCR, w3 C M I ;; cmp.PRED p8, p0 = w0, u0 C M I shrp x2 = w3, w2, 1 C I0 ld8 u0 = [up], 8 C M01 ADDSUB w1 = u1, v1 C M I ;; ld8 v1 = [vp], 8 C M01 cmp.PRED p9, p0 = w1, u1 C M I ld8 u1 = [up], 8 C M01 (p7) cmp.eq.or p8, p0 = LIM, w0 C M I (p7) add w0 = INCR, w0 C M I br .LL01 C B .Lb10: ld8 v2 = [vp], 8 C M01 ld8 u2 = [up], 8 C M01 shr.u n = n, 2 C I0 ADDSUB w1 = r10, r11 C M I (p15) br.dpnt .grt2 C B ;; cmp.PRED p9, p0 = w1, r10 C M I and r8 = 1, w1 C M I ADDSUB w2 = u2, v2 C M I ;; cmp.PRED p6, p0 = w2, u2 C M I ;; (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I ;; shrp x1 = w2, w1, 1 C I0 shr.u x2 = w2, 1 C I0 br .Lcj2 C B .grt2: ld8 v3 = [vp], 8 C M01 ld8 u3 = [up], 8 C M01 ;; ld8 v0 = [vp], 8 C M01 ld8 u0 = [up], 8 C M01 mov.i ar.lc = n C I0 ;; ld8 v1 = [vp], 8 C M01 cmp.PRED p9, p0 = w1, r10 C M I ld8 u1 = [up], 8 C M01 and r8 = 1, w1 C M I ;; ADDSUB w2 = u2, v2 C M I ld8 v2 = [vp], 8 C M01 ;; cmp.PRED p6, p0 = w2, u2 C M I ld8 u2 = [up], 8 C M01 ADDSUB w3 = u3, v3 C M I br.cloop.dptk .grt6 C B ;; cmp.PRED p7, p0 = w3, u3 C M I (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I ;; shrp x1 = w2, w1, 1 C I0 ADDSUB w0 = u0, v0 C M I (p6) cmp.eq.or p7, p0 = LIM, w3 C M I (p6) add w3 = INCR, w3 C M I br .Lcj6 C B .grt6: ld8 v3 = [vp], 8 C M01 cmp.PRED p7, p0 = w3, u3 C M I ld8 u3 = [up], 8 C M01 (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I ;; shrp x1 = w2, w1, 1 C I0 ADDSUB w0 = u0, v0 C M I ld8 v0 = [vp], 8 C M01 (p6) cmp.eq.or p7, p0 = LIM, w3 C M I (p6) add w3 = INCR, w3 C M I br .LL10 C B .Lb11: ld8 v1 = [vp], 8 C M01 ld8 u1 = [up], 8 C M01 shr.u n = n, 2 C I0 ;; ld8 v2 = [vp], 8 C M01 ld8 u2 = [up], 8 C M01 ADDSUB w0 = r10, r11 C M I (p15) br.dpnt .grt3 C B ;; cmp.PRED p8, p0 = w0, r10 C M I ADDSUB w1 = u1, v1 C M I and r8 = 1, w0 C M I ;; cmp.PRED p9, p0 = w1, u1 C M I ;; ADDSUB w2 = u2, v2 C M I (p8) cmp.eq.or p9, p0 = LIM, w1 C M I (p8) add w1 = INCR, w1 C M I ;; cmp.PRED p6, p0 = w2, u2 C M I shrp x0 = w1, w0, 1 C I0 ;; (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I br .Lcj3 C B .grt3: ld8 v3 = [vp], 8 C M01 ld8 u3 = [up], 8 C M01 ;; ld8 v0 = [vp], 8 C M01 mov.i ar.lc = n C I0 cmp.PRED p8, p0 = w0, r10 C M I ld8 u0 = [up], 8 C M01 ADDSUB w1 = u1, v1 C M I and r8 = 1, w0 C M I ;; ld8 v1 = [vp], 8 C M01 cmp.PRED p9, p0 = w1, u1 C M I ld8 u1 = [up], 8 C M01 ;; ADDSUB w2 = u2, v2 C M I ld8 v2 = [vp], 8 C M01 (p8) cmp.eq.or p9, p0 = LIM, w1 C M I (p8) add w1 = INCR, w1 C M I ;; cmp.PRED p6, p0 = w2, u2 C M I shrp x0 = w1, w0, 1 C I0 ld8 u2 = [up], 8 C M01 ADDSUB w3 = u3, v3 C M I br.cloop.dptk .grt7 C B ;; cmp.PRED p7, p0 = w3, u3 C M I (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I br .Lcj7 C B .grt7: ld8 v3 = [vp], 8 C M01 cmp.PRED p7, p0 = w3, u3 C M I ld8 u3 = [up], 8 C M01 (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I br .LL11 C B C *** MAIN LOOP START *** ALIGN(32) .Loop: st8 [rp] = x3, 8 C M23 ld8 v3 = [vp], 8 C M01 cmp.PRED p7, p0 = w3, u3 C M I ld8 u3 = [up], 8 C M01 (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I ;; .LL11: st8 [rp] = x0, 8 C M23 shrp x1 = w2, w1, 1 C I0 ADDSUB w0 = u0, v0 C M I ld8 v0 = [vp], 8 C M01 (p6) cmp.eq.or p7, p0 = LIM, w3 C M I (p6) add w3 = INCR, w3 C M I ;; .LL10: cmp.PRED p8, p0 = w0, u0 C M I shrp x2 = w3, w2, 1 C I0 nop.b 0 ld8 u0 = [up], 8 C M01 ADDSUB w1 = u1, v1 C M I nop.b 0 ;; st8 [rp] = x1, 8 C M23 ld8 v1 = [vp], 8 C M01 cmp.PRED p9, p0 = w1, u1 C M I ld8 u1 = [up], 8 C M01 (p7) cmp.eq.or p8, p0 = LIM, w0 C M I (p7) add w0 = INCR, w0 C M I ;; .LL01: st8 [rp] = x2, 8 C M23 shrp x3 = w0, w3, 1 C I0 ADDSUB w2 = u2, v2 C M I ld8 v2 = [vp], 8 C M01 (p8) cmp.eq.or p9, p0 = LIM, w1 C M I (p8) add w1 = INCR, w1 C M I ;; .LL00: cmp.PRED p6, p0 = w2, u2 C M I shrp x0 = w1, w0, 1 C I0 nop.b 0 ld8 u2 = [up], 8 C M01 ADDSUB w3 = u3, v3 C M I br.cloop.dptk .Loop C B ;; C *** MAIN LOOP END *** .Lskip: st8 [rp] = x3, 8 C M23 cmp.PRED p7, p0 = w3, u3 C M I (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I ;; .Lcj7: st8 [rp] = x0, 8 C M23 shrp x1 = w2, w1, 1 C I0 ADDSUB w0 = u0, v0 C M I (p6) cmp.eq.or p7, p0 = LIM, w3 C M I (p6) add w3 = INCR, w3 C M I ;; .Lcj6: cmp.PRED p8, p0 = w0, u0 C M I shrp x2 = w3, w2, 1 C I0 ADDSUB w1 = u1, v1 C M I ;; st8 [rp] = x1, 8 C M23 cmp.PRED p9, p0 = w1, u1 C M I (p7) cmp.eq.or p8, p0 = LIM, w0 C M I (p7) add w0 = INCR, w0 C M I ;; .Lcj5: st8 [rp] = x2, 8 C M23 shrp x3 = w0, w3, 1 C I0 ADDSUB w2 = u2, v2 C M I (p8) cmp.eq.or p9, p0 = LIM, w1 C M I (p8) add w1 = INCR, w1 C M I ;; .Lcj4: cmp.PRED p6, p0 = w2, u2 C M I shrp x0 = w1, w0, 1 C I0 ;; st8 [rp] = x3, 8 C M23 (p9) cmp.eq.or p6, p0 = LIM, w2 C M I (p9) add w2 = INCR, w2 C M I ;; .Lcj3: st8 [rp] = x0, 8 C M23 shrp x1 = w2, w1, 1 C I0 shr.u x2 = w2, 1 C I0 ;; .Lcj2: st8 [rp] = x1, 8 C M23 (p6) dep x2 = -1, x2, 63, 1 C I0 ;; .Lcj1: st8 [rp] = x2 C M23 mov.i ar.lc = r2 C I0 br.ret.sptk.many b0 C B EPILOGUE()