X-Git-Url: https://oss.titaniummirror.com/gitweb/?a=blobdiff_plain;f=gmp%2Fmpn%2Fia64%2Frsh1aors_n.asm;fp=gmp%2Fmpn%2Fia64%2Frsh1aors_n.asm;h=366b5c50bb38d30e768c8d05b551c917c515097c;hb=6fed43773c9b0ce596dca5686f37ac3fc0fa11c0;hp=0000000000000000000000000000000000000000;hpb=27b11d56b743098deb193d510b337ba22dc52e5c;p=msp430-gcc.git diff --git a/gmp/mpn/ia64/rsh1aors_n.asm b/gmp/mpn/ia64/rsh1aors_n.asm new file mode 100644 index 00000000..366b5c50 --- /dev/null +++ b/gmp/mpn/ia64/rsh1aors_n.asm @@ -0,0 +1,432 @@ +dnl IA-64 mpn_rsh1add_n/mpn_rsh1sub_n -- rp[] = (up[] +- vp[]) >> 1. + +dnl Copyright 2003, 2004, 2005 Free Software Foundation, Inc. + +dnl This file is part of the GNU MP Library. + +dnl The GNU MP Library is free software; you can redistribute it and/or modify +dnl it under the terms of the GNU Lesser General Public License as published +dnl by the Free Software Foundation; either version 3 of the License, or (at +dnl your option) any later version. + +dnl The GNU MP Library is distributed in the hope that it will be useful, but +dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY +dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public +dnl License for more details. + +dnl You should have received a copy of the GNU Lesser General Public License +dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. + +include(`../config.m4') + +C cycles/limb +C Itanium: 2.5 +C Itanium 2: 1.5 + +C TODO +C * Rewrite function entry code using aorslsh1_n.asm style. +C * Micro-optimize feed-in and wind-down code. + +C INPUT PARAMETERS +define(`rp',`r32') +define(`up',`r33') +define(`vp',`r34') +define(`n',`r35') + +ifdef(`OPERATION_rsh1add_n',` + define(ADDSUB, add) + define(PRED, ltu) + define(INCR, 1) + define(LIM, -1) + define(func, mpn_rsh1add_n) +') +ifdef(`OPERATION_rsh1sub_n',` + define(ADDSUB, sub) + define(PRED, gtu) + define(INCR, -1) + define(LIM, 0) + define(func, mpn_rsh1sub_n) +') + +C Some useful aliases for registers we use +define(`u0',`r14') define(`u1',`r15') define(`u2',`r16') define(`u3',`r17') +define(`v0',`r18') define(`v1',`r19') define(`v2',`r20') define(`v3',`r21') +define(`w0',`r22') define(`w1',`r23') define(`w2',`r24') define(`w3',`r25') +define(`x0',`r26') define(`x1',`r9') define(`x2',`r30') define(`x3',`r31') + +MULFUNC_PROLOGUE(mpn_rsh1add_n mpn_rsh1sub_n) + +ASM_START() +PROLOGUE(func) + .prologue + .save ar.lc, r2 + .body +ifdef(`HAVE_ABI_32',` + addp4 rp = 0, rp C M I + addp4 up = 0, up C M I + addp4 vp = 0, vp C M I + zxt4 n = n C I + ;; +') + {.mmi; ld8 r11 = [vp], 8 C M01 + ld8 r10 = [up], 8 C M01 + mov.i r2 = ar.lc C I0 +}{.mmi; and r14 = 3, n C M I + cmp.lt p15, p0 = 4, n C M I + add n = -4, n C M I + ;; +}{.mmi; cmp.eq p6, p0 = 1, r14 C M I + cmp.eq p7, p0 = 2, r14 C M I + cmp.eq p8, p0 = 3, r14 C M I +}{.bbb + (p6) br.dptk .Lb01 C B + (p7) br.dptk .Lb10 C B + (p8) br.dptk .Lb11 C B +} + +.Lb00: ld8 v0 = [vp], 8 C M01 + ld8 u0 = [up], 8 C M01 + shr.u n = n, 2 C I0 + ;; + ld8 v1 = [vp], 8 C M01 + ld8 u1 = [up], 8 C M01 + ADDSUB w3 = r10, r11 C M I + ;; + ld8 v2 = [vp], 8 C M01 + ld8 u2 = [up], 8 C M01 + (p15) br.dpnt .grt4 C B + ;; + + cmp.PRED p7, p0 = w3, r10 C M I + and r8 = 1, w3 C M I + ADDSUB w0 = u0, v0 C M I + ;; + cmp.PRED p8, p0 = w0, u0 C M I + ADDSUB w1 = u1, v1 C M I + ;; + cmp.PRED p9, p0 = w1, u1 C M I + (p7) cmp.eq.or p8, p0 = LIM, w0 C M I + (p7) add w0 = INCR, w0 C M I + ;; + shrp x3 = w0, w3, 1 C I0 + ADDSUB w2 = u2, v2 C M I + (p8) cmp.eq.or p9, p0 = LIM, w1 C M I + (p8) add w1 = INCR, w1 C M I + br .Lcj4 C B + +.grt4: ld8 v3 = [vp], 8 C M01 + cmp.PRED p7, p0 = w3, r10 C M I + ld8 u3 = [up], 8 C M01 + and r8 = 1, w3 C M I + ;; + ADDSUB w0 = u0, v0 C M I + ld8 v0 = [vp], 8 C M01 + add n = -1, n + ;; + cmp.PRED p8, p0 = w0, u0 C M I + ld8 u0 = [up], 8 C M01 + ADDSUB w1 = u1, v1 C M I + ;; + ld8 v1 = [vp], 8 C M01 + mov.i ar.lc = n C I0 + cmp.PRED p9, p0 = w1, u1 C M I + ld8 u1 = [up], 8 C M01 + (p7) cmp.eq.or p8, p0 = LIM, w0 C M I + (p7) add w0 = INCR, w0 C M I + ;; + ADDSUB w2 = u2, v2 C M I + ld8 v2 = [vp], 8 C M01 + shrp x3 = w0, w3, 1 C I0 + (p8) cmp.eq.or p9, p0 = LIM, w1 C M I + (p8) add w1 = INCR, w1 C M I + br .LL00 C B + + +.Lb01: ADDSUB w2 = r10, r11 C M I + shr.u n = n, 2 C I0 + (p15) br.dpnt .grt1 C B + ;; + + cmp.PRED p6, p7 = w2, r10 C M I + shr.u x2 = w2, 1 C I0 + and r8 = 1, w2 C M I + ;; + (p6) dep x2 = -1, x2, 63, 1 C I0 + br .Lcj1 C B + +.grt1: ld8 v3 = [vp], 8 C M01 + ld8 u3 = [up], 8 C M01 + ;; + ld8 v0 = [vp], 8 C M01 + ld8 u0 = [up], 8 C M01 + mov.i ar.lc = n C FIXME swap with next I0 + ;; + ld8 v1 = [vp], 8 C M01 + ld8 u1 = [up], 8 C M01 + ;; + ld8 v2 = [vp], 8 C M01 + ld8 u2 = [up], 8 C M01 + cmp.PRED p6, p0 = w2, r10 C M I + and r8 = 1, w2 C M I + ADDSUB w3 = u3, v3 C M I + br.cloop.dptk .grt5 C B + ;; + + cmp.PRED p7, p0 = w3, u3 C M I + ;; + ADDSUB w0 = u0, v0 C M I + (p6) cmp.eq.or p7, p0 = LIM, w3 C M I + (p6) add w3 = INCR, w3 C M I + ;; + cmp.PRED p8, p0 = w0, u0 C M I + shrp x2 = w3, w2, 1 C I0 + ADDSUB w1 = u1, v1 C M I + ;; + cmp.PRED p9, p0 = w1, u1 C M I + (p7) cmp.eq.or p8, p0 = LIM, w0 C M I + (p7) add w0 = INCR, w0 C M I + br .Lcj5 C B + +.grt5: ld8 v3 = [vp], 8 C M01 + cmp.PRED p7, p0 = w3, u3 C M I + ld8 u3 = [up], 8 C M01 + ;; + ADDSUB w0 = u0, v0 C M I + ld8 v0 = [vp], 8 C M01 + (p6) cmp.eq.or p7, p0 = LIM, w3 C M I + (p6) add w3 = INCR, w3 C M I + ;; + cmp.PRED p8, p0 = w0, u0 C M I + shrp x2 = w3, w2, 1 C I0 + ld8 u0 = [up], 8 C M01 + ADDSUB w1 = u1, v1 C M I + ;; + ld8 v1 = [vp], 8 C M01 + cmp.PRED p9, p0 = w1, u1 C M I + ld8 u1 = [up], 8 C M01 + (p7) cmp.eq.or p8, p0 = LIM, w0 C M I + (p7) add w0 = INCR, w0 C M I + br .LL01 C B + + +.Lb10: ld8 v2 = [vp], 8 C M01 + ld8 u2 = [up], 8 C M01 + shr.u n = n, 2 C I0 + ADDSUB w1 = r10, r11 C M I + (p15) br.dpnt .grt2 C B + ;; + + cmp.PRED p9, p0 = w1, r10 C M I + and r8 = 1, w1 C M I + ADDSUB w2 = u2, v2 C M I + ;; + cmp.PRED p6, p0 = w2, u2 C M I + ;; + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + ;; + shrp x1 = w2, w1, 1 C I0 + shr.u x2 = w2, 1 C I0 + br .Lcj2 C B + +.grt2: ld8 v3 = [vp], 8 C M01 + ld8 u3 = [up], 8 C M01 + ;; + ld8 v0 = [vp], 8 C M01 + ld8 u0 = [up], 8 C M01 + mov.i ar.lc = n C I0 + ;; + ld8 v1 = [vp], 8 C M01 + cmp.PRED p9, p0 = w1, r10 C M I + ld8 u1 = [up], 8 C M01 + and r8 = 1, w1 C M I + ;; + ADDSUB w2 = u2, v2 C M I + ld8 v2 = [vp], 8 C M01 + ;; + cmp.PRED p6, p0 = w2, u2 C M I + ld8 u2 = [up], 8 C M01 + ADDSUB w3 = u3, v3 C M I + br.cloop.dptk .grt6 C B + ;; + + cmp.PRED p7, p0 = w3, u3 C M I + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + ;; + shrp x1 = w2, w1, 1 C I0 + ADDSUB w0 = u0, v0 C M I + (p6) cmp.eq.or p7, p0 = LIM, w3 C M I + (p6) add w3 = INCR, w3 C M I + br .Lcj6 C B + +.grt6: ld8 v3 = [vp], 8 C M01 + cmp.PRED p7, p0 = w3, u3 C M I + ld8 u3 = [up], 8 C M01 + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + ;; + shrp x1 = w2, w1, 1 C I0 + ADDSUB w0 = u0, v0 C M I + ld8 v0 = [vp], 8 C M01 + (p6) cmp.eq.or p7, p0 = LIM, w3 C M I + (p6) add w3 = INCR, w3 C M I + br .LL10 C B + + +.Lb11: ld8 v1 = [vp], 8 C M01 + ld8 u1 = [up], 8 C M01 + shr.u n = n, 2 C I0 + ;; + ld8 v2 = [vp], 8 C M01 + ld8 u2 = [up], 8 C M01 + ADDSUB w0 = r10, r11 C M I + (p15) br.dpnt .grt3 C B + ;; + + cmp.PRED p8, p0 = w0, r10 C M I + ADDSUB w1 = u1, v1 C M I + and r8 = 1, w0 C M I + ;; + cmp.PRED p9, p0 = w1, u1 C M I + ;; + ADDSUB w2 = u2, v2 C M I + (p8) cmp.eq.or p9, p0 = LIM, w1 C M I + (p8) add w1 = INCR, w1 C M I + ;; + cmp.PRED p6, p0 = w2, u2 C M I + shrp x0 = w1, w0, 1 C I0 + ;; + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + br .Lcj3 C B + +.grt3: ld8 v3 = [vp], 8 C M01 + ld8 u3 = [up], 8 C M01 + ;; + ld8 v0 = [vp], 8 C M01 + mov.i ar.lc = n C I0 + cmp.PRED p8, p0 = w0, r10 C M I + ld8 u0 = [up], 8 C M01 + ADDSUB w1 = u1, v1 C M I + and r8 = 1, w0 C M I + ;; + ld8 v1 = [vp], 8 C M01 + cmp.PRED p9, p0 = w1, u1 C M I + ld8 u1 = [up], 8 C M01 + ;; + ADDSUB w2 = u2, v2 C M I + ld8 v2 = [vp], 8 C M01 + (p8) cmp.eq.or p9, p0 = LIM, w1 C M I + (p8) add w1 = INCR, w1 C M I + ;; + cmp.PRED p6, p0 = w2, u2 C M I + shrp x0 = w1, w0, 1 C I0 + ld8 u2 = [up], 8 C M01 + ADDSUB w3 = u3, v3 C M I + br.cloop.dptk .grt7 C B + ;; + + cmp.PRED p7, p0 = w3, u3 C M I + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + br .Lcj7 C B + +.grt7: ld8 v3 = [vp], 8 C M01 + cmp.PRED p7, p0 = w3, u3 C M I + ld8 u3 = [up], 8 C M01 + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + br .LL11 C B + + +C *** MAIN LOOP START *** + ALIGN(32) +.Loop: st8 [rp] = x3, 8 C M23 + ld8 v3 = [vp], 8 C M01 + cmp.PRED p7, p0 = w3, u3 C M I + ld8 u3 = [up], 8 C M01 + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + ;; +.LL11: st8 [rp] = x0, 8 C M23 + shrp x1 = w2, w1, 1 C I0 + ADDSUB w0 = u0, v0 C M I + ld8 v0 = [vp], 8 C M01 + (p6) cmp.eq.or p7, p0 = LIM, w3 C M I + (p6) add w3 = INCR, w3 C M I + ;; +.LL10: cmp.PRED p8, p0 = w0, u0 C M I + shrp x2 = w3, w2, 1 C I0 + nop.b 0 + ld8 u0 = [up], 8 C M01 + ADDSUB w1 = u1, v1 C M I + nop.b 0 + ;; + st8 [rp] = x1, 8 C M23 + ld8 v1 = [vp], 8 C M01 + cmp.PRED p9, p0 = w1, u1 C M I + ld8 u1 = [up], 8 C M01 + (p7) cmp.eq.or p8, p0 = LIM, w0 C M I + (p7) add w0 = INCR, w0 C M I + ;; +.LL01: st8 [rp] = x2, 8 C M23 + shrp x3 = w0, w3, 1 C I0 + ADDSUB w2 = u2, v2 C M I + ld8 v2 = [vp], 8 C M01 + (p8) cmp.eq.or p9, p0 = LIM, w1 C M I + (p8) add w1 = INCR, w1 C M I + ;; +.LL00: cmp.PRED p6, p0 = w2, u2 C M I + shrp x0 = w1, w0, 1 C I0 + nop.b 0 + ld8 u2 = [up], 8 C M01 + ADDSUB w3 = u3, v3 C M I + br.cloop.dptk .Loop C B + ;; +C *** MAIN LOOP END *** + +.Lskip: st8 [rp] = x3, 8 C M23 + cmp.PRED p7, p0 = w3, u3 C M I + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + ;; +.Lcj7: st8 [rp] = x0, 8 C M23 + shrp x1 = w2, w1, 1 C I0 + ADDSUB w0 = u0, v0 C M I + (p6) cmp.eq.or p7, p0 = LIM, w3 C M I + (p6) add w3 = INCR, w3 C M I + ;; +.Lcj6: cmp.PRED p8, p0 = w0, u0 C M I + shrp x2 = w3, w2, 1 C I0 + ADDSUB w1 = u1, v1 C M I + ;; + st8 [rp] = x1, 8 C M23 + cmp.PRED p9, p0 = w1, u1 C M I + (p7) cmp.eq.or p8, p0 = LIM, w0 C M I + (p7) add w0 = INCR, w0 C M I + ;; +.Lcj5: st8 [rp] = x2, 8 C M23 + shrp x3 = w0, w3, 1 C I0 + ADDSUB w2 = u2, v2 C M I + (p8) cmp.eq.or p9, p0 = LIM, w1 C M I + (p8) add w1 = INCR, w1 C M I + ;; +.Lcj4: cmp.PRED p6, p0 = w2, u2 C M I + shrp x0 = w1, w0, 1 C I0 + ;; + st8 [rp] = x3, 8 C M23 + (p9) cmp.eq.or p6, p0 = LIM, w2 C M I + (p9) add w2 = INCR, w2 C M I + ;; +.Lcj3: st8 [rp] = x0, 8 C M23 + shrp x1 = w2, w1, 1 C I0 + shr.u x2 = w2, 1 C I0 + ;; +.Lcj2: st8 [rp] = x1, 8 C M23 + (p6) dep x2 = -1, x2, 63, 1 C I0 + ;; +.Lcj1: st8 [rp] = x2 C M23 + mov.i ar.lc = r2 C I0 + br.ret.sptk.many b0 C B +EPILOGUE()