X-Git-Url: https://oss.titaniummirror.com/gitweb?a=blobdiff_plain;f=gmp%2Fmpn%2Fx86_64%2Faddaddmul_1msb0.asm;fp=gmp%2Fmpn%2Fx86_64%2Faddaddmul_1msb0.asm;h=89e7bed9800431eabc31891ae9ffe2ae4546735c;hb=6fed43773c9b0ce596dca5686f37ac3fc0fa11c0;hp=0000000000000000000000000000000000000000;hpb=27b11d56b743098deb193d510b337ba22dc52e5c;p=msp430-gcc.git diff --git a/gmp/mpn/x86_64/addaddmul_1msb0.asm b/gmp/mpn/x86_64/addaddmul_1msb0.asm new file mode 100644 index 00000000..89e7bed9 --- /dev/null +++ b/gmp/mpn/x86_64/addaddmul_1msb0.asm @@ -0,0 +1,155 @@ +dnl AMD64 mpn_addaddmul_1msb0, R = Au + Bv, u,v < 2^63. + +dnl Copyright 2008 Free Software Foundation, Inc. + +dnl This file is part of the GNU MP Library. + +dnl The GNU MP Library is free software; you can redistribute it and/or modify +dnl it under the terms of the GNU Lesser General Public License as published +dnl by the Free Software Foundation; either version 3 of the License, or (at +dnl your option) any later version. + +dnl The GNU MP Library is distributed in the hope that it will be useful, but +dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY +dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public +dnl License for more details. + +dnl You should have received a copy of the GNU Lesser General Public License +dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. + +include(`../config.m4') + +C cycles/limb +C K8: 2.167 +C P4: 12.0 +C P6-15: 4.0 + +C TODO +C * Perhaps handle various n mod 3 sizes better. The code now is too large. + +C INPUT PARAMETERS +define(`rp', `%rdi') +define(`ap', `%rsi') +define(`bp_param', `%rdx') +define(`n', `%rcx') +define(`u0', `%r8') +define(`v0', `%r9') + + +define(`bp', `%rbp') + +ASM_START() + TEXT + ALIGN(16) +PROLOGUE(mpn_addaddmul_1msb0) + push %r12 + push %rbp + + lea (ap,n,8), ap + lea (bp_param,n,8), bp + lea (rp,n,8), rp + neg n + + mov (ap,n,8), %rax + mul %r8 + mov %rax, %r12 + mov (bp,n,8), %rax + mov %rdx, %r10 + add $3, n + jns L(end) + + ALIGN(16) +L(top): mul %r9 + add %rax, %r12 + mov -16(ap,n,8), %rax + adc %rdx, %r10 + mov %r12, -24(rp,n,8) + mul %r8 + add %rax, %r10 + mov -16(bp,n,8), %rax + mov $0, %r11d + adc %rdx, %r11 + mul %r9 + add %rax, %r10 + mov -8(ap,n,8), %rax + adc %rdx, %r11 + mov %r10, -16(rp,n,8) + mul %r8 + add %rax, %r11 + mov -8(bp,n,8), %rax + mov $0, %r12d + adc %rdx, %r12 + mul %r9 + add %rax, %r11 + adc %rdx, %r12 + mov (ap,n,8), %rax + mul %r8 + add %rax, %r12 + mov %r11, -8(rp,n,8) + mov (bp,n,8), %rax + mov $0, %r10d + adc %rdx, %r10 + add $3, n + js L(top) + +L(end): cmp $1, R32(n) + ja 2f + jz 1f + + mul %r9 + add %rax, %r12 + mov -16(ap), %rax + adc %rdx, %r10 + mov %r12, -24(rp) + mul %r8 + add %rax, %r10 + mov -16(bp), %rax + mov $0, %r11d + adc %rdx, %r11 + mul %r9 + add %rax, %r10 + mov -8(ap), %rax + adc %rdx, %r11 + mov %r10, -16(rp) + mul %r8 + add %rax, %r11 + mov -8(bp), %rax + mov $0, %r12d + adc %rdx, %r12 + mul %r9 + add %rax, %r11 + adc %rdx, %r12 + mov %r11, -8(rp) + mov %r12, %rax + pop %rbp + pop %r12 + ret + +1: mul %r9 + add %rax, %r12 + mov -8(ap), %rax + adc %rdx, %r10 + mov %r12, -16(rp) + mul %r8 + add %rax, %r10 + mov -8(bp), %rax + mov $0, %r11d + adc %rdx, %r11 + mul %r9 + add %rax, %r10 + adc %rdx, %r11 + mov %r10, -8(rp) + mov %r11, %rax + pop %rbp + pop %r12 + ret + +2: mul %r9 + add %rax, %r12 + mov %r12, -8(rp) + adc %rdx, %r10 + mov %r10, %rax + pop %rbp + pop %r12 + ret +EPILOGUE()