dnl AMD64 mpn_rsh1add_n -- rp[] = (up[] + vp[]) >> 1 dnl Copyright 2003, 2005 Free Software Foundation, Inc. dnl This file is part of the GNU MP Library. dnl The GNU MP Library is free software; you can redistribute it and/or modify dnl it under the terms of the GNU Lesser General Public License as published dnl by the Free Software Foundation; either version 3 of the License, or (at dnl your option) any later version. dnl The GNU MP Library is distributed in the hope that it will be useful, but dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public dnl License for more details. dnl You should have received a copy of the GNU Lesser General Public License dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. include(`../config.m4') C cycles/limb C K8,K9: 2.14 (mpn_add_n + mpn_rshift need 4.125) C K10: 2.14 (mpn_add_n + mpn_rshift need 4.125) C P4: 12.75 C P6-15: 3.75 C TODO C * Rewrite to use indexed addressing, like addlsh1.asm and sublsh1.asm. C * Try to approach the cache bandwidth 1.5 c/l. It should be possible. C INPUT PARAMETERS define(`rp',`%rdi') define(`up',`%rsi') define(`vp',`%rdx') define(`n',`%rcx') define(`n32',`%ecx') ASM_START() TEXT ALIGN(16) .byte 0,0,0,0,0,0,0,0 PROLOGUE(mpn_rsh1add_n) pushq %rbx C 1 xorl %eax, %eax movq (up), %rbx addq (vp), %rbx rcrq %rbx C rotate, save acy adcl %eax, %eax C return value movl n32, %r11d andl $3, %r11d cmpl $1, %r11d je L(do) C jump if n = 1 5 9 ... L(n1): cmpl $2, %r11d jne L(n2) C jump unless n = 2 6 10 ... addq %rbx, %rbx C rotate carry limb, restore acy movq 8(up), %r10 adcq 8(vp), %r10 leaq 8(up), up leaq 8(vp), vp leaq 8(rp), rp rcrq %r10 rcrq %rbx movq %rbx, -8(rp) jmp L(cj1) L(n2): cmpl $3, %r11d jne L(n3) C jump unless n = 3 7 11 ... addq %rbx, %rbx C rotate carry limb, restore acy movq 8(up), %r9 movq 16(up), %r10 adcq 8(vp), %r9 adcq 16(vp), %r10 leaq 16(up), up leaq 16(vp), vp leaq 16(rp), rp rcrq %r10 rcrq %r9 rcrq %rbx movq %rbx, -16(rp) jmp L(cj2) L(n3): decq n C come here for n = 4 8 12 ... addq %rbx, %rbx C rotate carry limb, restore acy movq 8(up), %r8 movq 16(up), %r9 adcq 8(vp), %r8 adcq 16(vp), %r9 movq 24(up), %r10 adcq 24(vp), %r10 leaq 24(up), up leaq 24(vp), vp leaq 24(rp), rp rcrq %r10 rcrq %r9 rcrq %r8 rcrq %rbx movq %rbx, -24(rp) movq %r8, -16(rp) L(cj2): movq %r9, -8(rp) L(cj1): movq %r10, %rbx L(do): shrq $2, n C 4 je L(end) C 2 ALIGN(16) L(oop): addq %rbx, %rbx C rotate carry limb, restore acy movq 8(up), %r8 movq 16(up), %r9 adcq 8(vp), %r8 adcq 16(vp), %r9 movq 24(up), %r10 movq 32(up), %r11 adcq 24(vp), %r10 adcq 32(vp), %r11 leaq 32(up), up leaq 32(vp), vp rcrq %r11 C rotate, save acy rcrq %r10 rcrq %r9 rcrq %r8 rcrq %rbx movq %rbx, (rp) movq %r8, 8(rp) movq %r9, 16(rp) movq %r10, 24(rp) movq %r11, %rbx leaq 32(rp), rp decq n jne L(oop) L(end): movq %rbx, (rp) popq %rbx ret EPILOGUE()