| dnl x86-64 mpn_addlsh1_n and mpn_sublsh1_n, optimized for "Core" 2. |
| |
| dnl Copyright 2008 Free Software Foundation, Inc. |
| |
| dnl This file is part of the GNU MP Library. |
| |
| dnl The GNU MP Library is free software; you can redistribute it and/or modify |
| dnl it under the terms of the GNU Lesser General Public License as published |
| dnl by the Free Software Foundation; either version 3 of the License, or (at |
| dnl your option) any later version. |
| |
| dnl The GNU MP Library is distributed in the hope that it will be useful, but |
| dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY |
| dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public |
| dnl License for more details. |
| |
| dnl You should have received a copy of the GNU Lesser General Public License |
| dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. |
| |
| include(`../config.m4') |
| |
| C cycles/limb |
| C K8,K9: 4.25 |
| C K10: ? |
| C P4: ? |
| C P6-15: 3 |
| |
| C INPUT PARAMETERS |
| define(`rp',`%rdi') |
| define(`up',`%rsi') |
| define(`vp',`%rdx') |
| define(`n', `%rcx') |
| |
| ifdef(`OPERATION_addlsh1_n', ` |
| define(ADDSUB, add) |
| define(ADCSBB, adc) |
| define(func, mpn_addlsh1_n)') |
| ifdef(`OPERATION_sublsh1_n', ` |
| define(ADDSUB, sub) |
| define(ADCSBB, sbb) |
| define(func, mpn_sublsh1_n)') |
| |
| MULFUNC_PROLOGUE(mpn_addlsh1_n mpn_sublsh1_n) |
| |
| ASM_START() |
| TEXT |
| ALIGN(8) |
| PROLOGUE(func) |
| push %rbx |
| push %r12 |
| |
| mov R32(%rcx), R32(%rax) |
| lea 24(up,n,8), up |
| lea 24(vp,n,8), vp |
| lea 24(rp,n,8), rp |
| neg n |
| |
| xor R32(%r11), R32(%r11) |
| |
| mov -24(vp,n,8), %r8 C do first limb early |
| shrd $63, %r8, %r11 |
| |
| and $3, R32(%rax) |
| je L(b0) |
| cmp $2, R32(%rax) |
| jc L(b1) |
| je L(b2) |
| |
| L(b3): mov -16(vp,n,8), %r9 |
| shrd $63, %r9, %r8 |
| mov -8(vp,n,8), %r10 |
| shrd $63, %r10, %r9 |
| mov -24(up,n,8), %r12 |
| ADDSUB %r11, %r12 |
| mov %r12, -24(rp,n,8) |
| mov -16(up,n,8), %r12 |
| ADCSBB %r8, %r12 |
| mov %r12, -16(rp,n,8) |
| mov -8(up,n,8), %r12 |
| ADCSBB %r9, %r12 |
| mov %r12, -8(rp,n,8) |
| mov %r10, %r11 |
| sbb R32(%rax), R32(%rax) C save cy |
| add $3, n |
| js L(top) |
| jmp L(end) |
| |
| L(b1): mov -24(up,n,8), %r12 |
| ADDSUB %r11, %r12 |
| mov %r12, -24(rp,n,8) |
| mov %r8, %r11 |
| sbb R32(%rax), R32(%rax) C save cy |
| inc n |
| js L(top) |
| jmp L(end) |
| |
| L(b2): mov -16(vp,n,8), %r9 |
| shrd $63, %r9, %r8 |
| mov -24(up,n,8), %r12 |
| ADDSUB %r11, %r12 |
| mov %r12, -24(rp,n,8) |
| mov -16(up,n,8), %r12 |
| ADCSBB %r8, %r12 |
| mov %r12, -16(rp,n,8) |
| mov %r9, %r11 |
| sbb R32(%rax), R32(%rax) C save cy |
| add $2, n |
| js L(top) |
| jmp L(end) |
| |
| ALIGN(16) |
| L(top): mov -24(vp,n,8), %r8 |
| shrd $63, %r8, %r11 |
| L(b0): mov -16(vp,n,8), %r9 |
| shrd $63, %r9, %r8 |
| mov -8(vp,n,8), %r10 |
| shrd $63, %r10, %r9 |
| mov (vp,n,8), %rbx |
| shrd $63, %rbx, %r10 |
| |
| add R32(%rax), R32(%rax) C restore cy |
| |
| mov -24(up,n,8), %r12 |
| ADCSBB %r11, %r12 |
| mov %r12, -24(rp,n,8) |
| |
| mov -16(up,n,8), %r12 |
| ADCSBB %r8, %r12 |
| mov %r12, -16(rp,n,8) |
| |
| mov -8(up,n,8), %r12 |
| ADCSBB %r9, %r12 |
| mov %r12, -8(rp,n,8) |
| |
| mov (up,n,8), %r12 |
| ADCSBB %r10, %r12 |
| mov %r12, (rp,n,8) |
| |
| mov %rbx, %r11 |
| sbb R32(%rax), R32(%rax) C save cy |
| |
| add $4, n |
| js L(top) |
| |
| L(end): add %r11, %r11 |
| pop %r12 |
| pop %rbx |
| sbb $0, R32(%rax) |
| neg R32(%rax) |
| ret |
| EPILOGUE() |