| dnl mpn_addmul_1 for Pentium 4 and P6 models with SSE2 (i.e., 9,D,E,F). |
| |
| dnl Copyright 2005, 2007 Free Software Foundation, Inc. |
| dnl |
| dnl This file is part of the GNU MP Library. |
| dnl |
| dnl The GNU MP Library is free software; you can redistribute it and/or modify |
| dnl it under the terms of the GNU Lesser General Public License as published |
| dnl by the Free Software Foundation; either version 3 of the License, or (at |
| dnl your option) any later version. |
| dnl |
| dnl The GNU MP Library is distributed in the hope that it will be useful, but |
| dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY |
| dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public |
| dnl License for more details. |
| dnl |
| dnl You should have received a copy of the GNU Lesser General Public License |
| dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. |
| |
| |
| include(`../config.m4') |
| |
| C TODO: |
| C * Tweak eax/edx offsets in loop as to save some lea's |
| C * Perhaps software pipeline small-case code |
| |
| C cycles/limb |
| C P6 model 0-8,10-12) - |
| C P6 model 9 (Banias) ? |
| C P6 model 13 (Dothan) 5.24 |
| C P4 model 0-1 (Willamette): 5 |
| C P4 model 2 (Northwood): 5 |
| C P4 model 3-4 (Prescott): 5 |
| |
| C INPUT PARAMETERS |
| C rp sp + 4 |
| C up sp + 8 |
| C n sp + 12 |
| C v0 sp + 16 |
| |
| TEXT |
| ALIGN(16) |
| PROLOGUE(mpn_addmul_1c) |
| mov 4(%esp), %edx |
| mov 8(%esp), %eax |
| mov 12(%esp), %ecx |
| movd 16(%esp), %mm7 |
| movd 20(%esp), %mm6 |
| jmp L(ent) |
| EPILOGUE() |
| ALIGN(16) |
| PROLOGUE(mpn_addmul_1) |
| mov 4(%esp), %edx |
| mov 8(%esp), %eax |
| mov 12(%esp), %ecx |
| movd 16(%esp), %mm7 |
| pxor %mm6, %mm6 |
| L(ent): cmp $4, %ecx |
| jnc L(big) |
| |
| L(lp0): movd (%eax), %mm0 |
| lea 4(%eax), %eax |
| movd (%edx), %mm4 |
| lea 4(%edx), %edx |
| pmuludq %mm7, %mm0 |
| paddq %mm0, %mm4 |
| paddq %mm4, %mm6 |
| movd %mm6, -4(%edx) |
| psrlq $32, %mm6 |
| dec %ecx |
| jnz L(lp0) |
| movd %mm6, %eax |
| emms |
| ret |
| |
| L(big): and $3, %ecx |
| je L(0) |
| cmp $2, %ecx |
| jc L(1) |
| je L(2) |
| jmp L(3) C FIXME: one case should fall through |
| |
| L(0): movd (%eax), %mm3 |
| sub 12(%esp), %ecx C loop count |
| lea -16(%eax), %eax |
| lea -12(%edx), %edx |
| pmuludq %mm7, %mm3 |
| movd 20(%eax), %mm0 |
| movd 12(%edx), %mm5 |
| pmuludq %mm7, %mm0 |
| movd 24(%eax), %mm1 |
| paddq %mm3, %mm5 |
| movd 16(%edx), %mm4 |
| jmp L(00) |
| |
| L(1): movd (%eax), %mm2 |
| sub 12(%esp), %ecx |
| lea -12(%eax), %eax |
| lea -8(%edx), %edx |
| movd 8(%edx), %mm4 |
| pmuludq %mm7, %mm2 |
| movd 16(%eax), %mm3 |
| pmuludq %mm7, %mm3 |
| movd 20(%eax), %mm0 |
| paddq %mm2, %mm4 |
| movd 12(%edx), %mm5 |
| jmp L(01) |
| |
| L(2): movd (%eax), %mm1 |
| sub 12(%esp), %ecx |
| lea -8(%eax), %eax |
| lea -4(%edx), %edx |
| pmuludq %mm7, %mm1 |
| movd 12(%eax), %mm2 |
| movd 4(%edx), %mm5 |
| pmuludq %mm7, %mm2 |
| movd 16(%eax), %mm3 |
| paddq %mm1, %mm5 |
| movd 8(%edx), %mm4 |
| jmp L(10) |
| |
| L(3): movd (%eax), %mm0 |
| sub 12(%esp), %ecx |
| lea -4(%eax), %eax |
| pmuludq %mm7, %mm0 |
| movd 8(%eax), %mm1 |
| movd (%edx), %mm4 |
| pmuludq %mm7, %mm1 |
| movd 12(%eax), %mm2 |
| paddq %mm0, %mm4 |
| movd 4(%edx), %mm5 |
| |
| ALIGN(16) |
| L(top): pmuludq %mm7, %mm2 |
| paddq %mm4, %mm6 |
| movd 16(%eax), %mm3 |
| paddq %mm1, %mm5 |
| movd 8(%edx), %mm4 |
| movd %mm6, 0(%edx) |
| psrlq $32, %mm6 |
| L(10): pmuludq %mm7, %mm3 |
| paddq %mm5, %mm6 |
| movd 20(%eax), %mm0 |
| paddq %mm2, %mm4 |
| movd 12(%edx), %mm5 |
| movd %mm6, 4(%edx) |
| psrlq $32, %mm6 |
| L(01): pmuludq %mm7, %mm0 |
| paddq %mm4, %mm6 |
| movd 24(%eax), %mm1 |
| paddq %mm3, %mm5 |
| movd 16(%edx), %mm4 |
| movd %mm6, 8(%edx) |
| psrlq $32, %mm6 |
| L(00): pmuludq %mm7, %mm1 |
| paddq %mm5, %mm6 |
| movd 28(%eax), %mm2 |
| paddq %mm0, %mm4 |
| movd 20(%edx), %mm5 |
| movd %mm6, 12(%edx) |
| psrlq $32, %mm6 |
| lea 16(%eax), %eax |
| lea 16(%edx), %edx |
| add $4, %ecx |
| jnz L(top) |
| |
| L(end): pmuludq %mm7, %mm2 |
| paddq %mm4, %mm6 |
| paddq %mm1, %mm5 |
| movd 8(%edx), %mm4 |
| movd %mm6, 0(%edx) |
| psrlq $32, %mm6 |
| paddq %mm5, %mm6 |
| paddq %mm2, %mm4 |
| movd %mm6, 4(%edx) |
| psrlq $32, %mm6 |
| paddq %mm4, %mm6 |
| movd %mm6, 8(%edx) |
| psrlq $32, %mm6 |
| movd %mm6, %eax |
| emms |
| ret |
| EPILOGUE() |