Signed-off-by: Hao Chen <chenhao@loongson.cn>tags/v0.3.26
@@ -40,8 +40,10 @@ DCOPYKERNEL = copy_lsx.S | |||||
SSWAPKERNEL = swap_lsx.S | SSWAPKERNEL = swap_lsx.S | ||||
DSWAPKERNEL = swap_lsx.S | DSWAPKERNEL = swap_lsx.S | ||||
SAXPYKERNEL = saxpy_lsx.S | |||||
DAXPYKERNEL = daxpy_lsx.S | |||||
SAXPYKERNEL = axpy_lsx.S | |||||
DAXPYKERNEL = axpy_lsx.S | |||||
CAXPYKERNEL = caxpy_lsx.S | |||||
ZAXPYKERNEL = caxpy_lsx.S | |||||
SAXPBYKERNEL = saxpby_lsx.S | SAXPBYKERNEL = saxpby_lsx.S | ||||
DAXPBYKERNEL = daxpby_lsx.S | DAXPBYKERNEL = daxpby_lsx.S | ||||
@@ -40,8 +40,10 @@ DCOPYKERNEL = copy_lasx.S | |||||
SSWAPKERNEL = swap_lasx.S | SSWAPKERNEL = swap_lasx.S | ||||
DSWAPKERNEL = swap_lasx.S | DSWAPKERNEL = swap_lasx.S | ||||
SAXPYKERNEL = saxpy_lasx.S | |||||
DAXPYKERNEL = daxpy_lasx.S | |||||
SAXPYKERNEL = axpy_lasx.S | |||||
DAXPYKERNEL = axpy_lasx.S | |||||
CAXPYKERNEL = caxpy_lasx.S | |||||
ZAXPYKERNEL = caxpy_lasx.S | |||||
SAXPBYKERNEL = saxpby_lasx.S | SAXPBYKERNEL = saxpby_lasx.S | ||||
DAXPBYKERNEL = daxpby_lasx.S | DAXPBYKERNEL = daxpby_lasx.S | ||||
@@ -1,6 +1,33 @@ | |||||
#define ASSEMBLER | |||||
/*************************************************************************** | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*****************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | #include "common.h" | ||||
#define N $r4 | #define N $r4 | ||||
#define XX $r5 | #define XX $r5 | ||||
#define YY $r6 | #define YY $r6 | ||||
@@ -35,16 +62,20 @@ | |||||
bge $r0, N, .L999 | bge $r0, N, .L999 | ||||
li.d TEMP, 1 | li.d TEMP, 1 | ||||
movgr2fr.d a1, $r0 | movgr2fr.d a1, $r0 | ||||
ffint.d.l a1, a1 | |||||
FFINT a1, a1 | |||||
movgr2fr.d a2, TEMP | movgr2fr.d a2, TEMP | ||||
ffint.d.l a2, a2 | |||||
fcmp.ceq.d $fcc0, ALPHA, a1 | |||||
FFINT a2, a2 | |||||
CMPEQ $fcc0, ALPHA, a1 | |||||
bcnez $fcc0, .L999 | bcnez $fcc0, .L999 | ||||
slli.d TEMP, TEMP, BASE_SHIFT | slli.d TEMP, TEMP, BASE_SHIFT | ||||
slli.d INCX, INCX, BASE_SHIFT | slli.d INCX, INCX, BASE_SHIFT | ||||
slli.d INCY, INCY, BASE_SHIFT | slli.d INCY, INCY, BASE_SHIFT | ||||
movfr2gr.d t1, ALPHA | |||||
MTG t1, ALPHA | |||||
#ifdef DOUBLE | |||||
xvreplgr2vr.d VXA, t1 | xvreplgr2vr.d VXA, t1 | ||||
#else | |||||
xvreplgr2vr.w VXA, t1 | |||||
#endif | |||||
srai.d I, N, 3 | srai.d I, N, 3 | ||||
bne INCX, TEMP, .L20 | bne INCX, TEMP, .L20 | ||||
@@ -56,11 +87,12 @@ | |||||
.L11: | .L11: | ||||
bge $r0, I, .L113 | bge $r0, I, .L113 | ||||
fcmp.ceq.d $fcc0, ALPHA, a2 | |||||
CMPEQ $fcc0, ALPHA, a2 | |||||
bceqz $fcc0, .L112 | bceqz $fcc0, .L112 | ||||
.align 3 | .align 3 | ||||
.L111: | .L111: | ||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | xvld VX0, X, 0 * SIZE | ||||
xvld VX2, Y, 0 * SIZE | xvld VX2, Y, 0 * SIZE | ||||
xvld VX1, X, 4 * SIZE | xvld VX1, X, 4 * SIZE | ||||
@@ -70,6 +102,13 @@ | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
xvst VX2, Y, 0 * SIZE | xvst VX2, Y, 0 * SIZE | ||||
xvst VX3, Y, 4 * SIZE | xvst VX3, Y, 4 * SIZE | ||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
xvfadd.s VX2, VX0, VX2 | |||||
xvst VX2, Y, 0 * SIZE | |||||
#endif | |||||
addi.d X, X, 8 * SIZE | addi.d X, X, 8 * SIZE | ||||
addi.d Y, Y, 8 * SIZE | addi.d Y, Y, 8 * SIZE | ||||
blt $r0, I, .L111 | blt $r0, I, .L111 | ||||
@@ -77,6 +116,7 @@ | |||||
.align 3 | .align 3 | ||||
.L112: | .L112: | ||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | xvld VX0, X, 0 * SIZE | ||||
xvld VX2, Y, 0 * SIZE | xvld VX2, Y, 0 * SIZE | ||||
xvld VX1, X, 4 * SIZE | xvld VX1, X, 4 * SIZE | ||||
@@ -86,6 +126,13 @@ | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
xvst VX2, Y, 0 * SIZE | xvst VX2, Y, 0 * SIZE | ||||
xvst VX3, Y, 4 * SIZE | xvst VX3, Y, 4 * SIZE | ||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
xvst VX2, Y, 0 * SIZE | |||||
#endif | |||||
addi.d X, X, 8 * SIZE | addi.d X, X, 8 * SIZE | ||||
addi.d Y, Y, 8 * SIZE | addi.d Y, Y, 8 * SIZE | ||||
blt $r0, I, .L112 | blt $r0, I, .L112 | ||||
@@ -97,11 +144,11 @@ | |||||
.align 3 | .align 3 | ||||
.L114: | .L114: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | addi.d X, X, SIZE | ||||
addi.d Y, Y, SIZE | addi.d Y, Y, SIZE | ||||
blt $r0, I, .L114 | blt $r0, I, .L114 | ||||
@@ -114,6 +161,7 @@ | |||||
.align 3 | .align 3 | ||||
.L121: | .L121: | ||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | xvld VX0, X, 0 * SIZE | ||||
ld.d t1, Y, 0 * SIZE | ld.d t1, Y, 0 * SIZE | ||||
add.d Y, Y, INCY | add.d Y, Y, INCY | ||||
@@ -158,6 +206,50 @@ | |||||
xvstelm.d VX3, YY, 0, 2 | xvstelm.d VX3, YY, 0, 2 | ||||
add.d YY, YY, INCY | add.d YY, YY, INCY | ||||
xvstelm.d VX3, YY, 0, 3 | xvstelm.d VX3, YY, 0, 3 | ||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
xvinsgr2vr.w VX2, t1, 0 | |||||
xvinsgr2vr.w VX2, t2, 1 | |||||
xvinsgr2vr.w VX2, t3, 2 | |||||
xvinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
xvinsgr2vr.w VX2, t1, 4 | |||||
xvinsgr2vr.w VX2, t2, 5 | |||||
xvinsgr2vr.w VX2, t3, 6 | |||||
xvinsgr2vr.w VX2, t4, 7 | |||||
add.d Y, Y, INCY | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
addi.d I, I, -1 | |||||
xvstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 4 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 5 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 6 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 7 | |||||
#endif | |||||
add.d YY, YY, INCY | add.d YY, YY, INCY | ||||
addi.d X, X, 8 * SIZE | addi.d X, X, 8 * SIZE | ||||
blt $r0, I, .L121 | blt $r0, I, .L121 | ||||
@@ -169,11 +261,11 @@ | |||||
.align 3 | .align 3 | ||||
.L123: | .L123: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | addi.d X, X, SIZE | ||||
add.d Y, Y, INCY | add.d Y, Y, INCY | ||||
blt $r0, I, .L123 | blt $r0, I, .L123 | ||||
@@ -185,6 +277,7 @@ | |||||
.align 3 | .align 3 | ||||
.L211: | .L211: | ||||
#ifdef DOUBLE | |||||
xvld VX2, Y, 0 * SIZE | xvld VX2, Y, 0 * SIZE | ||||
ld.d t1, X, 0 * SIZE | ld.d t1, X, 0 * SIZE | ||||
add.d X, X, INCX | add.d X, X, INCX | ||||
@@ -217,6 +310,37 @@ | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
xvst VX3, Y, 4 * SIZE | xvst VX3, Y, 4 * SIZE | ||||
addi.d Y, Y, 8 * SIZE | addi.d Y, Y, 8 * SIZE | ||||
#else | |||||
xvld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
xvinsgr2vr.w VX0, t1, 0 | |||||
xvinsgr2vr.w VX0, t2, 1 | |||||
xvinsgr2vr.w VX0, t3, 2 | |||||
xvinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 4 | |||||
xvinsgr2vr.w VX0, t2, 5 | |||||
xvinsgr2vr.w VX0, t3, 6 | |||||
xvinsgr2vr.w VX0, t4, 7 | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
addi.d I, I, -1 | |||||
xvst VX2, Y, 0 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
#endif | |||||
blt $r0, I, .L211 | blt $r0, I, .L211 | ||||
.align 3 | .align 3 | ||||
@@ -226,11 +350,11 @@ | |||||
.align 3 | .align 3 | ||||
.L213: | .L213: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | add.d X, X, INCX | ||||
addi.d Y, Y, SIZE | addi.d Y, Y, SIZE | ||||
blt $r0, I, .L213 | blt $r0, I, .L213 | ||||
@@ -243,6 +367,7 @@ | |||||
.align 3 | .align 3 | ||||
.L222: | .L222: | ||||
#ifdef DOUBLE | |||||
ld.d t1, X, 0 * SIZE | ld.d t1, X, 0 * SIZE | ||||
add.d X, X, INCX | add.d X, X, INCX | ||||
ld.d t2, X, 0 * SIZE | ld.d t2, X, 0 * SIZE | ||||
@@ -309,6 +434,73 @@ | |||||
xvstelm.d VX3, YY, 0, 2 | xvstelm.d VX3, YY, 0, 2 | ||||
add.d YY, YY, INCY | add.d YY, YY, INCY | ||||
xvstelm.d VX3, YY, 0, 3 | xvstelm.d VX3, YY, 0, 3 | ||||
#else | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 0 | |||||
xvinsgr2vr.w VX0, t2, 1 | |||||
xvinsgr2vr.w VX0, t3, 2 | |||||
xvinsgr2vr.w VX0, t4, 3 | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w VX2, t1, 0 | |||||
xvinsgr2vr.w VX2, t2, 1 | |||||
xvinsgr2vr.w VX2, t3, 2 | |||||
xvinsgr2vr.w VX2, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 4 | |||||
xvinsgr2vr.w VX0, t2, 5 | |||||
xvinsgr2vr.w VX0, t3, 6 | |||||
xvinsgr2vr.w VX0, t4, 7 | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
xvinsgr2vr.w VX2, t1, 4 | |||||
xvinsgr2vr.w VX2, t2, 5 | |||||
xvinsgr2vr.w VX2, t3, 6 | |||||
xvinsgr2vr.w VX2, t4, 7 | |||||
add.d Y, Y, INCY | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
addi.d I, I, -1 | |||||
xvstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 4 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 5 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 6 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 7 | |||||
#endif | |||||
add.d YY, YY, INCY | add.d YY, YY, INCY | ||||
blt $r0, I, .L222 | blt $r0, I, .L222 | ||||
.align 3 | .align 3 | ||||
@@ -319,15 +511,14 @@ | |||||
.align 3 | .align 3 | ||||
.L224: | .L224: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | add.d X, X, INCX | ||||
add.d Y, Y, INCY | add.d Y, Y, INCY | ||||
blt $r0, I, .L224 | blt $r0, I, .L224 | ||||
b .L999 | |||||
.align 3 | .align 3 | ||||
.L999: | .L999: |
@@ -1,6 +1,33 @@ | |||||
#define ASSEMBLER | |||||
/*************************************************************************** | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*****************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | #include "common.h" | ||||
#define N $r4 | #define N $r4 | ||||
#define XX $r5 | #define XX $r5 | ||||
#define YY $r6 | #define YY $r6 | ||||
@@ -35,16 +62,20 @@ | |||||
bge $r0, N, .L999 | bge $r0, N, .L999 | ||||
li.d TEMP, 1 | li.d TEMP, 1 | ||||
movgr2fr.d a1, $r0 | movgr2fr.d a1, $r0 | ||||
ffint.d.l a1, a1 | |||||
FFINT a1, a1 | |||||
movgr2fr.d a2, TEMP | movgr2fr.d a2, TEMP | ||||
ffint.d.l a2, a2 | |||||
fcmp.ceq.d $fcc0, ALPHA, a1 | |||||
FFINT a2, a2 | |||||
CMPEQ $fcc0, ALPHA, a1 | |||||
bcnez $fcc0, .L999 | bcnez $fcc0, .L999 | ||||
slli.d TEMP, TEMP, BASE_SHIFT | slli.d TEMP, TEMP, BASE_SHIFT | ||||
slli.d INCX, INCX, BASE_SHIFT | slli.d INCX, INCX, BASE_SHIFT | ||||
slli.d INCY, INCY, BASE_SHIFT | slli.d INCY, INCY, BASE_SHIFT | ||||
movfr2gr.d t1, ALPHA | |||||
MTG t1, ALPHA | |||||
#ifdef DOUBLE | |||||
vreplgr2vr.d VXA, t1 | vreplgr2vr.d VXA, t1 | ||||
#else | |||||
vreplgr2vr.w VXA, t1 | |||||
#endif | |||||
srai.d I, N, 3 | srai.d I, N, 3 | ||||
bne INCX, TEMP, .L20 | bne INCX, TEMP, .L20 | ||||
@@ -56,11 +87,12 @@ | |||||
.L11: | .L11: | ||||
bge $r0, I, .L113 | bge $r0, I, .L113 | ||||
fcmp.ceq.d $fcc0, ALPHA, a2 | |||||
CMPEQ $fcc0, ALPHA, a2 | |||||
bceqz $fcc0, .L112 | bceqz $fcc0, .L112 | ||||
.align 3 | .align 3 | ||||
.L111: | .L111: | ||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | vld VX0, X, 0 * SIZE | ||||
vld VX2, Y, 0 * SIZE | vld VX2, Y, 0 * SIZE | ||||
vld VX1, X, 2 * SIZE | vld VX1, X, 2 * SIZE | ||||
@@ -75,16 +107,27 @@ | |||||
vld VX3, Y, 6 * SIZE | vld VX3, Y, 6 * SIZE | ||||
vfadd.d VX2, VX0, VX2 | vfadd.d VX2, VX0, VX2 | ||||
vfadd.d VX3, VX1, VX3 | vfadd.d VX3, VX1, VX3 | ||||
addi.d I, I, -1 | |||||
vst VX2, Y, 4 * SIZE | vst VX2, Y, 4 * SIZE | ||||
vst VX3, Y, 6 * SIZE | vst VX3, Y, 6 * SIZE | ||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
vld VX3, Y, 4 * SIZE | |||||
vfadd.s VX2, VX0, VX2 | |||||
vfadd.s VX3, VX1, VX3 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 4 * SIZE | |||||
#endif | |||||
addi.d X, X, 8 * SIZE | addi.d X, X, 8 * SIZE | ||||
addi.d Y, Y, 8 * SIZE | addi.d Y, Y, 8 * SIZE | ||||
addi.d I, I, -1 | |||||
blt $r0, I, .L111 | blt $r0, I, .L111 | ||||
b .L113 | b .L113 | ||||
.align 3 | .align 3 | ||||
.L112: | .L112: | ||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | vld VX0, X, 0 * SIZE | ||||
vld VX2, Y, 0 * SIZE | vld VX2, Y, 0 * SIZE | ||||
vld VX1, X, 2 * SIZE | vld VX1, X, 2 * SIZE | ||||
@@ -104,6 +147,19 @@ | |||||
vst VX2, Y, 4 * SIZE | vst VX2, Y, 4 * SIZE | ||||
vst VX3, Y, 6 * SIZE | vst VX3, Y, 6 * SIZE | ||||
addi.d Y, Y, 8 * SIZE | addi.d Y, Y, 8 * SIZE | ||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
vld VX3, Y, 4 * SIZE | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 4 * SIZE | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
#endif | |||||
blt $r0, I, .L112 | blt $r0, I, .L112 | ||||
.align 3 | .align 3 | ||||
@@ -113,11 +169,11 @@ | |||||
.align 3 | .align 3 | ||||
.L114: | .L114: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | addi.d X, X, SIZE | ||||
addi.d Y, Y, SIZE | addi.d Y, Y, SIZE | ||||
blt $r0, I, .L114 | blt $r0, I, .L114 | ||||
@@ -130,6 +186,7 @@ | |||||
.align 3 | .align 3 | ||||
.L121: | .L121: | ||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | vld VX0, X, 0 * SIZE | ||||
ld.d t1, Y, 0 * SIZE | ld.d t1, Y, 0 * SIZE | ||||
add.d Y, Y, INCY | add.d Y, Y, INCY | ||||
@@ -180,6 +237,54 @@ | |||||
add.d YY, YY, INCY | add.d YY, YY, INCY | ||||
addi.d X, X, 8 * SIZE | addi.d X, X, 8 * SIZE | ||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX2, t1, 0 | |||||
vinsgr2vr.w VX2, t2, 1 | |||||
vinsgr2vr.w VX2, t3, 2 | |||||
vinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
vld VX1, X, 4 * SIZE | |||||
vstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX3, t1, 0 | |||||
vinsgr2vr.w VX3, t2, 1 | |||||
vinsgr2vr.w VX3, t3, 2 | |||||
vinsgr2vr.w VX3, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
addi.d I, I, -1 | |||||
vstelm.w VX3, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
addi.d X, X, 8 * SIZE | |||||
#endif | |||||
blt $r0, I, .L121 | blt $r0, I, .L121 | ||||
.align 3 | .align 3 | ||||
@@ -189,11 +294,11 @@ | |||||
.align 3 | .align 3 | ||||
.L123: | .L123: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | addi.d X, X, SIZE | ||||
add.d Y, Y, INCY | add.d Y, Y, INCY | ||||
blt $r0, I, .L123 | blt $r0, I, .L123 | ||||
@@ -205,6 +310,7 @@ | |||||
.align 3 | .align 3 | ||||
.L211: | .L211: | ||||
#ifdef DOUBLE | |||||
vld VX2, Y, 0 * SIZE | vld VX2, Y, 0 * SIZE | ||||
ld.d t1, X, 0 * SIZE | ld.d t1, X, 0 * SIZE | ||||
add.d X, X, INCX | add.d X, X, INCX | ||||
@@ -242,6 +348,39 @@ | |||||
vfmadd.d VX3, VX1, VXA, VX3 | vfmadd.d VX3, VX1, VXA, VX3 | ||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
vst VX3, Y, 6 * SIZE | vst VX3, Y, 6 * SIZE | ||||
#else | |||||
vld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
vinsgr2vr.w VX0, t1, 0 | |||||
vinsgr2vr.w VX0, t2, 1 | |||||
vinsgr2vr.w VX0, t3, 2 | |||||
vinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
vld VX3, Y, 4 * SIZE | |||||
vst VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
vinsgr2vr.w VX1, t1, 0 | |||||
vinsgr2vr.w VX1, t2, 1 | |||||
vinsgr2vr.w VX1, t3, 2 | |||||
vinsgr2vr.w VX1, t4, 3 | |||||
add.d X, X, INCX | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
addi.d I, I, -1 | |||||
vst VX3, Y, 4 * SIZE | |||||
#endif | |||||
addi.d Y, Y, 8 * SIZE | addi.d Y, Y, 8 * SIZE | ||||
blt $r0, I, .L211 | blt $r0, I, .L211 | ||||
.align 3 | .align 3 | ||||
@@ -252,11 +391,11 @@ | |||||
.align 3 | .align 3 | ||||
.L213: | .L213: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | add.d X, X, INCX | ||||
addi.d Y, Y, SIZE | addi.d Y, Y, SIZE | ||||
blt $r0, I, .L213 | blt $r0, I, .L213 | ||||
@@ -269,6 +408,7 @@ | |||||
.align 3 | .align 3 | ||||
.L222: | .L222: | ||||
#ifdef DOUBLE | |||||
ld.d t1, X, 0 * SIZE | ld.d t1, X, 0 * SIZE | ||||
add.d X, X, INCX | add.d X, X, INCX | ||||
ld.d t2, X, 0 * SIZE | ld.d t2, X, 0 * SIZE | ||||
@@ -337,6 +477,74 @@ | |||||
vstelm.d VX3, YY, 0, 0 | vstelm.d VX3, YY, 0, 0 | ||||
add.d YY, YY, INCY | add.d YY, YY, INCY | ||||
vstelm.d VX3, YY, 0, 1 | vstelm.d VX3, YY, 0, 1 | ||||
#else | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
vinsgr2vr.w VX0, t1, 0 | |||||
vinsgr2vr.w VX0, t2, 1 | |||||
vinsgr2vr.w VX0, t3, 2 | |||||
vinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX2, t1, 0 | |||||
vinsgr2vr.w VX2, t2, 1 | |||||
vinsgr2vr.w VX2, t3, 2 | |||||
vinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w VX1, t1, 0 | |||||
vinsgr2vr.w VX1, t2, 1 | |||||
vinsgr2vr.w VX1, t3, 2 | |||||
vinsgr2vr.w VX1, t4, 3 | |||||
vstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX3, t1, 0 | |||||
vinsgr2vr.w VX3, t2, 1 | |||||
vinsgr2vr.w VX3, t3, 2 | |||||
vinsgr2vr.w VX3, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
addi.d I, I, -1 | |||||
vstelm.w VX3, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 3 | |||||
#endif | |||||
add.d YY, YY, INCY | add.d YY, YY, INCY | ||||
blt $r0, I, .L222 | blt $r0, I, .L222 | ||||
.align 3 | .align 3 | ||||
@@ -347,11 +555,11 @@ | |||||
.align 3 | .align 3 | ||||
.L224: | .L224: | ||||
fld.d $f12, X, 0 * SIZE | |||||
fld.d $f14, Y, 0 * SIZE | |||||
LD $f12, X, 0 * SIZE | |||||
LD $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | addi.d I, I, -1 | ||||
fmadd.d $f14, $f12, $f0, $f14 | |||||
fst.d $f14, Y, 0 * SIZE | |||||
MADD $f14, $f12, $f0, $f14 | |||||
ST $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | add.d X, X, INCX | ||||
add.d Y, Y, INCY | add.d Y, Y, INCY | ||||
blt $r0, I, .L224 | blt $r0, I, .L224 |
@@ -0,0 +1,707 @@ | |||||
/*************************************************************************** | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*****************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define XX $r5 | |||||
#define YY $r6 | |||||
#define ALPHAR $f0 | |||||
#define ALPHAI $f1 | |||||
#define X $r7 | |||||
#define INCX $r8 | |||||
#define Y $r9 | |||||
#define INCY $r10 | |||||
#define I $r12 | |||||
#define TEMP $r13 | |||||
#define t1 $r14 | |||||
#define t2 $r16 | |||||
#define t3 $r15 | |||||
#define t4 $r17 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define s1 $f16 | |||||
#define s2 $f17 | |||||
#define s3 $f18 | |||||
#define s4 $f19 | |||||
#define VX0 $xr8 | |||||
#define VX1 $xr20 | |||||
#define VX2 $xr21 | |||||
#define VX3 $xr22 | |||||
#define VXAR $xr23 | |||||
#define VXAI $xr19 | |||||
#define x1 $xr18 | |||||
#define x2 $xr17 | |||||
#define x3 $xr16 | |||||
#define x4 $xr15 | |||||
PROLOGUE | |||||
bge $r0, N, .L999 | |||||
li.d TEMP, 1 | |||||
movgr2fr.d a1, $r0 | |||||
FFINT a1, a1 | |||||
CMPEQ $fcc0, ALPHAR, a1 | |||||
CMPEQ $fcc1, ALPHAI, a1 | |||||
bceqz $fcc0, .L10 | |||||
bcnez $fcc1, .L999 | |||||
.L10: | |||||
slli.d TEMP, TEMP, ZBASE_SHIFT | |||||
slli.d INCX, INCX, ZBASE_SHIFT | |||||
slli.d INCY, INCY, ZBASE_SHIFT | |||||
MTG t1, ALPHAR | |||||
MTG t2, ALPHAI | |||||
#ifdef DOUBLE | |||||
xvreplgr2vr.d VXAR, t1 | |||||
xvreplgr2vr.d VXAI, t2 | |||||
srai.d I, N, 2 | |||||
#else | |||||
xvreplgr2vr.w VXAR, t1 | |||||
xvreplgr2vr.w VXAI, t2 | |||||
srai.d I, N, 3 | |||||
#endif | |||||
bne INCX, TEMP, .L20 | |||||
bne INCY, TEMP, .L12 // INCX==1 and INCY!=1 | |||||
b .L11 // INCX==1 and INCY==1 | |||||
.L20: | |||||
bne INCY, TEMP, .L22 // INCX!=1 and INCY!=1 | |||||
b .L21 // INCX!=1 and INCY==1 | |||||
.L11: | |||||
bge $r0, I, .L997 | |||||
.align 3 | |||||
.L111: | |||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
xvld VX1, X, 4 * SIZE | |||||
xvld VX3, Y, 4 * SIZE | |||||
xvpickev.d x1, VX1, VX0 | |||||
xvpickod.d x2, VX1, VX0 | |||||
xvpickev.d x3, VX3, VX2 | |||||
xvpickod.d x4, VX3, VX2 | |||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
xvld VX1, X, 8 * SIZE | |||||
xvld VX3, Y, 8 * SIZE | |||||
xvpickev.w x1, VX1, VX0 | |||||
xvpickod.w x2, VX1, VX0 | |||||
xvpickev.w x3, VX3, VX2 | |||||
xvpickod.w x4, VX3, VX2 | |||||
#endif | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmsub.d VX1, VXAR, x1, VX0 | |||||
xvfmadd.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfadd.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmsub.s VX1, VXAR, x1, VX0 | |||||
xvfmadd.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmadd.d VX1, VXAR, x1, VX0 | |||||
xvfmsub.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfsub.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmadd.s VX1, VXAR, x1, VX0 | |||||
xvfmsub.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
#ifdef DOUBLE | |||||
xvilvl.d VX2, x4 ,x3 | |||||
xvilvh.d VX3, x4, x3 | |||||
xvst VX2, Y, 0 * SIZE | |||||
xvst VX3, Y, 4 * SIZE | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
#else | |||||
xvilvl.w VX2, x4 ,x3 | |||||
xvilvh.w VX3, x4, x3 | |||||
xvst VX2, Y, 0 * SIZE | |||||
xvst VX3, Y, 8 * SIZE | |||||
addi.d X, X, 16 * SIZE | |||||
addi.d Y, Y, 16 * SIZE | |||||
#endif | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L111 | |||||
b .L997 | |||||
.align 3 | |||||
.L12: // INCX==1 and INCY!=1 | |||||
bge $r0, I, .L997 | |||||
move YY, Y | |||||
.align 3 | |||||
.L121: | |||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX1, X, 4 * SIZE | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.d x3, t1, 0 | |||||
xvinsgr2vr.d x4, t2, 0 | |||||
xvinsgr2vr.d x3, t3, 2 | |||||
xvinsgr2vr.d x4, t4, 2 | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
xvinsgr2vr.d x3, t1, 1 | |||||
xvinsgr2vr.d x4, t2, 1 | |||||
xvinsgr2vr.d x3, t3, 3 | |||||
xvinsgr2vr.d x4, t4, 3 | |||||
add.d Y, Y, INCY | |||||
xvpickev.d x1, VX1, VX0 | |||||
xvpickod.d x2, VX1, VX0 | |||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w x3, t1, 0 | |||||
xvinsgr2vr.w x4, t2, 0 | |||||
xvinsgr2vr.w x3, t3, 1 | |||||
xvinsgr2vr.w x4, t4, 1 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w x3, t1, 4 | |||||
xvinsgr2vr.w x4, t2, 4 | |||||
xvinsgr2vr.w x3, t3, 5 | |||||
xvinsgr2vr.w x4, t4, 5 | |||||
xvld VX1, X, 8 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w x3, t1, 2 | |||||
xvinsgr2vr.w x4, t2, 2 | |||||
xvinsgr2vr.w x3, t3, 3 | |||||
xvinsgr2vr.w x4, t4, 3 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
xvinsgr2vr.w x3, t1, 6 | |||||
xvinsgr2vr.w x4, t2, 6 | |||||
xvinsgr2vr.w x3, t3, 7 | |||||
xvinsgr2vr.w x4, t4, 7 | |||||
add.d Y, Y, INCY | |||||
xvpickev.w x1, VX1, VX0 | |||||
xvpickod.w x2, VX1, VX0 | |||||
#endif | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmsub.d VX1, VXAR, x1, VX0 | |||||
xvfmadd.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfadd.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmsub.s VX1, VXAR, x1, VX0 | |||||
xvfmadd.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmadd.d VX1, VXAR, x1, VX0 | |||||
xvfmsub.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfsub.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmadd.s VX1, VXAR, x1, VX0 | |||||
xvfmsub.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
#ifdef DOUBLE | |||||
xvstelm.d x3, YY, 0 * SIZE, 0 | |||||
xvstelm.d x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.d x3, YY, 0 * SIZE, 2 | |||||
xvstelm.d x4, YY, 1 * SIZE, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.d x3, YY, 0 * SIZE, 1 | |||||
xvstelm.d x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.d x3, YY, 0 * SIZE, 3 | |||||
xvstelm.d x4, YY, 1 * SIZE, 3 | |||||
add.d YY, YY, INCY | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
#else | |||||
addi.d I, I, -1 | |||||
xvstelm.w x3, YY, 0 * SIZE, 0 | |||||
xvstelm.w x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 1 | |||||
xvstelm.w x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 4 | |||||
xvstelm.w x4, YY, 1 * SIZE, 4 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 5 | |||||
xvstelm.w x4, YY, 1 * SIZE, 5 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 2 | |||||
xvstelm.w x4, YY, 1 * SIZE, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 3 | |||||
xvstelm.w x4, YY, 1 * SIZE, 3 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 6 | |||||
xvstelm.w x4, YY, 1 * SIZE, 6 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 7 | |||||
xvstelm.w x4, YY, 1 * SIZE, 7 | |||||
add.d YY, YY, INCY | |||||
addi.d X, X, 16 * SIZE | |||||
#endif | |||||
blt $r0, I, .L121 | |||||
b .L997 | |||||
.align 3 | |||||
.L21:// INCX!=1 and INCY==1 | |||||
bge $r0, I, .L997 | |||||
.align 3 | |||||
.L211: | |||||
#ifdef DOUBLE | |||||
xvld VX2, Y, 0 * SIZE | |||||
xvld VX3, Y, 4 * SIZE | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.d x1, t1, 0 | |||||
xvinsgr2vr.d x2, t2, 0 | |||||
xvinsgr2vr.d x1, t3, 2 | |||||
xvinsgr2vr.d x2, t4, 2 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
xvinsgr2vr.d x1, t1, 1 | |||||
xvinsgr2vr.d x2, t2, 1 | |||||
xvinsgr2vr.d x1, t3, 3 | |||||
xvinsgr2vr.d x2, t4, 3 | |||||
add.d X, X, INCX | |||||
xvpickev.d x3, VX3, VX2 | |||||
xvpickod.d x4, VX3, VX2 | |||||
#else | |||||
xvld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 0 | |||||
xvinsgr2vr.w x2, t2, 0 | |||||
xvinsgr2vr.w x1, t3, 1 | |||||
xvinsgr2vr.w x2, t4, 1 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 4 | |||||
xvinsgr2vr.w x2, t2, 4 | |||||
xvinsgr2vr.w x1, t3, 5 | |||||
xvinsgr2vr.w x2, t4, 5 | |||||
xvld VX3, Y, 8 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 2 | |||||
xvinsgr2vr.w x2, t2, 2 | |||||
xvinsgr2vr.w x1, t3, 3 | |||||
xvinsgr2vr.w x2, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 6 | |||||
xvinsgr2vr.w x2, t2, 6 | |||||
xvinsgr2vr.w x1, t3, 7 | |||||
xvinsgr2vr.w x2, t4, 7 | |||||
xvpickev.w x3, VX3, VX2 | |||||
xvpickod.w x4, VX3, VX2 | |||||
#endif | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmsub.d VX1, VXAR, x1, VX0 | |||||
xvfmadd.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfadd.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmsub.s VX1, VXAR, x1, VX0 | |||||
xvfmadd.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmadd.d VX1, VXAR, x1, VX0 | |||||
xvfmsub.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfsub.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmadd.s VX1, VXAR, x1, VX0 | |||||
xvfmsub.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
#ifdef DOUBLE | |||||
xvilvl.d VX2, x4 ,x3 | |||||
xvilvh.d VX3, x4, x3 | |||||
addi.d I, I, -1 | |||||
xvst VX2, Y, 0 * SIZE | |||||
xvst VX3, Y, 4 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
#else | |||||
xvilvl.w VX2, x4 ,x3 | |||||
xvilvh.w VX3, x4, x3 | |||||
addi.d I, I, -1 | |||||
xvst VX2, Y, 0 * SIZE | |||||
xvst VX3, Y, 8 * SIZE | |||||
addi.d Y, Y, 16 * SIZE | |||||
#endif | |||||
blt $r0, I, .L211 | |||||
b .L997 | |||||
.align 3 | |||||
.L22: | |||||
bge $r0, I, .L997 | |||||
move YY, Y | |||||
.align 3 | |||||
.L222: | |||||
#ifdef DOUBLE | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.d x1, t1, 0 | |||||
xvinsgr2vr.d x2, t2, 0 | |||||
xvinsgr2vr.d x1, t3, 1 | |||||
xvinsgr2vr.d x2, t4, 1 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.d x1, t1, 2 | |||||
xvinsgr2vr.d x2, t2, 2 | |||||
xvinsgr2vr.d x1, t3, 3 | |||||
xvinsgr2vr.d x2, t4, 3 | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.d x3, t1, 0 | |||||
xvinsgr2vr.d x4, t2, 0 | |||||
xvinsgr2vr.d x3, t3, 1 | |||||
xvinsgr2vr.d x4, t4, 1 | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.d x3, t1, 2 | |||||
xvinsgr2vr.d x4, t2, 2 | |||||
xvinsgr2vr.d x3, t3, 3 | |||||
xvinsgr2vr.d x4, t4, 3 | |||||
#else | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 0 | |||||
xvinsgr2vr.w x2, t2, 0 | |||||
xvinsgr2vr.w x1, t3, 1 | |||||
xvinsgr2vr.w x2, t4, 1 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 2 | |||||
xvinsgr2vr.w x2, t2, 2 | |||||
xvinsgr2vr.w x1, t3, 3 | |||||
xvinsgr2vr.w x2, t4, 3 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w x3, t1, 0 | |||||
xvinsgr2vr.w x4, t2, 0 | |||||
xvinsgr2vr.w x3, t3, 1 | |||||
xvinsgr2vr.w x4, t4, 1 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w x3, t1, 2 | |||||
xvinsgr2vr.w x4, t2, 2 | |||||
xvinsgr2vr.w x3, t3, 3 | |||||
xvinsgr2vr.w x4, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 4 | |||||
xvinsgr2vr.w x2, t2, 4 | |||||
xvinsgr2vr.w x1, t3, 5 | |||||
xvinsgr2vr.w x2, t4, 5 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w x1, t1, 6 | |||||
xvinsgr2vr.w x2, t2, 6 | |||||
xvinsgr2vr.w x1, t3, 7 | |||||
xvinsgr2vr.w x2, t4, 7 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w x3, t1, 4 | |||||
xvinsgr2vr.w x4, t2, 4 | |||||
xvinsgr2vr.w x3, t3, 5 | |||||
xvinsgr2vr.w x4, t4, 5 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w x3, t1, 6 | |||||
xvinsgr2vr.w x4, t2, 6 | |||||
xvinsgr2vr.w x3, t3, 7 | |||||
xvinsgr2vr.w x4, t4, 7 | |||||
#endif | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmsub.d VX1, VXAR, x1, VX0 | |||||
xvfmadd.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfadd.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmsub.s VX1, VXAR, x1, VX0 | |||||
xvfmadd.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
xvfmul.d VX0, VXAI, x2 | |||||
xvfmul.d VX2, VXAI, x1 | |||||
xvfmadd.d VX1, VXAR, x1, VX0 | |||||
xvfmsub.d VX3, x2, VXAR, VX2 | |||||
xvfadd.d x3, x3, VX1 | |||||
xvfsub.d x4, x4, VX3 | |||||
#else | |||||
xvfmul.s VX0, VXAI, x2 | |||||
xvfmul.s VX2, VXAI, x1 | |||||
xvfmadd.s VX1, VXAR, x1, VX0 | |||||
xvfmsub.s VX3, x2, VXAR, VX2 | |||||
xvfadd.s x3, x3, VX1 | |||||
xvfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
addi.d I, I, -1 | |||||
#ifdef DOUBLE | |||||
xvstelm.d x3, YY, 0 * SIZE, 0 | |||||
xvstelm.d x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.d x3, YY, 0 * SIZE, 1 | |||||
xvstelm.d x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.d x3, YY, 0 * SIZE, 2 | |||||
xvstelm.d x4, YY, 1 * SIZE, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.d x3, YY, 0 * SIZE, 3 | |||||
xvstelm.d x4, YY, 1 * SIZE, 3 | |||||
#else | |||||
xvstelm.w x3, YY, 0 * SIZE, 0 | |||||
xvstelm.w x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 1 | |||||
xvstelm.w x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 2 | |||||
xvstelm.w x4, YY, 1 * SIZE, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 3 | |||||
xvstelm.w x4, YY, 1 * SIZE, 3 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 4 | |||||
xvstelm.w x4, YY, 1 * SIZE, 4 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 5 | |||||
xvstelm.w x4, YY, 1 * SIZE, 5 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 6 | |||||
xvstelm.w x4, YY, 1 * SIZE, 6 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w x3, YY, 0 * SIZE, 7 | |||||
xvstelm.w x4, YY, 1 * SIZE, 7 | |||||
#endif | |||||
add.d YY, YY, INCY | |||||
blt $r0, I, .L222 | |||||
.align 3 | |||||
.L997: | |||||
#ifdef DOUBLE | |||||
andi I, N, 3 | |||||
#else | |||||
andi I, N, 7 | |||||
#endif | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L998: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
#if !defined(CONJ) | |||||
MUL s1, ALPHAI, a2 | |||||
MUL s2, ALPHAI, a1 | |||||
MSUB s3, ALPHAR, a1, s1 | |||||
MADD s4, a2, ALPHAR, s2 | |||||
ADD s3, s3, a3 | |||||
ADD s4, s4, a4 | |||||
#else | |||||
MUL s1, ALPHAI, a2 | |||||
MUL s2, ALPHAI, a1 | |||||
MADD s3, ALPHAR, a1, s1 | |||||
MSUB s4, a2, ALPHAR, s2 | |||||
ADD s3, s3, a3 | |||||
SUB s4, a4, s4 | |||||
#endif | |||||
ST s3, Y, 0 * SIZE | |||||
ST s4, Y, 1 * SIZE | |||||
add.d X, X, INCX | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L998 | |||||
.align 3 | |||||
.L999: | |||||
move $r4, $r12 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |
@@ -0,0 +1,679 @@ | |||||
/*************************************************************************** | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*****************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define XX $r5 | |||||
#define YY $r6 | |||||
#define ALPHAR $f0 | |||||
#define ALPHAI $f1 | |||||
#define X $r7 | |||||
#define INCX $r8 | |||||
#define Y $r9 | |||||
#define INCY $r10 | |||||
#define I $r12 | |||||
#define TEMP $r13 | |||||
#define t1 $r14 | |||||
#define t2 $r16 | |||||
#define t3 $r15 | |||||
#define t4 $r17 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define s1 $f16 | |||||
#define s2 $f17 | |||||
#define s3 $f18 | |||||
#define s4 $f19 | |||||
#define VX0 $vr8 | |||||
#define VX1 $vr20 | |||||
#define VX2 $vr21 | |||||
#define VX3 $vr22 | |||||
#define VXAR $vr23 | |||||
#define VXAI $vr19 | |||||
#define x1 $vr18 | |||||
#define x2 $vr17 | |||||
#define x3 $vr16 | |||||
#define x4 $vr15 | |||||
PROLOGUE | |||||
bge $r0, N, .L999 | |||||
li.d TEMP, 1 | |||||
movgr2fr.d a1, $r0 | |||||
FFINT a1, a1 | |||||
CMPEQ $fcc0, ALPHAR, a1 | |||||
CMPEQ $fcc1, ALPHAI, a1 | |||||
bceqz $fcc0, .L10 | |||||
bcnez $fcc1, .L999 | |||||
.L10: | |||||
slli.d TEMP, TEMP, ZBASE_SHIFT | |||||
slli.d INCX, INCX, ZBASE_SHIFT | |||||
slli.d INCY, INCY, ZBASE_SHIFT | |||||
MTG t1, ALPHAR | |||||
MTG t2, ALPHAI | |||||
#ifdef DOUBLE | |||||
vreplgr2vr.d VXAR, t1 | |||||
vreplgr2vr.d VXAI, t2 | |||||
#else | |||||
vreplgr2vr.w VXAR, t1 | |||||
vreplgr2vr.w VXAI, t2 | |||||
#endif | |||||
srai.d I, N, 2 | |||||
bne INCX, TEMP, .L20 | |||||
bne INCY, TEMP, .L12 // INCX==1 and INCY!=1 | |||||
b .L11 // INCX==1 and INCY==1 | |||||
.L20: | |||||
bne INCY, TEMP, .L22 // INCX!=1 and INCY!=1 | |||||
b .L21 // INCX!=1 and INCY==1 | |||||
.L11: | |||||
bge $r0, I, .L997 | |||||
.align 3 | |||||
.L111: | |||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX1, X, 2 * SIZE | |||||
vld VX3, Y, 2 * SIZE | |||||
vpickev.d x1, VX1, VX0 | |||||
vpickod.d x2, VX1, VX0 | |||||
vpickev.d x3, VX3, VX2 | |||||
vpickod.d x4, VX3, VX2 | |||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
vld VX3, Y, 4 * SIZE | |||||
vpickev.w x1, VX1, VX0 | |||||
vpickod.w x2, VX1, VX0 | |||||
vpickev.w x3, VX3, VX2 | |||||
vpickod.w x4, VX3, VX2 | |||||
#endif | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmsub.s VX1, VXAR, x1, VX0 | |||||
vfmadd.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmadd.s VX1, VXAR, x1, VX0 | |||||
vfmsub.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
#ifdef DOUBLE | |||||
vilvl.d VX2, x4 ,x3 | |||||
vilvh.d VX3, x4, x3 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 2 * SIZE | |||||
vld VX0, X, 4 * SIZE | |||||
vld VX2, Y, 4 * SIZE | |||||
vld VX1, X, 6 * SIZE | |||||
vld VX3, Y, 6 * SIZE | |||||
vpickev.d x1, VX1, VX0 | |||||
vpickod.d x2, VX1, VX0 | |||||
vpickev.d x3, VX3, VX2 | |||||
vpickod.d x4, VX3, VX2 | |||||
#if !defined(CONJ) | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#endif | |||||
vilvl.d VX2, x4 ,x3 | |||||
vilvh.d VX3, x4, x3 | |||||
vst VX2, Y, 4 * SIZE | |||||
vst VX3, Y, 6 * SIZE | |||||
#else | |||||
vilvl.w VX2, x4 ,x3 | |||||
vilvh.w VX3, x4, x3 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 4 * SIZE | |||||
#endif | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L111 | |||||
b .L997 | |||||
.align 3 | |||||
.L12: // INCX==1 and INCY!=1 | |||||
bge $r0, I, .L997 | |||||
move YY, Y | |||||
.align 3 | |||||
.L121: | |||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX1, X, 2 * SIZE | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
vinsgr2vr.d x3, t1, 0 | |||||
vinsgr2vr.d x4, t2, 0 | |||||
vinsgr2vr.d x3, t3, 1 | |||||
vinsgr2vr.d x4, t4, 1 | |||||
add.d Y, Y, INCY | |||||
vpickev.d x1, VX1, VX0 | |||||
vpickod.d x2, VX1, VX0 | |||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
vinsgr2vr.w x3, t1, 0 | |||||
vinsgr2vr.w x4, t2, 0 | |||||
vinsgr2vr.w x3, t3, 1 | |||||
vinsgr2vr.w x4, t4, 1 | |||||
vld VX1, X, 4 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
vinsgr2vr.w x3, t1, 2 | |||||
vinsgr2vr.w x4, t2, 2 | |||||
vinsgr2vr.w x3, t3, 3 | |||||
vinsgr2vr.w x4, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vpickev.w x1, VX1, VX0 | |||||
vpickod.w x2, VX1, VX0 | |||||
#endif | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmsub.s VX1, VXAR, x1, VX0 | |||||
vfmadd.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmadd.s VX1, VXAR, x1, VX0 | |||||
vfmsub.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
#ifdef DOUBLE | |||||
vstelm.d x3, YY, 0 * SIZE, 0 | |||||
vstelm.d x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.d x3, YY, 0 * SIZE, 1 | |||||
vstelm.d x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
vld VX0, X, 4 * SIZE | |||||
vld VX1, X, 6 * SIZE | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
vinsgr2vr.d x3, t1, 0 | |||||
vinsgr2vr.d x4, t2, 0 | |||||
vinsgr2vr.d x3, t3, 1 | |||||
vinsgr2vr.d x4, t4, 1 | |||||
add.d Y, Y, INCY | |||||
vpickev.d x1, VX1, VX0 | |||||
vpickod.d x2, VX1, VX0 | |||||
#if !defined(CONJ) | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#endif | |||||
addi.d I, I, -1 | |||||
vstelm.d x3, YY, 0 * SIZE, 0 | |||||
vstelm.d x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.d x3, YY, 0 * SIZE, 1 | |||||
vstelm.d x4, YY, 1 * SIZE, 1 | |||||
#else | |||||
addi.d I, I, -1 | |||||
vstelm.w x3, YY, 0 * SIZE, 0 | |||||
vstelm.w x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w x3, YY, 0 * SIZE, 1 | |||||
vstelm.w x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w x3, YY, 0 * SIZE, 2 | |||||
vstelm.w x4, YY, 1 * SIZE, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w x3, YY, 0 * SIZE, 3 | |||||
vstelm.w x4, YY, 1 * SIZE, 3 | |||||
#endif | |||||
add.d YY, YY, INCY | |||||
addi.d X, X, 8 * SIZE | |||||
blt $r0, I, .L121 | |||||
b .L997 | |||||
.align 3 | |||||
.L21:// INCX!=1 and INCY==1 | |||||
bge $r0, I, .L997 | |||||
.align 3 | |||||
.L211: | |||||
#ifdef DOUBLE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX3, Y, 2 * SIZE | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
vinsgr2vr.d x1, t1, 0 | |||||
vinsgr2vr.d x2, t2, 0 | |||||
vinsgr2vr.d x1, t3, 1 | |||||
vinsgr2vr.d x2, t4, 1 | |||||
add.d X, X, INCX | |||||
vpickev.d x3, VX3, VX2 | |||||
vpickod.d x4, VX3, VX2 | |||||
#else | |||||
vld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w x1, t1, 0 | |||||
vinsgr2vr.w x2, t2, 0 | |||||
vinsgr2vr.w x1, t3, 1 | |||||
vinsgr2vr.w x2, t4, 1 | |||||
vld VX3, Y, 4 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
vinsgr2vr.w x1, t1, 2 | |||||
vinsgr2vr.w x2, t2, 2 | |||||
vinsgr2vr.w x1, t3, 3 | |||||
vinsgr2vr.w x2, t4, 3 | |||||
add.d X, X, INCX | |||||
vpickev.w x3, VX3, VX2 | |||||
vpickod.w x4, VX3, VX2 | |||||
#endif | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmsub.s VX1, VXAR, x1, VX0 | |||||
vfmadd.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmadd.s VX1, VXAR, x1, VX0 | |||||
vfmsub.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
#ifdef DOUBLE | |||||
vilvl.d VX2, x4 ,x3 | |||||
vilvh.d VX3, x4, x3 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 2 * SIZE | |||||
vld VX2, Y, 4 * SIZE | |||||
vld VX3, Y, 6 * SIZE | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
vinsgr2vr.d x1, t1, 0 | |||||
vinsgr2vr.d x2, t2, 0 | |||||
vinsgr2vr.d x1, t3, 1 | |||||
vinsgr2vr.d x2, t4, 1 | |||||
add.d X, X, INCX | |||||
vpickev.d x3, VX3, VX2 | |||||
vpickod.d x4, VX3, VX2 | |||||
#if !defined(CONJ) | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#endif | |||||
vilvl.d VX2, x4 ,x3 | |||||
vilvh.d VX3, x4, x3 | |||||
addi.d I, I, -1 | |||||
vst VX2, Y, 4 * SIZE | |||||
vst VX3, Y, 6 * SIZE | |||||
#else | |||||
vilvl.w VX2, x4 ,x3 | |||||
vilvh.w VX3, x4, x3 | |||||
addi.d I, I, -1 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 4 * SIZE | |||||
#endif | |||||
addi.d Y, Y, 8 * SIZE | |||||
blt $r0, I, .L211 | |||||
b .L997 | |||||
.align 3 | |||||
.L22: | |||||
bge $r0, I, .L997 | |||||
move YY, Y | |||||
.align 3 | |||||
.L222: | |||||
#ifdef DOUBLE | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.d x1, t1, 0 | |||||
vinsgr2vr.d x2, t2, 0 | |||||
vinsgr2vr.d x1, t3, 1 | |||||
vinsgr2vr.d x2, t4, 1 | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
vinsgr2vr.d x3, t1, 0 | |||||
vinsgr2vr.d x4, t2, 0 | |||||
vinsgr2vr.d x3, t3, 1 | |||||
vinsgr2vr.d x4, t4, 1 | |||||
#else | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w x1, t1, 0 | |||||
vinsgr2vr.w x2, t2, 0 | |||||
vinsgr2vr.w x1, t3, 1 | |||||
vinsgr2vr.w x2, t4, 1 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
vinsgr2vr.w x3, t1, 0 | |||||
vinsgr2vr.w x4, t2, 0 | |||||
vinsgr2vr.w x3, t3, 1 | |||||
vinsgr2vr.w x4, t4, 1 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w x1, t1, 2 | |||||
vinsgr2vr.w x2, t2, 2 | |||||
vinsgr2vr.w x1, t3, 3 | |||||
vinsgr2vr.w x2, t4, 3 | |||||
ld.w t1, Y, 0 * SIZE | |||||
ld.w t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
ld.w t4, Y, 1 * SIZE | |||||
vinsgr2vr.w x3, t1, 2 | |||||
vinsgr2vr.w x4, t2, 2 | |||||
vinsgr2vr.w x3, t3, 3 | |||||
vinsgr2vr.w x4, t4, 3 | |||||
#endif | |||||
add.d Y, Y, INCY | |||||
#if !defined(CONJ) | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmsub.s VX1, VXAR, x1, VX0 | |||||
vfmadd.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfadd.s x4, x4, VX3 | |||||
#endif | |||||
#else | |||||
#ifdef DOUBLE | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#else | |||||
vfmul.s VX0, VXAI, x2 | |||||
vfmul.s VX2, VXAI, x1 | |||||
vfmadd.s VX1, VXAR, x1, VX0 | |||||
vfmsub.s VX3, x2, VXAR, VX2 | |||||
vfadd.s x3, x3, VX1 | |||||
vfsub.s x4, x4, VX3 | |||||
#endif | |||||
#endif | |||||
#ifdef DOUBLE | |||||
vstelm.d x3, YY, 0 * SIZE, 0 | |||||
vstelm.d x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.d x3, YY, 0 * SIZE, 1 | |||||
vstelm.d x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.d x1, t1, 0 | |||||
vinsgr2vr.d x2, t2, 0 | |||||
vinsgr2vr.d x1, t3, 1 | |||||
vinsgr2vr.d x2, t4, 1 | |||||
ld.d t1, Y, 0 * SIZE | |||||
ld.d t2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
ld.d t4, Y, 1 * SIZE | |||||
vinsgr2vr.d x3, t1, 0 | |||||
vinsgr2vr.d x4, t2, 0 | |||||
vinsgr2vr.d x3, t3, 1 | |||||
vinsgr2vr.d x4, t4, 1 | |||||
add.d Y, Y, INCY | |||||
#if !defined(CONJ) | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmsub.d VX1, VXAR, x1, VX0 | |||||
vfmadd.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfadd.d x4, x4, VX3 | |||||
#else | |||||
vfmul.d VX0, VXAI, x2 | |||||
vfmul.d VX2, VXAI, x1 | |||||
vfmadd.d VX1, VXAR, x1, VX0 | |||||
vfmsub.d VX3, x2, VXAR, VX2 | |||||
vfadd.d x3, x3, VX1 | |||||
vfsub.d x4, x4, VX3 | |||||
#endif | |||||
addi.d I, I, -1 | |||||
vstelm.d x3, YY, 0 * SIZE, 0 | |||||
vstelm.d x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.d x3, YY, 0 * SIZE, 1 | |||||
vstelm.d x4, YY, 1 * SIZE, 1 | |||||
#else | |||||
addi.d I, I, -1 | |||||
vstelm.w x3, YY, 0 * SIZE, 0 | |||||
vstelm.w x4, YY, 1 * SIZE, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w x3, YY, 0 * SIZE, 1 | |||||
vstelm.w x4, YY, 1 * SIZE, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w x3, YY, 0 * SIZE, 2 | |||||
vstelm.w x4, YY, 1 * SIZE, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w x3, YY, 0 * SIZE, 3 | |||||
vstelm.w x4, YY, 1 * SIZE, 3 | |||||
#endif | |||||
add.d YY, YY, INCY | |||||
blt $r0, I, .L222 | |||||
.align 3 | |||||
.L997: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L998: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
#if !defined(CONJ) | |||||
MUL s1, ALPHAI, a2 | |||||
MUL s2, ALPHAI, a1 | |||||
MSUB s3, ALPHAR, a1, s1 | |||||
MADD s4, a2, ALPHAR, s2 | |||||
ADD s3, s3, a3 | |||||
ADD s4, s4, a4 | |||||
#else | |||||
MUL s1, ALPHAI, a2 | |||||
MUL s2, ALPHAI, a1 | |||||
MADD s3, ALPHAR, a1, s1 | |||||
MSUB s4, a2, ALPHAR, s2 | |||||
ADD s3, s3, a3 | |||||
SUB s4, a4, s4 | |||||
#endif | |||||
ST s3, Y, 0 * SIZE | |||||
ST s4, Y, 1 * SIZE | |||||
add.d X, X, INCX | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L998 | |||||
.align 3 | |||||
.L999: | |||||
move $r4, $r12 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |
@@ -1,323 +0,0 @@ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define XX $r5 | |||||
#define YY $r6 | |||||
#define ALPHA $f0 | |||||
#define X $r7 | |||||
#define INCX $r8 | |||||
#define Y $r9 | |||||
#define INCY $r10 | |||||
#define I $r12 | |||||
#define TEMP $r13 | |||||
#define t1 $r14 | |||||
#define t2 $r16 | |||||
#define t3 $r15 | |||||
#define t4 $r17 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define b1 $f16 | |||||
#define b2 $f17 | |||||
#define b3 $f18 | |||||
#define b4 $f19 | |||||
#define VX0 $xr8 | |||||
#define VX1 $xr20 | |||||
#define VX2 $xr21 | |||||
#define VX3 $xr22 | |||||
#define VXA $xr23 | |||||
PROLOGUE | |||||
bge $r0, N, .L999 | |||||
li.d TEMP, 1 | |||||
movgr2fr.d a1, $r0 | |||||
ffint.s.l a1, a1 | |||||
movgr2fr.d a2, TEMP | |||||
ffint.s.l a2, a2 | |||||
fcmp.ceq.s $fcc0, ALPHA, a1 | |||||
bcnez $fcc0, .L999 | |||||
slli.d TEMP, TEMP, BASE_SHIFT | |||||
slli.d INCX, INCX, BASE_SHIFT | |||||
slli.d INCY, INCY, BASE_SHIFT | |||||
movfr2gr.s t1, ALPHA | |||||
xvreplgr2vr.w VXA, t1 | |||||
srai.d I, N, 3 | |||||
bne INCX, TEMP, .L20 | |||||
bne INCY, TEMP, .L12 // INCX==1 and INCY!=1 | |||||
b .L11 // INCX==1 and INCY==1 | |||||
.L20: | |||||
bne INCY, TEMP, .L22 // INCX!=1 and INCY!=1 | |||||
b .L21 // INCX!=1 and INCY==1 | |||||
.L11: | |||||
bge $r0, I, .L113 | |||||
fcmp.ceq.s $fcc0, ALPHA, a2 | |||||
bceqz $fcc0, .L112 | |||||
.align 3 | |||||
.L111: | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
xvfadd.s VX2, VX0, VX2 | |||||
xvst VX2, Y, 0 * SIZE | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
blt $r0, I, .L111 | |||||
b .L113 | |||||
.align 3 | |||||
.L112: | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
xvst VX2, Y, 0 * SIZE | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
blt $r0, I, .L112 | |||||
.align 3 | |||||
.L113: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L114: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | |||||
addi.d Y, Y, SIZE | |||||
blt $r0, I, .L114 | |||||
b .L999 | |||||
.align 3 | |||||
.L12: // INCX==1 and INCY!=1 | |||||
bge $r0, I, .L122 | |||||
move YY, Y | |||||
.align 3 | |||||
.L121: | |||||
xvld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
xvinsgr2vr.w VX2, t1, 0 | |||||
xvinsgr2vr.w VX2, t2, 1 | |||||
xvinsgr2vr.w VX2, t3, 2 | |||||
xvinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
xvinsgr2vr.w VX2, t1, 4 | |||||
xvinsgr2vr.w VX2, t2, 5 | |||||
xvinsgr2vr.w VX2, t3, 6 | |||||
xvinsgr2vr.w VX2, t4, 7 | |||||
add.d Y, Y, INCY | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
addi.d I, I, -1 | |||||
xvstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 4 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 5 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 6 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 7 | |||||
add.d YY, YY, INCY | |||||
addi.d X, X, 8 * SIZE | |||||
blt $r0, I, .L121 | |||||
.align 3 | |||||
.L122: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L123: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L123 | |||||
b .L999 | |||||
.align 3 | |||||
.L21:// INCX!=1 and INCY==1 | |||||
bge $r0, I, .L212 | |||||
.align 3 | |||||
.L211: | |||||
xvld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
xvinsgr2vr.w VX0, t1, 0 | |||||
xvinsgr2vr.w VX0, t2, 1 | |||||
xvinsgr2vr.w VX0, t3, 2 | |||||
xvinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 4 | |||||
xvinsgr2vr.w VX0, t2, 5 | |||||
xvinsgr2vr.w VX0, t3, 6 | |||||
xvinsgr2vr.w VX0, t4, 7 | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
addi.d I, I, -1 | |||||
xvst VX2, Y, 0 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
blt $r0, I, .L211 | |||||
.align 3 | |||||
.L212: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L213: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | |||||
addi.d Y, Y, SIZE | |||||
blt $r0, I, .L213 | |||||
b .L999 | |||||
.align 3 | |||||
.L22: | |||||
bge $r0, I, .L223 | |||||
move YY, Y | |||||
.align 3 | |||||
.L222: | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 0 | |||||
xvinsgr2vr.w VX0, t2, 1 | |||||
xvinsgr2vr.w VX0, t3, 2 | |||||
xvinsgr2vr.w VX0, t4, 3 | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
xvinsgr2vr.w VX2, t1, 0 | |||||
xvinsgr2vr.w VX2, t2, 1 | |||||
xvinsgr2vr.w VX2, t3, 2 | |||||
xvinsgr2vr.w VX2, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 4 | |||||
xvinsgr2vr.w VX0, t2, 5 | |||||
xvinsgr2vr.w VX0, t3, 6 | |||||
xvinsgr2vr.w VX0, t4, 7 | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
xvinsgr2vr.w VX2, t1, 4 | |||||
xvinsgr2vr.w VX2, t2, 5 | |||||
xvinsgr2vr.w VX2, t3, 6 | |||||
xvinsgr2vr.w VX2, t4, 7 | |||||
add.d Y, Y, INCY | |||||
xvfmadd.s VX2, VX0, VXA, VX2 | |||||
addi.d I, I, -1 | |||||
xvstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 4 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 5 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 6 | |||||
add.d YY, YY, INCY | |||||
xvstelm.w VX2, YY, 0, 7 | |||||
add.d YY, YY, INCY | |||||
blt $r0, I, .L222 | |||||
.align 3 | |||||
.L223: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L224: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L224 | |||||
.align 3 | |||||
.L999: | |||||
move $r4, $r12 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |
@@ -1,338 +0,0 @@ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define XX $r5 | |||||
#define YY $r6 | |||||
#define ALPHA $f0 | |||||
#define X $r7 | |||||
#define INCX $r8 | |||||
#define Y $r9 | |||||
#define INCY $r10 | |||||
#define I $r12 | |||||
#define TEMP $r13 | |||||
#define t1 $r14 | |||||
#define t2 $r16 | |||||
#define t3 $r15 | |||||
#define t4 $r17 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define b1 $f16 | |||||
#define b2 $f17 | |||||
#define b3 $f18 | |||||
#define b4 $f19 | |||||
#define VX0 $vr8 | |||||
#define VX1 $vr20 | |||||
#define VX2 $vr21 | |||||
#define VX3 $vr22 | |||||
#define VXA $vr23 | |||||
PROLOGUE | |||||
bge $r0, N, .L999 | |||||
li.d TEMP, 1 | |||||
movgr2fr.d a1, $r0 | |||||
ffint.s.l a1, a1 | |||||
movgr2fr.d a2, TEMP | |||||
ffint.s.l a2, a2 | |||||
fcmp.ceq.s $fcc0, ALPHA, a1 | |||||
bcnez $fcc0, .L999 | |||||
slli.d TEMP, TEMP, BASE_SHIFT | |||||
slli.d INCX, INCX, BASE_SHIFT | |||||
slli.d INCY, INCY, BASE_SHIFT | |||||
movfr2gr.s t1, ALPHA | |||||
vreplgr2vr.w VXA, t1 | |||||
srai.d I, N, 3 | |||||
bne INCX, TEMP, .L20 | |||||
bne INCY, TEMP, .L12 // INCX==1 and INCY!=1 | |||||
b .L11 // INCX==1 and INCY==1 | |||||
.L20: | |||||
bne INCY, TEMP, .L22 // INCX!=1 and INCY!=1 | |||||
b .L21 // INCX!=1 and INCY==1 | |||||
.L11: | |||||
bge $r0, I, .L113 | |||||
fcmp.ceq.s $fcc0, ALPHA, a2 | |||||
bceqz $fcc0, .L112 | |||||
.align 3 | |||||
.L111: | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
vld VX3, Y, 4 * SIZE | |||||
vfadd.s VX2, VX0, VX2 | |||||
vfadd.s VX3, VX1, VX3 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 4 * SIZE | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L111 | |||||
b .L113 | |||||
.align 3 | |||||
.L112: | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
vld VX3, Y, 4 * SIZE | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
vst VX2, Y, 0 * SIZE | |||||
vst VX3, Y, 4 * SIZE | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L112 | |||||
b .L113 | |||||
.align 3 | |||||
.L113: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L114: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | |||||
addi.d Y, Y, SIZE | |||||
blt $r0, I, .L114 | |||||
b .L999 | |||||
.align 3 | |||||
.L12: // INCX==1 and INCY!=1 | |||||
bge $r0, I, .L122 | |||||
move YY, Y | |||||
.align 3 | |||||
.L121: | |||||
vld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX2, t1, 0 | |||||
vinsgr2vr.w VX2, t2, 1 | |||||
vinsgr2vr.w VX2, t3, 2 | |||||
vinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
vld VX1, X, 4 * SIZE | |||||
vstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX3, t1, 0 | |||||
vinsgr2vr.w VX3, t2, 1 | |||||
vinsgr2vr.w VX3, t3, 2 | |||||
vinsgr2vr.w VX3, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
addi.d I, I, -1 | |||||
vstelm.w VX3, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
addi.d X, X, 8 * SIZE | |||||
blt $r0, I, .L121 | |||||
.align 3 | |||||
.L122: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L123: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
addi.d X, X, SIZE | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L123 | |||||
b .L999 | |||||
.align 3 | |||||
.L21:// INCX!=1 and INCY==1 | |||||
bge $r0, I, .L212 | |||||
.align 3 | |||||
.L211: | |||||
vld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
vinsgr2vr.w VX0, t1, 0 | |||||
vinsgr2vr.w VX0, t2, 1 | |||||
vinsgr2vr.w VX0, t3, 2 | |||||
vinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
vld VX3, Y, 4 * SIZE | |||||
vst VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
vinsgr2vr.w VX1, t1, 0 | |||||
vinsgr2vr.w VX1, t2, 1 | |||||
vinsgr2vr.w VX1, t3, 2 | |||||
vinsgr2vr.w VX1, t4, 3 | |||||
add.d X, X, INCX | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
addi.d I, I, -1 | |||||
vst VX3, Y, 4 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
blt $r0, I, .L211 | |||||
.align 3 | |||||
.L212: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L213: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | |||||
addi.d Y, Y, SIZE | |||||
blt $r0, I, .L213 | |||||
b .L999 | |||||
.align 3 | |||||
.L22: | |||||
bge $r0, I, .L223 | |||||
move YY, Y | |||||
.align 3 | |||||
.L222: | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
vinsgr2vr.w VX0, t1, 0 | |||||
vinsgr2vr.w VX0, t2, 1 | |||||
vinsgr2vr.w VX0, t3, 2 | |||||
vinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX2, t1, 0 | |||||
vinsgr2vr.w VX2, t2, 1 | |||||
vinsgr2vr.w VX2, t3, 2 | |||||
vinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX2, VX0, VXA, VX2 | |||||
ld.w t1, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t2, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t4, X, 0 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w VX1, t1, 0 | |||||
vinsgr2vr.w VX1, t2, 1 | |||||
vinsgr2vr.w VX1, t3, 2 | |||||
vinsgr2vr.w VX1, t4, 3 | |||||
vstelm.w VX2, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX2, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
ld.w t1, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t2, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
add.d Y, Y, INCY | |||||
ld.w t4, Y, 0 * SIZE | |||||
vinsgr2vr.w VX3, t1, 0 | |||||
vinsgr2vr.w VX3, t2, 1 | |||||
vinsgr2vr.w VX3, t3, 2 | |||||
vinsgr2vr.w VX3, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vfmadd.s VX3, VX1, VXA, VX3 | |||||
addi.d I, I, -1 | |||||
vstelm.w VX3, YY, 0, 0 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 1 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 2 | |||||
add.d YY, YY, INCY | |||||
vstelm.w VX3, YY, 0, 3 | |||||
add.d YY, YY, INCY | |||||
blt $r0, I, .L222 | |||||
.align 3 | |||||
.L223: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L224: | |||||
fld.s $f12, X, 0 * SIZE | |||||
fld.s $f14, Y, 0 * SIZE | |||||
addi.d I, I, -1 | |||||
fmadd.s $f14, $f12, $f0, $f14 | |||||
fst.s $f14, Y, 0 * SIZE | |||||
add.d X, X, INCX | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L224 | |||||
.align 3 | |||||
.L999: | |||||
move $r4, $r12 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |