@@ -75,6 +75,12 @@ DNRM2KERNEL = dnrm2_lsx.S | |||||
CNRM2KERNEL = cnrm2_lsx.S | CNRM2KERNEL = cnrm2_lsx.S | ||||
ZNRM2KERNEL = znrm2_lsx.S | ZNRM2KERNEL = znrm2_lsx.S | ||||
CSWAPKERNEL = cswap_lsx.S | |||||
ZSWAPKERNEL = cswap_lsx.S | |||||
CSUMKERNEL = csum_lsx.S | |||||
ZSUMKERNEL = csum_lsx.S | |||||
DGEMMKERNEL = dgemm_kernel_8x4.S | DGEMMKERNEL = dgemm_kernel_8x4.S | ||||
DGEMMINCOPY = dgemm_ncopy_8_lsx.S | DGEMMINCOPY = dgemm_ncopy_8_lsx.S | ||||
DGEMMITCOPY = dgemm_tcopy_8_lsx.S | DGEMMITCOPY = dgemm_tcopy_8_lsx.S | ||||
@@ -75,6 +75,12 @@ DNRM2KERNEL = dnrm2_lasx.S | |||||
CNRM2KERNEL = cnrm2_lasx.S | CNRM2KERNEL = cnrm2_lasx.S | ||||
ZNRM2KERNEL = znrm2_lasx.S | ZNRM2KERNEL = znrm2_lasx.S | ||||
CSWAPKERNEL = cswap_lasx.S | |||||
ZSWAPKERNEL = cswap_lasx.S | |||||
CSUMKERNEL = csum_lasx.S | |||||
ZSUMKERNEL = csum_lasx.S | |||||
DGEMMKERNEL = dgemm_kernel_16x4.S | DGEMMKERNEL = dgemm_kernel_16x4.S | ||||
DGEMMINCOPY = dgemm_ncopy_16.S | DGEMMINCOPY = dgemm_ncopy_16.S | ||||
DGEMMITCOPY = dgemm_tcopy_16.S | DGEMMITCOPY = dgemm_tcopy_16.S | ||||
@@ -0,0 +1,274 @@ | |||||
/******************************************************************************* | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*******************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define X $r5 | |||||
#define INCX $r6 | |||||
#define I $r17 | |||||
#define TEMP $r18 | |||||
#define t1 $r15 | |||||
#define t2 $r12 | |||||
#define t3 $r13 | |||||
#define t4 $r14 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define s1 $f16 | |||||
#define VX0 $xr12 | |||||
#define VX1 $xr13 | |||||
#define VX2 $xr14 | |||||
#define VX3 $xr15 | |||||
#define res1 $xr16 | |||||
#define res2 $xr17 | |||||
PROLOGUE | |||||
xvxor.v res1, res1, res1 | |||||
xvxor.v res2, res2, res2 | |||||
bge $r0, N, .L999 | |||||
bge $r0, INCX, .L999 | |||||
li.d TEMP, 1 | |||||
slli.d TEMP, TEMP, ZBASE_SHIFT | |||||
slli.d INCX, INCX, ZBASE_SHIFT | |||||
srai.d I, N, 3 | |||||
bne INCX, TEMP, .L20 | |||||
bge $r0, I, .L13 | |||||
.align 3 | |||||
.L11: | |||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX1, X, 4 * SIZE | |||||
xvfadd.d res2, VX0, VX1 | |||||
xvfadd.d res1, res1, res2 | |||||
xvld VX2, X, 8 * SIZE | |||||
xvld VX3, X, 12 * SIZE | |||||
xvfadd.d res2, VX2, VX3 | |||||
xvfadd.d res1, res1, res2 | |||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX1, X, 8 * SIZE | |||||
xvfadd.s res2, VX0, VX1 | |||||
xvfadd.s res1, res2, res1 | |||||
#endif | |||||
addi.d X, X, 16 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L11 | |||||
.align 3 | |||||
.L12: | |||||
#ifdef DOUBLE | |||||
xvpickve.d VX1, res1, 1 | |||||
xvpickve.d VX2, res1, 2 | |||||
xvpickve.d VX3, res1, 3 | |||||
xvfadd.d res1, VX1, res1 | |||||
xvfadd.d res1, VX2, res1 | |||||
xvfadd.d res1, VX3, res1 | |||||
#else | |||||
xvfadd.s res2, res1, res2 | |||||
xvpickve.w VX1, res1, 1 | |||||
xvpickve.w VX2, res1, 2 | |||||
xvpickve.w VX3, res1, 3 | |||||
xvfadd.s res1, VX1, res1 | |||||
xvfadd.s res1, VX2, res1 | |||||
xvfadd.s res1, VX3, res1 | |||||
xvpickve.w VX0, res2, 4 | |||||
xvpickve.w VX1, res2, 5 | |||||
xvpickve.w VX2, res2, 6 | |||||
xvpickve.w VX3, res2, 7 | |||||
xvfadd.s res1, VX0, res1 | |||||
xvfadd.s res1, VX1, res1 | |||||
xvfadd.s res1, VX2, res1 | |||||
xvfadd.s res1, VX2, res1 | |||||
#endif | |||||
.align 3 | |||||
.L13: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L14: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
ADD a1, a1, a2 | |||||
ADD s1, a1, s1 | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 2 * SIZE | |||||
blt $r0, I, .L14 | |||||
b .L999 | |||||
.align 3 | |||||
.L20: | |||||
bge $r0, I, .L23 | |||||
.align 3 | |||||
.L21: | |||||
#ifdef DOUBLE | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.d VX0, t1, 0 | |||||
xvinsgr2vr.d VX0, t2, 1 | |||||
xvinsgr2vr.d VX0, t3, 2 | |||||
xvinsgr2vr.d VX0, t4, 3 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.d VX1, t1, 0 | |||||
xvinsgr2vr.d VX1, t2, 1 | |||||
xvinsgr2vr.d VX1, t3, 2 | |||||
xvinsgr2vr.d VX1, t4, 3 | |||||
xvfadd.d res2, VX0, VX1 | |||||
xvfadd.d res1, res1, res2 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.d VX0, t1, 0 | |||||
xvinsgr2vr.d VX0, t2, 1 | |||||
xvinsgr2vr.d VX0, t3, 2 | |||||
xvinsgr2vr.d VX0, t4, 3 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.d VX1, t1, 0 | |||||
xvinsgr2vr.d VX1, t2, 1 | |||||
xvinsgr2vr.d VX1, t3, 2 | |||||
xvinsgr2vr.d VX1, t4, 3 | |||||
xvfadd.d res2, VX0, VX1 | |||||
xvfadd.d res1, res1, res2 | |||||
#else | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 0 | |||||
xvinsgr2vr.w VX0, t2, 1 | |||||
xvinsgr2vr.w VX0, t3, 2 | |||||
xvinsgr2vr.w VX0, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX0, t1, 4 | |||||
xvinsgr2vr.w VX0, t2, 5 | |||||
xvinsgr2vr.w VX0, t3, 6 | |||||
xvinsgr2vr.w VX0, t4, 7 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX1, t1, 0 | |||||
xvinsgr2vr.w VX1, t2, 1 | |||||
xvinsgr2vr.w VX1, t3, 2 | |||||
xvinsgr2vr.w VX1, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
xvinsgr2vr.w VX1, t1, 4 | |||||
xvinsgr2vr.w VX1, t2, 5 | |||||
xvinsgr2vr.w VX1, t3, 6 | |||||
xvinsgr2vr.w VX1, t4, 7 | |||||
xvfadd.s res2, VX0, VX1 | |||||
xvfadd.s res1, res2, res1 | |||||
#endif | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L21 | |||||
.align 3 | |||||
.L22: | |||||
#ifdef DOUBLE | |||||
xvpickve.d VX1, res1, 1 | |||||
xvpickve.d VX2, res1, 2 | |||||
xvpickve.d VX3, res1, 3 | |||||
xvfadd.d res1, VX1, res1 | |||||
xvfadd.d res1, VX2, res1 | |||||
xvfadd.d res1, VX3, res1 | |||||
#else | |||||
xvfadd.s res2, res1, res2 | |||||
xvpickve.w VX1, res1, 1 | |||||
xvpickve.w VX2, res1, 2 | |||||
xvpickve.w VX3, res1, 3 | |||||
xvfadd.s res1, VX1, res1 | |||||
xvfadd.s res1, VX2, res1 | |||||
xvfadd.s res1, VX3, res1 | |||||
xvpickve.w VX0, res2, 4 | |||||
xvpickve.w VX1, res2, 5 | |||||
xvpickve.w VX2, res2, 6 | |||||
xvpickve.w VX3, res2, 7 | |||||
xvfadd.s res1, VX0, res1 | |||||
xvfadd.s res1, VX1, res1 | |||||
xvfadd.s res1, VX2, res1 | |||||
xvfadd.s res1, VX2, res1 | |||||
#endif | |||||
.align 3 | |||||
.L23: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L24: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
ADD a1, a1, a2 | |||||
ADD s1, a1, s1 | |||||
addi.d I, I, -1 | |||||
add.d X, X, INCX | |||||
blt $r0, I, .L24 | |||||
.align 3 | |||||
.L999: | |||||
fmov.s $f0, $f16 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |
@@ -0,0 +1,266 @@ | |||||
/******************************************************************************* | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*******************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define X $r5 | |||||
#define INCX $r6 | |||||
#define I $r17 | |||||
#define TEMP $r18 | |||||
#define t1 $r15 | |||||
#define t2 $r12 | |||||
#define t3 $r13 | |||||
#define t4 $r14 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define s1 $f16 | |||||
#define VX0 $vr12 | |||||
#define VX1 $vr13 | |||||
#define VX2 $vr14 | |||||
#define VX3 $vr15 | |||||
#define res1 $vr16 | |||||
#define res2 $vr17 | |||||
#define res3 $vr18 | |||||
PROLOGUE | |||||
vxor.v res1, res1, res1 | |||||
vxor.v res2, res2, res2 | |||||
bge $r0, N, .L999 | |||||
bge $r0, INCX, .L999 | |||||
li.d TEMP, 1 | |||||
slli.d TEMP, TEMP, ZBASE_SHIFT | |||||
slli.d INCX, INCX, ZBASE_SHIFT | |||||
srai.d I, N, 3 | |||||
bne INCX, TEMP, .L20 | |||||
bge $r0, I, .L13 | |||||
.align 3 | |||||
.L11: | |||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX1, X, 2 * SIZE | |||||
vfadd.d res2, VX0, VX1 | |||||
vfadd.d res1, res1, res2 | |||||
vld VX2, X, 4 * SIZE | |||||
vld VX3, X, 6 * SIZE | |||||
vfadd.d res2, VX2, VX3 | |||||
vfadd.d res1, res1, res2 | |||||
vld VX0, X, 8 * SIZE | |||||
vld VX1, X, 10 * SIZE | |||||
vfadd.d res2, VX0, VX1 | |||||
vfadd.d res1, res1, res2 | |||||
vld VX2, X, 12 * SIZE | |||||
vld VX3, X, 14 * SIZE | |||||
vfadd.d res2, VX2, VX3 | |||||
vfadd.d res1, res1, res2 | |||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
vfadd.s res2, VX0, VX1 | |||||
vld VX2, X, 8 * SIZE | |||||
vld VX3, X, 12 * SIZE | |||||
vfadd.s res3, VX2, VX3 | |||||
vfadd.s res2, res3, res2 | |||||
vfadd.s res1, res1, res2 | |||||
#endif | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 16 * SIZE | |||||
blt $r0, I, .L11 | |||||
.align 3 | |||||
.L12: | |||||
#ifdef DOUBLE | |||||
vreplvei.d VX1, res1, 1 | |||||
vfadd.d res1, VX1, res1 | |||||
#else | |||||
vreplvei.w VX1, res1, 1 | |||||
vreplvei.w VX2, res1, 2 | |||||
vreplvei.w VX3, res1, 3 | |||||
vfadd.s res1, VX1, res1 | |||||
vfadd.s res1, VX2, res1 | |||||
vfadd.s res1, VX3, res1 | |||||
#endif | |||||
.align 3 | |||||
.L13: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L14: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
ADD a1, a1, a2 | |||||
ADD s1, a1, s1 | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 2 * SIZE | |||||
blt $r0, I, .L14 | |||||
b .L999 | |||||
.align 3 | |||||
.L20: | |||||
bge $r0, I, .L23 | |||||
.align 3 | |||||
.L21: | |||||
#ifdef DOUBLE | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.d VX0, t1, 0 | |||||
vinsgr2vr.d VX0, t2, 1 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
vinsgr2vr.d VX1, t1, 0 | |||||
vinsgr2vr.d VX1, t2, 1 | |||||
add.d X, X, INCX | |||||
vfadd.d res2, VX0, VX1 | |||||
vfadd.d res1, res1, res2 | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.d VX0, t3, 0 | |||||
vinsgr2vr.d VX0, t4, 1 | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
vinsgr2vr.d VX1, t3, 0 | |||||
vinsgr2vr.d VX1, t4, 1 | |||||
add.d X, X, INCX | |||||
vfadd.d res2, VX0, VX1 | |||||
vfadd.d res1, res1, res2 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.d VX0, t1, 0 | |||||
vinsgr2vr.d VX0, t2, 1 | |||||
ld.d t1, X, 0 * SIZE | |||||
ld.d t2, X, 1 * SIZE | |||||
vinsgr2vr.d VX1, t1, 0 | |||||
vinsgr2vr.d VX1, t2, 1 | |||||
add.d X, X, INCX | |||||
vfadd.d res2, VX0, VX1 | |||||
vfadd.d res1, res1, res2 | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.d VX0, t3, 0 | |||||
vinsgr2vr.d VX0, t4, 1 | |||||
ld.d t3, X, 0 * SIZE | |||||
ld.d t4, X, 1 * SIZE | |||||
vinsgr2vr.d VX1, t3, 0 | |||||
vinsgr2vr.d VX1, t4, 1 | |||||
add.d X, X, INCX | |||||
vfadd.d res2, VX0, VX1 | |||||
vfadd.d res1, res1, res2 | |||||
#else | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w VX0, t1, 0 | |||||
vinsgr2vr.w VX0, t2, 1 | |||||
vinsgr2vr.w VX0, t3, 2 | |||||
vinsgr2vr.w VX0, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w VX1, t1, 0 | |||||
vinsgr2vr.w VX1, t2, 1 | |||||
vinsgr2vr.w VX1, t3, 2 | |||||
vinsgr2vr.w VX1, t4, 3 | |||||
vfadd.s res2, VX0, VX1 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w VX2, t1, 0 | |||||
vinsgr2vr.w VX2, t2, 1 | |||||
vinsgr2vr.w VX2, t3, 2 | |||||
vinsgr2vr.w VX2, t4, 3 | |||||
ld.w t1, X, 0 * SIZE | |||||
ld.w t2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
ld.w t4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
vinsgr2vr.w VX3, t1, 0 | |||||
vinsgr2vr.w VX3, t2, 1 | |||||
vinsgr2vr.w VX3, t3, 2 | |||||
vinsgr2vr.w VX3, t4, 3 | |||||
vfadd.s res3, VX2, VX3 | |||||
vfadd.s res2, res3, res2 | |||||
vfadd.s res1, res1, res2 | |||||
#endif | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L21 | |||||
.align 3 | |||||
.L22: | |||||
#ifdef DOUBLE | |||||
vreplvei.d VX1, res1, 1 | |||||
vfadd.d res1, VX1, res1 | |||||
#else | |||||
vreplvei.w VX1, res1, 1 | |||||
vreplvei.w VX2, res1, 2 | |||||
vreplvei.w VX3, res1, 3 | |||||
vfadd.s res1, VX1, res1 | |||||
vfadd.s res1, VX2, res1 | |||||
vfadd.s res1, VX3, res1 | |||||
#endif | |||||
.align 3 | |||||
.L23: | |||||
andi I, N, 7 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L24: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
ADD a1, a1, a2 | |||||
ADD s1, a1, s1 | |||||
addi.d I, I, -1 | |||||
add.d X, X, INCX | |||||
blt $r0, I, .L24 | |||||
.align 3 | |||||
.L999: | |||||
fmov.s $f0, $f16 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |
@@ -0,0 +1,394 @@ | |||||
/******************************************************************************* | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*******************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define X $r7 | |||||
#define INCX $r8 | |||||
#define Y $r9 | |||||
#define INCY $r10 | |||||
#define I $r17 | |||||
#define TEMP $r18 | |||||
#define XX $r5 | |||||
#define YY $r6 | |||||
#define t1 $r14 | |||||
#define t2 $r15 | |||||
#define t3 $r16 | |||||
#define t4 $r19 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define b1 $f16 | |||||
#define b2 $f17 | |||||
#define b3 $f18 | |||||
#define b4 $f19 | |||||
#define VX0 $xr12 | |||||
#define VX1 $xr13 | |||||
#define VX2 $xr14 | |||||
#define VX3 $xr15 | |||||
PROLOGUE | |||||
bge $r0, N, .L999 | |||||
li.d TEMP, 1 | |||||
slli.d TEMP, TEMP, ZBASE_SHIFT | |||||
slli.d INCX, INCX, ZBASE_SHIFT | |||||
slli.d INCY, INCY, ZBASE_SHIFT | |||||
srai.d I, N, 2 | |||||
bne INCX, TEMP, .L20 | |||||
bne INCY, TEMP, .L12 // INCX==1 and INCY!=1 | |||||
b .L11 // INCX==1 and INCY==1 | |||||
.L20: | |||||
bne INCY, TEMP, .L22 // INCX!=1 and INCY!=1 | |||||
b .L21 // INCX!=1 and INCY==1 | |||||
.L11: | |||||
bge $r0, I, .L112 | |||||
.align 3 | |||||
.L111: | |||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX1, X, 4 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
xvld VX3, Y, 4 * SIZE | |||||
xvst VX2, X, 0 * SIZE | |||||
xvst VX3, X, 4 * SIZE | |||||
xvst VX0, Y, 0 * SIZE | |||||
xvst VX1, Y, 4 * SIZE | |||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
xvld VX2, Y, 0 * SIZE | |||||
xvst VX2, X, 0 * SIZE | |||||
xvst VX0, Y, 0 * SIZE | |||||
#endif | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L111 | |||||
.align 3 | |||||
.L112: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L113: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 2 * SIZE | |||||
addi.d Y, Y, 2 * SIZE | |||||
blt $r0, I, .L113 | |||||
b .L999 | |||||
.align 3 | |||||
.L12: // INCX==1 and INCY!=1 | |||||
bge $r0, I, .L122 | |||||
.align 3 | |||||
.L121: | |||||
#ifdef DOUBLE | |||||
xvld VX0, X, 0 * SIZE | |||||
ld.d t1, Y, 0 * SIZE | |||||
xvstelm.d VX0, Y, 0 * SIZE, 0 | |||||
ld.d t2, Y, 1 * SIZE | |||||
xvstelm.d VX0, Y, 1 * SIZE, 1 | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
xvstelm.d VX0, Y, 0 * SIZE, 2 | |||||
ld.d t4, Y, 1 * SIZE | |||||
xvstelm.d VX0, Y, 1 * SIZE, 3 | |||||
xvinsgr2vr.d VX2, t1, 0 | |||||
xvinsgr2vr.d VX2, t2, 1 | |||||
xvinsgr2vr.d VX2, t3, 2 | |||||
xvinsgr2vr.d VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
xvst VX2, X, 0 * SIZE | |||||
xvld VX1, X, 4 * SIZE | |||||
ld.d t1, Y, 0 * SIZE | |||||
xvstelm.d VX1, Y, 0 * SIZE, 0 | |||||
ld.d t2, Y, 1 * SIZE | |||||
xvstelm.d VX1, Y, 1 * SIZE, 1 | |||||
add.d Y, Y, INCY | |||||
ld.d t3, Y, 0 * SIZE | |||||
xvstelm.d VX1, Y, 0 * SIZE, 2 | |||||
ld.d t4, Y, 1 * SIZE | |||||
xvstelm.d VX1, Y, 1 * SIZE, 3 | |||||
xvinsgr2vr.d VX3, t1, 0 | |||||
xvinsgr2vr.d VX3, t2, 1 | |||||
xvinsgr2vr.d VX3, t3, 2 | |||||
xvinsgr2vr.d VX3, t4, 3 | |||||
add.d Y, Y, INCY | |||||
xvst VX3, X, 4 * SIZE | |||||
#else | |||||
xvld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
xvstelm.w VX0, Y, 0 * SIZE, 0 | |||||
ld.w t2, Y, 1 * SIZE | |||||
xvstelm.w VX0, Y, 1 * SIZE, 1 | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
xvstelm.w VX0, Y, 0 * SIZE, 2 | |||||
ld.w t4, Y, 1 * SIZE | |||||
xvstelm.w VX0, Y, 1 * SIZE, 3 | |||||
xvinsgr2vr.w VX2, t1, 0 | |||||
xvinsgr2vr.w VX2, t2, 1 | |||||
xvinsgr2vr.w VX2, t3, 2 | |||||
xvinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
ld.w t1, Y, 0 * SIZE | |||||
xvstelm.w VX0, Y, 0 * SIZE, 4 | |||||
ld.w t2, Y, 1 * SIZE | |||||
xvstelm.w VX0, Y, 1 * SIZE, 5 | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
xvstelm.w VX0, Y, 0 * SIZE, 6 | |||||
ld.w t4, Y, 1 * SIZE | |||||
xvstelm.w VX0, Y, 1 * SIZE, 7 | |||||
xvinsgr2vr.w VX2, t1, 4 | |||||
xvinsgr2vr.w VX2, t2, 5 | |||||
xvinsgr2vr.w VX2, t3, 6 | |||||
xvinsgr2vr.w VX2, t4, 7 | |||||
add.d Y, Y, INCY | |||||
xvst VX2, X, 0 * SIZE | |||||
#endif | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L121 | |||||
.align 3 | |||||
.L122: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L123: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 2 * SIZE | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L123 | |||||
b .L999 | |||||
.align 3 | |||||
.L21: | |||||
bge $r0, I, .L212 | |||||
.align 3 | |||||
.L211: | |||||
#ifdef DOUBLE | |||||
xvld VX2, Y, 0 * SIZE | |||||
ld.d t1, X, 0 * SIZE | |||||
xvstelm.d VX2, X, 0 * SIZE, 0 | |||||
ld.d t2, X, 1 * SIZE | |||||
xvstelm.d VX2, X, 1 * SIZE, 1 | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
xvstelm.d VX2, X, 0 * SIZE, 2 | |||||
ld.d t4, X, 1 * SIZE | |||||
xvstelm.d VX2, X, 1 * SIZE, 3 | |||||
xvinsgr2vr.d VX0, t1, 0 | |||||
xvinsgr2vr.d VX0, t2, 1 | |||||
xvinsgr2vr.d VX0, t3, 2 | |||||
xvinsgr2vr.d VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
xvst VX0, Y, 0 * SIZE | |||||
xvld VX3, Y, 4 * SIZE | |||||
ld.d t1, X, 0 * SIZE | |||||
xvstelm.d VX3, X, 0 * SIZE, 0 | |||||
ld.d t2, X, 1 * SIZE | |||||
xvstelm.d VX3, X, 1 * SIZE, 1 | |||||
add.d X, X, INCX | |||||
ld.d t3, X, 0 * SIZE | |||||
xvstelm.d VX3, X, 0 * SIZE, 2 | |||||
ld.d t4, X, 1 * SIZE | |||||
xvstelm.d VX3, X, 1 * SIZE, 3 | |||||
xvinsgr2vr.d VX1, t1, 0 | |||||
xvinsgr2vr.d VX1, t2, 1 | |||||
xvinsgr2vr.d VX1, t3, 2 | |||||
xvinsgr2vr.d VX1, t4, 3 | |||||
add.d X, X, INCX | |||||
xvst VX1, Y, 4 * SIZE | |||||
#else | |||||
xvld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
xvstelm.w VX2, X, 0 * SIZE, 0 | |||||
ld.w t2, X, 1 * SIZE | |||||
xvstelm.w VX2, X, 1 * SIZE, 1 | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
xvstelm.w VX2, X, 0 * SIZE, 2 | |||||
ld.w t4, X, 1 * SIZE | |||||
xvstelm.w VX2, X, 1 * SIZE, 3 | |||||
xvinsgr2vr.w VX0, t1, 0 | |||||
xvinsgr2vr.w VX0, t2, 1 | |||||
xvinsgr2vr.w VX0, t3, 2 | |||||
xvinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
ld.w t1, X, 0 * SIZE | |||||
xvstelm.w VX2, X, 0 * SIZE, 4 | |||||
ld.w t2, X, 1 * SIZE | |||||
xvstelm.w VX2, X, 1 * SIZE, 5 | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
xvstelm.w VX2, X, 0 * SIZE, 6 | |||||
ld.w t4, X, 1 * SIZE | |||||
xvstelm.w VX2, X, 1 * SIZE, 7 | |||||
xvinsgr2vr.w VX0, t1, 4 | |||||
xvinsgr2vr.w VX0, t2, 5 | |||||
xvinsgr2vr.w VX0, t3, 6 | |||||
xvinsgr2vr.w VX0, t4, 7 | |||||
add.d X, X, INCX | |||||
xvst VX0, Y, 0 * SIZE | |||||
#endif | |||||
addi.d Y, Y, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L211 | |||||
.align 3 | |||||
.L212: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L213: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
add.d X, X, INCX | |||||
addi.d Y, Y, 2 * SIZE | |||||
blt $r0, I, .L213 | |||||
b .L999 | |||||
.align 3 | |||||
.L22: | |||||
bge $r0, I, .L223 | |||||
.align 3 | |||||
move XX, X | |||||
.L222: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
LD a3, X, 0 * SIZE | |||||
LD a4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
LD b1, Y, 0 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
LD b2, Y, 1 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
LD b3, Y, 0 * SIZE | |||||
ST a3, Y, 0 * SIZE | |||||
LD b4, Y, 1 * SIZE | |||||
ST a4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
LD a1, X, 0 * SIZE | |||||
ST b1, XX, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ST b2, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
LD a3, X, 0 * SIZE | |||||
ST b3, XX, 0 * SIZE | |||||
LD a4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ST b4, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
LD b1, Y, 0 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
LD b2, Y, 1 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
LD b3, Y, 0 * SIZE | |||||
ST a3, Y, 0 * SIZE | |||||
LD b4, Y, 1 * SIZE | |||||
ST a4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ST b1, XX, 0 * SIZE | |||||
ST b2, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
ST b3, XX, 0 * SIZE | |||||
ST b4, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L222 | |||||
.align 3 | |||||
.L223: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L224: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
add.d X, X, INCX | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L224 | |||||
.align 3 | |||||
.L999: | |||||
move $r4, $r12 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |
@@ -0,0 +1,421 @@ | |||||
/******************************************************************************* | |||||
Copyright (c) 2023, The OpenBLAS Project | |||||
All rights reserved. | |||||
Redistribution and use in source and binary forms, with or without | |||||
modification, are permitted provided that the following conditions are | |||||
met: | |||||
1. Redistributions of source code must retain the above copyright | |||||
notice, this list of conditions and the following disclaimer. | |||||
2. Redistributions in binary form must reproduce the above copyright | |||||
notice, this list of conditions and the following disclaimer in | |||||
the documentation and/or other materials provided with the | |||||
distribution. | |||||
3. Neither the name of the OpenBLAS project nor the names of | |||||
its contributors may be used to endorse or promote products | |||||
derived from this software without specific prior written permission. | |||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||||
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE | |||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE | |||||
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |||||
*******************************************************************************/ | |||||
#define ASSEMBLER | |||||
#include "common.h" | |||||
#define N $r4 | |||||
#define X $r7 | |||||
#define INCX $r8 | |||||
#define Y $r9 | |||||
#define INCY $r10 | |||||
#define I $r17 | |||||
#define TEMP $r18 | |||||
#define XX $r5 | |||||
#define YY $r6 | |||||
#define t1 $r14 | |||||
#define t2 $r15 | |||||
#define t3 $r16 | |||||
#define t4 $r19 | |||||
#define a1 $f12 | |||||
#define a2 $f13 | |||||
#define a3 $f14 | |||||
#define a4 $f15 | |||||
#define b1 $f16 | |||||
#define b2 $f17 | |||||
#define b3 $f18 | |||||
#define b4 $f19 | |||||
#define VX0 $vr12 | |||||
#define VX1 $vr13 | |||||
#define VX2 $vr14 | |||||
#define VX3 $vr15 | |||||
PROLOGUE | |||||
bge $r0, N, .L999 | |||||
li.d TEMP, 1 | |||||
slli.d TEMP, TEMP, ZBASE_SHIFT | |||||
slli.d INCX, INCX, ZBASE_SHIFT | |||||
slli.d INCY, INCY, ZBASE_SHIFT | |||||
srai.d I, N, 2 | |||||
bne INCX, TEMP, .L20 | |||||
bne INCY, TEMP, .L12 // INCX==1 and INCY!=1 | |||||
b .L11 // INCX==1 and INCY==1 | |||||
.L20: | |||||
bne INCY, TEMP, .L22 // INCX!=1 and INCY!=1 | |||||
b .L21 // INCX!=1 and INCY==1 | |||||
.L11: | |||||
bge $r0, I, .L112 | |||||
.align 3 | |||||
.L111: | |||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX1, X, 2 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX3, Y, 2 * SIZE | |||||
vst VX2, X, 0 * SIZE | |||||
vst VX3, X, 2 * SIZE | |||||
vst VX0, Y, 0 * SIZE | |||||
vst VX1, Y, 2 * SIZE | |||||
vld VX0, X, 4 * SIZE | |||||
vld VX1, X, 6 * SIZE | |||||
vld VX2, Y, 4 * SIZE | |||||
vld VX3, Y, 6 * SIZE | |||||
vst VX2, X, 4 * SIZE | |||||
vst VX3, X, 6 * SIZE | |||||
vst VX0, Y, 4 * SIZE | |||||
vst VX1, Y, 6 * SIZE | |||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
vld VX2, Y, 0 * SIZE | |||||
vld VX3, Y, 4 * SIZE | |||||
vst VX2, X, 0 * SIZE | |||||
vst VX3, X, 4 * SIZE | |||||
vst VX0, Y, 0 * SIZE | |||||
vst VX1, Y, 4 * SIZE | |||||
#endif | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d Y, Y, 8 * SIZE | |||||
blt $r0, I, .L111 | |||||
.align 3 | |||||
.L112: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L113: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 2 * SIZE | |||||
addi.d Y, Y, 2 * SIZE | |||||
blt $r0, I, .L113 | |||||
b .L999 | |||||
.align 3 | |||||
.L12: // INCX==1 and INCY!=1 | |||||
bge $r0, I, .L122 | |||||
.align 3 | |||||
.L121: | |||||
#ifdef DOUBLE | |||||
vld VX0, X, 0 * SIZE | |||||
ld.d t1, Y, 0 * SIZE | |||||
vstelm.d VX0, Y, 0 * SIZE, 0 | |||||
ld.d t2, Y, 1 * SIZE | |||||
vstelm.d VX0, Y, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX2, t1, 0 | |||||
vinsgr2vr.d VX2, t2, 1 | |||||
add.d Y, Y, INCY | |||||
vst VX2, X, 0 * SIZE | |||||
vld VX1, X, 2 * SIZE | |||||
ld.d t3, Y, 0 * SIZE | |||||
vstelm.d VX1, Y, 0 * SIZE, 0 | |||||
ld.d t4, Y, 1 * SIZE | |||||
vstelm.d VX1, Y, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX3, t3, 0 | |||||
vinsgr2vr.d VX3, t4, 1 | |||||
add.d Y, Y, INCY | |||||
vst VX3, X, 2 * SIZE | |||||
vld VX0, X, 4 * SIZE | |||||
ld.d t1, Y, 0 * SIZE | |||||
vstelm.d VX0, Y, 0 * SIZE, 0 | |||||
ld.d t2, Y, 1 * SIZE | |||||
vstelm.d VX0, Y, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX2, t1, 0 | |||||
vinsgr2vr.d VX2, t2, 1 | |||||
add.d Y, Y, INCY | |||||
vst VX2, X, 4 * SIZE | |||||
vld VX1, X, 6 * SIZE | |||||
ld.d t3, Y, 0 * SIZE | |||||
vstelm.d VX1, Y, 0 * SIZE, 0 | |||||
ld.d t4, Y, 1 * SIZE | |||||
vstelm.d VX1, Y, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX3, t3, 0 | |||||
vinsgr2vr.d VX3, t4, 1 | |||||
add.d Y, Y, INCY | |||||
vst VX3, X, 6 * SIZE | |||||
#else | |||||
vld VX0, X, 0 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
vstelm.w VX0, Y, 0 * SIZE, 0 | |||||
ld.w t2, Y, 1 * SIZE | |||||
vstelm.w VX0, Y, 1 * SIZE, 1 | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
vstelm.w VX0, Y, 0 * SIZE, 2 | |||||
ld.w t4, Y, 1 * SIZE | |||||
vstelm.w VX0, Y, 1 * SIZE, 3 | |||||
vinsgr2vr.w VX2, t1, 0 | |||||
vinsgr2vr.w VX2, t2, 1 | |||||
vinsgr2vr.w VX2, t3, 2 | |||||
vinsgr2vr.w VX2, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vst VX2, X, 0 * SIZE | |||||
vld VX1, X, 4 * SIZE | |||||
ld.w t1, Y, 0 * SIZE | |||||
vstelm.w VX1, Y, 0 * SIZE, 0 | |||||
ld.w t2, Y, 1 * SIZE | |||||
vstelm.w VX1, Y, 1 * SIZE, 1 | |||||
add.d Y, Y, INCY | |||||
ld.w t3, Y, 0 * SIZE | |||||
vstelm.w VX1, Y, 0 * SIZE, 2 | |||||
ld.w t4, Y, 1 * SIZE | |||||
vstelm.w VX1, Y, 1 * SIZE, 3 | |||||
vinsgr2vr.w VX3, t1, 0 | |||||
vinsgr2vr.w VX3, t2, 1 | |||||
vinsgr2vr.w VX3, t3, 2 | |||||
vinsgr2vr.w VX3, t4, 3 | |||||
add.d Y, Y, INCY | |||||
vst VX3, X, 4 * SIZE | |||||
#endif | |||||
addi.d X, X, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L121 | |||||
.align 3 | |||||
.L122: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L123: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
addi.d X, X, 2 * SIZE | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L123 | |||||
b .L999 | |||||
.align 3 | |||||
.L21:// INCX!=1 and INCY==1 | |||||
bge $r0, I, .L212 | |||||
.align 3 | |||||
.L211: | |||||
#ifdef DOUBLE | |||||
vld VX2, Y, 0 * SIZE | |||||
ld.d t1, X, 0 * SIZE | |||||
vstelm.d VX2, X, 0 * SIZE, 0 | |||||
ld.d t2, X, 1 * SIZE | |||||
vstelm.d VX2, X, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX0, t1, 0 | |||||
vinsgr2vr.d VX0, t2, 1 | |||||
add.d X, X, INCX | |||||
vst VX0, Y, 0 * SIZE | |||||
vld VX3, Y, 2 * SIZE | |||||
ld.d t3, X, 0 * SIZE | |||||
vstelm.d VX3, X, 0 * SIZE, 0 | |||||
ld.d t4, X, 1 * SIZE | |||||
vstelm.d VX3, X, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX1, t3, 0 | |||||
vinsgr2vr.d VX1, t4, 1 | |||||
add.d X, X, INCX | |||||
vst VX1, Y, 2 * SIZE | |||||
vld VX2, Y, 4 * SIZE | |||||
ld.d t1, X, 0 * SIZE | |||||
vstelm.d VX2, X, 0 * SIZE, 0 | |||||
ld.d t2, X, 1 * SIZE | |||||
vstelm.d VX2, X, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX0, t1, 0 | |||||
vinsgr2vr.d VX0, t2, 1 | |||||
add.d X, X, INCX | |||||
vst VX0, Y, 4 * SIZE | |||||
vld VX3, Y, 6 * SIZE | |||||
ld.d t3, X, 0 * SIZE | |||||
vstelm.d VX3, X, 0 * SIZE, 0 | |||||
ld.d t4, X, 1 * SIZE | |||||
vstelm.d VX3, X, 1 * SIZE, 1 | |||||
vinsgr2vr.d VX1, t3, 0 | |||||
vinsgr2vr.d VX1, t4, 1 | |||||
add.d X, X, INCX | |||||
vst VX1, Y, 6 * SIZE | |||||
#else | |||||
vld VX2, Y, 0 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
vstelm.w VX2, X, 0 * SIZE, 0 | |||||
ld.w t2, X, 1 * SIZE | |||||
vstelm.w VX2, X, 1 * SIZE, 1 | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
vstelm.w VX2, X, 0 * SIZE, 2 | |||||
ld.w t4, X, 1 * SIZE | |||||
vstelm.w VX2, X, 1 * SIZE, 3 | |||||
vinsgr2vr.w VX0, t1, 0 | |||||
vinsgr2vr.w VX0, t2, 1 | |||||
vinsgr2vr.w VX0, t3, 2 | |||||
vinsgr2vr.w VX0, t4, 3 | |||||
add.d X, X, INCX | |||||
vst VX0, Y, 0 * SIZE | |||||
vld VX3, Y, 4 * SIZE | |||||
ld.w t1, X, 0 * SIZE | |||||
vstelm.w VX3, X, 0 * SIZE, 0 | |||||
ld.w t2, X, 1 * SIZE | |||||
vstelm.w VX3, X, 1 * SIZE, 1 | |||||
add.d X, X, INCX | |||||
ld.w t3, X, 0 * SIZE | |||||
vstelm.w VX3, X, 0 * SIZE, 2 | |||||
ld.w t4, X, 1 * SIZE | |||||
vstelm.w VX3, X, 1 * SIZE, 3 | |||||
vinsgr2vr.w VX1, t1, 0 | |||||
vinsgr2vr.w VX1, t2, 1 | |||||
vinsgr2vr.w VX1, t3, 2 | |||||
vinsgr2vr.w VX1, t4, 3 | |||||
add.d X, X, INCX | |||||
vst VX1, Y, 4 * SIZE | |||||
#endif | |||||
addi.d Y, Y, 8 * SIZE | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L211 | |||||
.align 3 | |||||
.L212: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L213: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
add.d X, X, INCX | |||||
addi.d Y, Y, 2 * SIZE | |||||
blt $r0, I, .L213 | |||||
b .L999 | |||||
.align 3 | |||||
.L22: | |||||
bge $r0, I, .L223 | |||||
.align 3 | |||||
move XX, X | |||||
.L222: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
LD a3, X, 0 * SIZE | |||||
LD a4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
LD b1, Y, 0 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
LD b2, Y, 1 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
LD b3, Y, 0 * SIZE | |||||
ST a3, Y, 0 * SIZE | |||||
LD b4, Y, 1 * SIZE | |||||
ST a4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
LD a1, X, 0 * SIZE | |||||
ST b1, XX, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ST b2, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
LD a3, X, 0 * SIZE | |||||
ST b3, XX, 0 * SIZE | |||||
LD a4, X, 1 * SIZE | |||||
add.d X, X, INCX | |||||
ST b4, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
LD b1, Y, 0 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
LD b2, Y, 1 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
LD b3, Y, 0 * SIZE | |||||
ST a3, Y, 0 * SIZE | |||||
LD b4, Y, 1 * SIZE | |||||
ST a4, Y, 1 * SIZE | |||||
add.d Y, Y, INCY | |||||
ST b1, XX, 0 * SIZE | |||||
ST b2, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
ST b3, XX, 0 * SIZE | |||||
ST b4, XX, 1 * SIZE | |||||
add.d XX, XX, INCX | |||||
addi.d I, I, -1 | |||||
blt $r0, I, .L222 | |||||
.align 3 | |||||
.L223: | |||||
andi I, N, 3 | |||||
bge $r0, I, .L999 | |||||
.align 3 | |||||
.L224: | |||||
LD a1, X, 0 * SIZE | |||||
LD a2, X, 1 * SIZE | |||||
LD a3, Y, 0 * SIZE | |||||
LD a4, Y, 1 * SIZE | |||||
ST a1, Y, 0 * SIZE | |||||
ST a2, Y, 1 * SIZE | |||||
ST a3, X, 0 * SIZE | |||||
ST a4, X, 1 * SIZE | |||||
addi.d I, I, -1 | |||||
add.d X, X, INCX | |||||
add.d Y, Y, INCY | |||||
blt $r0, I, .L224 | |||||
.align 3 | |||||
.L999: | |||||
move $r4, $r12 | |||||
jirl $r0, $r1, 0x0 | |||||
.align 3 | |||||
EPILOGUE |