|
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325 |
- /*******************************************************************************
- Copyright (c) 2024, The OpenBLAS Project
- All rights reserved.
- Redistribution and use in source and binary forms, with or without
- modification, are permitted provided that the following conditions are
- met:
- 1. Redistributions of source code must retain the above copyright
- notice, this list of conditions and the following disclaimer.
- 2. Redistributions in binary form must reproduce the above copyright
- notice, this list of conditions and the following disclaimer in
- the documentation and/or other materials provided with the
- distribution.
- 3. Neither the name of the OpenBLAS project nor the names of
- its contributors may be used to endorse or promote products
- derived from this software without specific prior written permission.
- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
- AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
- LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
- DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
- SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
- CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
- OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
- USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
- *******************************************************************************/
- #define ASSEMBLER
-
- #include "common.h"
-
- /* Function parameters */
- #define M $r4 // param 1: m
- #define N $r5 // param 2: n
- #define SRC $r6 // param 3: src
- #define LDA $r7 // param 4: lda
- #define DST $r8 // param 5: dst
-
- #define I $r9
- #define J $r10
- #define S1 $r12
- #define S2 $r13
- #define S3 $r14
- #define S4 $r15
- #define S5 $r16
- #define S6 $r17
- #define S7 $r18
- #define TD $r20
- #define TS $r11
- #define TL $r19
- #define T0 $r23
- #define ZERO $r0
-
- #define F0 $f0
- #define F1 $f1
- #define F2 $f2
- #define F3 $f3
- #define F4 $f4
- #define F5 $f5
- #define F6 $f6
- #define F7 $f7
-
- /* LASX vectors */
- #define U0 $xr0
- #define U1 $xr1
- #define U2 $xr2
- #define U3 $xr3
- #define U4 $xr4
- #define U5 $xr5
- #define U6 $xr6
- #define U7 $xr7
- #define D0 $xr8
- #define D1 $xr9
- #define D2 $xr10
- #define D3 $xr11
- #define D4 $xr12
- #define D5 $xr13
- #define D6 $xr14
- #define D7 $xr15
- #define D8 $xr16
-
- PROLOGUE
-
- addi.d $sp, $sp, -8
- SDARG $r23, $sp, 0
-
- move TD, DST //boffset
- move TS, SRC //aoffset
-
- slli.d TL, LDA, 0x02
- slli.d TL, TL, 0x01
-
- srai.d J, N, 0x02
- beq J, ZERO, .L_N0
-
- .L_J1: /* J-- */
- move S1, TS
- add.d S2, S1, TL
- add.d S3, S2, TL
- add.d S4, S3, TL
-
- slli.d T0, TL, 0x02
- add.d TS, TS, T0
-
- srai.d I, M, 0x02
- beq I, ZERO, .L_I3
-
- .L_I1: /* I-- */
- xvld U0, S1, 0x00 //1 2 3 4 5 6 7 8
- xvld U1, S2, 0x00 //9 10 11 12 13 14 15 16
- xvld U2, S3, 0x00 //17 18 19 20 21 22 23 24
- xvld U3, S4, 0x00 //25 26 27 28 29 30 31 32
-
- xvand.v D0, U0, U0
- xvand.v D1, U1, U1
- xvand.v D2, U2, U2
- xvand.v D3, U3, U3
-
- xvshuf4i.d D0, U1, 0x88 //1 2 9 10 5 6 13 14
- xvshuf4i.d D2, U3, 0x88 //17 18 25 26 21 22 29 30
- xvshuf4i.d D1, U0, 0x77 //3 4 11 12 7 8 15 16
- xvshuf4i.d D3, U2, 0x77 //19 20 27 28 23 24 31 32
-
- xvand.v U4, D0, D0
- xvand.v U5, D1, D1
-
- xvpermi.q U4, D2, 0x02 //1 2 9 10 17 18 25 26
- xvpermi.q U5, D3, 0x02 //3 4 11 12 19 20 27 28
- xvpermi.q D2, D0, 0x31 //5 6 13 14 21 22 29 30
- xvpermi.q D3, D1, 0x31 //7 8 15 16 23 24 31 32
-
- xvst U4, TD, 0x00
- xvst U5, TD, 0x20
- xvst D2, TD, 0x40
- xvst D3, TD, 0x60
-
- addi.d S1, S1, 0x20 // a_offset
- addi.d S2, S2, 0x20
- addi.d S3, S3, 0x20
- addi.d S4, S4, 0x20
- addi.d TD, TD, 0x80 // b_offset
-
- addi.d I, I, -1
- blt ZERO, I, .L_I1
-
- .L_I3: /* if(m&2) */
- andi I, M, 0x02
- beq I, ZERO, .L_II20
-
- vld $vr0, S1, 0x00
- vld $vr1, S2, 0x00
- vld $vr2, S3, 0x00
- vld $vr3, S4, 0x00
-
- vand.v $vr8, $vr1, $vr1
- vand.v $vr9, $vr1, $vr1
- vand.v $vr10, $vr3, $vr3
- vand.v $vr11, $vr3, $vr3
-
- vpermi.w $vr8, $vr0, 0x44
- vpermi.w $vr10, $vr2, 0x44
- vpermi.w $vr9, $vr0, 0xee
- vpermi.w $vr11, $vr2, 0xee
-
- vst $vr8, TD, 0x00
- vst $vr10, TD, 0x10
- vst $vr9, TD, 0x20
- vst $vr11, TD, 0x30
-
- addi.d S1, S1, 0x10
- addi.d S2, S2, 0x10
- addi.d S3, S3, 0x10
- addi.d S4, S4, 0x10
- addi.d TD, TD, 0x40
-
- .L_II20: /* if(m&1) */
- andi I, M, 0x01
- beq I, ZERO, .L_J0
-
- fld.s F0, S1, 0x00
- fld.s F1, S1, 0x04
-
- fld.s F2, S2, 0x00
- fld.s F3, S2, 0x04
-
- fld.s F4, S3, 0x00
- fld.s F5, S3, 0x04
-
- fld.s F6, S4, 0x00
- fld.s F7, S4, 0x04
-
- fst.s F0, TD, 0x00
- fst.s F1, TD, 0x04
- fst.s F2, TD, 0x08
- fst.s F3, TD, 0x0c
- fst.s F4, TD, 0x10
- fst.s F5, TD, 0x14
- fst.s F6, TD, 0x18
- fst.s F7, TD, 0x1c
-
- addi.d TD, TD, 0x20
-
- .L_J0:
- addi.d J, J, -1
- blt ZERO, J, .L_J1
-
- .L_N0: /* if(n&2) */
- andi I, N, 0x02
- beq ZERO, I, .L_N20
-
- move S1, TS
- add.d S2, S1, TL
-
- slli.d T0, TL, 0x01
- add.d TS, TS, T0
-
- srai.d I, M, 0x02
- beq ZERO, I, .L_N10
-
- .L_N11: /* if(i>0) */
- xvld U0, S1, 0x00 //1 2 3 4 5 6 7 8
- xvld U1, S2, 0x00 //9 10 11 12 13 14 15 16
-
- xvand.v D0, U0, U0
- xvand.v D1, U1, U1
-
- xvshuf4i.d D0, U1, 0x88 //1 2 9 10 5 6 13 14
- xvshuf4i.d D1, U0, 0x77 //3 4 11 12 7 8 15 16
-
- xvand.v U4, D0, D0
-
- xvpermi.q U4, D1, 0x02 //1 2 9 10 3 4 11 12
- xvpermi.q D1, D0, 0x31 //5 6 13 14 7 8 15 16
-
- xvst U4, TD, 0x00
- xvst D1, TD, 0x20
-
- addi.d S1, S1, 0x20 // a_offset
- addi.d S2, S2, 0x20
- addi.d TD, TD, 0x40 // b_offset
-
- addi.d I, I, -1
- blt ZERO, I, .L_N11
-
- .L_N10: /* if(m&2) */
- andi I, M, 0x02
- beq I, ZERO, .L_N130
-
- vld $vr0, S1, 0x00
- vld $vr1, S2, 0x00
- vand.v $vr8, $vr1, $vr1
-
- vpermi.w $vr8, $vr0, 0x44
- vpermi.w $vr1, $vr0, 0xee
-
- vst $vr8, TD, 0x00
- vst $vr1, TD, 0x10
-
- addi.d S1, S1, 0x10 // a_offset
- addi.d S2, S2, 0x10
- addi.d TD, TD, 0x20 // b_offset
-
- .L_N130: /* if(m&1) */
- andi I, M, 0x01
- beq I, ZERO, .L_N20
-
- fld.s F0, S1, 0x00
- fld.s F1, S1, 0x04
-
- fld.s F2, S2, 0x00
- fld.s F3, S2, 0x04
-
- fst.s F0, TD, 0x00
- fst.s F1, TD, 0x04
- fst.s F2, TD, 0x08
- fst.s F3, TD, 0x0c
-
- addi.d TD, TD, 0x10
-
- .L_N20: /* if(n&1) */
- andi I, N, 0x01
- beq I, ZERO, .L_N00
-
- move S1, TS
- srai.d I, M, 0x02
-
- beq I, ZERO, .L_N30
-
- .L_N21: /* if(i>0) */
- xvld U0, S1, 0x00
-
- xvst U0, TD, 0x00
-
- addi.d S1, S1, 0x20 // aoffset1
- addi.d TD, TD, 0x20 // b_offset
-
- addi.d I, I, -1
- blt ZERO, I, .L_N21
-
- .L_N30: /* if(m&2) */
- andi I, M, 0x02
- beq I, ZERO, .L_N330
-
- vld $vr0, S1, 0x00
-
- vst $vr0, TD, 0x00
-
- addi.d S1, S1, 0x10 // aoffset1
- addi.d TD, TD, 0x10 // b_offset
-
- .L_N330: /* if(m&1) */
- andi I, M, 0x01
- beq I, ZERO, .L_N00
-
- fld.s F0, S1, 0x00
- fld.s F1, S1, 0x04
-
- fst.s F0, TD, 0x00
- fst.s F1, TD, 0x04
-
- .L_N00:
- LDARG $r23, $sp, 0
- addi.d $sp, $sp, 8
- jirl $r0, $r1, 0x00
-
- EPILOGUE
|