Browse Source

Merge remote branch 'origin/loongson3a' into x86

tags/v0.1alpha1
Xianyi Zhang 14 years ago
parent
commit
04769bdf54
4 changed files with 1025 additions and 0 deletions
  1. +8
    -0
      common_mips64.h
  2. +2
    -0
      kernel/mips64/KERNEL.LOONGSON3A
  3. +453
    -0
      kernel/mips64/axpy_loongson3a.S
  4. +562
    -0
      kernel/mips64/daxpy_loongson3a_simd.S

+ 8
- 0
common_mips64.h View File

@@ -230,4 +230,12 @@ REALNAME: ;\
#ifndef MAP_ANONYMOUS #ifndef MAP_ANONYMOUS
#define MAP_ANONYMOUS MAP_ANON #define MAP_ANONYMOUS MAP_ANON
#endif #endif

#if defined(LOONGSON3A)
#define PREFETCHD_(x) ld $0, x
#define PREFETCHD(x) PREFETCHD_(x)
#else
#define PREFETCHD(x)
#endif

#endif #endif

+ 2
- 0
kernel/mips64/KERNEL.LOONGSON3A View File

@@ -0,0 +1,2 @@
SAXPYKERNEL=axpy_loongson3a.S
DAXPYKERNEL=daxpy_loongson3a_simd.S

+ 453
- 0
kernel/mips64/axpy_loongson3a.S View File

@@ -0,0 +1,453 @@
/*****************************************************************************
Copyright (c) 2011, Lab of Parallel Software and Computational Science,ICSAS
All rights reserved.

Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:

1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.

2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
3. Neither the name of the ISCAS nor the names of its contributors may
be used to endorse or promote products derived from this software
without specific prior written permission.

THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

**********************************************************************************/

/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/

#define ASSEMBLER
#include "common.h"

#define PREFETCH_DISTANCE 48
#define N $4

#define X $8
#define INCX $9

#define Y $10
#define INCY $11

#define I $2
#define TEMP $3

#define YY $5

#define ALPHA $f15

#define a1 $f0
#define a2 $f1
#define a3 $f2
#define a4 $f3
#define a5 $f4
#define a6 $f5
#define a7 $f6
#define a8 $f7

#define b1 $f8
#define b2 $f9
#define b3 $f10
#define b4 $f11
#define b5 $f12
#define b6 $f13
#define b7 $f14
#define b8 $f17

#define t1 $f18
#define t2 $f19
#define t3 $f20
#define t4 $f21

PROLOGUE
#ifndef __64BIT__
daddiu $sp, $sp, -16
sdc1 $f20, 0($sp)
sdc1 $f21, 8($sp)
#endif

li TEMP, SIZE

blez N, .L999
dsll INCX, INCX, BASE_SHIFT

bne INCX, TEMP, .L20
dsll INCY, INCY, BASE_SHIFT

bne INCY, TEMP, .L20
dsra I, N, 3

blez I, .L15
daddiu I, I, -1

LD a1, 0 * SIZE(X)
LD a2, 1 * SIZE(X)
LD a3, 2 * SIZE(X)
LD a4, 3 * SIZE(X)
LD a5, 4 * SIZE(X)
LD a6, 5 * SIZE(X)
LD a7, 6 * SIZE(X)
LD a8, 7 * SIZE(X)

LD b1, 0 * SIZE(Y)
LD b2, 1 * SIZE(Y)
LD b3, 2 * SIZE(Y)
LD b4, 3 * SIZE(Y)
LD b5, 4 * SIZE(Y)
LD b6, 5 * SIZE(Y)
LD b7, 6 * SIZE(Y)
LD b8, 7 * SIZE(Y)
blez I, .L13
NOP
.align 5

.L12:
PREFETCHD(PREFETCH_DISTANCE*SIZE(X))
PREFETCHD((PREFETCH_DISTANCE+4)*SIZE(X))
MADD t1, b1, ALPHA, a1
MADD t2, b2, ALPHA, a2
LD b1, 8 * SIZE(Y)
LD b2, 9 * SIZE(Y)
MADD t3, b3, ALPHA, a3
MADD t4, b4, ALPHA, a4
LD b3, 10 * SIZE(Y)
LD b4, 11 * SIZE(Y)
LD a1, 8 * SIZE(X)
LD a2, 9 * SIZE(X)
LD a3, 10 * SIZE(X)
LD a4, 11 * SIZE(X)


ST t1, 0 * SIZE(Y)
ST t2, 1 * SIZE(Y)
ST t3, 2 * SIZE(Y)
ST t4, 3 * SIZE(Y)


PREFETCHD(PREFETCH_DISTANCE*SIZE(Y))
PREFETCHD((PREFETCH_DISTANCE+4)*SIZE(Y))

MADD t1, b5, ALPHA, a5
MADD t2, b6, ALPHA, a6
LD b5, 12 * SIZE(Y)
LD b6, 13 * SIZE(Y)
MADD t3, b7, ALPHA, a7
MADD t4, b8, ALPHA, a8
LD b7, 14 * SIZE(Y)
LD b8, 15 * SIZE(Y)
LD a5, 12 * SIZE(X)
LD a6, 13 * SIZE(X)
LD a7, 14 * SIZE(X)
LD a8, 15 * SIZE(X)


ST t1, 4 * SIZE(Y)
ST t2, 5 * SIZE(Y)
ST t3, 6 * SIZE(Y)
ST t4, 7 * SIZE(Y)

daddiu I, I, -1
daddiu Y, Y, 8 * SIZE

bgtz I, .L12
daddiu X, X, 8 * SIZE
.align 5

.L13:
MADD t1, b1, ALPHA, a1
MADD t2, b2, ALPHA, a2
MADD t3, b3, ALPHA, a3
MADD t4, b4, ALPHA, a4

ST t1, 0 * SIZE(Y)
MADD t1, b5, ALPHA, a5
ST t2, 1 * SIZE(Y)
MADD t2, b6, ALPHA, a6
ST t3, 2 * SIZE(Y)
MADD t3, b7, ALPHA, a7
ST t4, 3 * SIZE(Y)
MADD t4, b8, ALPHA, a8

ST t1, 4 * SIZE(Y)
ST t2, 5 * SIZE(Y)
ST t3, 6 * SIZE(Y)
ST t4, 7 * SIZE(Y)

daddiu X, X, 8 * SIZE
daddiu Y, Y, 8 * SIZE
.align 5

.L15:
andi I, N, 7

blez I, .L999
NOP
.align 3

.L16:
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)

daddiu X, X, SIZE
daddiu Y, Y, SIZE

MADD t1, b1, ALPHA, a1
daddiu I, I, -1

bgtz I, .L16
ST t1, -1 * SIZE(Y)

#ifndef __64BIT__
ldc1 $f20, 0($sp)
ldc1 $f21, 8($sp)
daddiu $sp, $sp, 16
#endif

j $31
NOP
.align 5

.L20:
dsra I, N, 3
move YY, Y

blez I, .L25
daddiu I, I, -1

LD a1, 0 * SIZE(X)
daddu X, X, INCX
LD b1, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a2, 0 * SIZE(X)
daddu X, X, INCX
LD b2, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a3, 0 * SIZE(X)
daddu X, X, INCX
LD b3, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a4, 0 * SIZE(X)
daddu X, X, INCX
LD b4, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a5, 0 * SIZE(X)
daddu X, X, INCX
LD b5, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a6, 0 * SIZE(X)
daddu X, X, INCX
LD b6, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a7, 0 * SIZE(X)
daddu X, X, INCX
LD b7, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a8, 0 * SIZE(X)
daddu X, X, INCX
LD b8, 0 * SIZE(Y)
daddu Y, Y, INCY

blez I, .L23
NOP
.align 5

.L22:
MADD t1, b1, ALPHA, a1
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

MADD t2, b2, ALPHA, a2
LD a2, 0 * SIZE(X)
LD b2, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

MADD t3, b3, ALPHA, a3
LD a3, 0 * SIZE(X)
LD b3, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

MADD t4, b4, ALPHA, a4
LD a4, 0 * SIZE(X)
LD b4, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t1, b5, ALPHA, a5

LD a5, 0 * SIZE(X)
LD b5, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t2, b6, ALPHA, a6

LD a6, 0 * SIZE(X)
LD b6, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t3, b7, ALPHA, a7

LD a7, 0 * SIZE(X)
LD b7, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t4, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t4, b8, ALPHA, a8

LD a8, 0 * SIZE(X)
daddu X, X, INCX

LD b8, 0 * SIZE(Y)
daddu Y, Y, INCY

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t4, 0 * SIZE(YY)
daddiu I, I, -1

bgtz I, .L22
daddu YY, YY, INCY
.align 5

.L23:
MADD t1, b1, ALPHA, a1
MADD t2, b2, ALPHA, a2
MADD t3, b3, ALPHA, a3
MADD t4, b4, ALPHA, a4

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t1, b5, ALPHA, a5

ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t2, b6, ALPHA, a6

ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t3, b7, ALPHA, a7

ST t4, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t4, b8, ALPHA, a8

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t4, 0 * SIZE(YY)
daddu YY, YY, INCY
.align 5

.L25:
andi I, N, 7

blez I, .L999
NOP
.align 3

.L26:
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)

MADD t1, b1, ALPHA, a1
daddu X, X, INCX

ST t1, 0 * SIZE(Y)
daddiu I, I, -1

bgtz I, .L26
daddu Y, Y, INCY
.align 5

.L999:
#ifndef __64BIT__
ldc1 $f20, 0($sp)
ldc1 $f21, 8($sp)
daddiu $sp, $sp, 16
#endif

j $31
NOP

EPILOGUE

+ 562
- 0
kernel/mips64/daxpy_loongson3a_simd.S View File

@@ -0,0 +1,562 @@
/*****************************************************************************
Copyright (c) 2011, Lab of Parallel Software and Computational Science,ICSAS
All rights reserved.

Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:

1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.

2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
3. Neither the name of the ISCAS nor the names of its contributors may
be used to endorse or promote products derived from this software
without specific prior written permission.

THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

**********************************************************************************/

/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/

#define ASSEMBLER
#include "common.h"

#define PREFETCH_DISTANCE 48
#define N $4

#define X $8
#define INCX $9

#define Y $10
#define INCY $11

#define I $2
#define TEMP $3

#define YY $5

#define ALPHA $f15

#define a1 $f0
#define a2 $f1
#define a3 $f2
#define a4 $f3
#define a5 $f4
#define a6 $f5
#define a7 $f6
#define a8 $f7

#define b1 $f8
#define b2 $f9
#define b3 $f10
#define b4 $f11
#define b5 $f12
#define b6 $f13
#define b7 $f14
#define b8 $f17

#define t1 $f18
#define t2 $f19
#define t3 $f20
#define t4 $f21

#define t5 $f22
#define t6 $f23
#define t7 $f24
#define t8 $f25


#define A1 0
#define A2 1
#define A3 2
#define A4 3
#define A5 4
#define A6 5
#define A7 6
#define A8 7

#define B1 8
#define B2 9
#define B3 10
#define B4 11
#define B5 12
#define B6 13
#define B7 14
#define B8 17

#define T1 18
#define T2 19
#define T3 20
#define T4 21

#define T5 22
#define T6 23
#define T7 24
#define T8 25

#define X_BASE 8
#define Y_BASE 10
#define gsLQC1_(base,fq,ft,offset) .word (0x32<<26|base<<21|ft<<16|0x1<<15|offset<<6|0x1<<5|fq)
#define gsLQC1(base,fq,ft,offset) gsLQC1_((base), (fq), (ft), (offset))

#define gsSQC1_(base,fq,ft,offset) .word (0x3A<<26|base<<21|ft<<16|0x1<<15|offset<<6|0x1<<5|fq)
#define gsSQC1(base,fq,ft,offset) gsSQC1_((base), (fq), (ft), (offset))

PROLOGUE
#ifndef __64BIT__
daddiu $sp, $sp, -16
sdc1 $f20, 0($sp)
sdc1 $f21, 8($sp)
#endif

daddiu $sp, $sp, -16
sdc1 t7, 0($sp)
sdc1 t8, 8($sp)

li TEMP, SIZE

blez N, .L999
dsll INCX, INCX, BASE_SHIFT

bne INCX, TEMP, .L20
dsll INCY, INCY, BASE_SHIFT

bne INCY, TEMP, .L20
dsra I, N, 3

blez I, .L15
daddiu I, I, -1

gsLQC1(X_BASE,A2,A1,0*SIZE)
gsLQC1(X_BASE,A4,A3,2*SIZE)
gsLQC1(X_BASE,A6,A5,4*SIZE)
gsLQC1(X_BASE,A8,A7,6*SIZE)
/* LD a1, 0 * SIZE(X)
LD a2, 1 * SIZE(X)
LD a3, 2 * SIZE(X)
LD a4, 3 * SIZE(X)
LD a5, 4 * SIZE(X)
LD a6, 5 * SIZE(X)
LD a7, 6 * SIZE(X)
LD a8, 7 * SIZE(X)
*/
gsLQC1(Y_BASE,B2,B1,0*SIZE)
gsLQC1(Y_BASE,B4,B3,2*SIZE)
gsLQC1(Y_BASE,B6,B5,4*SIZE)
gsLQC1(Y_BASE,B8,B7,6*SIZE)
/*
LD b1, 0 * SIZE(Y)
LD b2, 1 * SIZE(Y)
LD b3, 2 * SIZE(Y)
LD b4, 3 * SIZE(Y)
LD b5, 4 * SIZE(Y)
LD b6, 5 * SIZE(Y)
LD b7, 6 * SIZE(Y)
LD b8, 7 * SIZE(Y)
*/
blez I, .L13
NOP
.align 5

.L12:


MADD t1, b1, ALPHA, a1
MADD t2, b2, ALPHA, a2
PREFETCHD(PREFETCH_DISTANCE*SIZE(X))
gsLQC1(Y_BASE,B2,B1,8*SIZE)

MADD t3, b3, ALPHA, a3
MADD t4, b4, ALPHA, a4
PREFETCHD((PREFETCH_DISTANCE+4)*SIZE(X))
gsLQC1(Y_BASE,B4,B3,10*SIZE)



MADD t5, b5, ALPHA, a5
MADD t6, b6, ALPHA, a6
PREFETCHD(PREFETCH_DISTANCE*SIZE(Y))
gsLQC1(Y_BASE,B6,B5,12*SIZE)
MADD t7, b7, ALPHA, a7
MADD t8, b8, ALPHA, a8
PREFETCHD((PREFETCH_DISTANCE+4)*SIZE(Y))
gsLQC1(Y_BASE,B8,B7,14*SIZE)


/* LD b1, 8 * SIZE(Y)
LD b2, 9 * SIZE(Y)
*/





/*
LD b3, 10 * SIZE(Y)
LD b4, 11 * SIZE(Y)
*/
gsLQC1(X_BASE,A2,A1,8*SIZE)
gsLQC1(X_BASE,A4,A3,10*SIZE)
gsLQC1(X_BASE,A6,A5,12*SIZE)
gsLQC1(X_BASE,A8,A7,14*SIZE)

/*
LD a1, 8 * SIZE(X)
LD a2, 9 * SIZE(X)
LD a3, 10 * SIZE(X)
LD a4, 11 * SIZE(X)
*/

/*
ST t1, 0 * SIZE(Y)
ST t2, 1 * SIZE(Y)
ST t3, 2 * SIZE(Y)
ST t4, 3 * SIZE(Y)
*/






/*
LD b5, 12 * SIZE(Y)
LD b6, 13 * SIZE(Y)
*/

/*
LD b7, 14 * SIZE(Y)
LD b8, 15 * SIZE(Y)
*/
/*
LD a5, 12 * SIZE(X)
LD a6, 13 * SIZE(X)
LD a7, 14 * SIZE(X)
LD a8, 15 * SIZE(X)
*/
gsSQC1(Y_BASE, T2, T1, 0*SIZE)
gsSQC1(Y_BASE, T4, T3, 2*SIZE)
gsSQC1(Y_BASE, T6, T5, 4*SIZE)
gsSQC1(Y_BASE, T8, T7, 6*SIZE)
/*
ST t1, 4 * SIZE(Y)
ST t2, 5 * SIZE(Y)
ST t3, 6 * SIZE(Y)
ST t4, 7 * SIZE(Y)
*/
daddiu I, I, -1
daddiu Y, Y, 8 * SIZE

bgtz I, .L12
daddiu X, X, 8 * SIZE
.align 5

.L13:
MADD t1, b1, ALPHA, a1
MADD t2, b2, ALPHA, a2
MADD t3, b3, ALPHA, a3
MADD t4, b4, ALPHA, a4

ST t1, 0 * SIZE(Y)
MADD t1, b5, ALPHA, a5
ST t2, 1 * SIZE(Y)
MADD t2, b6, ALPHA, a6
ST t3, 2 * SIZE(Y)
MADD t3, b7, ALPHA, a7
ST t4, 3 * SIZE(Y)
MADD t4, b8, ALPHA, a8

ST t1, 4 * SIZE(Y)
ST t2, 5 * SIZE(Y)
ST t3, 6 * SIZE(Y)
ST t4, 7 * SIZE(Y)

daddiu X, X, 8 * SIZE
daddiu Y, Y, 8 * SIZE
.align 5

.L15:
andi I, N, 7

blez I, .L999
NOP
.align 3

.L16:
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)

daddiu X, X, SIZE
daddiu Y, Y, SIZE

MADD t1, b1, ALPHA, a1
daddiu I, I, -1

bgtz I, .L16
ST t1, -1 * SIZE(Y)

ldc1 t7, 0($sp)
ldc1 t8, 8($sp)
daddiu $sp, $sp, 16

#ifndef __64BIT__
ldc1 $f20, 0($sp)
ldc1 $f21, 8($sp)
daddiu $sp, $sp, 16
#endif

j $31
NOP
.align 5

.L20:
dsra I, N, 3
move YY, Y

blez I, .L25
daddiu I, I, -1

LD a1, 0 * SIZE(X)
daddu X, X, INCX
LD b1, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a2, 0 * SIZE(X)
daddu X, X, INCX
LD b2, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a3, 0 * SIZE(X)
daddu X, X, INCX
LD b3, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a4, 0 * SIZE(X)
daddu X, X, INCX
LD b4, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a5, 0 * SIZE(X)
daddu X, X, INCX
LD b5, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a6, 0 * SIZE(X)
daddu X, X, INCX
LD b6, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a7, 0 * SIZE(X)
daddu X, X, INCX
LD b7, 0 * SIZE(Y)
daddu Y, Y, INCY
LD a8, 0 * SIZE(X)
daddu X, X, INCX
LD b8, 0 * SIZE(Y)
daddu Y, Y, INCY

blez I, .L23
NOP
.align 5

.L22:
MADD t1, b1, ALPHA, a1
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

MADD t2, b2, ALPHA, a2
LD a2, 0 * SIZE(X)
LD b2, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

MADD t3, b3, ALPHA, a3
LD a3, 0 * SIZE(X)
LD b3, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

MADD t4, b4, ALPHA, a4
LD a4, 0 * SIZE(X)
LD b4, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t1, b5, ALPHA, a5

LD a5, 0 * SIZE(X)
LD b5, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t2, b6, ALPHA, a6

LD a6, 0 * SIZE(X)
LD b6, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t3, b7, ALPHA, a7

LD a7, 0 * SIZE(X)
LD b7, 0 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY

ST t4, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t4, b8, ALPHA, a8

LD a8, 0 * SIZE(X)
daddu X, X, INCX

LD b8, 0 * SIZE(Y)
daddu Y, Y, INCY

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t4, 0 * SIZE(YY)
daddiu I, I, -1

bgtz I, .L22
daddu YY, YY, INCY
.align 5

.L23:
MADD t1, b1, ALPHA, a1
MADD t2, b2, ALPHA, a2
MADD t3, b3, ALPHA, a3
MADD t4, b4, ALPHA, a4

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t1, b5, ALPHA, a5

ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t2, b6, ALPHA, a6

ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t3, b7, ALPHA, a7

ST t4, 0 * SIZE(YY)
daddu YY, YY, INCY
MADD t4, b8, ALPHA, a8

ST t1, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t2, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
daddu YY, YY, INCY
ST t4, 0 * SIZE(YY)
daddu YY, YY, INCY
.align 5

.L25:
andi I, N, 7

blez I, .L999
NOP
.align 3

.L26:
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)

MADD t1, b1, ALPHA, a1
daddu X, X, INCX

ST t1, 0 * SIZE(Y)
daddiu I, I, -1

bgtz I, .L26
daddu Y, Y, INCY
.align 5

.L999:
ldc1 t7, 0($sp)
ldc1 t8, 8($sp)
daddiu $sp, $sp, 16

#ifndef __64BIT__
ldc1 $f20, 0($sp)
ldc1 $f21, 8($sp)
daddiu $sp, $sp, 16
#endif

j $31
NOP

EPILOGUE

Loading…
Cancel
Save