|
- /*********************************************************************/
- /* Copyright 2009, 2010 The University of Texas at Austin. */
- /* All rights reserved. */
- /* */
- /* Redistribution and use in source and binary forms, with or */
- /* without modification, are permitted provided that the following */
- /* conditions are met: */
- /* */
- /* 1. Redistributions of source code must retain the above */
- /* copyright notice, this list of conditions and the following */
- /* disclaimer. */
- /* */
- /* 2. Redistributions in binary form must reproduce the above */
- /* copyright notice, this list of conditions and the following */
- /* disclaimer in the documentation and/or other materials */
- /* provided with the distribution. */
- /* */
- /* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
- /* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
- /* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
- /* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
- /* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
- /* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
- /* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
- /* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
- /* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
- /* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
- /* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
- /* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
- /* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
- /* POSSIBILITY OF SUCH DAMAGE. */
- /* */
- /* The views and conclusions contained in the software and */
- /* documentation are those of the authors and should not be */
- /* interpreted as representing official policies, either expressed */
- /* or implied, of The University of Texas at Austin. */
- /*********************************************************************/
-
- #define ASSEMBLER
- #include "common.h"
-
- #define N r3
- #define X r6
- #define INCX r7
-
- #define INCX2 r4
- #define X2 r5
-
- #define ALPHA f1
-
- #define A1 f0
- #define A2 f16
- #define A3 f2
- #define A4 f3
- #define A5 f4
- #define A6 f5
- #define A7 f6
- #define A8 f7
-
- #define B1 f8
- #define B2 f9
- #define B3 f10
- #define B4 f11
- #define B5 f12
- #define B6 f13
- #define B7 f14
- #define B8 f15
-
- PROLOGUE
- PROFCODE
-
- li r10, -16
-
- stfpdux f14, SP, r10
- stfpdux f15, SP, r10
- stfpdux f16, SP, r10
-
- li r10, 0
- stwu r10, -4(SP)
- stwu r10, -4(SP)
- stwu r10, -4(SP)
- stwu r10, -4(SP)
-
- lfpdx A1, SP, r10 # Zero clear
- fsmfp ALPHA, ALPHA
-
- slwi INCX, INCX, BASE_SHIFT
- add INCX2, INCX, INCX
-
- cmpwi cr0, N, 0
- ble LL(999)
-
- cmpwi cr0, INCX, SIZE
- bne LL(100)
-
- fcmpu cr7, ALPHA, A1
- bne cr7, LL(50)
-
- sub X, X, INCX2
-
- andi. r0, X, 2 * SIZE - 1
- beq LL(11)
-
- STFDX A1, X, INCX2
- addi X, X, 1 * SIZE
- addi N, N, -1
- cmpwi cr0, N, 0
- ble LL(999)
- .align 4
-
- LL(11):
- srawi. r0, N, 4
- mtspr CTR, r0
- beq- LL(15)
- .align 4
-
- LL(12):
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- bdnz LL(12)
- .align 4
-
- LL(15):
- andi. r0, N, 15
- beq LL(999)
- andi. r0, N, 8
- beq LL(16)
-
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- .align 4
-
- LL(16):
- andi. r0, N, 4
- beq LL(17)
-
- STFPDUX A1, X, INCX2
- STFPDUX A1, X, INCX2
- .align 4
-
- LL(17):
- andi. r0, N, 2
- beq LL(18)
-
- STFPDUX A1, X, INCX2
- .align 4
-
- LL(18):
- andi. r0, N, 1
- beq LL(999)
- STFDUX A1, X, INCX2
- b LL(999)
- .align 4
-
- LL(50):
- sub X2, X, INCX2
- sub X, X, INCX2
-
- andi. r0, X, 2 * SIZE - 1
- beq LL(51)
-
- LFDX A1, X, INCX2
- addi X, X, 1 * SIZE
-
- fmul B1, ALPHA, A1
- addi N, N, -1
- cmpwi cr0, N, 0
-
- STFDX B1, X2, INCX2
- addi X2, X2, 1 * SIZE
- ble LL(999)
- .align 4
-
- LL(51):
- srawi. r0, N, 4
- mtspr CTR, r0
- beq- LL(55)
-
- LFPDUX A1, X, INCX2
- LFPDUX A2, X, INCX2
- LFPDUX A3, X, INCX2
- LFPDUX A4, X, INCX2
- LFPDUX A5, X, INCX2
- LFPDUX A6, X, INCX2
- LFPDUX A7, X, INCX2
- LFPDUX A8, X, INCX2
- bdz LL(53)
- .align 4
-
- LL(52):
- fpmul B1, ALPHA, A1
- LFPDUX A1, X, INCX2
- fpmul B2, ALPHA, A2
- LFPDUX A2, X, INCX2
- fpmul B3, ALPHA, A3
- LFPDUX A3, X, INCX2
- fpmul B4, ALPHA, A4
- LFPDUX A4, X, INCX2
- fpmul B5, ALPHA, A5
- LFPDUX A5, X, INCX2
- fpmul B6, ALPHA, A6
- LFPDUX A6, X, INCX2
- fpmul B7, ALPHA, A7
- LFPDUX A7, X, INCX2
- fpmul B8, ALPHA, A8
- LFPDUX A8, X, INCX2
-
- STFPDUX B1, X2, INCX2
- STFPDUX B2, X2, INCX2
- STFPDUX B3, X2, INCX2
- STFPDUX B4, X2, INCX2
- STFPDUX B5, X2, INCX2
- STFPDUX B6, X2, INCX2
- STFPDUX B7, X2, INCX2
- STFPDUX B8, X2, INCX2
- bdnz LL(52)
- .align 4
-
- LL(53):
- fpmul B1, ALPHA, A1
- fpmul B2, ALPHA, A2
- fpmul B3, ALPHA, A3
- fpmul B4, ALPHA, A4
- fpmul B5, ALPHA, A5
- fpmul B6, ALPHA, A6
- STFPDUX B1, X2, INCX2
- fpmul B7, ALPHA, A7
- STFPDUX B2, X2, INCX2
- fpmul B8, ALPHA, A8
- STFPDUX B3, X2, INCX2
-
- STFPDUX B4, X2, INCX2
- STFPDUX B5, X2, INCX2
- STFPDUX B6, X2, INCX2
- STFPDUX B7, X2, INCX2
- STFPDUX B8, X2, INCX2
- .align 4
-
- LL(55):
- andi. r0, N, 15
- beq LL(999)
- andi. r0, N, 8
- beq LL(56)
-
- LFPDUX A1, X, INCX2
- LFPDUX A2, X, INCX2
- LFPDUX A3, X, INCX2
- LFPDUX A4, X, INCX2
-
- fpmul B1, ALPHA, A1
- fpmul B2, ALPHA, A2
- fpmul B3, ALPHA, A3
- fpmul B4, ALPHA, A4
-
- STFPDUX B1, X2, INCX2
- STFPDUX B2, X2, INCX2
- STFPDUX B3, X2, INCX2
- STFPDUX B4, X2, INCX2
- .align 4
-
- LL(56):
- andi. r0, N, 4
- beq LL(57)
-
- LFPDUX A1, X, INCX2
- LFPDUX A2, X, INCX2
- fpmul B1, ALPHA, A1
- fpmul B2, ALPHA, A2
- STFPDUX B1, X2, INCX2
- STFPDUX B2, X2, INCX2
- .align 4
-
- LL(57):
- andi. r0, N, 2
- beq LL(58)
-
- LFPDUX A1, X, INCX2
- fpmul B1, ALPHA, A1
- STFPDUX B1, X2, INCX2
- .align 4
-
- LL(58):
- andi. r0, N, 1
- beq LL(999)
-
- LFDX A1, X, INCX2
- fmul B1, ALPHA, A1
- STFDX B1, X2, INCX2
- b LL(999)
- .align 4
-
-
- LL(100):
- fcmpu cr7, ALPHA, A1
- bne cr7, LL(200)
-
- sub X, X, INCX
-
- srawi. r0, N, 3
- mtspr CTR, r0
- beq- LL(115)
- .align 4
-
- LL(112):
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- bdnz LL(112)
- .align 4
-
- LL(115):
- andi. r0, N, 7
- beq LL(999)
- andi. r0, N, 4
- beq LL(117)
-
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- .align 4
-
- LL(117):
- andi. r0, N, 2
- beq LL(118)
-
- STFDUX A1, X, INCX
- STFDUX A1, X, INCX
- .align 4
-
- LL(118):
- andi. r0, N, 1
- beq LL(999)
- STFDUX A1, X, INCX
- b LL(999)
- .align 4
-
- LL(200):
- sub X2, X, INCX
- sub X, X, INCX
-
- srawi. r0, N, 3
- mtspr CTR, r0
- beq- LL(215)
-
- LFDUX A1, X, INCX
- LFDUX A2, X, INCX
- LFDUX A3, X, INCX
- LFDUX A4, X, INCX
- LFDUX A5, X, INCX
- LFDUX A6, X, INCX
- LFDUX A7, X, INCX
- LFDUX A8, X, INCX
- bdz LL(213)
- .align 4
-
- LL(212):
- fmul B1, ALPHA, A1
- LFDUX A1, X, INCX
- fmul B2, ALPHA, A2
- LFDUX A2, X, INCX
-
- fmul B3, ALPHA, A3
- LFDUX A3, X, INCX
- fmul B4, ALPHA, A4
- LFDUX A4, X, INCX
-
- fmul B5, ALPHA, A5
- LFDUX A5, X, INCX
- fmul B6, ALPHA, A6
- LFDUX A6, X, INCX
-
- fmul B7, ALPHA, A7
- LFDUX A7, X, INCX
- fmul B8, ALPHA, A8
- LFDUX A8, X, INCX
-
- STFDUX B1, X2, INCX
- STFDUX B2, X2, INCX
- STFDUX B3, X2, INCX
- STFDUX B4, X2, INCX
- STFDUX B5, X2, INCX
- STFDUX B6, X2, INCX
- STFDUX B7, X2, INCX
- STFDUX B8, X2, INCX
- bdnz LL(212)
- .align 4
-
- LL(213):
- fmul B1, ALPHA, A1
- fmul B2, ALPHA, A2
- fmul B3, ALPHA, A3
- fmul B4, ALPHA, A4
- fmul B5, ALPHA, A5
-
- fmul B6, ALPHA, A6
- STFDUX B1, X2, INCX
- fmul B7, ALPHA, A7
- STFDUX B2, X2, INCX
- fmul B8, ALPHA, A8
- STFDUX B3, X2, INCX
- STFDUX B4, X2, INCX
- STFDUX B5, X2, INCX
- STFDUX B6, X2, INCX
- STFDUX B7, X2, INCX
- STFDUX B8, X2, INCX
- .align 4
-
- LL(215):
- andi. r0, N, 7
- beq LL(999)
- andi. r0, N, 4
- beq LL(217)
-
- LFDUX A1, X, INCX
- LFDUX A2, X, INCX
- LFDUX A3, X, INCX
- LFDUX A4, X, INCX
-
- fmul B1, ALPHA, A1
- fmul B2, ALPHA, A2
- fmul B3, ALPHA, A3
- fmul B4, ALPHA, A4
-
- STFDUX B1, X2, INCX
- STFDUX B2, X2, INCX
- STFDUX B3, X2, INCX
- STFDUX B4, X2, INCX
- .align 4
-
- LL(217):
- andi. r0, N, 2
- beq LL(218)
-
- LFDUX A1, X, INCX
- LFDUX A2, X, INCX
-
- fmul B1, ALPHA, A1
- fmul B2, ALPHA, A2
-
- STFDUX B1, X2, INCX
- STFDUX B2, X2, INCX
- .align 4
-
- LL(218):
- andi. r0, N, 1
- beq LL(999)
-
- LFDUX A1, X, INCX
- fmul B1, ALPHA, A1
- STFDUX B1, X2, INCX
- .align 4
-
- LL(999):
- li r10, 16
-
- lfpdux f16, SP, r10
- lfpdux f15, SP, r10
- lfpdux f14, SP, r10
-
- addi SP, SP, 16
- blr
-
- EPILOGUE
|