tahoma2d/thirdparty/openblas/xianyi-OpenBLAS-e6e87a2/kernel/mips64/zaxpy.S
2016-03-24 02:47:04 +09:00

438 lines
8.9 KiB
ArmAsm

/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#define N $4
#define X $9
#define INCX $10
#define Y $11
#define INCY $8
#define I $2
#define TEMP $3
#define YY $5
#define ALPHA_R $f15
#define ALPHA_I $f16
#define a1 $f0
#define a2 $f1
#define a3 $f2
#define a4 $f3
#define a5 $f4
#define a6 $f5
#define a7 $f6
#define a8 $f7
#define b1 $f8
#define b2 $f9
#define b3 $f10
#define b4 $f11
#define b5 $f12
#define b6 $f13
#define b7 $f14
#define b8 $f17
#define t1 $f18
#define t2 $f19
#define t3 $f20
#define t4 $f21
#ifndef CONJ
#define MADD1 NMSUB
#define MADD2 MADD
#else
#define MADD1 MADD
#define MADD2 NMSUB
#endif
PROLOGUE
LDARG INCY, 0($sp)
li TEMP, 2 * SIZE
#ifndef __64BIT__
daddiu $sp, $sp, -16
sdc1 $f20, 0($sp)
sdc1 $f21, 8($sp)
#endif
blez N, .L999
dsll INCX, INCX, ZBASE_SHIFT
bne INCX, TEMP, .L20
dsll INCY, INCY, ZBASE_SHIFT
bne INCY, TEMP, .L20
dsra I, N, 2
blez I, .L15
daddiu I, I, -1
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)
LD a2, 1 * SIZE(X)
LD b2, 1 * SIZE(Y)
LD a3, 2 * SIZE(X)
LD b3, 2 * SIZE(Y)
LD a4, 3 * SIZE(X)
LD b4, 3 * SIZE(Y)
LD a5, 4 * SIZE(X)
LD b5, 4 * SIZE(Y)
LD a6, 5 * SIZE(X)
LD b6, 5 * SIZE(Y)
LD a7, 6 * SIZE(X)
LD b7, 6 * SIZE(Y)
LD a8, 7 * SIZE(X)
LD b8, 7 * SIZE(Y)
blez I, .L13
NOP
.align 3
.L12:
MADD t1, b1, ALPHA_R, a1
LD b1, 8 * SIZE(Y)
MADD t2, b2, ALPHA_I, a1
LD a1, 8 * SIZE(X)
MADD t3, b3, ALPHA_R, a3
LD b3, 10 * SIZE(Y)
MADD t4, b4, ALPHA_I, a3
LD a3, 10 * SIZE(X)
MADD1 t1, t1, ALPHA_I, a2
LD b2, 9 * SIZE(Y)
MADD2 t2, t2, ALPHA_R, a2
LD a2, 9 * SIZE(X)
MADD1 t3, t3, ALPHA_I, a4
LD b4, 11 * SIZE(Y)
MADD2 t4, t4, ALPHA_R, a4
LD a4, 11 * SIZE(X)
ST t1, 0 * SIZE(Y)
ST t2, 1 * SIZE(Y)
ST t3, 2 * SIZE(Y)
ST t4, 3 * SIZE(Y)
MADD t1, b5, ALPHA_R, a5
LD b5, 12 * SIZE(Y)
MADD t2, b6, ALPHA_I, a5
LD a5, 12 * SIZE(X)
MADD t3, b7, ALPHA_R, a7
LD b7, 14 * SIZE(Y)
MADD t4, b8, ALPHA_I, a7
LD a7, 14 * SIZE(X)
MADD1 t1, t1, ALPHA_I, a6
LD b6, 13 * SIZE(Y)
MADD2 t2, t2, ALPHA_R, a6
LD a6, 13 * SIZE(X)
MADD1 t3, t3, ALPHA_I, a8
LD b8, 15 * SIZE(Y)
MADD2 t4, t4, ALPHA_R, a8
LD a8, 15 * SIZE(X)
ST t1, 4 * SIZE(Y)
ST t2, 5 * SIZE(Y)
ST t3, 6 * SIZE(Y)
ST t4, 7 * SIZE(Y)
daddiu I, I, -1
daddiu Y, Y, 8 * SIZE
bgtz I, .L12
daddiu X, X, 8 * SIZE
.align 3
.L13:
MADD t1, b1, ALPHA_R, a1
MADD t2, b2, ALPHA_I, a1
MADD t3, b3, ALPHA_R, a3
MADD t4, b4, ALPHA_I, a3
MADD1 t1, t1, ALPHA_I, a2
MADD2 t2, t2, ALPHA_R, a2
MADD1 t3, t3, ALPHA_I, a4
MADD2 t4, t4, ALPHA_R, a4
ST t1, 0 * SIZE(Y)
MADD t1, b5, ALPHA_R, a5
ST t2, 1 * SIZE(Y)
MADD t2, b6, ALPHA_I, a5
ST t3, 2 * SIZE(Y)
MADD t3, b7, ALPHA_R, a7
ST t4, 3 * SIZE(Y)
MADD t4, b8, ALPHA_I, a7
MADD1 t1, t1, ALPHA_I, a6
MADD2 t2, t2, ALPHA_R, a6
MADD1 t3, t3, ALPHA_I, a8
MADD2 t4, t4, ALPHA_R, a8
ST t1, 4 * SIZE(Y)
ST t2, 5 * SIZE(Y)
ST t3, 6 * SIZE(Y)
ST t4, 7 * SIZE(Y)
daddiu X, X, 8 * SIZE
daddiu Y, Y, 8 * SIZE
.align 3
.L15:
andi I, N, 3
blez I, .L999
NOP
.align 3
.L16:
LD a1, 0 * SIZE(X)
LD a2, 1 * SIZE(X)
LD b1, 0 * SIZE(Y)
LD b2, 1 * SIZE(Y)
MADD t1, b1, ALPHA_R, a1
daddiu X, X, 2 * SIZE
MADD t2, b2, ALPHA_I, a1
MADD1 t1, t1, ALPHA_I, a2
daddiu I, I, -1
MADD2 t2, t2, ALPHA_R, a2
daddiu Y, Y, 2 * SIZE
ST t1, -2 * SIZE(Y)
bgtz I, .L16
ST t2, -1 * SIZE(Y)
#ifndef __64BIT__
ldc1 $f20, 0($sp)
ldc1 $f21, 8($sp)
daddiu $sp, $sp, 16
#endif
j $31
NOP
.align 3
.L20:
dsra I, N, 2
move YY, Y
blez I, .L25
daddiu I, I, -1
LD a1, 0 * SIZE(X)
LD b1, 0 * SIZE(Y)
LD a2, 1 * SIZE(X)
LD b2, 1 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY
LD a3, 0 * SIZE(X)
LD b3, 0 * SIZE(Y)
LD a4, 1 * SIZE(X)
LD b4, 1 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY
LD a5, 0 * SIZE(X)
LD b5, 0 * SIZE(Y)
LD a6, 1 * SIZE(X)
LD b6, 1 * SIZE(Y)
daddu X, X, INCX
daddu Y, Y, INCY
LD a7, 0 * SIZE(X)
blez I, .L23
LD b7, 0 * SIZE(Y)
.align 3
.L22:
MADD t1, b1, ALPHA_R, a1
LD b8, 1 * SIZE(Y)
daddu Y, Y, INCY
MADD t2, b2, ALPHA_I, a1
LD a8, 1 * SIZE(X)
daddu X, X, INCX
MADD t3, b3, ALPHA_R, a3
LD b1, 0 * SIZE(Y)
MADD t4, b4, ALPHA_I, a3
LD a1, 0 * SIZE(X)
MADD1 t1, t1, ALPHA_I, a2
LD b2, 1 * SIZE(Y)
daddu Y, Y, INCY
MADD2 t2, t2, ALPHA_R, a2
LD a2, 1 * SIZE(X)
daddu X, X, INCX
MADD1 t3, t3, ALPHA_I, a4
LD a3, 0 * SIZE(X)
MADD2 t4, t4, ALPHA_R, a4
LD b3, 0 * SIZE(Y)
ST t1, 0 * SIZE(YY)
ST t2, 1 * SIZE(YY)
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
ST t4, 1 * SIZE(YY)
daddu YY, YY, INCY
MADD t1, b5, ALPHA_R, a5
LD a4, 1 * SIZE(X)
daddu X, X, INCX
MADD t2, b6, ALPHA_I, a5
LD b4, 1 * SIZE(Y)
daddu Y, Y, INCY
MADD t3, b7, ALPHA_R, a7
LD b5, 0 * SIZE(Y)
MADD t4, b8, ALPHA_I, a7
LD a5, 0 * SIZE(X)
MADD1 t1, t1, ALPHA_I, a6
LD b6, 1 * SIZE(Y)
daddu Y, Y, INCY
MADD2 t2, t2, ALPHA_R, a6
LD a6, 1 * SIZE(X)
daddu X, X, INCX
MADD1 t3, t3, ALPHA_I, a8
LD b7, 0 * SIZE(Y)
MADD2 t4, t4, ALPHA_R, a8
LD a7, 0 * SIZE(X)
ST t1, 0 * SIZE(YY)
ST t2, 1 * SIZE(YY)
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
ST t4, 1 * SIZE(YY)
daddu YY, YY, INCY
daddiu I, I, -1
bgtz I, .L22
NOP
.align 3
.L23:
MADD t1, b1, ALPHA_R, a1
LD a8, 1 * SIZE(X)
MADD t2, b2, ALPHA_I, a1
LD b8, 1 * SIZE(Y)
MADD t3, b3, ALPHA_R, a3
daddu X, X, INCX
MADD t4, b4, ALPHA_I, a3
daddu Y, Y, INCY
MADD1 t1, t1, ALPHA_I, a2
MADD2 t2, t2, ALPHA_R, a2
MADD1 t3, t3, ALPHA_I, a4
MADD2 t4, t4, ALPHA_R, a4
ST t1, 0 * SIZE(YY)
MADD t1, b5, ALPHA_R, a5
ST t2, 1 * SIZE(YY)
MADD t2, b6, ALPHA_I, a5
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
MADD t3, b7, ALPHA_R, a7
ST t4, 1 * SIZE(YY)
MADD t4, b8, ALPHA_I, a7
daddu YY, YY, INCY
MADD1 t1, t1, ALPHA_I, a6
MADD2 t2, t2, ALPHA_R, a6
MADD1 t3, t3, ALPHA_I, a8
MADD2 t4, t4, ALPHA_R, a8
ST t1, 0 * SIZE(YY)
ST t2, 1 * SIZE(YY)
daddu YY, YY, INCY
ST t3, 0 * SIZE(YY)
ST t4, 1 * SIZE(YY)
daddu YY, YY, INCY
.align 3
.L25:
andi I, N, 3
blez I, .L999
NOP
.align 3
.L26:
LD a1, 0 * SIZE(X)
LD a2, 1 * SIZE(X)
LD b1, 0 * SIZE(Y)
LD b2, 1 * SIZE(Y)
MADD t1, b1, ALPHA_R, a1
MADD t2, b2, ALPHA_I, a1
daddu X, X, INCX
MADD1 t1, t1, ALPHA_I, a2
MADD2 t2, t2, ALPHA_R, a2
daddiu I, I, -1
ST t1, 0 * SIZE(Y)
ST t2, 1 * SIZE(Y)
bgtz I, .L26
daddu Y, Y, INCY
.align 3
.L999:
#ifndef __64BIT__
ldc1 $f20, 0($sp)
ldc1 $f21, 8($sp)
daddiu $sp, $sp, 16
#endif
j $31
NOP
EPILOGUE