tahoma2d/thirdparty/openblas/xianyi-OpenBLAS-e6e87a2/kernel/alpha/zgemv_t.S
2016-03-24 02:47:04 +09:00

922 lines
15 KiB
ArmAsm

/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#include "version.h"
#define STACKSIZE 64
#define PREFETCHSIZE 32
#define M $16
#define N $17
#define A $21
#define LDA $18
#define X $19
#define INCX $20
#define Y $22
#define INCY $23
#define BUFFER $24
#define I $25
#define J $27
#define X1 $3
#define Y1 $4
#define A1 $5
#define A2 $6
#define alpha_r $f19
#define alpha_i $f20
#define s0 $f0
#define s1 $f1
#define s2 $f10
#define s3 $f11
#define t0 $f12
#define t1 $f13
#define t2 $f14
#define t3 $f15
#define x0 $f16
#define x1 $f17
#define x2 $f18
#define x3 $f21
#define a0 $f22
#define a1 $f23
#define a2 $f24
#define a3 $f25
#define a4 $f26
#define a5 $f27
#define a6 $f28
#define a7 $f29
#define a8 $f2
#define a9 $f3
#define a10 $f4
#define a11 $f5
#define a12 $f6
#define a13 $f7
#define a14 $f8
#define a15 $f9
#if !defined(CONJ) && !defined(XCONJ)
#define ADD1 ADD
#define ADD2 ADD
#define ADD3 SUB
#define ADD4 ADD
#elif !defined(CONJ) && defined(XCONJ)
#define ADD1 ADD
#define ADD2 ADD
#define ADD3 ADD
#define ADD4 SUB
#elif defined(CONJ) && !defined(XCONJ)
#define ADD1 ADD
#define ADD2 SUB
#define ADD3 ADD
#define ADD4 ADD
#else
#define ADD1 ADD
#define ADD2 SUB
#define ADD3 SUB
#define ADD4 SUB
#endif
PROLOGUE
lda $sp, -STACKSIZE($sp)
ldq LDA, 0 + STACKSIZE($sp)
ldq X, 8 + STACKSIZE($sp)
ldq INCX, 16 + STACKSIZE($sp)
ldq Y, 24 + STACKSIZE($sp)
ldq INCY, 32 + STACKSIZE($sp)
ldq BUFFER, 40 + STACKSIZE($sp)
stt $f2, 0($sp)
stt $f3, 8($sp)
stt $f4, 16($sp)
stt $f5, 24($sp)
stt $f6, 32($sp)
stt $f7, 40($sp)
stt $f8, 48($sp)
stt $f9, 56($sp)
PROFCODE
cmple M, 0, $0
sll INCX, ZBASE_SHIFT, INCX
cmple N, 0, $1
sll INCY, ZBASE_SHIFT, INCY
or $0, $1, $0
bne $0, $L999
cmpeq INCX, 2 * SIZE, $0
mov X, X1
sll LDA, ZBASE_SHIFT,LDA
bne $0, $L10
sra M, 2, I
mov BUFFER, Y1
mov BUFFER, X
ble I, $L05
.align 4
$L02:
ldl $31, (PREFETCHSIZE + 0) * SIZE(X1)
lda I, -1(I)
LD a0, 0 * SIZE(X1)
LD a1, 1 * SIZE(X1)
addq X1, INCX, X1
LD a2, 0 * SIZE(X1)
LD a3, 1 * SIZE(X1)
addq X1, INCX, X1
ST a0, 0 * SIZE(Y1)
ST a1, 1 * SIZE(Y1)
ST a2, 2 * SIZE(Y1)
ST a3, 3 * SIZE(Y1)
LD a4, 0 * SIZE(X1)
LD a5, 1 * SIZE(X1)
addq X1, INCX, X1
LD a6, 0 * SIZE(X1)
LD a7, 1 * SIZE(X1)
addq X1, INCX, X1
ST a4, 4 * SIZE(Y1)
ST a5, 5 * SIZE(Y1)
ST a6, 6 * SIZE(Y1)
ST a7, 7 * SIZE(Y1)
lda Y1, 8 * SIZE(Y1)
bgt I, $L02
.align 4
$L05:
and M, 3, I
ble I, $L10
.align 4
$L06:
LD a0, 0 * SIZE(X1)
LD a1, 1 * SIZE(X1)
addq X1, INCX, X1
ST a0, 0 * SIZE(Y1)
ST a1, 1 * SIZE(Y1)
lda Y1, 2 * SIZE(Y1)
lda I, -1(I)
bgt I, $L06
.align 4
$L10:
mov Y, Y1
fclr t0
unop
fclr t1
sra N, 1, J
fclr t2
fclr t3
ble J, $L20
.align 4
$L11:
mov A, A1
fclr s0
addq A, LDA, A2
fclr s1
addq A2, LDA, A
unop
mov X, X1
lds $f31, 3 * SIZE(Y)
sra M, 2, I
fclr s2
fclr s3
ble I, $L15
LD a0, 0 * SIZE(A1)
LD a1, 1 * SIZE(A1)
LD a2, 0 * SIZE(A2)
LD a3, 1 * SIZE(A2)
LD a4, 2 * SIZE(A1)
LD a5, 3 * SIZE(A1)
LD a6, 2 * SIZE(A2)
LD a7, 3 * SIZE(A2)
LD a8, 4 * SIZE(A1)
LD a9, 5 * SIZE(A1)
LD a10, 4 * SIZE(A2)
LD a11, 5 * SIZE(A2)
LD a12, 6 * SIZE(A1)
LD a13, 7 * SIZE(A1)
LD a14, 6 * SIZE(A2)
LD a15, 7 * SIZE(A2)
LD x0, 0 * SIZE(X1)
LD x1, 1 * SIZE(X1)
LD x2, 2 * SIZE(X1)
lda I, -1(I)
ble I, $L13
.align 4
$L12:
ADD3 s0, t0, s0
unop
MUL x0, a0, t0
LD x3, 3 * SIZE(X1)
ADD4 s1, t1, s1
ldl $31, (PREFETCHSIZE + 0) * SIZE(A1)
MUL x0, a1, t1
unop
ADD3 s2, t2, s2
unop
MUL x0, a2, t2
unop
ADD4 s3, t3, s3
unop
MUL x0, a3, t3
LD x0, 4 * SIZE(X1)
ADD1 s0, t0, s0
unop
MUL x1, a1, t0
LD a1, 9 * SIZE(A1)
ADD2 s1, t1, s1
unop
MUL x1, a0, t1
LD a0, 8 * SIZE(A1)
ADD1 s2, t2, s2
unop
MUL x1, a3, t2
LD a3, 9 * SIZE(A2)
ADD2 s3, t3, s3
unop
MUL x1, a2, t3
LD a2, 8 * SIZE(A2)
ADD3 s0, t0, s0
unop
MUL x2, a4, t0
LD x1, 5 * SIZE(X1)
ADD4 s1, t1, s1
MUL x2, a5, t1
ADD3 s2, t2, s2
MUL x2, a6, t2
ADD4 s3, t3, s3
unop
MUL x2, a7, t3
LD x2, 6 * SIZE(X1)
ADD1 s0, t0, s0
unop
MUL x3, a5, t0
LD a5, 11 * SIZE(A1)
ADD2 s1, t1, s1
unop
MUL x3, a4, t1
LD a4, 10 * SIZE(A1)
ADD1 s2, t2, s2
unop
MUL x3, a7, t2
LD a7, 11 * SIZE(A2)
ADD2 s3, t3, s3
unop
MUL x3, a6, t3
LD a6, 10 * SIZE(A2)
ADD3 s0, t0, s0
unop
MUL x0, a8, t0
LD x3, 7 * SIZE(X1)
ADD4 s1, t1, s1
ldl $31, (PREFETCHSIZE + 0) * SIZE(A2)
MUL x0, a9, t1
unop
ADD3 s2, t2, s2
lda I, -1(I)
MUL x0, a10, t2
unop
ADD4 s3, t3, s3
unop
MUL x0, a11, t3
LD x0, 8 * SIZE(X1)
ADD1 s0, t0, s0
unop
MUL x1, a9, t0
LD a9, 13 * SIZE(A1)
ADD2 s1, t1, s1
unop
MUL x1, a8, t1
LD a8, 12 * SIZE(A1)
ADD1 s2, t2, s2
lda A1, 8 * SIZE(A1)
MUL x1, a11, t2
LD a11, 13 * SIZE(A2)
ADD2 s3, t3, s3
unop
MUL x1, a10, t3
LD a10, 12 * SIZE(A2)
ADD3 s0, t0, s0
unop
MUL x2, a12, t0
LD x1, 9 * SIZE(X1)
ADD4 s1, t1, s1
ldl $31, (PREFETCHSIZE + 0) * SIZE(X1)
MUL x2, a13, t1
lda A2, 8 * SIZE(A2)
ADD3 s2, t2, s2
unop
MUL x2, a14, t2
unop
ADD4 s3, t3, s3
unop
MUL x2, a15, t3
LD x2, 10 * SIZE(X1)
ADD1 s0, t0, s0
unop
MUL x3, a13, t0
LD a13, 7 * SIZE(A1)
ADD2 s1, t1, s1
lda X1, 8 * SIZE(X1)
MUL x3, a12, t1
LD a12, 6 * SIZE(A1)
ADD1 s2, t2, s2
unop
MUL x3, a15, t2
LD a15, 7 * SIZE(A2)
ADD2 s3, t3, s3
MUL x3, a14, t3
LD a14, 6 * SIZE(A2)
bgt I, $L12
.align 4
$L13:
ADD3 s0, t0, s0
unop
MUL x0, a0, t0
LD x3, 3 * SIZE(X1)
ADD4 s1, t1, s1
MUL x0, a1, t1
ADD3 s2, t2, s2
MUL x0, a2, t2
ADD4 s3, t3, s3
unop
MUL x0, a3, t3
LD x0, 4 * SIZE(X1)
ADD1 s0, t0, s0
MUL x1, a1, t0
ADD2 s1, t1, s1
MUL x1, a0, t1
ADD1 s2, t2, s2
unop
MUL x1, a3, t2
unop
ADD2 s3, t3, s3
lda A1, 8 * SIZE(A1)
MUL x1, a2, t3
LD x1, 5 * SIZE(X1)
ADD3 s0, t0, s0
MUL x2, a4, t0
ADD4 s1, t1, s1
MUL x2, a5, t1
ADD3 s2, t2, s2
unop
MUL x2, a6, t2
unop
ADD4 s3, t3, s3
lda A2, 8 * SIZE(A2)
MUL x2, a7, t3
LD x2, 6 * SIZE(X1)
ADD1 s0, t0, s0
MUL x3, a5, t0
ADD2 s1, t1, s1
MUL x3, a4, t1
ADD1 s2, t2, s2
unop
MUL x3, a7, t2
lda X1, 8 * SIZE(X1)
ADD2 s3, t3, s3
unop
MUL x3, a6, t3
LD x3, -1 * SIZE(X1)
ADD3 s0, t0, s0
MUL x0, a8, t0
ADD4 s1, t1, s1
MUL x0, a9, t1
ADD3 s2, t2, s2
MUL x0, a10, t2
ADD4 s3, t3, s3
MUL x0, a11, t3
ADD1 s0, t0, s0
MUL x1, a9, t0
ADD2 s1, t1, s1
MUL x1, a8, t1
ADD1 s2, t2, s2
MUL x1, a11, t2
ADD2 s3, t3, s3
MUL x1, a10, t3
ADD3 s0, t0, s0
MUL x2, a12, t0
ADD4 s1, t1, s1
MUL x2, a13, t1
ADD3 s2, t2, s2
MUL x2, a14, t2
ADD4 s3, t3, s3
MUL x2, a15, t3
ADD1 s0, t0, s0
MUL x3, a13, t0
ADD2 s1, t1, s1
MUL x3, a12, t1
ADD1 s2, t2, s2
MUL x3, a15, t2
ADD2 s3, t3, s3
MUL x3, a14, t3
.align 4
$L15:
and M, 3, I
ble I, $L18
LD a0, 0 * SIZE(A1)
LD a1, 1 * SIZE(A1)
LD a2, 0 * SIZE(A2)
LD a3, 1 * SIZE(A2)
LD x0, 0 * SIZE(X1)
lda I, -1(I)
ble I, $L17
.align 4
$L16:
ADD3 s0, t0, s0
lda I, -1(I)
MUL x0, a0, t0
LD x1, 1 * SIZE(X1)
ADD4 s1, t1, s1
MUL x0, a1, t1
ADD3 s2, t2, s2
MUL x0, a2, t2
ADD4 s3, t3, s3
unop
MUL x0, a3, t3
LD x0, 2 * SIZE(X1)
ADD1 s0, t0, s0
lda A2, 2 * SIZE(A2)
MUL x1, a1, t0
LD a1, 3 * SIZE(A1)
ADD2 s1, t1, s1
lda X1, 2 * SIZE(X1)
MUL x1, a0, t1
LD a0, 2 * SIZE(A1)
ADD1 s2, t2, s2
lda A1, 2 * SIZE(A1)
MUL x1, a3, t2
LD a3, 1 * SIZE(A2)
ADD2 s3, t3, s3
MUL x1, a2, t3
LD a2, 0 * SIZE(A2)
bgt I, $L16
.align 4
$L17:
ADD3 s0, t0, s0
unop
MUL x0, a0, t0
LD x1, 1 * SIZE(X1)
ADD4 s1, t1, s1
unop
MUL x0, a1, t1
unop
ADD3 s2, t2, s2
MUL x0, a2, t2
ADD4 s3, t3, s3
MUL x0, a3, t3
ADD1 s0, t0, s0
MUL x1, a1, t0
ADD2 s1, t1, s1
MUL x1, a0, t1
ADD1 s2, t2, s2
MUL x1, a3, t2
ADD2 s3, t3, s3
MUL x1, a2, t3
.align 4
$L18:
LD a0, 0 * SIZE(Y)
unop
LD a1, 1 * SIZE(Y)
addq Y, INCY, Y
LD a2, 0 * SIZE(Y)
unop
LD a3, 1 * SIZE(Y)
addq Y, INCY, Y
ADD3 s0, t0, s0
ADD4 s1, t1, s1
ADD3 s2, t2, s2
ADD4 s3, t3, s3
MUL alpha_r, s0, t0
MUL alpha_r, s1, t1
MUL alpha_r, s2, t2
MUL alpha_r, s3, t3
ADD a0, t0, a0
MUL alpha_i, s1, t0
ADD a1, t1, a1
MUL alpha_i, s0, t1
ADD a2, t2, a2
MUL alpha_i, s3, t2
ADD a3, t3, a3
MUL alpha_i, s2, t3
SUB a0, t0, a0
ADD a1, t1, a1
SUB a2, t2, a2
ADD a3, t3, a3
ST a0, 0 * SIZE(Y1)
fclr t0
ST a1, 1 * SIZE(Y1)
addq Y1, INCY, Y1
ST a2, 0 * SIZE(Y1)
fclr t1
ST a3, 1 * SIZE(Y1)
addq Y1, INCY, Y1
fclr t2
lda J, -1(J)
fclr t3
bgt J, $L11
.align 4
$L20:
blbc N, $L999
mov A, A1
fclr s0
fclr s1
mov X, X1
sra M, 2, I
fclr s2
fclr s3
ble I, $L25
LD a0, 0 * SIZE(A1)
LD a1, 1 * SIZE(A1)
LD a4, 2 * SIZE(A1)
LD a5, 3 * SIZE(A1)
LD a8, 4 * SIZE(A1)
LD a9, 5 * SIZE(A1)
LD a12, 6 * SIZE(A1)
LD a13, 7 * SIZE(A1)
LD x0, 0 * SIZE(X1)
LD x1, 1 * SIZE(X1)
LD x2, 2 * SIZE(X1)
lda I, -1(I)
ble I, $L23
.align 4
$L22:
ADD3 s0, t0, s0
ldl $31, (PREFETCHSIZE + 0) * SIZE(A1)
MUL x0, a0, t0
LD x3, 3 * SIZE(X1)
ADD4 s1, t1, s1
unop
MUL x0, a1, t1
LD x0, 4 * SIZE(X1)
ADD1 s2, t0, s2
lda I, -1(I)
MUL x1, a1, t0
LD a1, 9 * SIZE(A1)
ADD2 s3, t1, s3
unop
MUL x1, a0, t1
LD a0, 8 * SIZE(A1)
ADD3 s0, t0, s0
unop
MUL x2, a4, t0
LD x1, 5 * SIZE(X1)
ADD4 s1, t1, s1
unop
MUL x2, a5, t1
LD x2, 6 * SIZE(X1)
ADD1 s2, t0, s2
unop
MUL x3, a5, t0
LD a5, 11 * SIZE(A1)
ADD2 s3, t1, s3
unop
MUL x3, a4, t1
LD a4, 10 * SIZE(A1)
ADD3 s0, t0, s0
unop
MUL x0, a8, t0
LD x3, 7 * SIZE(X1)
ADD4 s1, t1, s1
unop
MUL x0, a9, t1
LD x0, 8 * SIZE(X1)
ADD1 s2, t0, s2
unop
MUL x1, a9, t0
LD a9, 13 * SIZE(A1)
ADD2 s3, t1, s3
unop
MUL x1, a8, t1
LD a8, 12 * SIZE(A1)
ADD3 s0, t0, s0
unop
MUL x2, a12, t0
LD x1, 9 * SIZE(X1)
ADD4 s1, t1, s1
lda A1, 8 * SIZE(A1)
MUL x2, a13, t1
LD x2, 10 * SIZE(X1)
ADD1 s2, t0, s2
lda X1, 8 * SIZE(X1)
MUL x3, a13, t0
LD a13, 7 * SIZE(A1)
ADD2 s3, t1, s3
MUL x3, a12, t1
LD a12, 6 * SIZE(A1)
bgt I, $L22
.align 4
$L23:
ADD3 s0, t0, s0
unop
MUL x0, a0, t0
LD x3, 3 * SIZE(X1)
ADD4 s1, t1, s1
unop
MUL x0, a1, t1
LD x0, 4 * SIZE(X1)
ADD1 s2, t0, s2
unop
MUL x1, a1, t0
lda A1, 8 * SIZE(A1)
ADD2 s3, t1, s3
unop
MUL x1, a0, t1
LD x1, 5 * SIZE(X1)
ADD3 s0, t0, s0
unop
MUL x2, a4, t0
unop
ADD4 s1, t1, s1
unop
MUL x2, a5, t1
LD x2, 6 * SIZE(X1)
ADD1 s2, t0, s2
unop
MUL x3, a5, t0
lda X1, 8 * SIZE(X1)
ADD2 s3, t1, s3
unop
MUL x3, a4, t1
LD x3, -1 * SIZE(X1)
ADD3 s0, t0, s0
MUL x0, a8, t0
ADD4 s1, t1, s1
MUL x0, a9, t1
ADD1 s2, t0, s2
MUL x1, a9, t0
ADD2 s3, t1, s3
MUL x1, a8, t1
ADD3 s0, t0, s0
MUL x2, a12, t0
ADD4 s1, t1, s1
MUL x2, a13, t1
ADD1 s2, t0, s2
MUL x3, a13, t0
ADD2 s3, t1, s3
MUL x3, a12, t1
.align 4
$L25:
and M, 3, I
ble I, $L28
LD a0, 0 * SIZE(A1)
LD a1, 1 * SIZE(A1)
LD x0, 0 * SIZE(X1)
lda I, -1(I)
ble I, $L27
.align 4
$L26:
ADD3 s0, t0, s0
lda A1, 2 * SIZE(A1)
MUL x0, a0, t0
LD x1, 1 * SIZE(X1)
ADD4 s1, t1, s1
lda I, -1(I)
MUL x0, a1, t1
LD x0, 2 * SIZE(X1)
ADD1 s0, t0, s0
lda X1, 2 * SIZE(X1)
MUL x1, a1, t0
LD a1, 1 * SIZE(A1)
ADD2 s1, t1, s1
MUL x1, a0, t1
LD a0, 0 * SIZE(A1)
bgt I, $L26
.align 4
$L27:
ADD3 s0, t0, s0
unop
MUL x0, a0, t0
LD x1, 1 * SIZE(X1)
ADD4 s1, t1, s1
unop
MUL x0, a1, t1
unop
ADD1 s0, t0, s0
MUL x1, a1, t0
ADD2 s1, t1, s1
MUL x1, a0, t1
.align 4
$L28:
LD a0, 0 * SIZE(Y)
LD a1, 1 * SIZE(Y)
ADD3 s0, t0, s0
ADD4 s1, t1, s1
ADD3 s2, t2, s2
ADD4 s3, t3, s3
ADD s0, s2, s0
ADD s1, s3, s1
MUL alpha_r, s0, t0
MUL alpha_r, s1, t1
ADD a0, t0, a0
MUL alpha_i, s1, t0
ADD a1, t1, a1
MUL alpha_i, s0, t1
SUB a0, t0, a0
ADD a1, t1, a1
ST a0, 0 * SIZE(Y1)
ST a1, 1 * SIZE(Y1)
.align 4
$L999:
ldt $f2, 0($sp)
ldt $f3, 8($sp)
ldt $f4, 16($sp)
ldt $f5, 24($sp)
ldt $f6, 32($sp)
ldt $f7, 40($sp)
ldt $f8, 48($sp)
ldt $f9, 56($sp)
lda $sp, STACKSIZE($sp)
ret
EPILOGUE