tahoma2d/thirdparty/openblas/xianyi-OpenBLAS-e6e87a2/kernel/x86_64/xgemm3m_kernel_2x2.S
2016-03-24 02:47:04 +09:00

877 lines
14 KiB
ArmAsm

/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#define M ARG1
#define N ARG2
#define K ARG3
#define A ARG4
#define B ARG5
#define C ARG6
#define LDC %r10
#define I %r12
#define J %r13
#define AO %r14
#define BO %r15
#define CO %rbp
#define KK %r11
#define KKK 48(%rsp)
#define STACKSIZE 64
#define ALPHA_R 8 + STACKSIZE(%rsp)
#define ALPHA_I 24 + STACKSIZE(%rsp)
#define OFFSET 48 + STACKSIZE(%rsp)
#ifdef OPTERON
#define PREFETCH prefetch
#define PREFETCHW prefetchw
#else
#define PREFETCH prefetcht0
#define PREFETCHW prefetcht0
#endif
#define PREFETCHSIZE (5 + 4 * 10)
#if defined(OS_LINUX) && defined(CORE_BARCELONA)
.align 32768
#endif
PROLOGUE
PROFCODE
subq $STACKSIZE, %rsp
movq %rbx, 0(%rsp)
movq %rbp, 8(%rsp)
movq %r12, 16(%rsp)
movq %r13, 24(%rsp)
movq %r14, 32(%rsp)
movq %r15, 40(%rsp)
movq 40 + STACKSIZE(%rsp), LDC
#if defined(TRMMKERNEL) && !defined(LEFT)
movq OFFSET, %rax
negq %rax
movq %rax, KK
#endif
addq $8 * SIZE, A
addq $8 * SIZE, B
salq $ZBASE_SHIFT, LDC
movq N, %rax
sarq $1, %rax
movq %rax, J
je .L30
ALIGN_4
.L01:
#if defined(TRMMKERNEL) && defined(LEFT)
movq OFFSET, %rax
movq %rax, KK
#endif
movq A, AO
movq C, CO
leaq (, LDC, 2), %rax
addq %rax, C
movq M, I
sarq $1, I
je .L20
ALIGN_4
.L11:
#if !defined(TRMMKERNEL) || \
(defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq B, BO
#else
movq KK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 2), AO
leaq (B, %rax, 2), BO
#endif
fldz
fldz
fldz
fldz
#if defined(HAVE_3DNOW)
prefetchw 2 * SIZE(CO)
prefetchw 2 * SIZE(CO, LDC, 1)
#elif defined(HAVE_SSE)
prefetchnta 2 * SIZE(CO)
prefetchnta 2 * SIZE(CO, LDC, 1)
#endif
#ifndef TRMMKERNEL
movq K, %rax
#elif (defined(LEFT) && !defined(TRANSA)) || (!defined(LEFT) && defined(TRANSA))
movq K, %rax
subq KK, %rax
movq %rax, KKK
#else
movq KK, %rax
#ifdef LEFT
addq $2, %rax
#else
addq $2, %rax
#endif
movq %rax, KKK
#endif
sarq $2, %rax
je .L15
ALIGN_4
.L12:
PREFETCH (PREFETCHSIZE + 0) * SIZE(AO)
FLD -8 * SIZE(AO)
FLD -8 * SIZE(BO)
fld %st(1)
fmul %st(1), %st
faddp %st, %st(3)
FLD -7 * SIZE(BO)
fmul %st, %st(2)
FLD -7 * SIZE(AO)
fmul %st, %st(2)
fmulp %st, %st(1)
faddp %st, %st(6)
faddp %st, %st(3)
faddp %st, %st(3)
FLD -6 * SIZE(AO)
FLD -6 * SIZE(BO)
fld %st(1)
fmul %st(1), %st
faddp %st, %st(3)
FLD -5 * SIZE(BO)
fmul %st, %st(2)
FLD -5 * SIZE(AO)
fmul %st, %st(2)
fmulp %st, %st(1)
faddp %st, %st(6)
faddp %st, %st(3)
faddp %st, %st(3)
PREFETCH (PREFETCHSIZE + 4) * SIZE(AO)
FLD -4 * SIZE(AO)
FLD -4 * SIZE(BO)
fld %st(1)
fmul %st(1), %st
faddp %st, %st(3)
FLD -3 * SIZE(BO)
fmul %st, %st(2)
FLD -3 * SIZE(AO)
fmul %st, %st(2)
fmulp %st, %st(1)
faddp %st, %st(6)
faddp %st, %st(3)
faddp %st, %st(3)
FLD -2 * SIZE(AO)
FLD -2 * SIZE(BO)
fld %st(1)
fmul %st(1), %st
faddp %st, %st(3)
FLD -1 * SIZE(BO)
fmul %st, %st(2)
FLD -1 * SIZE(AO)
fmul %st, %st(2)
fmulp %st, %st(1)
faddp %st, %st(6)
faddp %st, %st(3)
faddp %st, %st(3)
addq $8 * SIZE,AO
addq $8 * SIZE,BO
decq %rax
jne .L12
ALIGN_4
.L15:
#ifndef TRMMKERNEL
movq K, %rax
#else
movq KKK, %rax
#endif
and $3, %rax
je .L18
ALIGN_4
.L16:
FLD -8 * SIZE(AO)
FLD -8 * SIZE(BO)
fld %st(1)
fmul %st(1), %st
faddp %st, %st(3)
FLD -7 * SIZE(BO)
fmul %st, %st(2)
FLD -7 * SIZE(AO)
fmul %st, %st(2)
fmulp %st, %st(1)
faddp %st, %st(6)
faddp %st, %st(3)
faddp %st, %st(3)
addq $2 * SIZE,AO
addq $2 * SIZE,BO
decq %rax
jne .L16
ALIGN_4
.L18:
#ifndef TRMMKERNEL
FLD ALPHA_I
FLD ALPHA_R
fld %st(2)
fmul %st(1), %st
FLD 0 * SIZE(CO)
faddp %st, %st(1)
FST 0 * SIZE(CO)
fld %st(3)
fmul %st(1), %st
FLD 2 * SIZE(CO)
faddp %st, %st(1)
FST 2 * SIZE(CO)
fld %st(4)
fmul %st(1), %st
FLD 0 * SIZE(CO, LDC)
faddp %st, %st(1)
FST 0 * SIZE(CO, LDC)
fmul %st(5), %st
FLD 2 * SIZE(CO, LDC)
faddp %st, %st(1)
FST 2 * SIZE(CO, LDC)
fmul %st, %st(1)
fmul %st, %st(2)
fmul %st, %st(3)
fmulp %st, %st(4)
FLD 1 * SIZE(CO)
faddp %st, %st(1)
FST 1 * SIZE(CO)
FLD 3 * SIZE(CO)
faddp %st, %st(1)
FST 3 * SIZE(CO)
FLD 1 * SIZE(CO, LDC)
faddp %st, %st(1)
FST 1 * SIZE(CO, LDC)
FLD 3 * SIZE(CO, LDC)
faddp %st, %st(1)
FST 3 * SIZE(CO, LDC)
#else
FST 0 * SIZE(CO)
FST 1 * SIZE(CO)
FST 0 * SIZE(CO, LDC)
FST 1 * SIZE(CO, LDC)
#endif
#if (defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq K, %rax
subq KKK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 2), AO
leaq (BO, %rax, 2), BO
#endif
#if defined(TRMMKERNEL) && defined(LEFT)
addq $2, KK
#endif
addq $4 * SIZE, CO
decq I
jne .L11
ALIGN_4
.L20:
movq M, %rax
andq $1, %rax
je .L29
ALIGN_4
.L21:
#if !defined(TRMMKERNEL) || \
(defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq B, BO
#else
movq KK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 1), AO
leaq ( B, %rax, 2), BO
#endif
fldz
fldz
#ifndef TRMMKERNEL
movq K, %rax
#elif (defined(LEFT) && !defined(TRANSA)) || (!defined(LEFT) && defined(TRANSA))
movq K, %rax
subq KK, %rax
movq %rax, KKK
#else
movq KK, %rax
#ifdef LEFT
addq $1, %rax
#else
addq $2, %rax
#endif
movq %rax, KKK
#endif
sarq $2, %rax
je .L25
ALIGN_4
.L22:
PREFETCH (PREFETCHSIZE + 0) * SIZE(AO)
FLD -8 * SIZE(AO)
FLD -8 * SIZE(BO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -7 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(2)
FLD -7 * SIZE(AO)
FLD -6 * SIZE(BO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -5 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(2)
FLD -6 * SIZE(AO)
FLD -4 * SIZE(BO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -3 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(2)
FLD -5 * SIZE(AO)
FLD -2 * SIZE(BO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -1 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(2)
addq $4 * SIZE,AO
addq $8 * SIZE,BO
decq %rax
jne .L22
ALIGN_4
.L25:
#ifndef TRMMKERNEL
movq K, %rax
#else
movq KKK, %rax
#endif
and $3, %rax
je .L28
ALIGN_4
.L26:
FLD -8 * SIZE(AO)
FLD -8 * SIZE(BO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -7 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(2)
addq $1 * SIZE,AO
addq $2 * SIZE,BO
decq %rax
jne .L26
ALIGN_4
.L28:
#ifndef TRMMKERNEL
FLD ALPHA_I
FLD ALPHA_R
fld %st(2)
fmul %st(1), %st
FLD 0 * SIZE(CO)
faddp %st, %st(1)
FST 0 * SIZE(CO)
fmul %st(3), %st
FLD 0 * SIZE(CO, LDC)
faddp %st, %st(1)
FST 0 * SIZE(CO, LDC)
fmul %st, %st(1)
fmulp %st, %st(2)
FLD 1 * SIZE(CO)
faddp %st, %st(1)
FST 1 * SIZE(CO)
FLD 1 * SIZE(CO, LDC)
faddp %st, %st(1)
FST 1 * SIZE(CO, LDC)
#else
FST 0 * SIZE(CO)
FST 0 * SIZE(CO, LDC)
#endif
#if (defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq K, %rax
subq KKK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 1), AO
leaq (BO, %rax, 2), BO
#endif
#if defined(TRMMKERNEL) && defined(LEFT)
addq $1, KK
#endif
addq $1 * SIZE, CO
ALIGN_4
.L29:
#if defined(TRMMKERNEL) && !defined(LEFT)
addq $2, KK
#endif
movq BO, B
decq J
jne .L01
ALIGN_4
.L30:
movq N, %rax
testq $1, %rax
je .L999
#if defined(TRMMKERNEL) && defined(LEFT)
movq OFFSET, %rax
movq %rax, KK
#endif
movq A, AO
movq C, CO
addq LDC, C
movq M, I
sarq $1, I
je .L40
ALIGN_4
.L31:
#if !defined(TRMMKERNEL) || \
(defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq B, BO
#else
movq KK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 2), AO
leaq ( B, %rax, 1), BO
#endif
fldz
fldz
#if defined(HAVE_3DNOW)
prefetchw 2 * SIZE(CO)
#elif defined(HAVE_SSE)
prefetchnta 2 * SIZE(CO)
#endif
#ifndef TRMMKERNEL
movq K, %rax
#elif (defined(LEFT) && !defined(TRANSA)) || (!defined(LEFT) && defined(TRANSA))
movq K, %rax
subq KK, %rax
movq %rax, KKK
#else
movq KK, %rax
#ifdef LEFT
addq $2, %rax
#else
addq $1, %rax
#endif
movq %rax, KKK
#endif
sarq $2, %rax
je .L35
ALIGN_4
.L32:
PREFETCH (PREFETCHSIZE + 0) * SIZE(AO)
FLD -8 * SIZE(BO)
FLD -8 * SIZE(AO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -7 * SIZE(AO)
fmulp %st, %st(1)
faddp %st, %st(2)
FLD -7 * SIZE(BO)
FLD -6 * SIZE(AO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -5 * SIZE(AO)
fmulp %st, %st(1)
faddp %st, %st(2)
FLD -6 * SIZE(BO)
FLD -4 * SIZE(AO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -3 * SIZE(AO)
fmulp %st, %st(1)
faddp %st, %st(2)
FLD -5 * SIZE(BO)
FLD -2 * SIZE(AO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -1 * SIZE(AO)
fmulp %st, %st(1)
faddp %st, %st(2)
addq $8 * SIZE,AO
addq $4 * SIZE,BO
decq %rax
jne .L32
ALIGN_4
.L35:
#ifndef TRMMKERNEL
movq K, %rax
#else
movq KKK, %rax
#endif
and $3, %rax
je .L38
ALIGN_4
.L36:
FLD -8 * SIZE(BO)
FLD -8 * SIZE(AO)
fmul %st(1), %st
faddp %st, %st(2)
FLD -7 * SIZE(AO)
fmulp %st, %st(1)
faddp %st, %st(2)
addq $2 * SIZE,AO
addq $1 * SIZE,BO
decq %rax
jne .L36
ALIGN_4
.L38:
#ifndef TRMMKERNEL
FLD ALPHA_I
FLD ALPHA_R
fld %st(2)
fmul %st(1), %st
FLD 0 * SIZE(CO)
faddp %st, %st(1)
FST 0 * SIZE(CO)
fmul %st(3), %st
FLD 2 * SIZE(CO)
faddp %st, %st(1)
FST 2 * SIZE(CO)
fmul %st, %st(1)
fmulp %st, %st(2)
FLD 1 * SIZE(CO)
faddp %st, %st(1)
FST 1 * SIZE(CO)
FLD 3 * SIZE(CO)
faddp %st, %st(1)
FST 3 * SIZE(CO)
#else
FST 0 * SIZE(CO)
FST 1 * SIZE(CO)
#endif
#if (defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq K, %rax
subq KKK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 2), AO
leaq (BO, %rax, 1), BO
#endif
#if defined(TRMMKERNEL) && defined(LEFT)
addq $2, KK
#endif
addq $4 * SIZE, CO
decq I
jne .L31
ALIGN_4
.L40:
movq M, %rax
andq $1, %rax
je .L49
ALIGN_4
.L41:
#if !defined(TRMMKERNEL) || \
(defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq B, BO
#else
movq KK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 1), AO
leaq ( B, %rax, 1), BO
#endif
fldz
#ifndef TRMMKERNEL
movq K, %rax
#elif (defined(LEFT) && !defined(TRANSA)) || (!defined(LEFT) && defined(TRANSA))
movq K, %rax
subq KK, %rax
movq %rax, KKK
#else
movq KK, %rax
#ifdef LEFT
addq $1, %rax
#else
addq $1, %rax
#endif
movq %rax, KKK
#endif
sarq $2, %rax
je .L45
ALIGN_4
.L42:
PREFETCH (PREFETCHSIZE + 0) * SIZE(AO)
FLD -8 * SIZE(AO)
FLD -8 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(1)
FLD -7 * SIZE(AO)
FLD -7 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(1)
FLD -6 * SIZE(AO)
FLD -6 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(1)
FLD -5 * SIZE(AO)
FLD -5 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(1)
addq $4 * SIZE,AO
addq $4 * SIZE,BO
decq %rax
jne .L42
ALIGN_4
.L45:
#ifndef TRMMKERNEL
movq K, %rax
#else
movq KKK, %rax
#endif
and $3, %rax
je .L48
ALIGN_4
.L46:
FLD -8 * SIZE(AO)
FLD -8 * SIZE(BO)
fmulp %st, %st(1)
faddp %st, %st(1)
addq $1 * SIZE,AO
addq $1 * SIZE,BO
decq %rax
jne .L46
ALIGN_4
.L48:
#ifndef TRMMKERNEL
FLD ALPHA_I
FLD ALPHA_R
fmul %st(2), %st
FLD 0 * SIZE(CO)
faddp %st, %st(1)
FST 0 * SIZE(CO)
fmulp %st, %st(1)
FLD 1 * SIZE(CO)
faddp %st, %st(1)
FST 1 * SIZE(CO)
#else
FST 0 * SIZE(CO)
#endif
#if (defined(TRMMKERNEL) && defined(LEFT) && defined(TRANSA)) || \
(defined(TRMMKERNEL) && !defined(LEFT) && !defined(TRANSA))
movq K, %rax
subq KKK, %rax
salq $BASE_SHIFT, %rax
leaq (AO, %rax, 1), AO
leaq (BO, %rax, 1), BO
#endif
#if defined(TRMMKERNEL) && defined(LEFT)
addq $1, KK
#endif
addq $1 * SIZE, CO
ALIGN_4
.L49:
#if defined(TRMMKERNEL) && !defined(LEFT)
addq $1, KK
#endif
movq BO, B
ALIGN_4
.L999:
EMMS
movq 0(%rsp), %rbx
movq 8(%rsp), %rbp
movq 16(%rsp), %r12
movq 24(%rsp), %r13
movq 32(%rsp), %r14
movq 40(%rsp), %r15
addq $STACKSIZE, %rsp
ret
EPILOGUE