tahoma2d/thirdparty/openblas/xianyi-OpenBLAS-e6e87a2/kernel/x86_64/zcopy.S
2016-03-24 02:47:04 +09:00

389 lines
7.8 KiB
ArmAsm

/*********************************************************************/
/* Copyright 2009, 2010 The University of Texas at Austin. */
/* All rights reserved. */
/* */
/* Redistribution and use in source and binary forms, with or */
/* without modification, are permitted provided that the following */
/* conditions are met: */
/* */
/* 1. Redistributions of source code must retain the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer. */
/* */
/* 2. Redistributions in binary form must reproduce the above */
/* copyright notice, this list of conditions and the following */
/* disclaimer in the documentation and/or other materials */
/* provided with the distribution. */
/* */
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
/* POSSIBILITY OF SUCH DAMAGE. */
/* */
/* The views and conclusions contained in the software and */
/* documentation are those of the authors and should not be */
/* interpreted as representing official policies, either expressed */
/* or implied, of The University of Texas at Austin. */
/*********************************************************************/
#define ASSEMBLER
#include "common.h"
#define N ARG1 /* rdi */
#define X ARG2 /* rsi */
#define INCX ARG3 /* rdx */
#define Y ARG4 /* rcx */
#ifndef WINDOWS_ABI
#define INCY ARG5 /* r8 */
#define FLAG ARG6
#else
#define INCY %r10
#define FLAG %r11
#endif
#include "l1param.h"
PROLOGUE
PROFCODE
#ifdef WINDOWS_ABI
movq 40(%rsp), INCY
#endif
EMMS
salq $ZBASE_SHIFT, INCX
salq $ZBASE_SHIFT, INCY
testq N, N # if m == 0 goto End
jle .L999
cmpq $2 * SIZE, INCX # if incx != 1
jne .L100
cmpq $2 * SIZE, INCY # if incy != 1
jne .L100
movq N, %rax # i = m
sarq $2, %rax
jle .L20
ALIGN_2
.L11:
#ifdef XDOUBLE
#ifdef PREFETCH
PREFETCH (PREFETCHSIZE + 0) - PREOFFSET(X)
#endif
movq 0(X), %mm0
movq %mm0, 0(Y)
movq 8(X), %mm1
movq %mm1, 8(Y)
movq 16(X), %mm2
movq %mm2, 16(Y)
movq 24(X), %mm3
movq %mm3, 24(Y)
#ifdef PREFETCHW
PREFETCHW (PREFETCHSIZE + 0) - PREOFFSET(Y)
#endif
movq 32(X), %mm4
movq %mm4, 32(Y)
movq 40(X), %mm5
movq %mm5, 40(Y)
movq 48(X), %mm6
movq %mm6, 48(Y)
movq 56(X), %mm7
movq %mm7, 56(Y)
#ifdef PREFETCH
PREFETCH (PREFETCHSIZE + 64) - PREOFFSET(X)
#endif
movq 64(X), %mm0
movq %mm0, 64(Y)
movq 72(X), %mm1
movq %mm1, 72(Y)
movq 80(X), %mm2
movq %mm2, 80(Y)
movq 88(X), %mm3
movq %mm3, 88(Y)
#ifdef PREFETCHW
PREFETCHW (PREFETCHSIZE + 64) - PREOFFSET(Y)
#endif
movq 96(X), %mm4
movq %mm4, 96(Y)
movq 104(X), %mm5
movq %mm5, 104(Y)
movq 112(X), %mm6
movq %mm6, 112(Y)
movq 120(X), %mm7
movq %mm7, 120(Y)
#elif defined(DOUBLE)
#ifdef PREFETCH
PREFETCH (PREFETCHSIZE + 0) - PREOFFSET(X)
#endif
movq 0 * SIZE(X), %mm0
movq 1 * SIZE(X), %mm1
movq %mm0, 0 * SIZE(Y)
movq %mm1, 1 * SIZE(Y)
movq 2 * SIZE(X), %mm2
movq 3 * SIZE(X), %mm3
movq %mm2, 2 * SIZE(Y)
movq %mm3, 3 * SIZE(Y)
movq 4 * SIZE(X), %mm4
movq 5 * SIZE(X), %mm5
#ifdef PREFETCHW
PREFETCHW (PREFETCHSIZE + 0) - PREOFFSET(Y)
#endif
movq %mm4, 4 * SIZE(Y)
movq %mm5, 5 * SIZE(Y)
movq 6 * SIZE(X), %mm6
movq 7 * SIZE(X), %mm7
movq %mm6, 6 * SIZE(Y)
movq %mm7, 7 * SIZE(Y)
#else
movq 0 * SIZE(X), %mm0
movq 2 * SIZE(X), %mm2
#ifdef PREFETCH
PREFETCH (PREFETCHSIZE + 0) - PREOFFSET(X)
#endif
movq %mm0, 0 * SIZE(Y)
movq %mm2, 2 * SIZE(Y)
movq 4 * SIZE(X), %mm4
movq 6 * SIZE(X), %mm6
#ifdef PREFETCHW
PREFETCHW (PREFETCHSIZE + 0) - PREOFFSET(Y)
#endif
movq %mm4, 4 * SIZE(Y)
movq %mm6, 6 * SIZE(Y)
#endif
addq $8 * SIZE, X
addq $8 * SIZE, Y
decq %rax
jg .L11
ALIGN_2
.L20:
movq N, %rax # i = m
andq $3, %rax
jle .L99
ALIGN_2
.L21:
#ifdef XDOUBLE
movq 0(X), %mm0
movq %mm0, 0(Y)
movq 8(X), %mm1
movq %mm1, 8(Y)
movq 16(X), %mm2
movq %mm2, 16(Y)
movq 24(X), %mm3
movq %mm3, 24(Y)
#elif defined(DOUBLE)
movq 0 * SIZE(X), %mm0
movq %mm0, 0 * SIZE(Y)
movq 1 * SIZE(X), %mm1
movq %mm1, 1 * SIZE(Y)
#else
movq 0 * SIZE(X), %mm0
movq %mm0, 0 * SIZE(Y)
#endif
addq $2 * SIZE, X
addq $2 * SIZE, Y
decq %rax
jg .L21
.L99:
xorq %rax,%rax
EMMS
ret
ALIGN_3
.L100:
movq N, %rax
sarq $2, %rax
jle .L120
ALIGN_2
.L111:
#ifdef XDOUBLE
movq 0(X), %mm0
movq %mm0, 0(Y)
movq 8(X), %mm1
movq %mm1, 8(Y)
movq 16(X), %mm2
movq %mm2, 16(Y)
movq 24(X), %mm3
movq %mm3, 24(Y)
addq INCX, X
addq INCY, Y
movq 0(X), %mm0
movq %mm0, 0(Y)
movq 8(X), %mm1
movq %mm1, 8(Y)
movq 16(X), %mm2
movq %mm2, 16(Y)
movq 24(X), %mm3
movq %mm3, 24(Y)
addq INCX, X
addq INCY, Y
movq 0(X), %mm0
movq %mm0, 0(Y)
movq 8(X), %mm1
movq %mm1, 8(Y)
movq 16(X), %mm2
movq %mm2, 16(Y)
movq 24(X), %mm3
movq %mm3, 24(Y)
addq INCX, X
addq INCY, Y
movq 0(X), %mm0
movq %mm0, 0(Y)
movq 8(X), %mm1
movq %mm1, 8(Y)
movq 16(X), %mm2
movq %mm2, 16(Y)
movq 24(X), %mm3
movq %mm3, 24(Y)
addq INCX, X
addq INCY, Y
#elif defined(DOUBLE)
movq 0 * SIZE(X), %mm0
movq %mm0, 0 * SIZE(Y)
movq 1 * SIZE(X), %mm1
movq %mm1, 1 * SIZE(Y)
addq INCX, X
addq INCY, Y
movq 0 * SIZE(X), %mm2
movq %mm2, 0 * SIZE(Y)
movq 1 * SIZE(X), %mm3
movq %mm3, 1 * SIZE(Y)
addq INCX, X
addq INCY, Y
movq 0 * SIZE(X), %mm4
movq %mm4, 0 * SIZE(Y)
movq 1 * SIZE(X), %mm5
movq %mm5, 1 * SIZE(Y)
addq INCX, X
addq INCY, Y
movq 0 * SIZE(X), %mm6
movq %mm6, 0 * SIZE(Y)
movq 1 * SIZE(X), %mm7
movq %mm7, 1 * SIZE(Y)
addq INCX, X
addq INCY, Y
#else
movq 0 * SIZE(X), %mm0
movq %mm0, 0 * SIZE(Y)
addq INCX, X
addq INCY, Y
movq 0 * SIZE(X), %mm2
movq %mm2, 0 * SIZE(Y)
addq INCX, X
addq INCY, Y
movq 0 * SIZE(X), %mm4
movq %mm4, 0 * SIZE(Y)
addq INCX, X
addq INCY, Y
movq 0 * SIZE(X), %mm6
movq %mm6, 0 * SIZE(Y)
addq INCX, X
addq INCY, Y
#endif
decq %rax
jg .L111
.L120:
movq N, %rax
andq $3, %rax
jle .L999
ALIGN_2
.L121:
#ifdef XDOUBLE
movq 0(X), %mm0
movq %mm0, 0(Y)
movq 8(X), %mm1
movq %mm1, 8(Y)
movq 16(X), %mm2
movq %mm2, 16(Y)
movq 24(X), %mm3
movq %mm3, 24(Y)
addq INCX, X
addq INCY, Y
#elif defined(DOUBLE)
movq 0 * SIZE(X), %mm0
movq %mm0, 0 * SIZE(Y)
movq 1 * SIZE(X), %mm1
movq %mm1, 1 * SIZE(Y)
addq INCX, X
addq INCY, Y
#else
movq 0 * SIZE(X), %mm0
movq %mm0, 0 * SIZE(Y)
addq INCX, X
addq INCY, Y
#endif
decq %rax
jg .L121
.L999:
xorq %rax,%rax
EMMS
ret
EPILOGUE