|
- /***************************************************************************
- * Copyright (c) 2022,2025 The OpenBLAS Project
- * All rights reserved.
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions are
- * met:
- * 1. Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- * 2. Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in
- * the documentation and/or other materials provided with the
- * distribution.
- * 3. Neither the name of the OpenBLAS project nor the names of
- * its contributors may be used to endorse or promote products
- * derived from this software without specific prior written permission.
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
- * POSSIBILITY OF SUCH DAMAGE.
- * *****************************************************************************/
-
- #include <arm_sve.h>
- #include <arm_neon.h>
-
- #include "common.h"
-
- #define INIT_C(M, N) mc##M##N = svdup_f32(0);
-
- #define MATMUL(M, N) mc##M##N = svbfmmla(mc##M##N, ma##M, mb##N);
-
- #define INIT_C_8x4 \
- do { \
- INIT_C(0, 0); \
- INIT_C(0, 1); \
- INIT_C(1, 0); \
- INIT_C(1, 1); \
- INIT_C(2, 0); \
- INIT_C(2, 1); \
- INIT_C(3, 0); \
- INIT_C(3, 1); \
- } while (0);
-
- #ifdef BGEMM
- #ifdef ALPHA_ONE
- #define UPDATE_C(PG16, PG32, PTR, SRC) \
- do { \
- tmp32 = svreinterpret_f32_u32(svld1uh_u32((PG16), (uint16_t*)PTR)); \
- tmp32 = svadd_z((PG32), SRC, tmp32); \
- tmp16 = svcvt_bf16_f32_z((PG32), tmp32); \
- tmp16 = svuzp1_bf16(tmp16, tmp16); \
- svst1_bf16((PG16), (PTR), tmp16); \
- } while (0)
- #else
- #define UPDATE_C(PG16, PG32, PTR, SRC) \
- do { \
- tmp32 = svreinterpret_f32_u32(svld1uh_u32((PG16), (uint16_t*)PTR)); \
- tmp32 = svmad_z((PG32), svalpha, SRC, tmp32); \
- tmp16 = svcvt_bf16_f32_z((PG32), tmp32); \
- tmp16 = svuzp1_bf16(tmp16, tmp16); \
- svst1_bf16((PG16), (PTR), tmp16); \
- } while (0)
- #endif
- #else
- #ifdef ALPHA_ONE
- #define UPDATE_C(PG16, PG32, PTR, SRC) \
- do { \
- tmp32 = svld1_f32((PG32), (PTR)); \
- tmp32 = svadd_z((PG32), SRC, tmp32); \
- svst1_f32((PG32), (PTR), tmp32); \
- } while (0);
- #else
- #define UPDATE_C(PG16, PG32, PTR, SRC) \
- do { \
- tmp32 = svld1_f32((PG32), (PTR)); \
- tmp32 = svmad_z((PG32), svalpha, SRC, tmp32); \
- svst1_f32((PG32), (PTR), tmp32); \
- } while (0);
- #endif
- #endif
-
- #ifdef BGEMM
- #define OUTPUT_FLOAT bfloat16_t
- #else
- #define OUTPUT_FLOAT float
- #endif
-
- #ifdef ALPHA_ONE
- static int gemm_kernel_neoversen2_alpha_one(BLASLONG m, BLASLONG n, BLASLONG k, FLOAT alpha, IFLOAT * A, IFLOAT * B, FLOAT * C, BLASLONG ldc)
- #else
- static int gemm_kernel_neoversen2_alpha(BLASLONG m, BLASLONG n, BLASLONG k, FLOAT alpha, IFLOAT * A, IFLOAT * B, FLOAT * C, BLASLONG ldc)
- #endif
- {
- BLASLONG pad_k = (k + 3) & ~3;
-
- svbfloat16_t ma0, ma1, ma2, ma3, mb0, mb1;
- svfloat32_t mc00, mc01, mc10, mc11, mc20, mc21, mc30, mc31,
- vc0, vc1, vc2, vc3, vc4, vc5, vc6, vc7;
-
- #ifndef ALPHA_ONE
- #ifdef BGEMM
- bfloat16_t alpha_bf16;
- memcpy(&alpha_bf16, &alpha, sizeof(bfloat16_t));
- svfloat32_t svalpha = svdup_f32(vcvtah_f32_bf16(alpha_bf16));
- #else
- svfloat32_t svalpha = svdup_f32(alpha);
- #endif
- #endif
-
- svbool_t pg32_first_4 = svdupq_b32(1, 1, 1, 1);
- svbool_t pg32_first_2 = svdupq_b32(1, 1, 0, 0);
- svbool_t pg32_first_1 = svdupq_b32(1, 0, 0, 0);
- svbool_t pg16_first_8 = svdupq_b16(1, 1, 1, 1, 1, 1, 1, 1);
- svbool_t pg16_first_4 = svdupq_b16(1, 1, 1, 1, 0, 0, 0, 0);
- #ifdef BGEMM
- svbool_t pg16_first_2 = svdupq_b16(1, 1, 0, 0, 0, 0, 0, 0);
- svbool_t pg16_first_1 = svdupq_b16(1, 0, 0, 0, 0, 0, 0, 0);
- #endif
-
- bfloat16_t *ptr_a = (bfloat16_t *)A;
- bfloat16_t *ptr_b = (bfloat16_t *)B;
- OUTPUT_FLOAT *ptr_c = (OUTPUT_FLOAT*)C;
-
- bfloat16_t *ptr_a0;
- bfloat16_t *ptr_b0;
- OUTPUT_FLOAT *ptr_c0, *ptr_c1, *ptr_c2, *ptr_c3;
-
- svfloat32_t tmp32;
- #ifdef BGEMM
- svbfloat16_t tmp16;
- #endif
-
- for (BLASLONG j = 0; j < n / 4; j++) {
- ptr_c0 = ptr_c;
- ptr_c1 = ptr_c0 + ldc;
- ptr_c2 = ptr_c1 + ldc;
- ptr_c3 = ptr_c2 + ldc;
- ptr_c += 4 * ldc;
- ptr_a = (bfloat16_t *)A;
-
- for (BLASLONG i = 0; i < m / 8; i++) {
- ptr_a0 = ptr_a;
- ptr_a += 8 * pad_k;
-
- ptr_b0 = ptr_b;
-
- INIT_C_8x4;
-
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- ma1 = svld1_bf16(pg16_first_8, ptr_a0 + 8);
- ma2 = svld1_bf16(pg16_first_8, ptr_a0 + 16);
- ma3 = svld1_bf16(pg16_first_8, ptr_a0 + 24);
-
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
- mb1 = svld1_bf16(pg16_first_8, ptr_b0 + 8);
-
- MATMUL(0, 0); MATMUL(0, 1);
- MATMUL(1, 0); MATMUL(1, 1);
- MATMUL(2, 0); MATMUL(2, 1);
- MATMUL(3, 0); MATMUL(3, 1);
-
- ptr_a0 += 32;
- ptr_b0 += 16;
- }
-
- vc0 = svuzp1(mc00, mc10);
- vc1 = svuzp1(mc20, mc30);
- vc2 = svuzp2(mc00, mc10);
- vc3 = svuzp2(mc20, mc30);
- vc4 = svuzp1(mc01, mc11);
- vc5 = svuzp1(mc21, mc31);
- vc6 = svuzp2(mc01, mc11);
- vc7 = svuzp2(mc21, mc31);
-
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0, vc0);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0+4, vc1);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c1, vc2);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c1+4, vc3);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c2, vc4);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c2+4, vc5);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c3, vc6);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c3+4, vc7);
-
- ptr_c0 += 8;
- ptr_c1 += 8;
- ptr_c2 += 8;
- ptr_c3 += 8;
- }
-
- if (m & 4) {
- ptr_a0 = ptr_a;
- ptr_a += 4 * pad_k;
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0); INIT_C(0, 1);
- INIT_C(1, 0); INIT_C(1, 1);
-
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- ma1 = svld1_bf16(pg16_first_8, ptr_a0 + 8);
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
- mb1 = svld1_bf16(pg16_first_8, ptr_b0 + 8);
-
- MATMUL(0, 0); MATMUL(0, 1);
- MATMUL(1, 0); MATMUL(1, 1);
-
- ptr_a0 += 16;
- ptr_b0 += 16;
- }
-
- vc0 = svuzp1(mc00, mc10);
- vc1 = svuzp2(mc00, mc10);
- vc2 = svuzp1(mc01, mc11);
- vc3 = svuzp2(mc01, mc11);
-
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0, vc0);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c1, vc1);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c2, vc2);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c3, vc3);
-
- ptr_c0 += 4;
- ptr_c1 += 4;
- ptr_c2 += 4;
- ptr_c3 += 4;
- }
-
- if (m & 2) {
- ptr_a0 = ptr_a;
- ptr_a += 2 * pad_k;
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0); INIT_C(0, 1);
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
- mb1 = svld1_bf16(pg16_first_8, ptr_b0 + 8);
-
- MATMUL(0, 0); MATMUL(0, 1);
-
- ptr_a0 += 8;
- ptr_b0 += 16;
- }
-
- vc0 = svuzp1(mc00, mc00);
- vc1 = svuzp2(mc00, mc00);
- vc2 = svuzp1(mc01, mc01);
- vc3 = svuzp2(mc01, mc01);
-
- UPDATE_C(pg16_first_2, pg32_first_2, ptr_c0, vc0);
- UPDATE_C(pg16_first_2, pg32_first_2, ptr_c1, vc1);
- UPDATE_C(pg16_first_2, pg32_first_2, ptr_c2, vc2);
- UPDATE_C(pg16_first_2, pg32_first_2, ptr_c3, vc3);
-
- ptr_c0 += 2;
- ptr_c1 += 2;
- ptr_c2 += 2;
- ptr_c3 += 2;
- }
-
- if (m & 1) {
- ptr_a0 = ptr_a;
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0); INIT_C(0, 1);
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_4, ptr_a0);
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
- mb1 = svld1_bf16(pg16_first_8, ptr_b0 + 8);
-
- MATMUL(0, 0); MATMUL(0, 1);
-
- ptr_a0 += 4;
- ptr_b0 += 16;
- }
-
- vc1 = svuzp2(mc00, mc00);
- vc3 = svuzp2(mc01, mc01);
-
- UPDATE_C(pg16_first_1, pg32_first_1, ptr_c0, mc00);
- UPDATE_C(pg16_first_1, pg32_first_1, ptr_c1, vc1);
- UPDATE_C(pg16_first_1, pg32_first_1, ptr_c2, mc01);
- UPDATE_C(pg16_first_1, pg32_first_1, ptr_c3, vc3);
-
- }
-
- ptr_b += 4 * pad_k;
- }
-
- if (n & 2) {
- ptr_c0 = ptr_c;
- ptr_c1 = ptr_c0 + ldc;
- ptr_c += 2 * ldc;
- ptr_a = (bfloat16_t *)A;
-
- for (BLASLONG i = 0; i < m / 8; i++) {
- ptr_a0 = ptr_a;
- ptr_a += 8 * pad_k;
-
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0);
- INIT_C(1, 0);
- INIT_C(2, 0);
- INIT_C(3, 0);
-
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- ma1 = svld1_bf16(pg16_first_8, ptr_a0 + 8);
- ma2 = svld1_bf16(pg16_first_8, ptr_a0 + 16);
- ma3 = svld1_bf16(pg16_first_8, ptr_a0 + 24);
-
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
-
- MATMUL(0, 0);
- MATMUL(1, 0);
- MATMUL(2, 0);
- MATMUL(3, 0);
-
- ptr_a0 += 32;
- ptr_b0 += 8;
- }
-
- vc0 = svuzp1(mc00, mc10);
- vc1 = svuzp1(mc20, mc30);
- vc2 = svuzp2(mc00, mc10);
- vc3 = svuzp2(mc20, mc30);
-
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0, vc0);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0 + 4, vc1);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c1, vc2);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c1 + 4, vc3);
-
- ptr_c0 += 8;
- ptr_c1 += 8;
- }
-
- if (m & 4) {
- ptr_a0 = ptr_a;
- ptr_a += 4 * pad_k;
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0);
- INIT_C(1, 0);
-
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- ma1 = svld1_bf16(pg16_first_8, ptr_a0 + 8);
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
- MATMUL(0, 0);
- MATMUL(1, 0);
- ptr_a0 += 16;
- ptr_b0 += 8;
- }
-
- vc0 = svuzp1(mc00, mc10);
- vc1 = svuzp2(mc00, mc10);
-
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0, vc0);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c1, vc1);
-
- ptr_c0 += 4;
- ptr_c1 += 4;
- }
-
- if (m & 2) {
- ptr_a0 = ptr_a;
- ptr_a += 2 * pad_k;
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0);
-
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
-
- MATMUL(0, 0);
-
- ptr_a0 += 8;
- ptr_b0 += 8;
- }
-
- vc0 = svuzp1(mc00, mc00);
- vc1 = svuzp2(mc00, mc00);
- UPDATE_C(pg16_first_2, pg32_first_2, ptr_c0, vc0);
- UPDATE_C(pg16_first_2, pg32_first_2, ptr_c1, vc1);
-
- ptr_c0 += 2;
- ptr_c1 += 2;
-
- }
-
- if (m & 1) {
- ptr_a0 = ptr_a;
- ptr_b0 = ptr_b;
- INIT_C(0, 0);
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_4, ptr_a0);
- mb0 = svld1_bf16(pg16_first_8, ptr_b0);
- MATMUL(0, 0);
- ptr_a0 += 4;
- ptr_b0 += 8;
- }
- vc1 = svuzp2(mc00, mc00);
-
- UPDATE_C(pg16_first_1, pg32_first_1, ptr_c0, mc00);
- UPDATE_C(pg16_first_1, pg32_first_1, ptr_c1, vc1);
- }
-
- ptr_b += 2 * pad_k;
- }
-
- if (n & 1) {
- ptr_c0 = ptr_c;
- ptr_a = (bfloat16_t *)A;
-
- for (BLASLONG i = 0; i < m / 8; i++) {
- ptr_a0 = ptr_a;
- ptr_a += 8 * pad_k;
-
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0);
- INIT_C(1, 0);
- INIT_C(2, 0);
- INIT_C(3, 0);
-
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- ma1 = svld1_bf16(pg16_first_8, ptr_a0 + 8);
- ma2 = svld1_bf16(pg16_first_8, ptr_a0 + 16);
- ma3 = svld1_bf16(pg16_first_8, ptr_a0 + 24);
-
- mb0 = svld1_bf16(pg16_first_4, ptr_b0);
-
- MATMUL(0, 0);
- MATMUL(1, 0);
- MATMUL(2, 0);
- MATMUL(3, 0);
-
- ptr_a0 += 32;
- ptr_b0 += 4;
- }
-
- vc0 = svuzp1(mc00, mc10);
- vc1 = svuzp1(mc20, mc30);
-
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0, vc0);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0 + 4, vc1);
-
- ptr_c0 += 8;
- }
-
- if (m & 4) {
- ptr_a0 = ptr_a;
- ptr_a += 4 * pad_k;
- ptr_b0 = ptr_b;
- INIT_C(0, 0);
- INIT_C(1, 0);
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- ma1 = svld1_bf16(pg16_first_8, ptr_a0 + 8);
- mb0 = svld1_bf16(pg16_first_4, ptr_b0);
- MATMUL(0, 0);
- MATMUL(1, 0);
- ptr_a0 += 16;
- ptr_b0 += 4;
- }
- vc0 = svuzp1(mc00, mc10);
- UPDATE_C(pg16_first_4, pg32_first_4, ptr_c0, vc0);
- ptr_c0 += 4;
- }
-
- if (m & 2) {
- ptr_a0 = ptr_a;
- ptr_a += 2 * pad_k;
- ptr_b0 = ptr_b;
-
- INIT_C(0, 0);
-
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_8, ptr_a0);
- mb0 = svld1_bf16(pg16_first_4, ptr_b0);
-
- MATMUL(0, 0);
-
- ptr_a0 += 8;
- ptr_b0 += 4;
- }
- vc0 = svuzp1(mc00, mc00);
- UPDATE_C(pg16_first_2, pg32_first_2, ptr_c0, vc0);
- ptr_c0 += 2;
- }
-
- if (m & 1) {
- ptr_a0 = ptr_a;
- ptr_b0 = ptr_b;
- INIT_C(0, 0);
- for (BLASLONG p = 0; p < pad_k; p += 4) {
- ma0 = svld1_bf16(pg16_first_4, ptr_a0);
- mb0 = svld1_bf16(pg16_first_4, ptr_b0);
- MATMUL(0, 0);
- ptr_a0 += 4;
- ptr_b0 += 4;
- }
- UPDATE_C(pg16_first_1, pg32_first_1, ptr_c0, mc00);
- }
- }
-
- return 0;
- }
|