|
|
|
|
|
|
|
|
|
|
|
#include <xnnpack/assembly.h> |
|
|
|
# void xnn_f16_gemm${"inc" if INC else ""}_minmax_ukernel_8x8__asm_aarch64_neonfp16arith_ld64( |
|
# size_t mr, x0 |
|
# size_t nc, x1 |
|
# size_t kc, x2 / x0 |
|
# const void* restrict a, x3 |
|
# size_t a_stride, x4 |
|
# const void* restrict w, x5 |
|
# void* restrict c, x6 |
|
# size_t cm_stride, x7 |
|
# size_t cn_stride, [sp] -> (x8) |
|
|
|
$if INC: |
|
# const float* restrict acc, [sp + 8] -> x15 |
|
# const union xnn_f16_minmax_params params[restrict XNN_MIN_ELEMENTS(1)]) [sp + 16] -> (x8) |
|
$else: |
|
# const union xnn_f16_minmax_params params[restrict XNN_MIN_ELEMENTS(1)]) [sp + 8] -> (x8) |
|
|
|
# d8-d15, x19-x30 need to be preserved if used. x18 is reserved by the OS. |
|
|
|
# Register usage |
|
# A0 x3 v0 |
|
# A1 x9 v1 |
|
# A2 x10 v2 |
|
# A3 x11 v3 |
|
# A4 x12 v4 |
|
# A5 x19 v5 |
|
# A6 x20 v6 |
|
# A7 x4 v7 |
|
# B x5 v16 v17 v18 v19 |
|
# C0 x6 v24 |
|
# C1 x16 v25 |
|
# C2 x17 v26 |
|
# C3 x14 v27 |
|
# C4 x13 v28 |
|
# C5 x21 v29 |
|
# C6 x22 v30 |
|
# C7 x7 v31 |
|
# Clamp v20 v21 |
|
# unused A v8 v9 v10 v11 |
|
# unused B v12 v13 v14 v15 |
|
|
|
BEGIN_FUNCTION xnn_f16_gemm${"inc" if INC else ""}_minmax_ukernel_8x8__asm_aarch64_neonfp16arith_ld64 |
|
|
|
$if INC: |
|
# Load acc, params pointer |
|
LDP x15, x8, [sp, 8] |
|
$else: |
|
# Load params pointer |
|
LDR x8, [sp, 8] |
|
|
|
# Save x19,x20,x21,x22 on stack |
|
STP x19, x20, [sp, -32]! |
|
STP x21, x22, [sp, 16] |
|
|
|
# Clamp A and C pointers |
|
CMP x0, 2 |
|
ADD x9, x3, x4 |
|
ADD x16, x6, x7 |
|
CSEL x9, x3, x9, LO |
|
CSEL x16, x6, x16, LO |
|
|
|
# Load params |
|
LD2R {v20.8h, v21.8h}, [x8] |
|
|
|
ADD x10, x9, x4 |
|
ADD x17, x16, x7 |
|
|
|
CSEL x10, x9, x10, LS |
|
CSEL x17, x16, x17, LS |
|
|
|
CMP x0, 4 |
|
ADD x11, x10, x4 |
|
ADD x14, x17, x7 |
|
CSEL x11, x10, x11, LO |
|
CSEL x14, x17, x14, LO |
|
|
|
ADD x12, x11, x4 |
|
ADD x13, x14, x7 |
|
|
|
CSEL x12, x11, x12, LS |
|
CSEL x13, x14, x13, LS |
|
|
|
CMP x0, 6 |
|
ADD x19, x12, x4 |
|
ADD x21, x13, x7 |
|
CSEL x19, x12, x19, LO |
|
CSEL x21, x13, x21, LO |
|
|
|
ADD x20, x19, x4 |
|
ADD x22, x21, x7 |
|
|
|
CSEL x20, x19, x20, LS |
|
CSEL x22, x21, x22, LS |
|
|
|
CMP x0, 8 |
|
ADD x4, x20, x4 |
|
ADD x7, x22, x7 |
|
CSEL x4, x20, x4, LO |
|
CSEL x7, x22, x7, LO |
|
|
|
LDR x8, [sp, 32] |
|
|
|
0: |
|
$if INC: |
|
# Load initial accumulators |
|
LDP q24, q25, [x15], 32 |
|
LDP q26, q27, [x15], 32 |
|
LDP q28, q29, [x15], 32 |
|
LDP q30, q31, [x15], 32 |
|
$else: |
|
# Load initial bias from w into accumulators |
|
LDR q24, [x5], 16 |
|
MOV v25.16b, v24.16b |
|
MOV v26.16b, v24.16b |
|
MOV v27.16b, v24.16b |
|
MOV v28.16b, v24.16b |
|
MOV v29.16b, v24.16b |
|
MOV v30.16b, v24.16b |
|
MOV v31.16b, v24.16b |
|
|
|
# Is there at least 4 halffloats (8 bytes)? |
|
SUBS x0, x2, 8 |
|
B.LO 3f |
|
|
|
# Main loop - 4 halffloats of A (8 bytes) |
|
# 32 FMA + 8 ld64 A + 4 LDR B |
|
1: |
|
LDR d0, [x3], 8 |
|
LDR q16, [x5], 16 |
|
LDR q17, [x5], 16 |
|
LDR d1, [x9], 8 |
|
LDR d2, [x10], 8 |
|
LDR d3, [x11], 8 |
|
LDR d4, [x12], 8 |
|
LDR d5, [x19], 8 |
|
LDR d6, [x20], 8 |
|
LDR d7, [x4], 8 |
|
LDR q18, [x5], 16 |
|
LDR q19, [x5], 16 |
|
SUBS x0, x0, 8 |
|
FMLA v24.8h, v16.8h, v0.h[0] |
|
FMLA v25.8h, v16.8h, v1.h[0] |
|
FMLA v26.8h, v16.8h, v2.h[0] |
|
FMLA v27.8h, v16.8h, v3.h[0] |
|
FMLA v28.8h, v16.8h, v4.h[0] |
|
FMLA v29.8h, v16.8h, v5.h[0] |
|
FMLA v30.8h, v16.8h, v6.h[0] |
|
FMLA v31.8h, v16.8h, v7.h[0] |
|
|
|
FMLA v24.8h, v17.8h, v0.h[1] |
|
FMLA v25.8h, v17.8h, v1.h[1] |
|
FMLA v26.8h, v17.8h, v2.h[1] |
|
FMLA v27.8h, v17.8h, v3.h[1] |
|
FMLA v28.8h, v17.8h, v4.h[1] |
|
FMLA v29.8h, v17.8h, v5.h[1] |
|
FMLA v30.8h, v17.8h, v6.h[1] |
|
FMLA v31.8h, v17.8h, v7.h[1] |
|
|
|
FMLA v24.8h, v18.8h, v0.h[2] |
|
FMLA v25.8h, v18.8h, v1.h[2] |
|
FMLA v26.8h, v18.8h, v2.h[2] |
|
FMLA v27.8h, v18.8h, v3.h[2] |
|
FMLA v28.8h, v18.8h, v4.h[2] |
|
FMLA v29.8h, v18.8h, v5.h[2] |
|
FMLA v30.8h, v18.8h, v6.h[2] |
|
FMLA v31.8h, v18.8h, v7.h[2] |
|
|
|
FMLA v24.8h, v19.8h, v0.h[3] |
|
FMLA v25.8h, v19.8h, v1.h[3] |
|
FMLA v26.8h, v19.8h, v2.h[3] |
|
FMLA v27.8h, v19.8h, v3.h[3] |
|
FMLA v28.8h, v19.8h, v4.h[3] |
|
FMLA v29.8h, v19.8h, v5.h[3] |
|
FMLA v30.8h, v19.8h, v6.h[3] |
|
FMLA v31.8h, v19.8h, v7.h[3] |
|
B.HS 1b |
|
|
|
# Is there a remainder?- 2 halffloats of A (4 bytes) |
|
TBNZ x0, 2, 4f |
|
# Is there a remainder?- 1 halffloat of A (2 bytes) |
|
TBNZ x0, 1, 5f |
|
2: |
|
# Clamp |
|
FMAX v24.8h, v24.8h, v20.8h |
|
FMAX v25.8h, v25.8h, v20.8h |
|
FMAX v26.8h, v26.8h, v20.8h |
|
FMAX v27.8h, v27.8h, v20.8h |
|
FMAX v28.8h, v28.8h, v20.8h |
|
FMAX v29.8h, v29.8h, v20.8h |
|
FMAX v30.8h, v30.8h, v20.8h |
|
FMAX v31.8h, v31.8h, v20.8h |
|
SUBS x1, x1, 8 |
|
FMIN v24.8h, v24.8h, v21.8h |
|
FMIN v25.8h, v25.8h, v21.8h |
|
FMIN v26.8h, v26.8h, v21.8h |
|
FMIN v27.8h, v27.8h, v21.8h |
|
FMIN v28.8h, v28.8h, v21.8h |
|
FMIN v29.8h, v29.8h, v21.8h |
|
FMIN v30.8h, v30.8h, v21.8h |
|
FMIN v31.8h, v31.8h, v21.8h |
|
|
|
# Store full 8 x 8 |
|
B.LO 6f |
|
|
|
$if INC: |
|
ST1 {v31.16b}, [x7], x8 |
|
SUB x3, x3, x2 |
|
ST1 {v30.16b}, [x22], x8 |
|
SUB x9, x9, x2 |
|
ST1 {v29.16b}, [x21], x8 |
|
SUB x10, x10, x2 |
|
ST1 {v28.16b}, [x13], x8 |
|
SUB x11, x11, x2 |
|
ST1 {v27.16b}, [x14], x8 |
|
SUB x12, x12, x2 |
|
ST1 {v26.16b}, [x17], x8 |
|
SUB x19, x19, x2 |
|
ST1 {v25.16b}, [x16], x8 |
|
SUB x20, x20, x2 |
|
ST1 {v24.16b}, [x6], x8 |
|
SUB x4, x4, x2 |
|
$else: |
|
ST1 {v24.16b}, [x6], x8 |
|
SUB x3, x3, x2 |
|
ST1 {v25.16b}, [x16], x8 |
|
SUB x9, x9, x2 |
|
ST1 {v26.16b}, [x17], x8 |
|
SUB x10, x10, x2 |
|
ST1 {v27.16b}, [x14], x8 |
|
SUB x11, x11, x2 |
|
ST1 {v28.16b}, [x13], x8 |
|
SUB x12, x12, x2 |
|
ST1 {v29.16b}, [x21], x8 |
|
SUB x19, x19, x2 |
|
ST1 {v30.16b}, [x22], x8 |
|
SUB x20, x20, x2 |
|
ST1 {v31.16b}, [x7], x8 |
|
SUB x4, x4, x2 |
|
|
|
B.HI 0b |
|
|
|
# Restore x19,x20,x21,x22 from stack |
|
LDP x21, x22, [sp, 16] |
|
LDP x19, x20, [sp], 32 |
|
RET |
|
|
|
3: |
|
TBZ x0, 2, 5f |
|
4: |
|
# Remainder- 2 halffloats of A (4 bytes) |
|
LDR s0, [x3], 4 |
|
LDR q16, [x5], 16 |
|
LDR q17, [x5], 16 |
|
LDR s1, [x9], 4 |
|
LDR s2, [x10], 4 |
|
LDR s3, [x11], 4 |
|
LDR s4, [x12], 4 |
|
LDR s5, [x19], 4 |
|
LDR s6, [x20], 4 |
|
LDR s7, [x4], 4 |
|
|
|
FMLA v24.8h, v16.8h, v0.h[0] |
|
FMLA v25.8h, v16.8h, v1.h[0] |
|
FMLA v26.8h, v16.8h, v2.h[0] |
|
FMLA v27.8h, v16.8h, v3.h[0] |
|
FMLA v28.8h, v16.8h, v4.h[0] |
|
FMLA v29.8h, v16.8h, v5.h[0] |
|
FMLA v30.8h, v16.8h, v6.h[0] |
|
FMLA v31.8h, v16.8h, v7.h[0] |
|
|
|
FMLA v24.8h, v17.8h, v0.h[1] |
|
FMLA v25.8h, v17.8h, v1.h[1] |
|
FMLA v26.8h, v17.8h, v2.h[1] |
|
FMLA v27.8h, v17.8h, v3.h[1] |
|
FMLA v28.8h, v17.8h, v4.h[1] |
|
FMLA v29.8h, v17.8h, v5.h[1] |
|
FMLA v30.8h, v17.8h, v6.h[1] |
|
FMLA v31.8h, v17.8h, v7.h[1] |
|
TBZ x0, 1, 2b |
|
|
|
5: |
|
# Remainder- 1 halffloat of A (2 bytes) |
|
LDR h0, [x3], 2 |
|
LDR q16, [x5], 16 |
|
LDR h1, [x9], 2 |
|
LDR h2, [x10], 2 |
|
LDR h3, [x11], 2 |
|
LDR h4, [x12], 2 |
|
LDR h5, [x19], 2 |
|
LDR h6, [x20], 2 |
|
LDR h7, [x4], 2 |
|
|
|
FMLA v24.8h, v16.8h, v0.h[0] |
|
FMLA v25.8h, v16.8h, v1.h[0] |
|
FMLA v26.8h, v16.8h, v2.h[0] |
|
FMLA v27.8h, v16.8h, v3.h[0] |
|
FMLA v28.8h, v16.8h, v4.h[0] |
|
FMLA v29.8h, v16.8h, v5.h[0] |
|
FMLA v30.8h, v16.8h, v6.h[0] |
|
FMLA v31.8h, v16.8h, v7.h[0] |
|
B 2b |
|
|
|
# Store odd width |
|
6: |
|
TBZ x1, 2, 7f |
|
$if INC: |
|
STR d31, [x7], 8 |
|
STR d30, [x22], 8 |
|
DUP d31, v31.d[1] |
|
DUP d30, v30.d[1] |
|
STR d29, [x21], 8 |
|
STR d28, [x13], 8 |
|
DUP d29, v29.d[1] |
|
DUP d28, v28.d[1] |
|
STR d27, [x14], 8 |
|
STR d26, [x17], 8 |
|
DUP d27, v27.d[1] |
|
DUP d26, v26.d[1] |
|
STR d25, [x16], 8 |
|
STR d24, [x6], 8 |
|
DUP d25, v25.d[1] |
|
DUP d24, v24.d[1] |
|
$else: |
|
STR d24, [x6], 8 |
|
STR d25, [x16], 8 |
|
DUP d24, v24.d[1] |
|
DUP d25, v25.d[1] |
|
STR d26, [x17], 8 |
|
STR d27, [x14], 8 |
|
DUP d26, v26.d[1] |
|
DUP d27, v27.d[1] |
|
STR d28, [x13], 8 |
|
STR d29, [x21], 8 |
|
DUP d28, v28.d[1] |
|
DUP d29, v29.d[1] |
|
STR d30, [x22], 8 |
|
STR d31, [x7], 8 |
|
DUP d30, v30.d[1] |
|
DUP d31, v31.d[1] |
|
7: |
|
TBZ x1, 1, 8f |
|
$if INC: |
|
STR s31, [x7], 4 |
|
STR s30, [x22], 4 |
|
DUP s31, v31.s[1] |
|
DUP s30, v30.s[1] |
|
STR s29, [x21], 4 |
|
STR s28, [x13], 4 |
|
DUP s29, v29.s[1] |
|
DUP s28, v28.s[1] |
|
STR s27, [x14], 4 |
|
STR s26, [x17], 4 |
|
DUP s27, v27.s[1] |
|
DUP s26, v26.s[1] |
|
STR s25, [x16], 4 |
|
STR s24, [x6], 4 |
|
DUP s25, v25.s[1] |
|
DUP s24, v24.s[1] |
|
$else: |
|
STR s24, [x6], 4 |
|
STR s25, [x16], 4 |
|
DUP s24, v24.s[1] |
|
DUP s25, v25.s[1] |
|
STR s26, [x17], 4 |
|
STR s27, [x14], 4 |
|
DUP s26, v26.s[1] |
|
DUP s27, v27.s[1] |
|
STR s28, [x13], 4 |
|
STR s29, [x21], 4 |
|
DUP s28, v28.s[1] |
|
DUP s29, v29.s[1] |
|
STR s30, [x22], 4 |
|
STR s31, [x7], 4 |
|
DUP s30, v30.s[1] |
|
DUP s31, v31.s[1] |
|
|
|
8: |
|
TBZ x1, 0, 9f |
|
$if INC: |
|
STR h31, [x7] |
|
STR h30, [x22] |
|
STR h29, [x21] |
|
STR h28, [x13] |
|
STR h27, [x14] |
|
STR h26, [x17] |
|
STR h25, [x16] |
|
STR h24, [x6] |
|
$else: |
|
STR h24, [x6] |
|
STR h25, [x16] |
|
STR h26, [x17] |
|
STR h27, [x14] |
|
STR h28, [x13] |
|
STR h29, [x21] |
|
STR h30, [x22] |
|
STR h31, [x7] |
|
9: |
|
# Restore x19,x20,x21,x22 from stack |
|
LDP x21, x22, [sp, 16] |
|
LDP x19, x20, [sp], 32 |
|
RET |
|
|
|
END_FUNCTION xnn_f16_gemm${"inc" if INC else ""}_minmax_ukernel_8x8__asm_aarch64_neonfp16arith_ld64 |
|
|
|
#ifdef __ELF__ |
|
.section ".note.GNU-stack","",%progbits |
|
#endif |
|
|