Spaces:
Running
Running
// Copyright (C) 2018-2022 Intel Corporation | |
// SPDX-License-Identifier: Apache-2.0 | |
// | |
__kernel void Convolution1x1_NCHW( | |
const __global half *in, | |
const __global half *out, | |
const __global half *w, | |
int IW, | |
int IH, | |
int IC, | |
int OW, | |
int OH, | |
int OC) | |
{ | |
__local half in_local[8 * 1024]; | |
__local half out_local[8 * 1024]; | |
event_t e1 = async_work_group_copy_2D2D( | |
in_local, // dst | |
in + get_group_id(0) * IW, // src | |
IW, // num_elements_per_line, | |
IC, // num_lines, | |
IW * IH - IW, // src_line_stride, | |
0, // dst_line_stride, | |
0); | |
wait_group_events(1, &e1); | |
int oh = get_global_id(0); | |
int oc = get_global_id(1); | |
int stride; | |
int write_output = 0; | |
__global half *src; | |
__global half8 *w8 = (__global half8 *)(&w[oc * IC]); | |
__global half *w1 = (__global half *)(&w[oc * IC]); | |
for (uint ow = 0; ow < (OW & (~0x7)); ow += 8) { | |
uint iw = ow; | |
uint ih = oh; | |
half8 val8_0 = 0.0f; | |
__local half8 *in8_0 = (__local half8 *)(&in_local[iw + 0 * IW]); | |
__local half8 *in8_1 = (__local half8 *)(&in_local[iw + 1 * IW]); | |
__local half8 *in8_2 = (__local half8 *)(&in_local[iw + 2 * IW]); | |
__local half8 *in8_3 = (__local half8 *)(&in_local[iw + 3 * IW]); | |
__local half8 *in8_4 = (__local half8 *)(&in_local[iw + 4 * IW]); | |
__local half8 *in8_5 = (__local half8 *)(&in_local[iw + 5 * IW]); | |
__local half8 *in8_6 = (__local half8 *)(&in_local[iw + 6 * IW]); | |
__local half8 *in8_7 = (__local half8 *)(&in_local[iw + 7 * IW]); | |
for (uint ic = 0; ic < IC / 8; ic++) { | |
val8_0 += (in8_0[ic * IW]) * ((half8)w8[ic].s0); | |
val8_0 += (in8_1[ic * IW]) * ((half8)w8[ic].s1); | |
val8_0 += (in8_2[ic * IW]) * ((half8)w8[ic].s2); | |
val8_0 += (in8_3[ic * IW]) * ((half8)w8[ic].s3); | |
val8_0 += (in8_4[ic * IW]) * ((half8)w8[ic].s4); | |
val8_0 += (in8_5[ic * IW]) * ((half8)w8[ic].s5); | |
val8_0 += (in8_6[ic * IW]) * ((half8)w8[ic].s6); | |
val8_0 += (in8_7[ic * IW]) * ((half8)w8[ic].s7); | |
} | |
for (uint ic = (IC & (~0x7)); ic < IC; ++ic) { | |
val8_0 += *((__local half8 *)(&in_local[iw + ic * IW])) * ((half8)w1[ic]); | |
} | |
*((__local half8 *)&out_local[ow + 0]) = (val8_0); | |
} | |
uint iw = (OW & (~0x7)); | |
uint ih = oh; | |
half8 val8_0 = 0.0f; | |
__local half8 *in8_0 = (__local half8 *)(&in_local[iw + 0 * IW]); | |
__local half8 *in8_1 = (__local half8 *)(&in_local[iw + 1 * IW]); | |
__local half8 *in8_2 = (__local half8 *)(&in_local[iw + 2 * IW]); | |
__local half8 *in8_3 = (__local half8 *)(&in_local[iw + 3 * IW]); | |
__local half8 *in8_4 = (__local half8 *)(&in_local[iw + 4 * IW]); | |
__local half8 *in8_5 = (__local half8 *)(&in_local[iw + 5 * IW]); | |
__local half8 *in8_6 = (__local half8 *)(&in_local[iw + 6 * IW]); | |
__local half8 *in8_7 = (__local half8 *)(&in_local[iw + 7 * IW]); | |
for (uint ic = 0; ic < IC / 8; ic++) { | |
val8_0 += (in8_0[ic * IW]) * ((half8)w8[ic].s0); | |
val8_0 += (in8_1[ic * IW]) * ((half8)w8[ic].s1); | |
val8_0 += (in8_2[ic * IW]) * ((half8)w8[ic].s2); | |
val8_0 += (in8_3[ic * IW]) * ((half8)w8[ic].s3); | |
val8_0 += (in8_4[ic * IW]) * ((half8)w8[ic].s4); | |
val8_0 += (in8_5[ic * IW]) * ((half8)w8[ic].s5); | |
val8_0 += (in8_6[ic * IW]) * ((half8)w8[ic].s6); | |
val8_0 += (in8_7[ic * IW]) * ((half8)w8[ic].s7); | |
} | |
for (uint ic = (IC & (~0x7)); ic < IC; ++ic) { | |
val8_0 += *((__local half8 *)(&in_local[iw + ic * IW])) * ((half8)w1[ic]); | |
} | |
for (uint ow = (OW & (~0x7)); ow < OW; ow++) { | |
out_local[ow + 0] = (val8_0[ow % 8]); | |
} | |
barrier(CLK_LOCAL_MEM_FENCE); | |
event_t e2 = async_work_group_copy( | |
out + get_group_id(1) * OW * OH + get_group_id(0) * OW, | |
out_local, | |
OW, | |
0); | |
wait_group_events(1, &e2); | |
} | |