// Copyright 2021 Google LLC // // This source code is licensed under the BSD-style license found in the // LICENSE file in the root directory of this source tree. #include #include #include #include #include "xnnpack/common.h" #include "xnnpack/microparams.h" #include "xnnpack/unaligned.h" #include "xnnpack/vunary.h" void xnn_u8_vclamp_ukernel__avx2_u128( size_t batch, const uint8_t* input, uint8_t* output, const struct xnn_u8_minmax_params params[restrict XNN_MIN_ELEMENTS(1)]) XNN_OOB_READS { assert(batch != 0); assert(batch % sizeof(uint8_t) == 0); assert(input != NULL); assert(output != NULL); const __m256i voutput_min = _mm256_set1_epi8(params->scalar.min); const __m256i voutput_max = _mm256_set1_epi8(params->scalar.max); XNN_FORCE_REALIZATION(voutput_min); XNN_FORCE_REALIZATION(voutput_max); for (; batch >= 128; batch -= 128) { __m256i vacc0 = _mm256_loadu_si256((const __m256i*) input); __m256i vacc1 = _mm256_loadu_si256((const __m256i*) input + 1); __m256i vacc2 = _mm256_loadu_si256((const __m256i*) input + 2); __m256i vacc3 = _mm256_loadu_si256((const __m256i*) input + 3); input += 128; vacc0 = _mm256_max_epu8(vacc0, voutput_min); vacc1 = _mm256_max_epu8(vacc1, voutput_min); vacc2 = _mm256_max_epu8(vacc2, voutput_min); vacc3 = _mm256_max_epu8(vacc3, voutput_min); vacc0 = _mm256_min_epu8(vacc0, voutput_max); vacc1 = _mm256_min_epu8(vacc1, voutput_max); vacc2 = _mm256_min_epu8(vacc2, voutput_max); vacc3 = _mm256_min_epu8(vacc3, voutput_max); _mm256_storeu_si256((__m256i*) output, vacc0); _mm256_storeu_si256((__m256i*) output + 1, vacc1); _mm256_storeu_si256((__m256i*) output + 2, vacc2); _mm256_storeu_si256((__m256i*) output + 3, vacc3); output += 128; } for (; batch >= 32; batch -= 32) { __m256i vacc = _mm256_loadu_si256((const __m256i*) input); input += 32; vacc = _mm256_max_epu8(vacc, voutput_min); vacc = _mm256_min_epu8(vacc, voutput_max); _mm256_storeu_si256((__m256i*) output, vacc); output += 32; } if (batch >= 16) { __m128i vacc = _mm_loadu_si128((const __m128i*) input); input += 16; vacc = _mm_max_epu8(vacc, _mm256_castsi256_si128(voutput_min)); vacc = _mm_min_epu8(vacc, _mm256_castsi256_si128(voutput_max)); _mm_storeu_si128((__m128i*) output, vacc); output += 16; batch -= 16; } if XNN_UNLIKELY(batch != 0) { assert(batch >= 1 && batch <= 15); __m128i vacc = _mm_loadu_si128((const __m128i*) input); vacc = _mm_max_epu8(vacc, _mm256_castsi256_si128(voutput_min)); vacc = _mm_min_epu8(vacc, _mm256_castsi256_si128(voutput_max)); if (batch & 8) { _mm_storel_epi64((__m128i*) output, vacc); output += 8; vacc = _mm_unpackhi_epi64(vacc, vacc); } if (batch & 4) { unaligned_store_u32(output, (uint32_t) _mm_cvtsi128_si32(vacc)); output += 4; vacc = _mm_srli_epi64(vacc, 32); } if (batch & 2) { unaligned_store_u16(output, (uint16_t) _mm_cvtsi128_si32(vacc)); output += 2; vacc = _mm_srli_epi32(vacc, 16); } if (batch & 1) { *output = (uint8_t) _mm_cvtsi128_si32(vacc); } } }