// Auto-generated file. Do not edit! // Template: src/f32-velu/scalar-rr2-lut16-p3.c.in // Generator: tools/xngen // // Copyright 2020 Google LLC // // This source code is licensed under the BSD-style license found in the // LICENSE file in the root directory of this source tree. #include #include #include "xnnpack/common.h" #include "xnnpack/math.h" #include "xnnpack/vunary.h" extern XNN_INTERNAL const uint32_t xnn_table_exp2minus_k_over_16[16]; void xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_u1( size_t batch, const float* input, float* output, const struct xnn_f32_elu_params params[restrict XNN_MIN_ELEMENTS(1)]) { assert(batch != 0); assert(batch % sizeof(float) == 0); assert(input != NULL); assert(output != NULL); const float vsat_cutoff = -0x1.154246p+4f; const float vmagic_bias = 0x1.800000p19f; const float vlog2e = 0x1.715476p+0f; const float vminus_ln2_hi = -0x1.62E400p-1f; const float vminus_ln2_lo = -0x1.7F7D1Cp-20f; const uint32_t vindex_mask = UINT32_C(0xF); const float vc3 = 0x1.55561Cp-3f; const float vc2 = 0x1.0001ECp-1f; const float vone = 1.0f; const float vprescale = params->scalar.prescale; const float valpha = params->scalar.alpha; const float vbeta = params->scalar.beta; do { float vx = *input++; const float vz = __builtin_wasm_min_f32(__builtin_wasm_max_f32(vx * vprescale, vsat_cutoff), 0.0f); float vn = vz * vlog2e + vmagic_bias; const uint32_t ven = float_as_uint32(vn) << 19; const uint32_t vidx = float_as_uint32(vn) & vindex_mask; vn -= vmagic_bias; float vt = vn * vminus_ln2_hi + vz; float vs = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx] + ven); vt = vn * vminus_ln2_lo + vt; float vp = vc3 * vt + vc2; vp *= vt; vt *= vs; vs -= vone; vp = vp * vt + vt; const float ve = (vp + vs) * valpha; float vy = __builtin_wasm_max_f32(vx * vbeta, 0.0f); vy += __builtin_wasm_min_f32(ve, 0.0f); *output++ = vy; batch -= sizeof(float); } while (batch != 0); }