blob: 1ce12ec23f69ce6bad3a8bb3e2f4d9486b6f0349 [file] [log] [blame]
/*
* Copyright (c) 2020, Alliance for Open Media. All rights reserved
*
* This source code is subject to the terms of the BSD 2 Clause License and
* the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License
* was not distributed with this source code in the LICENSE file, you can
* obtain it at www.aomedia.org/license/software. If the Alliance for Open
* Media Patent License 1.0 was not distributed with this source code in the
* PATENTS file, you can obtain it at www.aomedia.org/license/patent.
*/
#include <arm_neon.h>
#include <assert.h>
#include "av1/common/arm/mem_neon.h"
#include "config/aom_dsp_rtcd.h"
static INLINE uint32x4_t sum_squares_i16_4x4_neon(const int16_t *src,
int stride) {
const int16x4_t v_val_01_lo = vld1_s16(src + 0 * stride);
const int16x4_t v_val_01_hi = vld1_s16(src + 1 * stride);
const int16x4_t v_val_23_lo = vld1_s16(src + 2 * stride);
const int16x4_t v_val_23_hi = vld1_s16(src + 3 * stride);
int32x4_t v_sq_01_d = vmull_s16(v_val_01_lo, v_val_01_lo);
v_sq_01_d = vmlal_s16(v_sq_01_d, v_val_01_hi, v_val_01_hi);
int32x4_t v_sq_23_d = vmull_s16(v_val_23_lo, v_val_23_lo);
v_sq_23_d = vmlal_s16(v_sq_23_d, v_val_23_hi, v_val_23_hi);
#if defined(__aarch64__)
return vreinterpretq_u32_s32(vpaddq_s32(v_sq_01_d, v_sq_23_d));
#else
return vreinterpretq_u32_s32(vcombine_s32(
vqmovn_s64(vpaddlq_s32(v_sq_01_d)), vqmovn_s64(vpaddlq_s32(v_sq_23_d))));
#endif
}
uint64_t aom_sum_squares_2d_i16_4x4_neon(const int16_t *src, int stride) {
const uint32x4_t v_sum_0123_d = sum_squares_i16_4x4_neon(src, stride);
#if defined(__aarch64__)
return (uint64_t)vaddvq_u32(v_sum_0123_d);
#else
uint64x2_t v_sum_d = vpaddlq_u32(v_sum_0123_d);
v_sum_d = vaddq_u64(v_sum_d, vextq_u64(v_sum_d, v_sum_d, 1));
return vgetq_lane_u64(v_sum_d, 0);
#endif
}
uint64_t aom_sum_squares_2d_i16_4xn_neon(const int16_t *src, int stride,
int height) {
int r = 0;
uint32x4_t v_acc_q = vdupq_n_u32(0);
do {
const uint32x4_t v_acc_d = sum_squares_i16_4x4_neon(src, stride);
v_acc_q = vaddq_u32(v_acc_q, v_acc_d);
src += stride << 2;
r += 4;
} while (r < height);
uint64x2_t v_acc_64 = vpaddlq_u32(v_acc_q);
#if defined(__aarch64__)
return vaddvq_u64(v_acc_64);
#else
v_acc_64 = vaddq_u64(v_acc_64, vextq_u64(v_acc_64, v_acc_64, 1));
return vgetq_lane_u64(v_acc_64, 0);
#endif
}
uint64_t aom_sum_squares_2d_i16_nxn_neon(const int16_t *src, int stride,
int width, int height) {
int r = 0;
const int32x4_t zero = vdupq_n_s32(0);
uint64x2_t v_acc_q = vreinterpretq_u64_s32(zero);
do {
int32x4_t v_sum = zero;
int c = 0;
do {
const int16_t *b = src + c;
const int16x8_t v_val_0 = vld1q_s16(b + 0 * stride);
const int16x8_t v_val_1 = vld1q_s16(b + 1 * stride);
const int16x8_t v_val_2 = vld1q_s16(b + 2 * stride);
const int16x8_t v_val_3 = vld1q_s16(b + 3 * stride);
const int16x4_t v_val_0_lo = vget_low_s16(v_val_0);
const int16x4_t v_val_1_lo = vget_low_s16(v_val_1);
const int16x4_t v_val_2_lo = vget_low_s16(v_val_2);
const int16x4_t v_val_3_lo = vget_low_s16(v_val_3);
int32x4_t v_sum_01 = vmull_s16(v_val_0_lo, v_val_0_lo);
v_sum_01 = vmlal_s16(v_sum_01, v_val_1_lo, v_val_1_lo);
int32x4_t v_sum_23 = vmull_s16(v_val_2_lo, v_val_2_lo);
v_sum_23 = vmlal_s16(v_sum_23, v_val_3_lo, v_val_3_lo);
#if defined(__aarch64__)
v_sum_01 = vmlal_high_s16(v_sum_01, v_val_0, v_val_0);
v_sum_01 = vmlal_high_s16(v_sum_01, v_val_1, v_val_1);
v_sum_23 = vmlal_high_s16(v_sum_23, v_val_2, v_val_2);
v_sum_23 = vmlal_high_s16(v_sum_23, v_val_3, v_val_3);
v_sum = vaddq_s32(v_sum, vpaddq_s32(v_sum_01, v_sum_23));
#else
const int16x4_t v_val_0_hi = vget_high_s16(v_val_0);
const int16x4_t v_val_1_hi = vget_high_s16(v_val_1);
const int16x4_t v_val_2_hi = vget_high_s16(v_val_2);
const int16x4_t v_val_3_hi = vget_high_s16(v_val_3);
v_sum_01 = vmlal_s16(v_sum_01, v_val_0_hi, v_val_0_hi);
v_sum_01 = vmlal_s16(v_sum_01, v_val_1_hi, v_val_1_hi);
v_sum_23 = vmlal_s16(v_sum_23, v_val_2_hi, v_val_2_hi);
v_sum_23 = vmlal_s16(v_sum_23, v_val_3_hi, v_val_3_hi);
v_sum = vaddq_s32(v_sum, vcombine_s32(vqmovn_s64(vpaddlq_s32(v_sum_01)),
vqmovn_s64(vpaddlq_s32(v_sum_23))));
#endif
c += 8;
} while (c < width);
v_acc_q = vpadalq_u32(v_acc_q, vreinterpretq_u32_s32(v_sum));
src += 4 * stride;
r += 4;
} while (r < height);
#if defined(__aarch64__)
return vaddvq_u64(v_acc_q);
#else
v_acc_q = vaddq_u64(v_acc_q, vextq_u64(v_acc_q, v_acc_q, 1));
return vgetq_lane_u64(v_acc_q, 0);
#endif
}
uint64_t aom_sum_squares_2d_i16_neon(const int16_t *src, int stride, int width,
int height) {
// 4 elements per row only requires half an SIMD register, so this
// must be a special case, but also note that over 75% of all calls
// are with size == 4, so it is also the common case.
if (LIKELY(width == 4 && height == 4)) {
return aom_sum_squares_2d_i16_4x4_neon(src, stride);
} else if (LIKELY(width == 4 && (height & 3) == 0)) {
return aom_sum_squares_2d_i16_4xn_neon(src, stride, height);
} else if (LIKELY((width & 7) == 0 && (height & 3) == 0)) {
// Generic case
return aom_sum_squares_2d_i16_nxn_neon(src, stride, width, height);
} else {
return aom_sum_squares_2d_i16_c(src, stride, width, height);
}
}