blob: 4bf7ae6ce4a8a0e0340800362f5ad75f74361bf4 [file] [log] [blame]
/*
* Copyright (c) 2023, Alliance for Open Media. All rights reserved
*
* This source code is subject to the terms of the BSD 2 Clause License and
* the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License
* was not distributed with this source code in the LICENSE file, you can
* obtain it at www.aomedia.org/license/software. If the Alliance for Open
* Media Patent License 1.0 was not distributed with this source code in the
* PATENTS file, you can obtain it at www.aomedia.org/license/patent.
*/
#include <assert.h>
#include <arm_neon.h>
#include "config/av1_rtcd.h"
#include "aom_dsp/arm/sum_neon.h"
int64_t av1_highbd_block_error_neon(const tran_low_t *coeff,
const tran_low_t *dqcoeff,
intptr_t block_size, int64_t *ssz, int bd) {
uint64x2_t err_u64 = vdupq_n_u64(0);
int64x2_t ssz_s64 = vdupq_n_s64(0);
const int shift = 2 * (bd - 8);
const int rounding = shift > 0 ? 1 << (shift - 1) : 0;
assert(block_size >= 16);
assert((block_size % 16) == 0);
do {
const int32x4_t c = vld1q_s32(coeff);
const int32x4_t d = vld1q_s32(dqcoeff);
const uint32x4_t diff = vreinterpretq_u32_s32(vabdq_s32(c, d));
err_u64 = vmlal_u32(err_u64, vget_low_u32(diff), vget_low_u32(diff));
err_u64 = vmlal_u32(err_u64, vget_high_u32(diff), vget_high_u32(diff));
ssz_s64 = vmlal_s32(ssz_s64, vget_low_s32(c), vget_low_s32(c));
ssz_s64 = vmlal_s32(ssz_s64, vget_high_s32(c), vget_high_s32(c));
coeff += 4;
dqcoeff += 4;
block_size -= 4;
} while (block_size != 0);
*ssz = (horizontal_add_s64x2(ssz_s64) + rounding) >> shift;
return ((int64_t)horizontal_add_u64x2(err_u64) + rounding) >> shift;
}