| /* | 
 |  * Copyright (c) 2023, Alliance for Open Media. All rights reserved. | 
 |  * | 
 |  * This source code is subject to the terms of the BSD 2 Clause License and | 
 |  * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License | 
 |  * was not distributed with this source code in the LICENSE file, you can | 
 |  * obtain it at www.aomedia.org/license/software. If the Alliance for Open | 
 |  * Media Patent License 1.0 was not distributed with this source code in the | 
 |  * PATENTS file, you can obtain it at www.aomedia.org/license/patent. | 
 |  */ | 
 |  | 
 | #include <arm_neon.h> | 
 | #include <assert.h> | 
 |  | 
 | #include "config/aom_dsp_rtcd.h" | 
 | #include "config/aom_config.h" | 
 |  | 
 | #include "aom_dsp/arm/aom_neon_sve_bridge.h" | 
 | #include "aom_dsp/arm/mem_neon.h" | 
 |  | 
 | static inline void get_blk_sse_sum_4xh_sve(const int16_t *data, int stride, | 
 |                                            int bh, int *x_sum, | 
 |                                            int64_t *x2_sum) { | 
 |   int32x4_t sum = vdupq_n_s32(0); | 
 |   int64x2_t sse = vdupq_n_s64(0); | 
 |  | 
 |   do { | 
 |     int16x8_t d = vcombine_s16(vld1_s16(data), vld1_s16(data + stride)); | 
 |  | 
 |     sum = vpadalq_s16(sum, d); | 
 |  | 
 |     sse = aom_sdotq_s16(sse, d, d); | 
 |  | 
 |     data += 2 * stride; | 
 |     bh -= 2; | 
 |   } while (bh != 0); | 
 |  | 
 |   *x_sum = vaddvq_s32(sum); | 
 |   *x2_sum = vaddvq_s64(sse); | 
 | } | 
 |  | 
 | static inline void get_blk_sse_sum_8xh_sve(const int16_t *data, int stride, | 
 |                                            int bh, int *x_sum, | 
 |                                            int64_t *x2_sum) { | 
 |   int32x4_t sum[2] = { vdupq_n_s32(0), vdupq_n_s32(0) }; | 
 |   int64x2_t sse[2] = { vdupq_n_s64(0), vdupq_n_s64(0) }; | 
 |  | 
 |   do { | 
 |     int16x8_t d0 = vld1q_s16(data); | 
 |     int16x8_t d1 = vld1q_s16(data + stride); | 
 |  | 
 |     sum[0] = vpadalq_s16(sum[0], d0); | 
 |     sum[1] = vpadalq_s16(sum[1], d1); | 
 |  | 
 |     sse[0] = aom_sdotq_s16(sse[0], d0, d0); | 
 |     sse[1] = aom_sdotq_s16(sse[1], d1, d1); | 
 |  | 
 |     data += 2 * stride; | 
 |     bh -= 2; | 
 |   } while (bh != 0); | 
 |  | 
 |   *x_sum = vaddvq_s32(vaddq_s32(sum[0], sum[1])); | 
 |   *x2_sum = vaddvq_s64(vaddq_s64(sse[0], sse[1])); | 
 | } | 
 |  | 
 | static inline void get_blk_sse_sum_large_sve(const int16_t *data, int stride, | 
 |                                              int bw, int bh, int *x_sum, | 
 |                                              int64_t *x2_sum) { | 
 |   int32x4_t sum[2] = { vdupq_n_s32(0), vdupq_n_s32(0) }; | 
 |   int64x2_t sse[2] = { vdupq_n_s64(0), vdupq_n_s64(0) }; | 
 |  | 
 |   do { | 
 |     int j = bw; | 
 |     const int16_t *data_ptr = data; | 
 |     do { | 
 |       int16x8_t d0 = vld1q_s16(data_ptr); | 
 |       int16x8_t d1 = vld1q_s16(data_ptr + 8); | 
 |  | 
 |       sum[0] = vpadalq_s16(sum[0], d0); | 
 |       sum[1] = vpadalq_s16(sum[1], d1); | 
 |  | 
 |       sse[0] = aom_sdotq_s16(sse[0], d0, d0); | 
 |       sse[1] = aom_sdotq_s16(sse[1], d1, d1); | 
 |  | 
 |       data_ptr += 16; | 
 |       j -= 16; | 
 |     } while (j != 0); | 
 |  | 
 |     data += stride; | 
 |   } while (--bh != 0); | 
 |  | 
 |   *x_sum = vaddvq_s32(vaddq_s32(sum[0], sum[1])); | 
 |   *x2_sum = vaddvq_s64(vaddq_s64(sse[0], sse[1])); | 
 | } | 
 |  | 
 | void aom_get_blk_sse_sum_sve(const int16_t *data, int stride, int bw, int bh, | 
 |                              int *x_sum, int64_t *x2_sum) { | 
 |   if (bw == 4) { | 
 |     get_blk_sse_sum_4xh_sve(data, stride, bh, x_sum, x2_sum); | 
 |   } else if (bw == 8) { | 
 |     get_blk_sse_sum_8xh_sve(data, stride, bh, x_sum, x2_sum); | 
 |   } else { | 
 |     assert(bw % 16 == 0); | 
 |     get_blk_sse_sum_large_sve(data, stride, bw, bh, x_sum, x2_sum); | 
 |   } | 
 | } |