| /* |
| * Copyright (c) 2023, Alliance for Open Media. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| #ifndef AOM_AV1_COMMON_ARM_HIGHBD_CONVOLVE_NEON_H_ |
| #define AOM_AV1_COMMON_ARM_HIGHBD_CONVOLVE_NEON_H_ |
| |
| #include <arm_neon.h> |
| |
| static INLINE int32x4_t highbd_convolve8_4_s32( |
| const int16x4_t s0, const int16x4_t s1, const int16x4_t s2, |
| const int16x4_t s3, const int16x4_t s4, const int16x4_t s5, |
| const int16x4_t s6, const int16x4_t s7, const int16x8_t y_filter) { |
| const int16x4_t y_filter_lo = vget_low_s16(y_filter); |
| const int16x4_t y_filter_hi = vget_high_s16(y_filter); |
| |
| int32x4_t sum = vmull_lane_s16(s0, y_filter_lo, 0); |
| sum = vmlal_lane_s16(sum, s1, y_filter_lo, 1); |
| sum = vmlal_lane_s16(sum, s2, y_filter_lo, 2); |
| sum = vmlal_lane_s16(sum, s3, y_filter_lo, 3); |
| sum = vmlal_lane_s16(sum, s4, y_filter_hi, 0); |
| sum = vmlal_lane_s16(sum, s5, y_filter_hi, 1); |
| sum = vmlal_lane_s16(sum, s6, y_filter_hi, 2); |
| sum = vmlal_lane_s16(sum, s7, y_filter_hi, 3); |
| |
| return sum; |
| } |
| |
| static INLINE uint16x4_t highbd_convolve8_4_s32_s16( |
| const int16x4_t s0, const int16x4_t s1, const int16x4_t s2, |
| const int16x4_t s3, const int16x4_t s4, const int16x4_t s5, |
| const int16x4_t s6, const int16x4_t s7, const int16x8_t y_filter) { |
| int32x4_t sum = |
| highbd_convolve8_4_s32(s0, s1, s2, s3, s4, s5, s6, s7, y_filter); |
| |
| return vqrshrun_n_s32(sum, COMPOUND_ROUND1_BITS); |
| } |
| |
| static INLINE void highbd_convolve8_8_s32( |
| const int16x8_t s0, const int16x8_t s1, const int16x8_t s2, |
| const int16x8_t s3, const int16x8_t s4, const int16x8_t s5, |
| const int16x8_t s6, const int16x8_t s7, const int16x8_t y_filter, |
| int32x4_t *sum0, int32x4_t *sum1) { |
| const int16x4_t y_filter_lo = vget_low_s16(y_filter); |
| const int16x4_t y_filter_hi = vget_high_s16(y_filter); |
| |
| *sum0 = vmull_lane_s16(vget_low_s16(s0), y_filter_lo, 0); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s1), y_filter_lo, 1); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s2), y_filter_lo, 2); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s3), y_filter_lo, 3); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s4), y_filter_hi, 0); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s5), y_filter_hi, 1); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s6), y_filter_hi, 2); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s7), y_filter_hi, 3); |
| |
| *sum1 = vmull_lane_s16(vget_high_s16(s0), y_filter_lo, 0); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s1), y_filter_lo, 1); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s2), y_filter_lo, 2); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s3), y_filter_lo, 3); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s4), y_filter_hi, 0); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s5), y_filter_hi, 1); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s6), y_filter_hi, 2); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s7), y_filter_hi, 3); |
| } |
| |
| static INLINE uint16x8_t highbd_convolve8_8_s32_s16( |
| const int16x8_t s0, const int16x8_t s1, const int16x8_t s2, |
| const int16x8_t s3, const int16x8_t s4, const int16x8_t s5, |
| const int16x8_t s6, const int16x8_t s7, const int16x8_t y_filter) { |
| int32x4_t sum0; |
| int32x4_t sum1; |
| highbd_convolve8_8_s32(s0, s1, s2, s3, s4, s5, s6, s7, y_filter, &sum0, |
| &sum1); |
| |
| return vcombine_u16(vqrshrun_n_s32(sum0, COMPOUND_ROUND1_BITS), |
| vqrshrun_n_s32(sum1, COMPOUND_ROUND1_BITS)); |
| } |
| |
| static INLINE int32x4_t highbd_convolve12_y_4x4_s32( |
| const int16x4_t s0, const int16x4_t s1, const int16x4_t s2, |
| const int16x4_t s3, const int16x4_t s4, const int16x4_t s5, |
| const int16x4_t s6, const int16x4_t s7, const int16x4_t s8, |
| const int16x4_t s9, const int16x4_t s10, const int16x4_t s11, |
| const int16x8_t y_filter_0_7, const int16x4_t y_filter_8_11) { |
| const int16x4_t y_filter_0_3 = vget_low_s16(y_filter_0_7); |
| const int16x4_t y_filter_4_7 = vget_high_s16(y_filter_0_7); |
| |
| int32x4_t sum = vmull_lane_s16(s0, y_filter_0_3, 0); |
| sum = vmlal_lane_s16(sum, s1, y_filter_0_3, 1); |
| sum = vmlal_lane_s16(sum, s2, y_filter_0_3, 2); |
| sum = vmlal_lane_s16(sum, s3, y_filter_0_3, 3); |
| sum = vmlal_lane_s16(sum, s4, y_filter_4_7, 0); |
| sum = vmlal_lane_s16(sum, s5, y_filter_4_7, 1); |
| sum = vmlal_lane_s16(sum, s6, y_filter_4_7, 2); |
| sum = vmlal_lane_s16(sum, s7, y_filter_4_7, 3); |
| sum = vmlal_lane_s16(sum, s8, y_filter_8_11, 0); |
| sum = vmlal_lane_s16(sum, s9, y_filter_8_11, 1); |
| sum = vmlal_lane_s16(sum, s10, y_filter_8_11, 2); |
| sum = vmlal_lane_s16(sum, s11, y_filter_8_11, 3); |
| |
| return sum; |
| } |
| |
| static INLINE uint16x4_t highbd_convolve12_y_4x4_s32_s16( |
| const int16x4_t s0, const int16x4_t s1, const int16x4_t s2, |
| const int16x4_t s3, const int16x4_t s4, const int16x4_t s5, |
| const int16x4_t s6, const int16x4_t s7, const int16x4_t s8, |
| const int16x4_t s9, const int16x4_t s10, const int16x4_t s11, |
| const int16x8_t y_filter_0_7, const int16x4_t y_filter_8_11) { |
| int32x4_t sum = |
| highbd_convolve12_y_4x4_s32(s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, |
| s11, y_filter_0_7, y_filter_8_11); |
| |
| return vqrshrun_n_s32(sum, FILTER_BITS); |
| } |
| |
| static INLINE void highbd_convolve12_y_8x4_s32( |
| const int16x8_t s0, const int16x8_t s1, const int16x8_t s2, |
| const int16x8_t s3, const int16x8_t s4, const int16x8_t s5, |
| const int16x8_t s6, const int16x8_t s7, const int16x8_t s8, |
| const int16x8_t s9, const int16x8_t s10, const int16x8_t s11, |
| const int16x8_t y_filter_0_7, const int16x4_t y_filter_8_11, |
| int32x4_t *sum0, int32x4_t *sum1) { |
| const int16x4_t y_filter_0_3 = vget_low_s16(y_filter_0_7); |
| const int16x4_t y_filter_4_7 = vget_high_s16(y_filter_0_7); |
| |
| *sum0 = vmull_lane_s16(vget_low_s16(s0), y_filter_0_3, 0); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s1), y_filter_0_3, 1); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s2), y_filter_0_3, 2); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s3), y_filter_0_3, 3); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s4), y_filter_4_7, 0); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s5), y_filter_4_7, 1); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s6), y_filter_4_7, 2); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s7), y_filter_4_7, 3); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s8), y_filter_8_11, 0); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s9), y_filter_8_11, 1); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s10), y_filter_8_11, 2); |
| *sum0 = vmlal_lane_s16(*sum0, vget_low_s16(s11), y_filter_8_11, 3); |
| |
| *sum1 = vmull_lane_s16(vget_high_s16(s0), y_filter_0_3, 0); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s1), y_filter_0_3, 1); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s2), y_filter_0_3, 2); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s3), y_filter_0_3, 3); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s4), y_filter_4_7, 0); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s5), y_filter_4_7, 1); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s6), y_filter_4_7, 2); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s7), y_filter_4_7, 3); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s8), y_filter_8_11, 0); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s9), y_filter_8_11, 1); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s10), y_filter_8_11, 2); |
| *sum1 = vmlal_lane_s16(*sum1, vget_high_s16(s11), y_filter_8_11, 3); |
| } |
| |
| static INLINE uint16x8_t highbd_convolve12_y_8x4_s32_s16( |
| const int16x8_t s0, const int16x8_t s1, const int16x8_t s2, |
| const int16x8_t s3, const int16x8_t s4, const int16x8_t s5, |
| const int16x8_t s6, const int16x8_t s7, const int16x8_t s8, |
| const int16x8_t s9, const int16x8_t s10, const int16x8_t s11, |
| const int16x8_t y_filter_0_7, const int16x4_t y_filter_8_11) { |
| int32x4_t sum0; |
| int32x4_t sum1; |
| highbd_convolve12_y_8x4_s32(s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, |
| y_filter_0_7, y_filter_8_11, &sum0, &sum1); |
| |
| return vcombine_u16(vqrshrun_n_s32(sum0, FILTER_BITS), |
| vqrshrun_n_s32(sum1, FILTER_BITS)); |
| } |
| |
| #endif // AOM_AV1_COMMON_ARM_HIGHBD_CONVOLVE_NEON_H_ |