|  | /* | 
|  | * Copyright (c) 2024, Alliance for Open Media. All rights reserved. | 
|  | * | 
|  | * This source code is subject to the terms of the BSD 2 Clause License and | 
|  | * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License | 
|  | * was not distributed with this source code in the LICENSE file, you can | 
|  | * obtain it at www.aomedia.org/license/software. If the Alliance for Open | 
|  | * Media Patent License 1.0 was not distributed with this source code in the | 
|  | * PATENTS file, you can obtain it at www.aomedia.org/license/patent. | 
|  | */ | 
|  |  | 
|  | #ifndef AOM_AOM_DSP_ARM_AOM_CONVOLVE8_NEON_H_ | 
|  | #define AOM_AOM_DSP_ARM_AOM_CONVOLVE8_NEON_H_ | 
|  |  | 
|  | #include <arm_neon.h> | 
|  |  | 
|  | #include "aom_dsp/aom_filter.h" | 
|  | #include "aom_dsp/arm/mem_neon.h" | 
|  | #include "config/aom_config.h" | 
|  |  | 
|  | static inline int16x4_t convolve8_4(const int16x4_t s0, const int16x4_t s1, | 
|  | const int16x4_t s2, const int16x4_t s3, | 
|  | const int16x4_t s4, const int16x4_t s5, | 
|  | const int16x4_t s6, const int16x4_t s7, | 
|  | const int16x8_t filter) { | 
|  | const int16x4_t filter_lo = vget_low_s16(filter); | 
|  | const int16x4_t filter_hi = vget_high_s16(filter); | 
|  |  | 
|  | int16x4_t sum = vmul_lane_s16(s0, filter_lo, 0); | 
|  | sum = vmla_lane_s16(sum, s1, filter_lo, 1); | 
|  | sum = vmla_lane_s16(sum, s2, filter_lo, 2); | 
|  | sum = vmla_lane_s16(sum, s3, filter_lo, 3); | 
|  | sum = vmla_lane_s16(sum, s4, filter_hi, 0); | 
|  | sum = vmla_lane_s16(sum, s5, filter_hi, 1); | 
|  | sum = vmla_lane_s16(sum, s6, filter_hi, 2); | 
|  | sum = vmla_lane_s16(sum, s7, filter_hi, 3); | 
|  |  | 
|  | return sum; | 
|  | } | 
|  |  | 
|  | static inline uint8x8_t convolve8_8(const int16x8_t s0, const int16x8_t s1, | 
|  | const int16x8_t s2, const int16x8_t s3, | 
|  | const int16x8_t s4, const int16x8_t s5, | 
|  | const int16x8_t s6, const int16x8_t s7, | 
|  | const int16x8_t filter) { | 
|  | const int16x4_t filter_lo = vget_low_s16(filter); | 
|  | const int16x4_t filter_hi = vget_high_s16(filter); | 
|  |  | 
|  | int16x8_t sum = vmulq_lane_s16(s0, filter_lo, 0); | 
|  | sum = vmlaq_lane_s16(sum, s1, filter_lo, 1); | 
|  | sum = vmlaq_lane_s16(sum, s2, filter_lo, 2); | 
|  | sum = vmlaq_lane_s16(sum, s3, filter_lo, 3); | 
|  | sum = vmlaq_lane_s16(sum, s4, filter_hi, 0); | 
|  | sum = vmlaq_lane_s16(sum, s5, filter_hi, 1); | 
|  | sum = vmlaq_lane_s16(sum, s6, filter_hi, 2); | 
|  | sum = vmlaq_lane_s16(sum, s7, filter_hi, 3); | 
|  |  | 
|  | // We halved the filter values so -1 from right shift. | 
|  | return vqrshrun_n_s16(sum, FILTER_BITS - 1); | 
|  | } | 
|  |  | 
|  | static inline void convolve8_horiz_2tap_neon(const uint8_t *src, | 
|  | ptrdiff_t src_stride, uint8_t *dst, | 
|  | ptrdiff_t dst_stride, | 
|  | const int16_t *filter_x, int w, | 
|  | int h) { | 
|  | // Bilinear filter values are all positive. | 
|  | const uint8x8_t f0 = vdup_n_u8((uint8_t)filter_x[3]); | 
|  | const uint8x8_t f1 = vdup_n_u8((uint8_t)filter_x[4]); | 
|  |  | 
|  | if (w == 4) { | 
|  | do { | 
|  | uint8x8_t s0 = | 
|  | load_unaligned_u8(src + 0 * src_stride + 0, (int)src_stride); | 
|  | uint8x8_t s1 = | 
|  | load_unaligned_u8(src + 0 * src_stride + 1, (int)src_stride); | 
|  | uint8x8_t s2 = | 
|  | load_unaligned_u8(src + 2 * src_stride + 0, (int)src_stride); | 
|  | uint8x8_t s3 = | 
|  | load_unaligned_u8(src + 2 * src_stride + 1, (int)src_stride); | 
|  |  | 
|  | uint16x8_t sum0 = vmull_u8(s0, f0); | 
|  | sum0 = vmlal_u8(sum0, s1, f1); | 
|  | uint16x8_t sum1 = vmull_u8(s2, f0); | 
|  | sum1 = vmlal_u8(sum1, s3, f1); | 
|  |  | 
|  | uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); | 
|  | uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); | 
|  |  | 
|  | store_u8x4_strided_x2(dst + 0 * dst_stride, dst_stride, d0); | 
|  | store_u8x4_strided_x2(dst + 2 * dst_stride, dst_stride, d1); | 
|  |  | 
|  | src += 4 * src_stride; | 
|  | dst += 4 * dst_stride; | 
|  | h -= 4; | 
|  | } while (h > 0); | 
|  | } else if (w == 8) { | 
|  | do { | 
|  | uint8x8_t s0 = vld1_u8(src + 0 * src_stride + 0); | 
|  | uint8x8_t s1 = vld1_u8(src + 0 * src_stride + 1); | 
|  | uint8x8_t s2 = vld1_u8(src + 1 * src_stride + 0); | 
|  | uint8x8_t s3 = vld1_u8(src + 1 * src_stride + 1); | 
|  |  | 
|  | uint16x8_t sum0 = vmull_u8(s0, f0); | 
|  | sum0 = vmlal_u8(sum0, s1, f1); | 
|  | uint16x8_t sum1 = vmull_u8(s2, f0); | 
|  | sum1 = vmlal_u8(sum1, s3, f1); | 
|  |  | 
|  | uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); | 
|  | uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); | 
|  |  | 
|  | vst1_u8(dst + 0 * dst_stride, d0); | 
|  | vst1_u8(dst + 1 * dst_stride, d1); | 
|  |  | 
|  | src += 2 * src_stride; | 
|  | dst += 2 * dst_stride; | 
|  | h -= 2; | 
|  | } while (h > 0); | 
|  | } else { | 
|  | do { | 
|  | int width = w; | 
|  | const uint8_t *s = src; | 
|  | uint8_t *d = dst; | 
|  |  | 
|  | do { | 
|  | uint8x16_t s0 = vld1q_u8(s + 0); | 
|  | uint8x16_t s1 = vld1q_u8(s + 1); | 
|  |  | 
|  | uint16x8_t sum0 = vmull_u8(vget_low_u8(s0), f0); | 
|  | sum0 = vmlal_u8(sum0, vget_low_u8(s1), f1); | 
|  | uint16x8_t sum1 = vmull_u8(vget_high_u8(s0), f0); | 
|  | sum1 = vmlal_u8(sum1, vget_high_u8(s1), f1); | 
|  |  | 
|  | uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); | 
|  | uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); | 
|  |  | 
|  | vst1q_u8(d, vcombine_u8(d0, d1)); | 
|  |  | 
|  | s += 16; | 
|  | d += 16; | 
|  | width -= 16; | 
|  | } while (width != 0); | 
|  | src += src_stride; | 
|  | dst += dst_stride; | 
|  | } while (--h > 0); | 
|  | } | 
|  | } | 
|  |  | 
|  | static inline uint8x8_t convolve4_8(const int16x8_t s0, const int16x8_t s1, | 
|  | const int16x8_t s2, const int16x8_t s3, | 
|  | const int16x4_t filter) { | 
|  | int16x8_t sum = vmulq_lane_s16(s0, filter, 0); | 
|  | sum = vmlaq_lane_s16(sum, s1, filter, 1); | 
|  | sum = vmlaq_lane_s16(sum, s2, filter, 2); | 
|  | sum = vmlaq_lane_s16(sum, s3, filter, 3); | 
|  |  | 
|  | // We halved the filter values so -1 from right shift. | 
|  | return vqrshrun_n_s16(sum, FILTER_BITS - 1); | 
|  | } | 
|  |  | 
|  | static inline void convolve8_vert_4tap_neon(const uint8_t *src, | 
|  | ptrdiff_t src_stride, uint8_t *dst, | 
|  | ptrdiff_t dst_stride, | 
|  | const int16_t *filter_y, int w, | 
|  | int h) { | 
|  | // All filter values are even, halve to reduce intermediate precision | 
|  | // requirements. | 
|  | const int16x4_t filter = vshr_n_s16(vld1_s16(filter_y + 2), 1); | 
|  |  | 
|  | if (w == 4) { | 
|  | uint8x8_t t01 = load_unaligned_u8(src + 0 * src_stride, (int)src_stride); | 
|  | uint8x8_t t12 = load_unaligned_u8(src + 1 * src_stride, (int)src_stride); | 
|  |  | 
|  | int16x8_t s01 = vreinterpretq_s16_u16(vmovl_u8(t01)); | 
|  | int16x8_t s12 = vreinterpretq_s16_u16(vmovl_u8(t12)); | 
|  |  | 
|  | src += 2 * src_stride; | 
|  |  | 
|  | do { | 
|  | uint8x8_t t23 = load_unaligned_u8(src + 0 * src_stride, (int)src_stride); | 
|  | uint8x8_t t34 = load_unaligned_u8(src + 1 * src_stride, (int)src_stride); | 
|  | uint8x8_t t45 = load_unaligned_u8(src + 2 * src_stride, (int)src_stride); | 
|  | uint8x8_t t56 = load_unaligned_u8(src + 3 * src_stride, (int)src_stride); | 
|  |  | 
|  | int16x8_t s23 = vreinterpretq_s16_u16(vmovl_u8(t23)); | 
|  | int16x8_t s34 = vreinterpretq_s16_u16(vmovl_u8(t34)); | 
|  | int16x8_t s45 = vreinterpretq_s16_u16(vmovl_u8(t45)); | 
|  | int16x8_t s56 = vreinterpretq_s16_u16(vmovl_u8(t56)); | 
|  |  | 
|  | uint8x8_t d01 = convolve4_8(s01, s12, s23, s34, filter); | 
|  | uint8x8_t d23 = convolve4_8(s23, s34, s45, s56, filter); | 
|  |  | 
|  | store_u8x4_strided_x2(dst + 0 * dst_stride, dst_stride, d01); | 
|  | store_u8x4_strided_x2(dst + 2 * dst_stride, dst_stride, d23); | 
|  |  | 
|  | s01 = s45; | 
|  | s12 = s56; | 
|  |  | 
|  | src += 4 * src_stride; | 
|  | dst += 4 * dst_stride; | 
|  | h -= 4; | 
|  | } while (h != 0); | 
|  | } else { | 
|  | do { | 
|  | uint8x8_t t0, t1, t2; | 
|  | load_u8_8x3(src, src_stride, &t0, &t1, &t2); | 
|  |  | 
|  | int16x8_t s0 = vreinterpretq_s16_u16(vmovl_u8(t0)); | 
|  | int16x8_t s1 = vreinterpretq_s16_u16(vmovl_u8(t1)); | 
|  | int16x8_t s2 = vreinterpretq_s16_u16(vmovl_u8(t2)); | 
|  |  | 
|  | int height = h; | 
|  | const uint8_t *s = src + 3 * src_stride; | 
|  | uint8_t *d = dst; | 
|  |  | 
|  | do { | 
|  | uint8x8_t t3; | 
|  | load_u8_8x4(s, src_stride, &t0, &t1, &t2, &t3); | 
|  |  | 
|  | int16x8_t s3 = vreinterpretq_s16_u16(vmovl_u8(t0)); | 
|  | int16x8_t s4 = vreinterpretq_s16_u16(vmovl_u8(t1)); | 
|  | int16x8_t s5 = vreinterpretq_s16_u16(vmovl_u8(t2)); | 
|  | int16x8_t s6 = vreinterpretq_s16_u16(vmovl_u8(t3)); | 
|  |  | 
|  | uint8x8_t d0 = convolve4_8(s0, s1, s2, s3, filter); | 
|  | uint8x8_t d1 = convolve4_8(s1, s2, s3, s4, filter); | 
|  | uint8x8_t d2 = convolve4_8(s2, s3, s4, s5, filter); | 
|  | uint8x8_t d3 = convolve4_8(s3, s4, s5, s6, filter); | 
|  |  | 
|  | store_u8_8x4(d, dst_stride, d0, d1, d2, d3); | 
|  |  | 
|  | s0 = s4; | 
|  | s1 = s5; | 
|  | s2 = s6; | 
|  |  | 
|  | s += 4 * src_stride; | 
|  | d += 4 * dst_stride; | 
|  | height -= 4; | 
|  | } while (height != 0); | 
|  | src += 8; | 
|  | dst += 8; | 
|  | w -= 8; | 
|  | } while (w != 0); | 
|  | } | 
|  | } | 
|  |  | 
|  | static inline void convolve8_vert_2tap_neon(const uint8_t *src, | 
|  | ptrdiff_t src_stride, uint8_t *dst, | 
|  | ptrdiff_t dst_stride, | 
|  | const int16_t *filter_y, int w, | 
|  | int h) { | 
|  | // Bilinear filter values are all positive. | 
|  | uint8x8_t f0 = vdup_n_u8((uint8_t)filter_y[3]); | 
|  | uint8x8_t f1 = vdup_n_u8((uint8_t)filter_y[4]); | 
|  |  | 
|  | if (w == 4) { | 
|  | do { | 
|  | uint8x8_t s0 = load_unaligned_u8(src + 0 * src_stride, (int)src_stride); | 
|  | uint8x8_t s1 = load_unaligned_u8(src + 1 * src_stride, (int)src_stride); | 
|  | uint8x8_t s2 = load_unaligned_u8(src + 2 * src_stride, (int)src_stride); | 
|  | uint8x8_t s3 = load_unaligned_u8(src + 3 * src_stride, (int)src_stride); | 
|  |  | 
|  | uint16x8_t sum0 = vmull_u8(s0, f0); | 
|  | sum0 = vmlal_u8(sum0, s1, f1); | 
|  | uint16x8_t sum1 = vmull_u8(s2, f0); | 
|  | sum1 = vmlal_u8(sum1, s3, f1); | 
|  |  | 
|  | uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); | 
|  | uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); | 
|  |  | 
|  | store_u8x4_strided_x2(dst + 0 * dst_stride, dst_stride, d0); | 
|  | store_u8x4_strided_x2(dst + 2 * dst_stride, dst_stride, d1); | 
|  |  | 
|  | src += 4 * src_stride; | 
|  | dst += 4 * dst_stride; | 
|  | h -= 4; | 
|  | } while (h > 0); | 
|  | } else if (w == 8) { | 
|  | do { | 
|  | uint8x8_t s0, s1, s2; | 
|  | load_u8_8x3(src, src_stride, &s0, &s1, &s2); | 
|  |  | 
|  | uint16x8_t sum0 = vmull_u8(s0, f0); | 
|  | sum0 = vmlal_u8(sum0, s1, f1); | 
|  | uint16x8_t sum1 = vmull_u8(s1, f0); | 
|  | sum1 = vmlal_u8(sum1, s2, f1); | 
|  |  | 
|  | uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); | 
|  | uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); | 
|  |  | 
|  | vst1_u8(dst + 0 * dst_stride, d0); | 
|  | vst1_u8(dst + 1 * dst_stride, d1); | 
|  |  | 
|  | src += 2 * src_stride; | 
|  | dst += 2 * dst_stride; | 
|  | h -= 2; | 
|  | } while (h > 0); | 
|  | } else { | 
|  | do { | 
|  | int width = w; | 
|  | const uint8_t *s = src; | 
|  | uint8_t *d = dst; | 
|  |  | 
|  | do { | 
|  | uint8x16_t s0 = vld1q_u8(s + 0 * src_stride); | 
|  | uint8x16_t s1 = vld1q_u8(s + 1 * src_stride); | 
|  |  | 
|  | uint16x8_t sum0 = vmull_u8(vget_low_u8(s0), f0); | 
|  | sum0 = vmlal_u8(sum0, vget_low_u8(s1), f1); | 
|  | uint16x8_t sum1 = vmull_u8(vget_high_u8(s0), f0); | 
|  | sum1 = vmlal_u8(sum1, vget_high_u8(s1), f1); | 
|  |  | 
|  | uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); | 
|  | uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); | 
|  |  | 
|  | vst1q_u8(d, vcombine_u8(d0, d1)); | 
|  |  | 
|  | s += 16; | 
|  | d += 16; | 
|  | width -= 16; | 
|  | } while (width != 0); | 
|  | src += src_stride; | 
|  | dst += dst_stride; | 
|  | } while (--h > 0); | 
|  | } | 
|  | } | 
|  |  | 
|  | #endif  // AOM_AOM_DSP_ARM_AOM_CONVOLVE8_NEON_H_ |