| /* |
| * Copyright (c) 2024, Alliance for Open Media. All rights reserved. |
| * |
| * This source code is subject to the terms of the BSD 2 Clause License and |
| * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License |
| * was not distributed with this source code in the LICENSE file, you can |
| * obtain it at www.aomedia.org/license/software. If the Alliance for Open |
| * Media Patent License 1.0 was not distributed with this source code in the |
| * PATENTS file, you can obtain it at www.aomedia.org/license/patent. |
| */ |
| |
| #ifndef AOM_AOM_DSP_ARM_AOM_CONVOLVE8_NEON_H_ |
| #define AOM_AOM_DSP_ARM_AOM_CONVOLVE8_NEON_H_ |
| |
| #include <arm_neon.h> |
| |
| #include "aom_dsp/aom_filter.h" |
| #include "aom_dsp/arm/mem_neon.h" |
| #include "config/aom_config.h" |
| |
| static inline int16x4_t convolve8_4(const int16x4_t s0, const int16x4_t s1, |
| const int16x4_t s2, const int16x4_t s3, |
| const int16x4_t s4, const int16x4_t s5, |
| const int16x4_t s6, const int16x4_t s7, |
| const int16x8_t filter) { |
| const int16x4_t filter_lo = vget_low_s16(filter); |
| const int16x4_t filter_hi = vget_high_s16(filter); |
| |
| int16x4_t sum = vmul_lane_s16(s0, filter_lo, 0); |
| sum = vmla_lane_s16(sum, s1, filter_lo, 1); |
| sum = vmla_lane_s16(sum, s2, filter_lo, 2); |
| sum = vmla_lane_s16(sum, s3, filter_lo, 3); |
| sum = vmla_lane_s16(sum, s4, filter_hi, 0); |
| sum = vmla_lane_s16(sum, s5, filter_hi, 1); |
| sum = vmla_lane_s16(sum, s6, filter_hi, 2); |
| sum = vmla_lane_s16(sum, s7, filter_hi, 3); |
| |
| return sum; |
| } |
| |
| static inline uint8x8_t convolve8_8(const int16x8_t s0, const int16x8_t s1, |
| const int16x8_t s2, const int16x8_t s3, |
| const int16x8_t s4, const int16x8_t s5, |
| const int16x8_t s6, const int16x8_t s7, |
| const int16x8_t filter) { |
| const int16x4_t filter_lo = vget_low_s16(filter); |
| const int16x4_t filter_hi = vget_high_s16(filter); |
| |
| int16x8_t sum = vmulq_lane_s16(s0, filter_lo, 0); |
| sum = vmlaq_lane_s16(sum, s1, filter_lo, 1); |
| sum = vmlaq_lane_s16(sum, s2, filter_lo, 2); |
| sum = vmlaq_lane_s16(sum, s3, filter_lo, 3); |
| sum = vmlaq_lane_s16(sum, s4, filter_hi, 0); |
| sum = vmlaq_lane_s16(sum, s5, filter_hi, 1); |
| sum = vmlaq_lane_s16(sum, s6, filter_hi, 2); |
| sum = vmlaq_lane_s16(sum, s7, filter_hi, 3); |
| |
| // We halved the filter values so -1 from right shift. |
| return vqrshrun_n_s16(sum, FILTER_BITS - 1); |
| } |
| |
| static inline void convolve8_horiz_2tap_neon(const uint8_t *src, |
| ptrdiff_t src_stride, uint8_t *dst, |
| ptrdiff_t dst_stride, |
| const int16_t *filter_x, int w, |
| int h) { |
| // Bilinear filter values are all positive. |
| const uint8x8_t f0 = vdup_n_u8((uint8_t)filter_x[3]); |
| const uint8x8_t f1 = vdup_n_u8((uint8_t)filter_x[4]); |
| |
| if (w == 4) { |
| do { |
| uint8x8_t s0 = |
| load_unaligned_u8(src + 0 * src_stride + 0, (int)src_stride); |
| uint8x8_t s1 = |
| load_unaligned_u8(src + 0 * src_stride + 1, (int)src_stride); |
| uint8x8_t s2 = |
| load_unaligned_u8(src + 2 * src_stride + 0, (int)src_stride); |
| uint8x8_t s3 = |
| load_unaligned_u8(src + 2 * src_stride + 1, (int)src_stride); |
| |
| uint16x8_t sum0 = vmull_u8(s0, f0); |
| sum0 = vmlal_u8(sum0, s1, f1); |
| uint16x8_t sum1 = vmull_u8(s2, f0); |
| sum1 = vmlal_u8(sum1, s3, f1); |
| |
| uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); |
| uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); |
| |
| store_u8x4_strided_x2(dst + 0 * dst_stride, dst_stride, d0); |
| store_u8x4_strided_x2(dst + 2 * dst_stride, dst_stride, d1); |
| |
| src += 4 * src_stride; |
| dst += 4 * dst_stride; |
| h -= 4; |
| } while (h > 0); |
| } else if (w == 8) { |
| do { |
| uint8x8_t s0 = vld1_u8(src + 0 * src_stride + 0); |
| uint8x8_t s1 = vld1_u8(src + 0 * src_stride + 1); |
| uint8x8_t s2 = vld1_u8(src + 1 * src_stride + 0); |
| uint8x8_t s3 = vld1_u8(src + 1 * src_stride + 1); |
| |
| uint16x8_t sum0 = vmull_u8(s0, f0); |
| sum0 = vmlal_u8(sum0, s1, f1); |
| uint16x8_t sum1 = vmull_u8(s2, f0); |
| sum1 = vmlal_u8(sum1, s3, f1); |
| |
| uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); |
| uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); |
| |
| vst1_u8(dst + 0 * dst_stride, d0); |
| vst1_u8(dst + 1 * dst_stride, d1); |
| |
| src += 2 * src_stride; |
| dst += 2 * dst_stride; |
| h -= 2; |
| } while (h > 0); |
| } else { |
| do { |
| int width = w; |
| const uint8_t *s = src; |
| uint8_t *d = dst; |
| |
| do { |
| uint8x16_t s0 = vld1q_u8(s + 0); |
| uint8x16_t s1 = vld1q_u8(s + 1); |
| |
| uint16x8_t sum0 = vmull_u8(vget_low_u8(s0), f0); |
| sum0 = vmlal_u8(sum0, vget_low_u8(s1), f1); |
| uint16x8_t sum1 = vmull_u8(vget_high_u8(s0), f0); |
| sum1 = vmlal_u8(sum1, vget_high_u8(s1), f1); |
| |
| uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); |
| uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); |
| |
| vst1q_u8(d, vcombine_u8(d0, d1)); |
| |
| s += 16; |
| d += 16; |
| width -= 16; |
| } while (width != 0); |
| src += src_stride; |
| dst += dst_stride; |
| } while (--h > 0); |
| } |
| } |
| |
| static inline uint8x8_t convolve4_8(const int16x8_t s0, const int16x8_t s1, |
| const int16x8_t s2, const int16x8_t s3, |
| const int16x4_t filter) { |
| int16x8_t sum = vmulq_lane_s16(s0, filter, 0); |
| sum = vmlaq_lane_s16(sum, s1, filter, 1); |
| sum = vmlaq_lane_s16(sum, s2, filter, 2); |
| sum = vmlaq_lane_s16(sum, s3, filter, 3); |
| |
| // We halved the filter values so -1 from right shift. |
| return vqrshrun_n_s16(sum, FILTER_BITS - 1); |
| } |
| |
| static inline void convolve8_vert_4tap_neon(const uint8_t *src, |
| ptrdiff_t src_stride, uint8_t *dst, |
| ptrdiff_t dst_stride, |
| const int16_t *filter_y, int w, |
| int h) { |
| // All filter values are even, halve to reduce intermediate precision |
| // requirements. |
| const int16x4_t filter = vshr_n_s16(vld1_s16(filter_y + 2), 1); |
| |
| if (w == 4) { |
| uint8x8_t t01 = load_unaligned_u8(src + 0 * src_stride, (int)src_stride); |
| uint8x8_t t12 = load_unaligned_u8(src + 1 * src_stride, (int)src_stride); |
| |
| int16x8_t s01 = vreinterpretq_s16_u16(vmovl_u8(t01)); |
| int16x8_t s12 = vreinterpretq_s16_u16(vmovl_u8(t12)); |
| |
| src += 2 * src_stride; |
| |
| do { |
| uint8x8_t t23 = load_unaligned_u8(src + 0 * src_stride, (int)src_stride); |
| uint8x8_t t34 = load_unaligned_u8(src + 1 * src_stride, (int)src_stride); |
| uint8x8_t t45 = load_unaligned_u8(src + 2 * src_stride, (int)src_stride); |
| uint8x8_t t56 = load_unaligned_u8(src + 3 * src_stride, (int)src_stride); |
| |
| int16x8_t s23 = vreinterpretq_s16_u16(vmovl_u8(t23)); |
| int16x8_t s34 = vreinterpretq_s16_u16(vmovl_u8(t34)); |
| int16x8_t s45 = vreinterpretq_s16_u16(vmovl_u8(t45)); |
| int16x8_t s56 = vreinterpretq_s16_u16(vmovl_u8(t56)); |
| |
| uint8x8_t d01 = convolve4_8(s01, s12, s23, s34, filter); |
| uint8x8_t d23 = convolve4_8(s23, s34, s45, s56, filter); |
| |
| store_u8x4_strided_x2(dst + 0 * dst_stride, dst_stride, d01); |
| store_u8x4_strided_x2(dst + 2 * dst_stride, dst_stride, d23); |
| |
| s01 = s45; |
| s12 = s56; |
| |
| src += 4 * src_stride; |
| dst += 4 * dst_stride; |
| h -= 4; |
| } while (h != 0); |
| } else { |
| do { |
| uint8x8_t t0, t1, t2; |
| load_u8_8x3(src, src_stride, &t0, &t1, &t2); |
| |
| int16x8_t s0 = vreinterpretq_s16_u16(vmovl_u8(t0)); |
| int16x8_t s1 = vreinterpretq_s16_u16(vmovl_u8(t1)); |
| int16x8_t s2 = vreinterpretq_s16_u16(vmovl_u8(t2)); |
| |
| int height = h; |
| const uint8_t *s = src + 3 * src_stride; |
| uint8_t *d = dst; |
| |
| do { |
| uint8x8_t t3; |
| load_u8_8x4(s, src_stride, &t0, &t1, &t2, &t3); |
| |
| int16x8_t s3 = vreinterpretq_s16_u16(vmovl_u8(t0)); |
| int16x8_t s4 = vreinterpretq_s16_u16(vmovl_u8(t1)); |
| int16x8_t s5 = vreinterpretq_s16_u16(vmovl_u8(t2)); |
| int16x8_t s6 = vreinterpretq_s16_u16(vmovl_u8(t3)); |
| |
| uint8x8_t d0 = convolve4_8(s0, s1, s2, s3, filter); |
| uint8x8_t d1 = convolve4_8(s1, s2, s3, s4, filter); |
| uint8x8_t d2 = convolve4_8(s2, s3, s4, s5, filter); |
| uint8x8_t d3 = convolve4_8(s3, s4, s5, s6, filter); |
| |
| store_u8_8x4(d, dst_stride, d0, d1, d2, d3); |
| |
| s0 = s4; |
| s1 = s5; |
| s2 = s6; |
| |
| s += 4 * src_stride; |
| d += 4 * dst_stride; |
| height -= 4; |
| } while (height != 0); |
| src += 8; |
| dst += 8; |
| w -= 8; |
| } while (w != 0); |
| } |
| } |
| |
| static inline void convolve8_vert_2tap_neon(const uint8_t *src, |
| ptrdiff_t src_stride, uint8_t *dst, |
| ptrdiff_t dst_stride, |
| const int16_t *filter_y, int w, |
| int h) { |
| // Bilinear filter values are all positive. |
| uint8x8_t f0 = vdup_n_u8((uint8_t)filter_y[3]); |
| uint8x8_t f1 = vdup_n_u8((uint8_t)filter_y[4]); |
| |
| if (w == 4) { |
| do { |
| uint8x8_t s0 = load_unaligned_u8(src + 0 * src_stride, (int)src_stride); |
| uint8x8_t s1 = load_unaligned_u8(src + 1 * src_stride, (int)src_stride); |
| uint8x8_t s2 = load_unaligned_u8(src + 2 * src_stride, (int)src_stride); |
| uint8x8_t s3 = load_unaligned_u8(src + 3 * src_stride, (int)src_stride); |
| |
| uint16x8_t sum0 = vmull_u8(s0, f0); |
| sum0 = vmlal_u8(sum0, s1, f1); |
| uint16x8_t sum1 = vmull_u8(s2, f0); |
| sum1 = vmlal_u8(sum1, s3, f1); |
| |
| uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); |
| uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); |
| |
| store_u8x4_strided_x2(dst + 0 * dst_stride, dst_stride, d0); |
| store_u8x4_strided_x2(dst + 2 * dst_stride, dst_stride, d1); |
| |
| src += 4 * src_stride; |
| dst += 4 * dst_stride; |
| h -= 4; |
| } while (h > 0); |
| } else if (w == 8) { |
| do { |
| uint8x8_t s0, s1, s2; |
| load_u8_8x3(src, src_stride, &s0, &s1, &s2); |
| |
| uint16x8_t sum0 = vmull_u8(s0, f0); |
| sum0 = vmlal_u8(sum0, s1, f1); |
| uint16x8_t sum1 = vmull_u8(s1, f0); |
| sum1 = vmlal_u8(sum1, s2, f1); |
| |
| uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); |
| uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); |
| |
| vst1_u8(dst + 0 * dst_stride, d0); |
| vst1_u8(dst + 1 * dst_stride, d1); |
| |
| src += 2 * src_stride; |
| dst += 2 * dst_stride; |
| h -= 2; |
| } while (h > 0); |
| } else { |
| do { |
| int width = w; |
| const uint8_t *s = src; |
| uint8_t *d = dst; |
| |
| do { |
| uint8x16_t s0 = vld1q_u8(s + 0 * src_stride); |
| uint8x16_t s1 = vld1q_u8(s + 1 * src_stride); |
| |
| uint16x8_t sum0 = vmull_u8(vget_low_u8(s0), f0); |
| sum0 = vmlal_u8(sum0, vget_low_u8(s1), f1); |
| uint16x8_t sum1 = vmull_u8(vget_high_u8(s0), f0); |
| sum1 = vmlal_u8(sum1, vget_high_u8(s1), f1); |
| |
| uint8x8_t d0 = vqrshrn_n_u16(sum0, FILTER_BITS); |
| uint8x8_t d1 = vqrshrn_n_u16(sum1, FILTER_BITS); |
| |
| vst1q_u8(d, vcombine_u8(d0, d1)); |
| |
| s += 16; |
| d += 16; |
| width -= 16; |
| } while (width != 0); |
| src += src_stride; |
| dst += dst_stride; |
| } while (--h > 0); |
| } |
| } |
| |
| #endif // AOM_AOM_DSP_ARM_AOM_CONVOLVE8_NEON_H_ |