| /* |
| * Copyright (c) 2025, Alliance for Open Media. All rights reserved. |
| * |
| * This source code is subject to the terms of the BSD 2 Clause License and |
| * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License |
| * was not distributed with this source code in the LICENSE file, you can |
| * obtain it at www.aomedia.org/license/software. If the Alliance for Open |
| * Media Patent License 1.0 was not distributed with this source code in the |
| * PATENTS file, you can obtain it at www.aomedia.org/license/patent. |
| */ |
| |
| #ifndef AOM_AV1_COMMON_RISCV_CONVOLVE_RVV_H_ |
| #define AOM_AV1_COMMON_RISCV_CONVOLVE_RVV_H_ |
| |
| #include "config/aom_config.h" |
| |
| #include "av1/common/convolve.h" |
| #include "av1/common/filter.h" |
| |
| // load_strided_u8_4xN |
| static inline vuint8mf2_t load_strided_u8_4xN(uint8_t *addr, ptrdiff_t stride, |
| size_t vl) { |
| const vuint8mf2_t px_l1 = __riscv_vle8_v_u8mf2(addr + stride, vl); |
| const vuint8mf2_t px_l0 = __riscv_vle8_v_u8mf2(addr, vl); |
| return __riscv_vslideup_vx_u8mf2(px_l0, px_l1, vl >> 1, vl); |
| } |
| |
| // store_strided_u8_4xN |
| static inline void store_strided_u8_4xN(uint8_t *addr, vuint8mf2_t vdst, |
| ptrdiff_t stride, size_t vl) { |
| __riscv_vse8_v_u8mf2(addr, vdst, vl >> 1); |
| vdst = __riscv_vslidedown_vx_u8mf2(vdst, vl >> 1, vl); |
| __riscv_vse8_v_u8mf2(addr + stride, vdst, vl >> 1); |
| } |
| |
| // load_strided_i16_4xN |
| static inline vint16m1_t load_strided_i16_4xN(int16_t *addr, ptrdiff_t stride, |
| size_t vl) { |
| const vint16m1_t px_l1 = __riscv_vle16_v_i16m1(addr + stride, vl >> 1); |
| const vint16m1_t px_l0 = __riscv_vle16_v_i16m1(addr, vl >> 1); |
| return __riscv_vslideup_vx_i16m1(px_l0, px_l1, vl >> 1, vl); |
| } |
| |
| // store_strided_i16_4xN |
| static inline void store_strided_i16_4xN(int16_t *addr, vint16m1_t vdst, |
| ptrdiff_t stride, size_t vl) { |
| __riscv_vse16_v_i16m1(addr, vdst, vl >> 1); |
| vdst = __riscv_vslidedown_vx_i16m1(vdst, vl >> 1, vl); |
| __riscv_vse16_v_i16m1(addr + stride, vdst, vl >> 1); |
| } |
| |
| static inline vuint8mf2_t convolve12_2d_v_rvv( |
| const vint16m1_t s0, const vint16m1_t s1, const vint16m1_t s2, |
| const vint16m1_t s3, const vint16m1_t s4, const vint16m1_t s5, |
| const vint16m1_t s6, const vint16m1_t s7, const vint16m1_t s8, |
| const vint16m1_t s9, const vint16m1_t s10, const vint16m1_t s11, |
| const int16_t *y_filter, const int16_t sub_const, const int vert_const, |
| size_t vl) { |
| vint32m2_t sum = __riscv_vwmul_vx_i32m2(s0, y_filter[0], vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[1], s1, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[2], s2, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[3], s3, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[4], s4, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[5], s5, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[6], s6, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[7], s7, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[8], s8, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[9], s9, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[10], s10, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[11], s11, vl); |
| sum = __riscv_vadd_vx_i32m2(sum, vert_const, vl); |
| |
| vint16m1_t i16_sum = |
| __riscv_vnsra_wx_i16m1(sum, ((FILTER_BITS << 1) - ROUND0_BITS), vl); |
| i16_sum = __riscv_vsub_vx_i16m1(i16_sum, sub_const, vl); |
| vint16m1_t iclip_sum = |
| __riscv_vmin_vx_i16m1(__riscv_vmax_vx_i16m1(i16_sum, 0, vl), 255, vl); |
| |
| return __riscv_vncvt_x_x_w_u8mf2( |
| __riscv_vreinterpret_v_i16m1_u16m1(iclip_sum), vl); |
| } |
| |
| static inline void convolve_2d_sr_vert_12tap_rvv( |
| int16_t *src_ptr, int src_stride, uint8_t *dst_ptr, int dst_stride, int w, |
| int h, const int16_t *y_filter_ptr, size_t vl) { |
| const int vert_const = (1 << ((FILTER_BITS << 1) - ROUND0_BITS)) >> 1; |
| const int16_t sub_const = 1 << FILTER_BITS; |
| |
| if (w == 4) { |
| vl = vl << 1; |
| |
| vint16m1_t s0 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s1 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s2 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s3 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s4 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s5 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s6 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s7 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s8 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s9 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| do { |
| vint16m1_t s10 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s11 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s12 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s13 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| vuint8mf2_t d0 = |
| convolve12_2d_v_rvv(s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, |
| y_filter_ptr, sub_const, vert_const, vl); |
| vuint8mf2_t d1 = |
| convolve12_2d_v_rvv(s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, |
| s13, y_filter_ptr, sub_const, vert_const, vl); |
| |
| store_strided_u8_4xN(dst_ptr, d0, dst_stride, vl); |
| dst_ptr += dst_stride << 1; |
| store_strided_u8_4xN(dst_ptr, d1, dst_stride, vl); |
| dst_ptr += dst_stride << 1; |
| |
| s0 = s4; |
| s1 = s5; |
| s2 = s6; |
| s3 = s7; |
| s4 = s8; |
| s5 = s9; |
| s6 = s10; |
| s7 = s11; |
| s8 = s12; |
| s9 = s13; |
| |
| h -= 4; |
| } while (h != 0); |
| } else { |
| do { |
| int height = h; |
| int16_t *s = src_ptr; |
| uint8_t *d = dst_ptr; |
| |
| vint16m1_t s0 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s1 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s2 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s3 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s4 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s5 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s6 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s7 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s8 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s9 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s10 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| |
| do { |
| vint16m1_t s11 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s12 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s13 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s14 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| |
| vuint8mf2_t d0 = |
| convolve12_2d_v_rvv(s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, |
| s11, y_filter_ptr, sub_const, vert_const, vl); |
| vuint8mf2_t d1 = |
| convolve12_2d_v_rvv(s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, |
| s12, y_filter_ptr, sub_const, vert_const, vl); |
| vuint8mf2_t d2 = |
| convolve12_2d_v_rvv(s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, |
| s13, y_filter_ptr, sub_const, vert_const, vl); |
| vuint8mf2_t d3 = |
| convolve12_2d_v_rvv(s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, |
| s14, y_filter_ptr, sub_const, vert_const, vl); |
| |
| __riscv_vse8_v_u8mf2(d, d0, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d1, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d2, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d3, vl); |
| d += dst_stride; |
| |
| s0 = s4; |
| s1 = s5; |
| s2 = s6; |
| s3 = s7; |
| s4 = s8; |
| s5 = s9; |
| s6 = s10; |
| s7 = s11; |
| s8 = s12; |
| s9 = s13; |
| s10 = s14; |
| |
| height -= 4; |
| } while (height != 0); |
| |
| src_ptr += vl; |
| dst_ptr += vl; |
| w -= vl; |
| } while (w != 0); |
| } |
| } |
| |
| static inline vuint8mf2_t convolve8_2d_v_rvv( |
| const vint16m1_t s0, const vint16m1_t s1, const vint16m1_t s2, |
| const vint16m1_t s3, const vint16m1_t s4, const vint16m1_t s5, |
| const vint16m1_t s6, const vint16m1_t s7, const int16_t *y_filter, |
| const int16_t sub_const, const int vert_const, size_t vl) { |
| vint32m2_t sum = __riscv_vwmul_vx_i32m2(s0, y_filter[0], vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[1], s1, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[2], s2, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[3], s3, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[4], s4, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[5], s5, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[6], s6, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[7], s7, vl); |
| sum = __riscv_vadd_vx_i32m2(sum, vert_const, vl); |
| |
| vint16m1_t i16_sum = |
| __riscv_vnsra_wx_i16m1(sum, ((FILTER_BITS << 1) - ROUND0_BITS), vl); |
| i16_sum = __riscv_vsub_vx_i16m1(i16_sum, sub_const, vl); |
| vint16m1_t iclip_sum = |
| __riscv_vmin_vx_i16m1(__riscv_vmax_vx_i16m1(i16_sum, 0, vl), 255, vl); |
| |
| return __riscv_vncvt_x_x_w_u8mf2( |
| __riscv_vreinterpret_v_i16m1_u16m1(iclip_sum), vl); |
| } |
| |
| static inline void convolve_2d_sr_vert_8tap_rvv( |
| int16_t *src_ptr, int src_stride, uint8_t *dst_ptr, int dst_stride, int w, |
| int h, const int16_t *y_filter_ptr, size_t vl) { |
| const int vert_const = (1 << ((FILTER_BITS << 1) - ROUND0_BITS)) >> 1; |
| const int16_t sub_const = 1 << FILTER_BITS; |
| |
| if (w <= 4) { |
| vl = vl << 1; |
| |
| vint16m1_t s0 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s1 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s2 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s3 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s4 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s5 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| do { |
| vint16m1_t s6 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s7 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| vuint8mf2_t d0 = |
| convolve8_2d_v_rvv(s0, s1, s2, s3, s4, s5, s6, s7, y_filter_ptr, |
| sub_const, vert_const, vl); |
| |
| store_strided_u8_4xN(dst_ptr, d0, dst_stride, vl); |
| dst_ptr += dst_stride << 1; |
| |
| s0 = s2; |
| s1 = s3; |
| s2 = s4; |
| s3 = s5; |
| s4 = s6; |
| s5 = s7; |
| |
| h -= 2; |
| } while (h != 0); |
| } else { |
| do { |
| int height = h; |
| int16_t *s = src_ptr; |
| uint8_t *d = dst_ptr; |
| |
| vint16m1_t s0 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s1 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s2 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s3 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s4 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s5 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s6 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| |
| do { |
| vint16m1_t s7 = __riscv_vle16_v_i16m1(s, vl); |
| vuint8mf2_t d0 = |
| convolve8_2d_v_rvv(s0, s1, s2, s3, s4, s5, s6, s7, y_filter_ptr, |
| sub_const, vert_const, vl); |
| __riscv_vse8_v_u8mf2(d, d0, vl); |
| |
| s0 = s1; |
| s1 = s2; |
| s2 = s3; |
| s3 = s4; |
| s4 = s5; |
| s5 = s6; |
| s6 = s7; |
| s += src_stride; |
| d += dst_stride; |
| height--; |
| } while (height != 0); |
| |
| src_ptr += vl; |
| dst_ptr += vl; |
| w -= vl; |
| } while (w != 0); |
| } |
| } |
| |
| static inline vuint8mf2_t convolve6_2d_v_rvv( |
| const vint16m1_t s0, const vint16m1_t s1, const vint16m1_t s2, |
| const vint16m1_t s3, const vint16m1_t s4, const vint16m1_t s5, |
| const int16_t *y_filter, const int16_t sub_const, const int vert_const, |
| size_t vl) { |
| vint32m2_t sum = __riscv_vwmul_vx_i32m2(s0, y_filter[0], vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[1], s1, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[2], s2, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[3], s3, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[4], s4, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[5], s5, vl); |
| sum = __riscv_vadd_vx_i32m2(sum, vert_const, vl); |
| |
| vint16m1_t i16_sum = |
| __riscv_vnsra_wx_i16m1(sum, ((FILTER_BITS << 1) - ROUND0_BITS), vl); |
| i16_sum = __riscv_vsub_vx_i16m1(i16_sum, sub_const, vl); |
| vint16m1_t iclip_sum = |
| __riscv_vmin_vx_i16m1(__riscv_vmax_vx_i16m1(i16_sum, 0, vl), 255, vl); |
| |
| return __riscv_vncvt_x_x_w_u8mf2( |
| __riscv_vreinterpret_v_i16m1_u16m1(iclip_sum), vl); |
| } |
| |
| static inline void convolve_2d_sr_vert_6tap_rvv( |
| int16_t *src_ptr, int src_stride, uint8_t *dst_ptr, int dst_stride, int w, |
| int h, const int16_t *y_filter_ptr, size_t vl) { |
| const int vert_const = (1 << ((FILTER_BITS << 1) - ROUND0_BITS)) >> 1; |
| const int16_t sub_const = 1 << FILTER_BITS; |
| |
| const int16_t *filter = y_filter_ptr + 1; |
| |
| if (w <= 4) { |
| vl = vl << 1; |
| |
| vint16m1_t s0 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s1 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s2 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s3 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| do { |
| vint16m1_t s4 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s5 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s6 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s7 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| vuint8mf2_t d0 = convolve6_2d_v_rvv(s0, s1, s2, s3, s4, s5, filter, |
| sub_const, vert_const, vl); |
| vuint8mf2_t d1 = convolve6_2d_v_rvv(s2, s3, s4, s5, s6, s7, filter, |
| sub_const, vert_const, vl); |
| |
| store_strided_u8_4xN(dst_ptr, d0, dst_stride, vl); |
| dst_ptr += dst_stride << 1; |
| store_strided_u8_4xN(dst_ptr, d1, dst_stride, vl); |
| dst_ptr += dst_stride << 1; |
| |
| s0 = s4; |
| s1 = s5; |
| s2 = s6; |
| s3 = s7; |
| |
| h -= 4; |
| } while (h != 0); |
| } else { |
| do { |
| int height = h; |
| int16_t *s = src_ptr; |
| uint8_t *d = dst_ptr; |
| |
| vint16m1_t s0 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s1 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s2 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s3 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s4 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| |
| do { |
| vint16m1_t s5 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s6 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s7 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s8 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| |
| vuint8mf2_t d0 = convolve6_2d_v_rvv(s0, s1, s2, s3, s4, s5, filter, |
| sub_const, vert_const, vl); |
| vuint8mf2_t d1 = convolve6_2d_v_rvv(s1, s2, s3, s4, s5, s6, filter, |
| sub_const, vert_const, vl); |
| vuint8mf2_t d2 = convolve6_2d_v_rvv(s2, s3, s4, s5, s6, s7, filter, |
| sub_const, vert_const, vl); |
| vuint8mf2_t d3 = convolve6_2d_v_rvv(s3, s4, s5, s6, s7, s8, filter, |
| sub_const, vert_const, vl); |
| |
| __riscv_vse8_v_u8mf2(d, d0, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d1, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d2, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d3, vl); |
| d += dst_stride; |
| |
| s0 = s4; |
| s1 = s5; |
| s2 = s6; |
| s3 = s7; |
| s4 = s8; |
| |
| height -= 4; |
| } while (height != 0); |
| |
| src_ptr += vl; |
| dst_ptr += vl; |
| w -= vl; |
| } while (w != 0); |
| } |
| } |
| |
| static inline vuint8mf2_t convolve4_2d_v_rvv( |
| const vint16m1_t s0, const vint16m1_t s1, const vint16m1_t s2, |
| const vint16m1_t s3, const int16_t *y_filter, const int16_t sub_const, |
| const int vert_const, size_t vl) { |
| vint32m2_t sum = __riscv_vwmul_vx_i32m2(s0, y_filter[0], vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[1], s1, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[2], s2, vl); |
| sum = __riscv_vwmacc_vx_i32m2(sum, y_filter[3], s3, vl); |
| sum = __riscv_vadd_vx_i32m2(sum, vert_const, vl); |
| |
| vint16m1_t i16_sum = |
| __riscv_vnsra_wx_i16m1(sum, ((FILTER_BITS << 1) - ROUND0_BITS), vl); |
| i16_sum = __riscv_vsub_vx_i16m1(i16_sum, sub_const, vl); |
| vint16m1_t iclip_sum = |
| __riscv_vmin_vx_i16m1(__riscv_vmax_vx_i16m1(i16_sum, 0, vl), 255, vl); |
| |
| return __riscv_vncvt_x_x_w_u8mf2( |
| __riscv_vreinterpret_v_i16m1_u16m1(iclip_sum), vl); |
| } |
| |
| static inline void convolve_2d_sr_vert_4tap_rvv( |
| int16_t *src_ptr, int src_stride, uint8_t *dst_ptr, int dst_stride, int w, |
| int h, const int16_t *y_filter_ptr, size_t vl) { |
| const int vert_const = (1 << ((FILTER_BITS << 1) - ROUND0_BITS)) >> 1; |
| const int16_t sub_const = 1 << FILTER_BITS; |
| // Filter values are at offset 2 |
| const int16_t *filter = y_filter_ptr + 2; |
| |
| if (w <= 4) { |
| vl = vl << 1; |
| |
| vint16m1_t s0 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s1 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| do { |
| vint16m1_t s2 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s3 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s4 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| vint16m1_t s5 = load_strided_i16_4xN(src_ptr, src_stride, vl); |
| src_ptr += src_stride; |
| |
| vuint8mf2_t d0 = |
| convolve4_2d_v_rvv(s0, s1, s2, s3, filter, sub_const, vert_const, vl); |
| vuint8mf2_t d1 = |
| convolve4_2d_v_rvv(s2, s3, s4, s5, filter, sub_const, vert_const, vl); |
| |
| store_strided_u8_4xN(dst_ptr, d0, dst_stride, vl); |
| dst_ptr += dst_stride << 1; |
| store_strided_u8_4xN(dst_ptr, d1, dst_stride, vl); |
| dst_ptr += dst_stride << 1; |
| |
| s0 = s4; |
| s1 = s5; |
| |
| h -= 4; |
| } while (h != 0); |
| } else { |
| do { |
| int height = h; |
| int16_t *s = src_ptr; |
| uint8_t *d = dst_ptr; |
| |
| vint16m1_t s0 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s1 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s2 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| |
| do { |
| vint16m1_t s3 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s4 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s5 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| vint16m1_t s6 = __riscv_vle16_v_i16m1(s, vl); |
| s += src_stride; |
| |
| vuint8mf2_t d0 = convolve4_2d_v_rvv(s0, s1, s2, s3, filter, sub_const, |
| vert_const, vl); |
| vuint8mf2_t d1 = convolve4_2d_v_rvv(s1, s2, s3, s4, filter, sub_const, |
| vert_const, vl); |
| vuint8mf2_t d2 = convolve4_2d_v_rvv(s2, s3, s4, s5, filter, sub_const, |
| vert_const, vl); |
| vuint8mf2_t d3 = convolve4_2d_v_rvv(s3, s4, s5, s6, filter, sub_const, |
| vert_const, vl); |
| |
| __riscv_vse8_v_u8mf2(d, d0, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d1, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d2, vl); |
| d += dst_stride; |
| __riscv_vse8_v_u8mf2(d, d3, vl); |
| d += dst_stride; |
| |
| s0 = s4; |
| s1 = s5; |
| s2 = s6; |
| |
| height -= 4; |
| } while (height != 0); |
| |
| src_ptr += vl; |
| dst_ptr += vl; |
| w -= vl; |
| } while (w != 0); |
| } |
| } |
| |
| #endif // AOM_AV1_COMMON_RISCV_CONVOLVE_RVV_H_ |