| /* |
| * Copyright (c) 2018, Alliance for Open Media. All rights reserved. |
| * |
| * This source code is subject to the terms of the BSD 2 Clause License and |
| * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License |
| * was not distributed with this source code in the LICENSE file, you can |
| * obtain it at www.aomedia.org/license/software. If the Alliance for Open |
| * Media Patent License 1.0 was not distributed with this source code in the |
| * PATENTS file, you can obtain it at www.aomedia.org/license/patent. |
| */ |
| |
| #ifndef AOM_AOM_DSP_X86_CONVOLVE_AVX2_H_ |
| #define AOM_AOM_DSP_X86_CONVOLVE_AVX2_H_ |
| |
| #include <immintrin.h> |
| |
| #include "aom_ports/mem.h" |
| |
| #include "av1/common/convolve.h" |
| #include "av1/common/filter.h" |
| |
| // filters for 16 |
| DECLARE_ALIGNED(32, static const uint8_t, filt_global_avx2[]) = { |
| 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6, 7, 7, 8, 0, 1, 1, |
| 2, 2, 3, 3, 4, 4, 5, 5, 6, 6, 7, 7, 8, 2, 3, 3, 4, 4, 5, |
| 5, 6, 6, 7, 7, 8, 8, 9, 9, 10, 2, 3, 3, 4, 4, 5, 5, 6, 6, |
| 7, 7, 8, 8, 9, 9, 10, 4, 5, 5, 6, 6, 7, 7, 8, 8, 9, 9, 10, |
| 10, 11, 11, 12, 4, 5, 5, 6, 6, 7, 7, 8, 8, 9, 9, 10, 10, 11, 11, |
| 12, 6, 7, 7, 8, 8, 9, 9, 10, 10, 11, 11, 12, 12, 13, 13, 14, 6, 7, |
| 7, 8, 8, 9, 9, 10, 10, 11, 11, 12, 12, 13, 13, 14 |
| }; |
| |
| DECLARE_ALIGNED(32, static const uint8_t, filt_d4_global_avx2[]) = { |
| 0, 1, 2, 3, 1, 2, 3, 4, 2, 3, 4, 5, 3, 4, 5, 6, 0, 1, 2, 3, 1, 2, |
| 3, 4, 2, 3, 4, 5, 3, 4, 5, 6, 4, 5, 6, 7, 5, 6, 7, 8, 6, 7, 8, 9, |
| 7, 8, 9, 10, 4, 5, 6, 7, 5, 6, 7, 8, 6, 7, 8, 9, 7, 8, 9, 10, |
| }; |
| |
| DECLARE_ALIGNED(32, static const uint8_t, filt4_d4_global_avx2[]) = { |
| 2, 3, 4, 5, 3, 4, 5, 6, 4, 5, 6, 7, 5, 6, 7, 8, |
| 2, 3, 4, 5, 3, 4, 5, 6, 4, 5, 6, 7, 5, 6, 7, 8, |
| }; |
| |
| DECLARE_ALIGNED(32, static const uint8_t, filt_center_global_avx2[32]) = { |
| 3, 255, 4, 255, 5, 255, 6, 255, 7, 255, 8, 255, 9, 255, 10, 255, |
| 3, 255, 4, 255, 5, 255, 6, 255, 7, 255, 8, 255, 9, 255, 10, 255 |
| }; |
| |
| DECLARE_ALIGNED(32, static const uint8_t, |
| filt1_global_avx2[32]) = { 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, |
| 6, 6, 7, 7, 8, 0, 1, 1, 2, 2, 3, |
| 3, 4, 4, 5, 5, 6, 6, 7, 7, 8 }; |
| |
| DECLARE_ALIGNED(32, static const uint8_t, |
| filt2_global_avx2[32]) = { 2, 3, 3, 4, 4, 5, 5, 6, 6, 7, 7, |
| 8, 8, 9, 9, 10, 2, 3, 3, 4, 4, 5, |
| 5, 6, 6, 7, 7, 8, 8, 9, 9, 10 }; |
| |
| DECLARE_ALIGNED(32, static const uint8_t, filt3_global_avx2[32]) = { |
| 4, 5, 5, 6, 6, 7, 7, 8, 8, 9, 9, 10, 10, 11, 11, 12, |
| 4, 5, 5, 6, 6, 7, 7, 8, 8, 9, 9, 10, 10, 11, 11, 12 |
| }; |
| |
| DECLARE_ALIGNED(32, static const uint8_t, filt4_global_avx2[32]) = { |
| 6, 7, 7, 8, 8, 9, 9, 10, 10, 11, 11, 12, 12, 13, 13, 14, |
| 6, 7, 7, 8, 8, 9, 9, 10, 10, 11, 11, 12, 12, 13, 13, 14 |
| }; |
| |
| #define CONVOLVE_SR_HORIZONTAL_FILTER_4TAP \ |
| for (i = 0; i < (im_h - 2); i += 2) { \ |
| __m256i data = _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)&src_ptr[(i * src_stride) + j])); \ |
| data = _mm256_inserti128_si256( \ |
| data, \ |
| _mm_loadu_si128( \ |
| (__m128i *)&src_ptr[(i * src_stride) + j + src_stride]), \ |
| 1); \ |
| __m256i res = convolve_lowbd_x_4tap(data, coeffs_h + 1, filt); \ |
| res = \ |
| _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ |
| _mm256_store_si256((__m256i *)&im_block[i * im_stride], res); \ |
| } \ |
| __m256i data_1 = _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)&src_ptr[(i * src_stride) + j])); \ |
| __m256i res = convolve_lowbd_x_4tap(data_1, coeffs_h + 1, filt); \ |
| res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ |
| _mm256_store_si256((__m256i *)&im_block[i * im_stride], res); |
| |
| #define CONVOLVE_SR_VERTICAL_FILTER_4TAP \ |
| __m256i s[6]; \ |
| __m256i src_0 = _mm256_loadu_si256((__m256i *)(im_block + 0 * im_stride)); \ |
| __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); \ |
| __m256i src_2 = _mm256_loadu_si256((__m256i *)(im_block + 2 * im_stride)); \ |
| __m256i src_3 = _mm256_loadu_si256((__m256i *)(im_block + 3 * im_stride)); \ |
| \ |
| s[0] = _mm256_unpacklo_epi16(src_0, src_1); \ |
| s[1] = _mm256_unpacklo_epi16(src_2, src_3); \ |
| s[3] = _mm256_unpackhi_epi16(src_0, src_1); \ |
| s[4] = _mm256_unpackhi_epi16(src_2, src_3); \ |
| \ |
| for (i = 0; i < h; i += 2) { \ |
| const int16_t *data = &im_block[i * im_stride]; \ |
| const __m256i s4 = _mm256_loadu_si256((__m256i *)(data + 4 * im_stride)); \ |
| const __m256i s5 = _mm256_loadu_si256((__m256i *)(data + 5 * im_stride)); \ |
| s[2] = _mm256_unpacklo_epi16(s4, s5); \ |
| s[5] = _mm256_unpackhi_epi16(s4, s5); \ |
| \ |
| __m256i res_a = convolve_4tap(s, coeffs_v + 1); \ |
| __m256i res_b = convolve_4tap(s + 3, coeffs_v + 1); \ |
| \ |
| res_a = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_a, sum_round_v), sum_shift_v); \ |
| res_b = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_b, sum_round_v), sum_shift_v); \ |
| const __m256i res_a_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_a, round_const_v), round_shift_v); \ |
| const __m256i res_b_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_b, round_const_v), round_shift_v); \ |
| const __m256i res_16bit = _mm256_packs_epi32(res_a_round, res_b_round); \ |
| const __m256i res_8b = _mm256_packus_epi16(res_16bit, res_16bit); \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_8b); \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_8b, 1); \ |
| \ |
| __m128i *const p_0 = (__m128i *)&dst[i * dst_stride + j]; \ |
| __m128i *const p_1 = (__m128i *)&dst[i * dst_stride + j + dst_stride]; \ |
| if (w - j > 4) { \ |
| _mm_storel_epi64(p_0, res_0); \ |
| _mm_storel_epi64(p_1, res_1); \ |
| } else if (w == 4) { \ |
| xx_storel_32(p_0, res_0); \ |
| xx_storel_32(p_1, res_1); \ |
| } else { \ |
| *(uint16_t *)p_0 = (uint16_t)_mm_cvtsi128_si32(res_0); \ |
| *(uint16_t *)p_1 = (uint16_t)_mm_cvtsi128_si32(res_1); \ |
| } \ |
| \ |
| s[0] = s[1]; \ |
| s[1] = s[2]; \ |
| s[3] = s[4]; \ |
| s[4] = s[5]; \ |
| } |
| |
| #define CONVOLVE_SR_HORIZONTAL_FILTER_6TAP \ |
| for (i = 0; i < (im_h - 2); i += 2) { \ |
| __m256i data = _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)&src_ptr[(i * src_stride) + j])); \ |
| data = _mm256_inserti128_si256( \ |
| data, \ |
| _mm_loadu_si128( \ |
| (__m128i *)&src_ptr[(i * src_stride) + j + src_stride]), \ |
| 1); \ |
| \ |
| __m256i res = convolve_lowbd_x_6tap(data, coeffs_h, filt); \ |
| res = \ |
| _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ |
| _mm256_store_si256((__m256i *)&im_block[i * im_stride], res); \ |
| } \ |
| \ |
| __m256i data_1 = _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)&src_ptr[(i * src_stride) + j])); \ |
| \ |
| __m256i res = convolve_lowbd_x_6tap(data_1, coeffs_h, filt); \ |
| \ |
| res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ |
| \ |
| _mm256_store_si256((__m256i *)&im_block[i * im_stride], res); |
| |
| #define CONVOLVE_SR_VERTICAL_FILTER_6TAP \ |
| __m256i src_0 = _mm256_loadu_si256((__m256i *)(im_block + 0 * im_stride)); \ |
| __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); \ |
| __m256i src_2 = _mm256_loadu_si256((__m256i *)(im_block + 2 * im_stride)); \ |
| __m256i src_3 = _mm256_loadu_si256((__m256i *)(im_block + 3 * im_stride)); \ |
| \ |
| __m256i s[8]; \ |
| s[0] = _mm256_unpacklo_epi16(src_0, src_1); \ |
| s[1] = _mm256_unpacklo_epi16(src_2, src_3); \ |
| \ |
| s[3] = _mm256_unpackhi_epi16(src_0, src_1); \ |
| s[4] = _mm256_unpackhi_epi16(src_2, src_3); \ |
| \ |
| for (i = 0; i < h; i += 2) { \ |
| const int16_t *data = &im_block[i * im_stride]; \ |
| \ |
| const __m256i s6 = _mm256_loadu_si256((__m256i *)(data + 4 * im_stride)); \ |
| const __m256i s7 = _mm256_loadu_si256((__m256i *)(data + 5 * im_stride)); \ |
| \ |
| s[2] = _mm256_unpacklo_epi16(s6, s7); \ |
| s[5] = _mm256_unpackhi_epi16(s6, s7); \ |
| \ |
| __m256i res_a = convolve_6tap(s, coeffs_v); \ |
| __m256i res_b = convolve_6tap(s + 3, coeffs_v); \ |
| \ |
| res_a = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_a, sum_round_v), sum_shift_v); \ |
| res_b = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_b, sum_round_v), sum_shift_v); \ |
| \ |
| const __m256i res_a_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_a, round_const_v), round_shift_v); \ |
| const __m256i res_b_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_b, round_const_v), round_shift_v); \ |
| \ |
| const __m256i res_16bit = _mm256_packs_epi32(res_a_round, res_b_round); \ |
| const __m256i res_8b = _mm256_packus_epi16(res_16bit, res_16bit); \ |
| \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_8b); \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_8b, 1); \ |
| \ |
| __m128i *const p_0 = (__m128i *)&dst[i * dst_stride + j]; \ |
| __m128i *const p_1 = (__m128i *)&dst[i * dst_stride + j + dst_stride]; \ |
| if (w - j > 4) { \ |
| _mm_storel_epi64(p_0, res_0); \ |
| _mm_storel_epi64(p_1, res_1); \ |
| } else if (w == 4) { \ |
| xx_storel_32(p_0, res_0); \ |
| xx_storel_32(p_1, res_1); \ |
| } else { \ |
| *(uint16_t *)p_0 = (uint16_t)_mm_cvtsi128_si32(res_0); \ |
| *(uint16_t *)p_1 = (uint16_t)_mm_cvtsi128_si32(res_1); \ |
| } \ |
| \ |
| s[0] = s[1]; \ |
| s[1] = s[2]; \ |
| \ |
| s[3] = s[4]; \ |
| s[4] = s[5]; \ |
| } |
| |
| #define CONVOLVE_SR_HORIZONTAL_FILTER_8TAP \ |
| for (i = 0; i < (im_h - 2); i += 2) { \ |
| __m256i data = _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)&src_ptr[(i * src_stride) + j])); \ |
| data = _mm256_inserti128_si256( \ |
| data, \ |
| _mm_loadu_si128( \ |
| (__m128i *)&src_ptr[(i * src_stride) + j + src_stride]), \ |
| 1); \ |
| \ |
| __m256i res = convolve_lowbd_x(data, coeffs_h, filt); \ |
| res = \ |
| _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ |
| _mm256_store_si256((__m256i *)&im_block[i * im_stride], res); \ |
| } \ |
| \ |
| __m256i data_1 = _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)&src_ptr[(i * src_stride) + j])); \ |
| \ |
| __m256i res = convolve_lowbd_x(data_1, coeffs_h, filt); \ |
| \ |
| res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ |
| \ |
| _mm256_store_si256((__m256i *)&im_block[i * im_stride], res); |
| |
| #define CONVOLVE_SR_VERTICAL_FILTER_8TAP \ |
| __m256i src_0 = _mm256_loadu_si256((__m256i *)(im_block + 0 * im_stride)); \ |
| __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); \ |
| __m256i src_2 = _mm256_loadu_si256((__m256i *)(im_block + 2 * im_stride)); \ |
| __m256i src_3 = _mm256_loadu_si256((__m256i *)(im_block + 3 * im_stride)); \ |
| __m256i src_4 = _mm256_loadu_si256((__m256i *)(im_block + 4 * im_stride)); \ |
| __m256i src_5 = _mm256_loadu_si256((__m256i *)(im_block + 5 * im_stride)); \ |
| \ |
| __m256i s[8]; \ |
| s[0] = _mm256_unpacklo_epi16(src_0, src_1); \ |
| s[1] = _mm256_unpacklo_epi16(src_2, src_3); \ |
| s[2] = _mm256_unpacklo_epi16(src_4, src_5); \ |
| \ |
| s[4] = _mm256_unpackhi_epi16(src_0, src_1); \ |
| s[5] = _mm256_unpackhi_epi16(src_2, src_3); \ |
| s[6] = _mm256_unpackhi_epi16(src_4, src_5); \ |
| \ |
| for (i = 0; i < h; i += 2) { \ |
| const int16_t *data = &im_block[i * im_stride]; \ |
| \ |
| const __m256i s6 = _mm256_loadu_si256((__m256i *)(data + 6 * im_stride)); \ |
| const __m256i s7 = _mm256_loadu_si256((__m256i *)(data + 7 * im_stride)); \ |
| \ |
| s[3] = _mm256_unpacklo_epi16(s6, s7); \ |
| s[7] = _mm256_unpackhi_epi16(s6, s7); \ |
| \ |
| __m256i res_a = convolve(s, coeffs_v); \ |
| __m256i res_b = convolve(s + 4, coeffs_v); \ |
| \ |
| res_a = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_a, sum_round_v), sum_shift_v); \ |
| res_b = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_b, sum_round_v), sum_shift_v); \ |
| \ |
| const __m256i res_a_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_a, round_const_v), round_shift_v); \ |
| const __m256i res_b_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_b, round_const_v), round_shift_v); \ |
| \ |
| const __m256i res_16bit = _mm256_packs_epi32(res_a_round, res_b_round); \ |
| const __m256i res_8b = _mm256_packus_epi16(res_16bit, res_16bit); \ |
| \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_8b); \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_8b, 1); \ |
| \ |
| __m128i *const p_0 = (__m128i *)&dst[i * dst_stride + j]; \ |
| __m128i *const p_1 = (__m128i *)&dst[i * dst_stride + j + dst_stride]; \ |
| if (w - j > 4) { \ |
| _mm_storel_epi64(p_0, res_0); \ |
| _mm_storel_epi64(p_1, res_1); \ |
| } else if (w == 4) { \ |
| xx_storel_32(p_0, res_0); \ |
| xx_storel_32(p_1, res_1); \ |
| } else { \ |
| *(uint16_t *)p_0 = (uint16_t)_mm_cvtsi128_si32(res_0); \ |
| *(uint16_t *)p_1 = (uint16_t)_mm_cvtsi128_si32(res_1); \ |
| } \ |
| \ |
| s[0] = s[1]; \ |
| s[1] = s[2]; \ |
| s[2] = s[3]; \ |
| \ |
| s[4] = s[5]; \ |
| s[5] = s[6]; \ |
| s[6] = s[7]; \ |
| } |
| |
| #define CONVOLVE_SR_HORIZONTAL_FILTER_12TAP \ |
| const __m256i v_zero = _mm256_setzero_si256(); \ |
| __m256i s[12]; \ |
| if (w <= 4) { \ |
| for (i = 0; i < im_h; i += 2) { \ |
| const __m256i data = _mm256_permute2x128_si256( \ |
| _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)(&src_ptr[i * src_stride + j]))), \ |
| _mm256_castsi128_si256(_mm_loadu_si128( \ |
| (__m128i *)(&src_ptr[i * src_stride + src_stride + j]))), \ |
| 0x20); \ |
| const __m256i s_16lo = _mm256_unpacklo_epi8(data, v_zero); \ |
| const __m256i s_16hi = _mm256_unpackhi_epi8(data, v_zero); \ |
| const __m256i s_lolo = _mm256_unpacklo_epi16(s_16lo, s_16lo); \ |
| const __m256i s_lohi = _mm256_unpackhi_epi16(s_16lo, s_16lo); \ |
| \ |
| const __m256i s_hilo = _mm256_unpacklo_epi16(s_16hi, s_16hi); \ |
| const __m256i s_hihi = _mm256_unpackhi_epi16(s_16hi, s_16hi); \ |
| \ |
| s[0] = _mm256_alignr_epi8(s_lohi, s_lolo, 2); \ |
| s[1] = _mm256_alignr_epi8(s_lohi, s_lolo, 10); \ |
| s[2] = _mm256_alignr_epi8(s_hilo, s_lohi, 2); \ |
| s[3] = _mm256_alignr_epi8(s_hilo, s_lohi, 10); \ |
| s[4] = _mm256_alignr_epi8(s_hihi, s_hilo, 2); \ |
| s[5] = _mm256_alignr_epi8(s_hihi, s_hilo, 10); \ |
| \ |
| const __m256i res_lo = convolve_12taps(s, coeffs_h); \ |
| \ |
| __m256i res_32b_lo = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_lo, round_const_h12), round_shift_h12); \ |
| __m256i res_16b_lo = _mm256_packs_epi32(res_32b_lo, res_32b_lo); \ |
| const __m128i res_0 = _mm256_extracti128_si256(res_16b_lo, 0); \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_16b_lo, 1); \ |
| if (w > 2) { \ |
| _mm_storel_epi64((__m128i *)&im_block[i * im_stride], res_0); \ |
| _mm_storel_epi64((__m128i *)&im_block[i * im_stride + im_stride], \ |
| res_1); \ |
| } else { \ |
| uint32_t horiz_2; \ |
| horiz_2 = (uint32_t)_mm_cvtsi128_si32(res_0); \ |
| im_block[i * im_stride] = (uint16_t)horiz_2; \ |
| im_block[i * im_stride + 1] = (uint16_t)(horiz_2 >> 16); \ |
| horiz_2 = (uint32_t)_mm_cvtsi128_si32(res_1); \ |
| im_block[i * im_stride + im_stride] = (uint16_t)horiz_2; \ |
| im_block[i * im_stride + im_stride + 1] = (uint16_t)(horiz_2 >> 16); \ |
| } \ |
| } \ |
| } else { \ |
| for (i = 0; i < im_h; i++) { \ |
| const __m256i data = _mm256_permute2x128_si256( \ |
| _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)(&src_ptr[i * src_stride + j]))), \ |
| _mm256_castsi128_si256( \ |
| _mm_loadu_si128((__m128i *)(&src_ptr[i * src_stride + j + 4]))), \ |
| 0x20); \ |
| const __m256i s_16lo = _mm256_unpacklo_epi8(data, v_zero); \ |
| const __m256i s_16hi = _mm256_unpackhi_epi8(data, v_zero); \ |
| \ |
| const __m256i s_lolo = _mm256_unpacklo_epi16(s_16lo, s_16lo); \ |
| const __m256i s_lohi = _mm256_unpackhi_epi16(s_16lo, s_16lo); \ |
| \ |
| const __m256i s_hilo = _mm256_unpacklo_epi16(s_16hi, s_16hi); \ |
| const __m256i s_hihi = _mm256_unpackhi_epi16(s_16hi, s_16hi); \ |
| \ |
| s[0] = _mm256_alignr_epi8(s_lohi, s_lolo, 2); \ |
| s[1] = _mm256_alignr_epi8(s_lohi, s_lolo, 10); \ |
| s[2] = _mm256_alignr_epi8(s_hilo, s_lohi, 2); \ |
| s[3] = _mm256_alignr_epi8(s_hilo, s_lohi, 10); \ |
| s[4] = _mm256_alignr_epi8(s_hihi, s_hilo, 2); \ |
| s[5] = _mm256_alignr_epi8(s_hihi, s_hilo, 10); \ |
| \ |
| const __m256i res_lo = convolve_12taps(s, coeffs_h); \ |
| \ |
| __m256i res_32b_lo = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_lo, round_const_h12), round_shift_h12); \ |
| \ |
| __m256i res_16b_lo = _mm256_packs_epi32(res_32b_lo, res_32b_lo); \ |
| _mm_store_si128((__m128i *)&im_block[i * im_stride], \ |
| _mm256_extracti128_si256( \ |
| _mm256_permute4x64_epi64(res_16b_lo, 0x88), 0)); \ |
| } \ |
| } |
| |
| #define CONVOLVE_SR_VERTICAL_FILTER_12TAP \ |
| __m256i src_0 = _mm256_loadu_si256((__m256i *)(im_block + 0 * im_stride)); \ |
| __m256i src_1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); \ |
| __m256i src_2 = _mm256_loadu_si256((__m256i *)(im_block + 2 * im_stride)); \ |
| __m256i src_3 = _mm256_loadu_si256((__m256i *)(im_block + 3 * im_stride)); \ |
| __m256i src_4 = _mm256_loadu_si256((__m256i *)(im_block + 4 * im_stride)); \ |
| __m256i src_5 = _mm256_loadu_si256((__m256i *)(im_block + 5 * im_stride)); \ |
| __m256i src_6 = _mm256_loadu_si256((__m256i *)(im_block + 6 * im_stride)); \ |
| __m256i src_7 = _mm256_loadu_si256((__m256i *)(im_block + 7 * im_stride)); \ |
| __m256i src_8 = _mm256_loadu_si256((__m256i *)(im_block + 8 * im_stride)); \ |
| __m256i src_9 = _mm256_loadu_si256((__m256i *)(im_block + 9 * im_stride)); \ |
| \ |
| s[0] = _mm256_unpacklo_epi16(src_0, src_1); \ |
| s[1] = _mm256_unpacklo_epi16(src_2, src_3); \ |
| s[2] = _mm256_unpacklo_epi16(src_4, src_5); \ |
| s[3] = _mm256_unpacklo_epi16(src_6, src_7); \ |
| s[4] = _mm256_unpacklo_epi16(src_8, src_9); \ |
| \ |
| s[6] = _mm256_unpackhi_epi16(src_0, src_1); \ |
| s[7] = _mm256_unpackhi_epi16(src_2, src_3); \ |
| s[8] = _mm256_unpackhi_epi16(src_4, src_5); \ |
| s[9] = _mm256_unpackhi_epi16(src_6, src_7); \ |
| s[10] = _mm256_unpackhi_epi16(src_8, src_9); \ |
| \ |
| for (i = 0; i < h; i += 2) { \ |
| const int16_t *data = &im_block[i * im_stride]; \ |
| \ |
| const __m256i s6 = _mm256_loadu_si256((__m256i *)(data + 10 * im_stride)); \ |
| const __m256i s7 = _mm256_loadu_si256((__m256i *)(data + 11 * im_stride)); \ |
| \ |
| s[5] = _mm256_unpacklo_epi16(s6, s7); \ |
| s[11] = _mm256_unpackhi_epi16(s6, s7); \ |
| \ |
| __m256i res_a = convolve_12taps(s, coeffs_v); \ |
| __m256i res_b = convolve_12taps(s + 6, coeffs_v); \ |
| \ |
| res_a = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_a, sum_round_v), sum_shift_v); \ |
| res_b = \ |
| _mm256_sra_epi32(_mm256_add_epi32(res_b, sum_round_v), sum_shift_v); \ |
| \ |
| const __m256i res_a_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_a, round_const_v), round_shift_v); \ |
| const __m256i res_b_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_b, round_const_v), round_shift_v); \ |
| \ |
| const __m256i res_16bit = _mm256_packs_epi32(res_a_round, res_b_round); \ |
| const __m256i res_8b = _mm256_packus_epi16(res_16bit, res_16bit); \ |
| \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_8b); \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_8b, 1); \ |
| \ |
| __m128i *const p_0 = (__m128i *)&dst[i * dst_stride + j]; \ |
| __m128i *const p_1 = (__m128i *)&dst[i * dst_stride + j + dst_stride]; \ |
| if (w - j > 4) { \ |
| _mm_storel_epi64(p_0, res_0); \ |
| _mm_storel_epi64(p_1, res_1); \ |
| } else if (w == 4) { \ |
| xx_storel_32(p_0, res_0); \ |
| xx_storel_32(p_1, res_1); \ |
| } else { \ |
| *(uint16_t *)p_0 = (uint16_t)_mm_cvtsi128_si32(res_0); \ |
| *(uint16_t *)p_1 = (uint16_t)_mm_cvtsi128_si32(res_1); \ |
| } \ |
| \ |
| s[0] = s[1]; \ |
| s[1] = s[2]; \ |
| s[2] = s[3]; \ |
| s[3] = s[4]; \ |
| s[4] = s[5]; \ |
| \ |
| s[6] = s[7]; \ |
| s[7] = s[8]; \ |
| s[8] = s[9]; \ |
| s[9] = s[10]; \ |
| s[10] = s[11]; \ |
| } |
| |
| #define DIST_WTD_CONVOLVE_HORIZONTAL_FILTER_8TAP \ |
| do { \ |
| for (i = 0; i < im_h; i += 2) { \ |
| __m256i data = \ |
| _mm256_castsi128_si256(_mm_loadu_si128((__m128i *)src_h)); \ |
| if (i + 1 < im_h) \ |
| data = _mm256_inserti128_si256( \ |
| data, _mm_loadu_si128((__m128i *)(src_h + src_stride)), 1); \ |
| src_h += (src_stride << 1); \ |
| __m256i res = convolve_lowbd_x(data, coeffs_x, filt); \ |
| \ |
| res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), \ |
| round_shift_h); \ |
| \ |
| _mm256_store_si256((__m256i *)&im_block[i * im_stride], res); \ |
| } \ |
| } while (0) |
| |
| #define DIST_WTD_CONVOLVE_VERTICAL_FILTER_8TAP \ |
| do { \ |
| __m256i s[8]; \ |
| __m256i s0 = _mm256_loadu_si256((__m256i *)(im_block + 0 * im_stride)); \ |
| __m256i s1 = _mm256_loadu_si256((__m256i *)(im_block + 1 * im_stride)); \ |
| __m256i s2 = _mm256_loadu_si256((__m256i *)(im_block + 2 * im_stride)); \ |
| __m256i s3 = _mm256_loadu_si256((__m256i *)(im_block + 3 * im_stride)); \ |
| __m256i s4 = _mm256_loadu_si256((__m256i *)(im_block + 4 * im_stride)); \ |
| __m256i s5 = _mm256_loadu_si256((__m256i *)(im_block + 5 * im_stride)); \ |
| \ |
| s[0] = _mm256_unpacklo_epi16(s0, s1); \ |
| s[1] = _mm256_unpacklo_epi16(s2, s3); \ |
| s[2] = _mm256_unpacklo_epi16(s4, s5); \ |
| \ |
| s[4] = _mm256_unpackhi_epi16(s0, s1); \ |
| s[5] = _mm256_unpackhi_epi16(s2, s3); \ |
| s[6] = _mm256_unpackhi_epi16(s4, s5); \ |
| \ |
| for (i = 0; i < h; i += 2) { \ |
| const int16_t *data = &im_block[i * im_stride]; \ |
| \ |
| const __m256i s6 = \ |
| _mm256_loadu_si256((__m256i *)(data + 6 * im_stride)); \ |
| const __m256i s7 = \ |
| _mm256_loadu_si256((__m256i *)(data + 7 * im_stride)); \ |
| \ |
| s[3] = _mm256_unpacklo_epi16(s6, s7); \ |
| s[7] = _mm256_unpackhi_epi16(s6, s7); \ |
| \ |
| const __m256i res_a = convolve(s, coeffs_y); \ |
| const __m256i res_a_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_a, round_const_v), round_shift_v); \ |
| \ |
| if (w - j > 4) { \ |
| const __m256i res_b = convolve(s + 4, coeffs_y); \ |
| const __m256i res_b_round = _mm256_sra_epi32( \ |
| _mm256_add_epi32(res_b, round_const_v), round_shift_v); \ |
| const __m256i res_16b = _mm256_packs_epi32(res_a_round, res_b_round); \ |
| const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \ |
| \ |
| if (do_average) { \ |
| const __m256i data_ref_0 = \ |
| load_line2_avx2(&dst[i * dst_stride + j], \ |
| &dst[i * dst_stride + j + dst_stride]); \ |
| const __m256i comp_avg_res = comp_avg(&data_ref_0, &res_unsigned, \ |
| &wt, use_dist_wtd_comp_avg); \ |
| \ |
| const __m256i round_result = convolve_rounding( \ |
| &comp_avg_res, &offset_const, &rounding_const, rounding_shift); \ |
| \ |
| const __m256i res_8 = \ |
| _mm256_packus_epi16(round_result, round_result); \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_8); \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_8, 1); \ |
| \ |
| _mm_storel_epi64((__m128i *)(&dst0[i * dst_stride0 + j]), res_0); \ |
| _mm_storel_epi64( \ |
| (__m128i *)((&dst0[i * dst_stride0 + j + dst_stride0])), res_1); \ |
| } else { \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_unsigned); \ |
| _mm_store_si128((__m128i *)(&dst[i * dst_stride + j]), res_0); \ |
| \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_unsigned, 1); \ |
| _mm_store_si128((__m128i *)(&dst[i * dst_stride + j + dst_stride]), \ |
| res_1); \ |
| } \ |
| } else { \ |
| const __m256i res_16b = _mm256_packs_epi32(res_a_round, res_a_round); \ |
| const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \ |
| \ |
| if (do_average) { \ |
| const __m256i data_ref_0 = \ |
| load_line2_avx2(&dst[i * dst_stride + j], \ |
| &dst[i * dst_stride + j + dst_stride]); \ |
| \ |
| const __m256i comp_avg_res = comp_avg(&data_ref_0, &res_unsigned, \ |
| &wt, use_dist_wtd_comp_avg); \ |
| \ |
| const __m256i round_result = convolve_rounding( \ |
| &comp_avg_res, &offset_const, &rounding_const, rounding_shift); \ |
| \ |
| const __m256i res_8 = \ |
| _mm256_packus_epi16(round_result, round_result); \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_8); \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_8, 1); \ |
| \ |
| *(int *)(&dst0[i * dst_stride0 + j]) = _mm_cvtsi128_si32(res_0); \ |
| *(int *)(&dst0[i * dst_stride0 + j + dst_stride0]) = \ |
| _mm_cvtsi128_si32(res_1); \ |
| \ |
| } else { \ |
| const __m128i res_0 = _mm256_castsi256_si128(res_unsigned); \ |
| _mm_store_si128((__m128i *)(&dst[i * dst_stride + j]), res_0); \ |
| \ |
| const __m128i res_1 = _mm256_extracti128_si256(res_unsigned, 1); \ |
| _mm_store_si128((__m128i *)(&dst[i * dst_stride + j + dst_stride]), \ |
| res_1); \ |
| } \ |
| } \ |
| \ |
| s[0] = s[1]; \ |
| s[1] = s[2]; \ |
| s[2] = s[3]; \ |
| \ |
| s[4] = s[5]; \ |
| s[5] = s[6]; \ |
| s[6] = s[7]; \ |
| } \ |
| } while (0) |
| |
| static inline void prepare_coeffs_lowbd( |
| const InterpFilterParams *const filter_params, const int subpel_q4, |
| __m256i *const coeffs /* [4] */) { |
| const int16_t *const filter = av1_get_interp_filter_subpel_kernel( |
| filter_params, subpel_q4 & SUBPEL_MASK); |
| const __m128i coeffs_8 = _mm_loadu_si128((__m128i *)filter); |
| const __m256i filter_coeffs = _mm256_broadcastsi128_si256(coeffs_8); |
| |
| // right shift all filter co-efficients by 1 to reduce the bits required. |
| // This extra right shift will be taken care of at the end while rounding |
| // the result. |
| // Since all filter co-efficients are even, this change will not affect the |
| // end result |
| assert(_mm_test_all_zeros(_mm_and_si128(coeffs_8, _mm_set1_epi16(1)), |
| _mm_set1_epi16((short)0xffff))); |
| |
| const __m256i coeffs_1 = _mm256_srai_epi16(filter_coeffs, 1); |
| |
| // coeffs 0 1 0 1 0 1 0 1 |
| coeffs[0] = _mm256_shuffle_epi8(coeffs_1, _mm256_set1_epi16(0x0200u)); |
| // coeffs 2 3 2 3 2 3 2 3 |
| coeffs[1] = _mm256_shuffle_epi8(coeffs_1, _mm256_set1_epi16(0x0604u)); |
| // coeffs 4 5 4 5 4 5 4 5 |
| coeffs[2] = _mm256_shuffle_epi8(coeffs_1, _mm256_set1_epi16(0x0a08u)); |
| // coeffs 6 7 6 7 6 7 6 7 |
| coeffs[3] = _mm256_shuffle_epi8(coeffs_1, _mm256_set1_epi16(0x0e0cu)); |
| } |
| |
| static inline void prepare_coeffs_6t_lowbd( |
| const InterpFilterParams *const filter_params, const int subpel_q4, |
| __m256i *const coeffs /* [4] */) { |
| const int16_t *const filter = av1_get_interp_filter_subpel_kernel( |
| filter_params, subpel_q4 & SUBPEL_MASK); |
| const __m128i coeffs_8 = _mm_loadu_si128((__m128i *)filter); |
| const __m256i filter_coeffs = _mm256_broadcastsi128_si256(coeffs_8); |
| |
| // right shift all filter co-efficients by 1 to reduce the bits required. |
| // This extra right shift will be taken care of at the end while rounding |
| // the result. |
| // Since all filter co-efficients are even, this change will not affect the |
| // end result |
| assert(_mm_test_all_zeros(_mm_and_si128(coeffs_8, _mm_set1_epi16(1)), |
| _mm_set1_epi16((int16_t)0xffff))); |
| |
| const __m256i coeffs_1 = _mm256_srai_epi16(filter_coeffs, 1); |
| |
| // coeffs 1 2 1 2 1 2 1 2 |
| coeffs[0] = _mm256_shuffle_epi8(coeffs_1, _mm256_set1_epi16(0x0402u)); |
| // coeffs 3 4 3 4 3 4 3 4 |
| coeffs[1] = _mm256_shuffle_epi8(coeffs_1, _mm256_set1_epi16(0x0806u)); |
| // coeffs 5 6 5 6 5 6 5 6 |
| coeffs[2] = _mm256_shuffle_epi8(coeffs_1, _mm256_set1_epi16(0x0c0au)); |
| } |
| |
| static inline void prepare_coeffs_6t( |
| const InterpFilterParams *const filter_params, const int subpel_q4, |
| __m256i *const coeffs /* [4] */) { |
| const int16_t *filter = av1_get_interp_filter_subpel_kernel( |
| filter_params, subpel_q4 & SUBPEL_MASK); |
| |
| const __m128i coeff_8 = _mm_loadu_si128((__m128i *)(filter + 1)); |
| const __m256i coeff = _mm256_broadcastsi128_si256(coeff_8); |
| |
| // coeffs 1 2 1 2 1 2 1 2 |
| coeffs[0] = _mm256_shuffle_epi32(coeff, 0x00); |
| // coeffs 3 4 3 4 3 4 3 4 |
| coeffs[1] = _mm256_shuffle_epi32(coeff, 0x55); |
| // coeffs 5 6 5 6 5 6 5 6 |
| coeffs[2] = _mm256_shuffle_epi32(coeff, 0xaa); |
| } |
| |
| static inline void prepare_coeffs(const InterpFilterParams *const filter_params, |
| const int subpel_q4, |
| __m256i *const coeffs /* [4] */) { |
| const int16_t *filter = av1_get_interp_filter_subpel_kernel( |
| filter_params, subpel_q4 & SUBPEL_MASK); |
| |
| const __m128i coeff_8 = _mm_loadu_si128((__m128i *)filter); |
| const __m256i coeff = _mm256_broadcastsi128_si256(coeff_8); |
| |
| // coeffs 0 1 0 1 0 1 0 1 |
| coeffs[0] = _mm256_shuffle_epi32(coeff, 0x00); |
| // coeffs 2 3 2 3 2 3 2 3 |
| coeffs[1] = _mm256_shuffle_epi32(coeff, 0x55); |
| // coeffs 4 5 4 5 4 5 4 5 |
| coeffs[2] = _mm256_shuffle_epi32(coeff, 0xaa); |
| // coeffs 6 7 6 7 6 7 6 7 |
| coeffs[3] = _mm256_shuffle_epi32(coeff, 0xff); |
| } |
| |
| static inline void prepare_coeffs_12taps( |
| const InterpFilterParams *const filter_params, const int subpel_q4, |
| __m256i *const coeffs /* [4] */) { |
| const int16_t *filter = av1_get_interp_filter_subpel_kernel( |
| filter_params, subpel_q4 & SUBPEL_MASK); |
| |
| __m128i coeff_8 = _mm_loadu_si128((__m128i *)filter); |
| __m256i coeff = _mm256_broadcastsi128_si256(coeff_8); |
| |
| // coeffs 0 1 0 1 0 1 0 1 |
| coeffs[0] = _mm256_shuffle_epi32(coeff, 0x00); |
| // coeffs 2 3 2 3 2 3 2 3 |
| coeffs[1] = _mm256_shuffle_epi32(coeff, 0x55); |
| // coeffs 4 5 4 5 4 5 4 5 |
| coeffs[2] = _mm256_shuffle_epi32(coeff, 0xaa); |
| // coeffs 6 7 6 7 6 7 6 7 |
| coeffs[3] = _mm256_shuffle_epi32(coeff, 0xff); |
| // coeffs 8 9 10 11 0 0 0 0 |
| coeff_8 = _mm_loadl_epi64((__m128i *)(filter + 8)); |
| coeff = _mm256_broadcastq_epi64(coeff_8); |
| coeffs[4] = _mm256_shuffle_epi32(coeff, 0x00); // coeffs 8 9 8 9 8 9 8 9 |
| coeffs[5] = _mm256_shuffle_epi32(coeff, 0x55); // coeffs 10 11 10 11.. 10 11 |
| } |
| |
| static inline __m256i convolve_lowbd(const __m256i *const s, |
| const __m256i *const coeffs) { |
| const __m256i res_01 = _mm256_maddubs_epi16(s[0], coeffs[0]); |
| const __m256i res_23 = _mm256_maddubs_epi16(s[1], coeffs[1]); |
| const __m256i res_45 = _mm256_maddubs_epi16(s[2], coeffs[2]); |
| const __m256i res_67 = _mm256_maddubs_epi16(s[3], coeffs[3]); |
| |
| // order: 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
| const __m256i res = _mm256_add_epi16(_mm256_add_epi16(res_01, res_45), |
| _mm256_add_epi16(res_23, res_67)); |
| |
| return res; |
| } |
| |
| static inline __m256i convolve_lowbd_6tap(const __m256i *const s, |
| const __m256i *const coeffs) { |
| const __m256i res_01 = _mm256_maddubs_epi16(s[0], coeffs[0]); |
| const __m256i res_23 = _mm256_maddubs_epi16(s[1], coeffs[1]); |
| const __m256i res_45 = _mm256_maddubs_epi16(s[2], coeffs[2]); |
| |
| // order: 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
| const __m256i res = |
| _mm256_add_epi16(_mm256_add_epi16(res_01, res_45), res_23); |
| |
| return res; |
| } |
| |
| static inline __m256i convolve_lowbd_4tap(const __m256i *const s, |
| const __m256i *const coeffs) { |
| const __m256i res_23 = _mm256_maddubs_epi16(s[0], coeffs[0]); |
| const __m256i res_45 = _mm256_maddubs_epi16(s[1], coeffs[1]); |
| |
| // order: 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
| const __m256i res = _mm256_add_epi16(res_45, res_23); |
| |
| return res; |
| } |
| |
| static inline __m256i convolve_6tap(const __m256i *const s, |
| const __m256i *const coeffs) { |
| const __m256i res_0 = _mm256_madd_epi16(s[0], coeffs[0]); |
| const __m256i res_1 = _mm256_madd_epi16(s[1], coeffs[1]); |
| const __m256i res_2 = _mm256_madd_epi16(s[2], coeffs[2]); |
| |
| const __m256i res = _mm256_add_epi32(_mm256_add_epi32(res_0, res_1), res_2); |
| |
| return res; |
| } |
| |
| static inline __m256i convolve_12taps(const __m256i *const s, |
| const __m256i *const coeffs) { |
| const __m256i res_0 = _mm256_madd_epi16(s[0], coeffs[0]); |
| const __m256i res_1 = _mm256_madd_epi16(s[1], coeffs[1]); |
| const __m256i res_2 = _mm256_madd_epi16(s[2], coeffs[2]); |
| const __m256i res_3 = _mm256_madd_epi16(s[3], coeffs[3]); |
| const __m256i res_4 = _mm256_madd_epi16(s[4], coeffs[4]); |
| const __m256i res_5 = _mm256_madd_epi16(s[5], coeffs[5]); |
| |
| const __m256i res1 = _mm256_add_epi32(_mm256_add_epi32(res_0, res_1), |
| _mm256_add_epi32(res_2, res_3)); |
| const __m256i res = _mm256_add_epi32(_mm256_add_epi32(res_4, res_5), res1); |
| |
| return res; |
| } |
| |
| static inline __m256i convolve(const __m256i *const s, |
| const __m256i *const coeffs) { |
| const __m256i res_0 = _mm256_madd_epi16(s[0], coeffs[0]); |
| const __m256i res_1 = _mm256_madd_epi16(s[1], coeffs[1]); |
| const __m256i res_2 = _mm256_madd_epi16(s[2], coeffs[2]); |
| const __m256i res_3 = _mm256_madd_epi16(s[3], coeffs[3]); |
| |
| const __m256i res = _mm256_add_epi32(_mm256_add_epi32(res_0, res_1), |
| _mm256_add_epi32(res_2, res_3)); |
| |
| return res; |
| } |
| |
| static inline __m256i convolve_4tap(const __m256i *const s, |
| const __m256i *const coeffs) { |
| const __m256i res_1 = _mm256_madd_epi16(s[0], coeffs[0]); |
| const __m256i res_2 = _mm256_madd_epi16(s[1], coeffs[1]); |
| |
| const __m256i res = _mm256_add_epi32(res_1, res_2); |
| return res; |
| } |
| |
| static inline __m256i convolve_lowbd_x(const __m256i data, |
| const __m256i *const coeffs, |
| const __m256i *const filt) { |
| __m256i s[4]; |
| |
| s[0] = _mm256_shuffle_epi8(data, filt[0]); |
| s[1] = _mm256_shuffle_epi8(data, filt[1]); |
| s[2] = _mm256_shuffle_epi8(data, filt[2]); |
| s[3] = _mm256_shuffle_epi8(data, filt[3]); |
| |
| return convolve_lowbd(s, coeffs); |
| } |
| |
| static inline __m256i convolve_lowbd_x_6tap(const __m256i data, |
| const __m256i *const coeffs, |
| const __m256i *const filt) { |
| __m256i s[4]; |
| |
| s[0] = _mm256_shuffle_epi8(data, filt[0]); |
| s[1] = _mm256_shuffle_epi8(data, filt[1]); |
| s[2] = _mm256_shuffle_epi8(data, filt[2]); |
| |
| return convolve_lowbd_6tap(s, coeffs); |
| } |
| |
| static inline __m256i convolve_lowbd_x_4tap(const __m256i data, |
| const __m256i *const coeffs, |
| const __m256i *const filt) { |
| __m256i s[2]; |
| |
| s[0] = _mm256_shuffle_epi8(data, filt[0]); |
| s[1] = _mm256_shuffle_epi8(data, filt[1]); |
| |
| return convolve_lowbd_4tap(s, coeffs); |
| } |
| |
| static inline void add_store_aligned_256(CONV_BUF_TYPE *const dst, |
| const __m256i *const res, |
| const int do_average) { |
| __m256i d; |
| if (do_average) { |
| d = _mm256_load_si256((__m256i *)dst); |
| d = _mm256_add_epi32(d, *res); |
| d = _mm256_srai_epi32(d, 1); |
| } else { |
| d = *res; |
| } |
| _mm256_store_si256((__m256i *)dst, d); |
| } |
| |
| static inline __m256i comp_avg(const __m256i *const data_ref_0, |
| const __m256i *const res_unsigned, |
| const __m256i *const wt, |
| const int use_dist_wtd_comp_avg) { |
| __m256i res; |
| if (use_dist_wtd_comp_avg) { |
| const __m256i data_lo = _mm256_unpacklo_epi16(*data_ref_0, *res_unsigned); |
| const __m256i data_hi = _mm256_unpackhi_epi16(*data_ref_0, *res_unsigned); |
| |
| const __m256i wt_res_lo = _mm256_madd_epi16(data_lo, *wt); |
| const __m256i wt_res_hi = _mm256_madd_epi16(data_hi, *wt); |
| |
| const __m256i res_lo = _mm256_srai_epi32(wt_res_lo, DIST_PRECISION_BITS); |
| const __m256i res_hi = _mm256_srai_epi32(wt_res_hi, DIST_PRECISION_BITS); |
| |
| res = _mm256_packs_epi32(res_lo, res_hi); |
| } else { |
| const __m256i wt_res = _mm256_add_epi16(*data_ref_0, *res_unsigned); |
| res = _mm256_srai_epi16(wt_res, 1); |
| } |
| return res; |
| } |
| |
| static inline __m256i convolve_rounding(const __m256i *const res_unsigned, |
| const __m256i *const offset_const, |
| const __m256i *const round_const, |
| const int round_shift) { |
| const __m256i res_signed = _mm256_sub_epi16(*res_unsigned, *offset_const); |
| const __m256i res_round = _mm256_srai_epi16( |
| _mm256_add_epi16(res_signed, *round_const), round_shift); |
| return res_round; |
| } |
| |
| static inline __m256i highbd_comp_avg(const __m256i *const data_ref_0, |
| const __m256i *const res_unsigned, |
| const __m256i *const wt0, |
| const __m256i *const wt1, |
| const int use_dist_wtd_comp_avg) { |
| __m256i res; |
| if (use_dist_wtd_comp_avg) { |
| const __m256i wt0_res = _mm256_mullo_epi32(*data_ref_0, *wt0); |
| const __m256i wt1_res = _mm256_mullo_epi32(*res_unsigned, *wt1); |
| const __m256i wt_res = _mm256_add_epi32(wt0_res, wt1_res); |
| res = _mm256_srai_epi32(wt_res, DIST_PRECISION_BITS); |
| } else { |
| const __m256i wt_res = _mm256_add_epi32(*data_ref_0, *res_unsigned); |
| res = _mm256_srai_epi32(wt_res, 1); |
| } |
| return res; |
| } |
| |
| static inline __m256i highbd_convolve_rounding( |
| const __m256i *const res_unsigned, const __m256i *const offset_const, |
| const __m256i *const round_const, const int round_shift) { |
| const __m256i res_signed = _mm256_sub_epi32(*res_unsigned, *offset_const); |
| const __m256i res_round = _mm256_srai_epi32( |
| _mm256_add_epi32(res_signed, *round_const), round_shift); |
| |
| return res_round; |
| } |
| |
| #endif // AOM_AOM_DSP_X86_CONVOLVE_AVX2_H_ |