|  | /* | 
|  | * Copyright (c) 2016, Alliance for Open Media. All rights reserved | 
|  | * | 
|  | * This source code is subject to the terms of the BSD 2 Clause License and | 
|  | * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License | 
|  | * was not distributed with this source code in the LICENSE file, you can | 
|  | * obtain it at www.aomedia.org/license/software. If the Alliance for Open | 
|  | * Media Patent License 1.0 was not distributed with this source code in the | 
|  | * PATENTS file, you can obtain it at www.aomedia.org/license/patent. | 
|  | */ | 
|  |  | 
|  | #ifndef AOM_DSP_X86_BLEND_SSE4_H_ | 
|  | #define AOM_DSP_X86_BLEND_SSE4_H_ | 
|  |  | 
|  | #include "aom_dsp/blend.h" | 
|  | #include "aom_dsp/x86/synonyms.h" | 
|  |  | 
|  | ////////////////////////////////////////////////////////////////////////////// | 
|  | // Common kernels | 
|  | ////////////////////////////////////////////////////////////////////////////// | 
|  |  | 
|  | static INLINE __m128i blend_4(const uint8_t *src0, const uint8_t *src1, | 
|  | const __m128i v_m0_w, const __m128i v_m1_w) { | 
|  | const __m128i v_s0_b = xx_loadl_32(src0); | 
|  | const __m128i v_s1_b = xx_loadl_32(src1); | 
|  | const __m128i v_s0_w = _mm_cvtepu8_epi16(v_s0_b); | 
|  | const __m128i v_s1_w = _mm_cvtepu8_epi16(v_s1_b); | 
|  |  | 
|  | const __m128i v_p0_w = _mm_mullo_epi16(v_s0_w, v_m0_w); | 
|  | const __m128i v_p1_w = _mm_mullo_epi16(v_s1_w, v_m1_w); | 
|  |  | 
|  | const __m128i v_sum_w = _mm_add_epi16(v_p0_w, v_p1_w); | 
|  |  | 
|  | const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); | 
|  |  | 
|  | return v_res_w; | 
|  | } | 
|  |  | 
|  | static INLINE __m128i blend_8(const uint8_t *src0, const uint8_t *src1, | 
|  | const __m128i v_m0_w, const __m128i v_m1_w) { | 
|  | const __m128i v_s0_b = xx_loadl_64(src0); | 
|  | const __m128i v_s1_b = xx_loadl_64(src1); | 
|  | const __m128i v_s0_w = _mm_cvtepu8_epi16(v_s0_b); | 
|  | const __m128i v_s1_w = _mm_cvtepu8_epi16(v_s1_b); | 
|  |  | 
|  | const __m128i v_p0_w = _mm_mullo_epi16(v_s0_w, v_m0_w); | 
|  | const __m128i v_p1_w = _mm_mullo_epi16(v_s1_w, v_m1_w); | 
|  |  | 
|  | const __m128i v_sum_w = _mm_add_epi16(v_p0_w, v_p1_w); | 
|  |  | 
|  | const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); | 
|  |  | 
|  | return v_res_w; | 
|  | } | 
|  |  | 
|  | #if CONFIG_HIGHBITDEPTH | 
|  | typedef __m128i (*blend_unit_fn)(const uint16_t *src0, const uint16_t *src1, | 
|  | const __m128i v_m0_w, const __m128i v_m1_w); | 
|  |  | 
|  | static INLINE __m128i blend_4_b10(const uint16_t *src0, const uint16_t *src1, | 
|  | const __m128i v_m0_w, const __m128i v_m1_w) { | 
|  | const __m128i v_s0_w = xx_loadl_64(src0); | 
|  | const __m128i v_s1_w = xx_loadl_64(src1); | 
|  |  | 
|  | const __m128i v_p0_w = _mm_mullo_epi16(v_s0_w, v_m0_w); | 
|  | const __m128i v_p1_w = _mm_mullo_epi16(v_s1_w, v_m1_w); | 
|  |  | 
|  | const __m128i v_sum_w = _mm_add_epi16(v_p0_w, v_p1_w); | 
|  |  | 
|  | const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); | 
|  |  | 
|  | return v_res_w; | 
|  | } | 
|  |  | 
|  | static INLINE __m128i blend_8_b10(const uint16_t *src0, const uint16_t *src1, | 
|  | const __m128i v_m0_w, const __m128i v_m1_w) { | 
|  | const __m128i v_s0_w = xx_loadu_128(src0); | 
|  | const __m128i v_s1_w = xx_loadu_128(src1); | 
|  |  | 
|  | const __m128i v_p0_w = _mm_mullo_epi16(v_s0_w, v_m0_w); | 
|  | const __m128i v_p1_w = _mm_mullo_epi16(v_s1_w, v_m1_w); | 
|  |  | 
|  | const __m128i v_sum_w = _mm_add_epi16(v_p0_w, v_p1_w); | 
|  |  | 
|  | const __m128i v_res_w = xx_roundn_epu16(v_sum_w, AOM_BLEND_A64_ROUND_BITS); | 
|  |  | 
|  | return v_res_w; | 
|  | } | 
|  |  | 
|  | static INLINE __m128i blend_4_b12(const uint16_t *src0, const uint16_t *src1, | 
|  | const __m128i v_m0_w, const __m128i v_m1_w) { | 
|  | const __m128i v_s0_w = xx_loadl_64(src0); | 
|  | const __m128i v_s1_w = xx_loadl_64(src1); | 
|  |  | 
|  | // Interleave | 
|  | const __m128i v_m01_w = _mm_unpacklo_epi16(v_m0_w, v_m1_w); | 
|  | const __m128i v_s01_w = _mm_unpacklo_epi16(v_s0_w, v_s1_w); | 
|  |  | 
|  | // Multiply-Add | 
|  | const __m128i v_sum_d = _mm_madd_epi16(v_s01_w, v_m01_w); | 
|  |  | 
|  | // Scale | 
|  | const __m128i v_ssum_d = | 
|  | _mm_srli_epi32(v_sum_d, AOM_BLEND_A64_ROUND_BITS - 1); | 
|  |  | 
|  | // Pack | 
|  | const __m128i v_pssum_d = _mm_packs_epi32(v_ssum_d, v_ssum_d); | 
|  |  | 
|  | // Round | 
|  | const __m128i v_res_w = xx_round_epu16(v_pssum_d); | 
|  |  | 
|  | return v_res_w; | 
|  | } | 
|  |  | 
|  | static INLINE __m128i blend_8_b12(const uint16_t *src0, const uint16_t *src1, | 
|  | const __m128i v_m0_w, const __m128i v_m1_w) { | 
|  | const __m128i v_s0_w = xx_loadu_128(src0); | 
|  | const __m128i v_s1_w = xx_loadu_128(src1); | 
|  |  | 
|  | // Interleave | 
|  | const __m128i v_m01l_w = _mm_unpacklo_epi16(v_m0_w, v_m1_w); | 
|  | const __m128i v_m01h_w = _mm_unpackhi_epi16(v_m0_w, v_m1_w); | 
|  | const __m128i v_s01l_w = _mm_unpacklo_epi16(v_s0_w, v_s1_w); | 
|  | const __m128i v_s01h_w = _mm_unpackhi_epi16(v_s0_w, v_s1_w); | 
|  |  | 
|  | // Multiply-Add | 
|  | const __m128i v_suml_d = _mm_madd_epi16(v_s01l_w, v_m01l_w); | 
|  | const __m128i v_sumh_d = _mm_madd_epi16(v_s01h_w, v_m01h_w); | 
|  |  | 
|  | // Scale | 
|  | const __m128i v_ssuml_d = | 
|  | _mm_srli_epi32(v_suml_d, AOM_BLEND_A64_ROUND_BITS - 1); | 
|  | const __m128i v_ssumh_d = | 
|  | _mm_srli_epi32(v_sumh_d, AOM_BLEND_A64_ROUND_BITS - 1); | 
|  |  | 
|  | // Pack | 
|  | const __m128i v_pssum_d = _mm_packs_epi32(v_ssuml_d, v_ssumh_d); | 
|  |  | 
|  | // Round | 
|  | const __m128i v_res_w = xx_round_epu16(v_pssum_d); | 
|  |  | 
|  | return v_res_w; | 
|  | } | 
|  | #endif  // CONFIG_HIGHBITDEPTH | 
|  |  | 
|  | #endif  // AOM_DSP_X86_BLEND_SSE4_H_ |