|  | /* | 
|  | * Copyright (c) 2017, Alliance for Open Media. All rights reserved | 
|  | * | 
|  | * This source code is subject to the terms of the BSD 2 Clause License and | 
|  | * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License | 
|  | * was not distributed with this source code in the LICENSE file, you can | 
|  | * obtain it at www.aomedia.org/license/software. If the Alliance for Open | 
|  | * Media Patent License 1.0 was not distributed with this source code in the | 
|  | * PATENTS file, you can obtain it at www.aomedia.org/license/patent. | 
|  | */ | 
|  |  | 
|  | #include <immintrin.h> | 
|  |  | 
|  | #include "config/aom_dsp_rtcd.h" | 
|  |  | 
|  | #include "aom/aom_integer.h" | 
|  |  | 
|  | static INLINE void init_one_qp(const __m128i *p, __m256i *qp) { | 
|  | const __m128i sign = _mm_srai_epi16(*p, 15); | 
|  | const __m128i dc = _mm_unpacklo_epi16(*p, sign); | 
|  | const __m128i ac = _mm_unpackhi_epi16(*p, sign); | 
|  | *qp = _mm256_insertf128_si256(_mm256_castsi128_si256(dc), ac, 1); | 
|  | } | 
|  |  | 
|  | static INLINE void update_qp(__m256i *qp) { | 
|  | int i; | 
|  | for (i = 0; i < 5; ++i) { | 
|  | qp[i] = _mm256_permute2x128_si256(qp[i], qp[i], 0x11); | 
|  | } | 
|  | } | 
|  |  | 
|  | static INLINE void init_qp(const int16_t *zbin_ptr, const int16_t *round_ptr, | 
|  | const int16_t *quant_ptr, const int16_t *dequant_ptr, | 
|  | const int16_t *quant_shift_ptr, __m256i *qp, | 
|  | int log_scale) { | 
|  | const __m128i zbin = _mm_loadu_si128((const __m128i *)zbin_ptr); | 
|  | const __m128i round = _mm_loadu_si128((const __m128i *)round_ptr); | 
|  | const __m128i quant = _mm_loadu_si128((const __m128i *)quant_ptr); | 
|  | const __m128i dequant = _mm_loadu_si128((const __m128i *)dequant_ptr); | 
|  | const __m128i quant_shift = _mm_loadu_si128((const __m128i *)quant_shift_ptr); | 
|  | init_one_qp(&zbin, &qp[0]); | 
|  | init_one_qp(&round, &qp[1]); | 
|  | init_one_qp(&quant, &qp[2]); | 
|  | init_one_qp(&dequant, &qp[3]); | 
|  | init_one_qp(&quant_shift, &qp[4]); | 
|  | if (log_scale > 0) { | 
|  | const __m256i rnd = _mm256_set1_epi32((int16_t)(1 << (log_scale - 1))); | 
|  | qp[0] = _mm256_add_epi32(qp[0], rnd); | 
|  | qp[0] = _mm256_srai_epi32(qp[0], log_scale); | 
|  |  | 
|  | qp[1] = _mm256_add_epi32(qp[1], rnd); | 
|  | qp[1] = _mm256_srai_epi32(qp[1], log_scale); | 
|  | } | 
|  | // Subtracting 1 here eliminates a _mm256_cmpeq_epi32() instruction when | 
|  | // calculating the zbin mask. | 
|  | qp[0] = _mm256_sub_epi32(qp[0], _mm256_set1_epi32(1)); | 
|  | } | 
|  |  | 
|  | // Note: | 
|  | // *x is vector multiplied by *y which is 16 int32_t parallel multiplication | 
|  | // and right shift 16.  The output, 16 int32_t is save in *p. | 
|  | static INLINE __m256i mm256_mul_shift_epi32(const __m256i *x, | 
|  | const __m256i *y) { | 
|  | __m256i prod_lo = _mm256_mul_epi32(*x, *y); | 
|  | __m256i prod_hi = _mm256_srli_epi64(*x, 32); | 
|  | const __m256i mult_hi = _mm256_srli_epi64(*y, 32); | 
|  | prod_hi = _mm256_mul_epi32(prod_hi, mult_hi); | 
|  |  | 
|  | prod_lo = _mm256_srli_epi64(prod_lo, 16); | 
|  | const __m256i mask = _mm256_set_epi32(0, -1, 0, -1, 0, -1, 0, -1); | 
|  | prod_lo = _mm256_and_si256(prod_lo, mask); | 
|  | prod_hi = _mm256_srli_epi64(prod_hi, 16); | 
|  |  | 
|  | prod_hi = _mm256_slli_epi64(prod_hi, 32); | 
|  | return _mm256_or_si256(prod_lo, prod_hi); | 
|  | } | 
|  |  | 
|  | static AOM_FORCE_INLINE __m256i get_max_lane_eob(const int16_t *iscan_ptr, | 
|  | __m256i eobmax, | 
|  | __m256i nz_mask) { | 
|  | const __m256i packed_nz_mask = _mm256_packs_epi32(nz_mask, nz_mask); | 
|  | const __m256i packed_nz_mask_perm = | 
|  | _mm256_permute4x64_epi64(packed_nz_mask, 0xD8); | 
|  | const __m256i iscan = | 
|  | _mm256_castsi128_si256(_mm_loadu_si128((const __m128i *)iscan_ptr)); | 
|  | const __m256i iscan_plus1 = _mm256_sub_epi16(iscan, packed_nz_mask_perm); | 
|  | const __m256i nz_iscan = _mm256_and_si256(iscan_plus1, packed_nz_mask_perm); | 
|  | return _mm256_max_epi16(eobmax, nz_iscan); | 
|  | } | 
|  |  | 
|  | // Get the max eob from the lower 128 bits. | 
|  | static AOM_FORCE_INLINE uint16_t get_max_eob(__m256i eob) { | 
|  | __m256i eob_s; | 
|  | eob_s = _mm256_shuffle_epi32(eob, 0xe); | 
|  | eob = _mm256_max_epi16(eob, eob_s); | 
|  | eob_s = _mm256_shufflelo_epi16(eob, 0xe); | 
|  | eob = _mm256_max_epi16(eob, eob_s); | 
|  | eob_s = _mm256_shufflelo_epi16(eob, 1); | 
|  | eob = _mm256_max_epi16(eob, eob_s); | 
|  | return (uint16_t)_mm256_extract_epi16(eob, 0); | 
|  | } | 
|  |  | 
|  | static AOM_FORCE_INLINE __m256i mm256_mul_shift_epi32_logscale(const __m256i *x, | 
|  | const __m256i *y, | 
|  | int log_scale) { | 
|  | __m256i prod_lo = _mm256_mul_epi32(*x, *y); | 
|  | __m256i prod_hi = _mm256_srli_epi64(*x, 32); | 
|  | const __m256i mult_hi = _mm256_srli_epi64(*y, 32); | 
|  | prod_hi = _mm256_mul_epi32(prod_hi, mult_hi); | 
|  | prod_lo = _mm256_srli_epi64(prod_lo, 16 - log_scale); | 
|  | const __m256i mask = _mm256_set_epi32(0, -1, 0, -1, 0, -1, 0, -1); | 
|  | prod_lo = _mm256_and_si256(prod_lo, mask); | 
|  | prod_hi = _mm256_srli_epi64(prod_hi, 16 - log_scale); | 
|  | prod_hi = _mm256_slli_epi64(prod_hi, 32); | 
|  | return _mm256_or_si256(prod_lo, prod_hi); | 
|  | } | 
|  |  | 
|  | static AOM_FORCE_INLINE void quantize_logscale( | 
|  | const __m256i *qp, const tran_low_t *coeff_ptr, const int16_t *iscan_ptr, | 
|  | tran_low_t *qcoeff, tran_low_t *dqcoeff, __m256i *eob, int log_scale) { | 
|  | const __m256i coeff = _mm256_loadu_si256((const __m256i *)coeff_ptr); | 
|  | const __m256i abs_coeff = _mm256_abs_epi32(coeff); | 
|  | const __m256i zbin_mask = _mm256_cmpgt_epi32(abs_coeff, qp[0]); | 
|  |  | 
|  | if (UNLIKELY(_mm256_movemask_epi8(zbin_mask) == 0)) { | 
|  | const __m256i zero = _mm256_setzero_si256(); | 
|  | _mm256_storeu_si256((__m256i *)qcoeff, zero); | 
|  | _mm256_storeu_si256((__m256i *)dqcoeff, zero); | 
|  | return; | 
|  | } | 
|  |  | 
|  | const __m256i tmp_rnd = | 
|  | _mm256_and_si256(_mm256_add_epi32(abs_coeff, qp[1]), zbin_mask); | 
|  | // const int64_t tmp2 = ((tmpw * quant_ptr[rc != 0]) >> 16) + tmpw; | 
|  | const __m256i tmp = mm256_mul_shift_epi32_logscale(&tmp_rnd, &qp[2], 0); | 
|  | const __m256i tmp2 = _mm256_add_epi32(tmp, tmp_rnd); | 
|  | // const int abs_qcoeff = (int)((tmp2 * quant_shift_ptr[rc != 0]) >> | 
|  | //                              (16 - log_scale + AOM_QM_BITS)); | 
|  | const __m256i abs_q = | 
|  | mm256_mul_shift_epi32_logscale(&tmp2, &qp[4], log_scale); | 
|  | const __m256i abs_dq = | 
|  | _mm256_srli_epi32(_mm256_mullo_epi32(abs_q, qp[3]), log_scale); | 
|  | const __m256i nz_mask = _mm256_cmpgt_epi32(abs_q, _mm256_setzero_si256()); | 
|  | const __m256i q = _mm256_sign_epi32(abs_q, coeff); | 
|  | const __m256i dq = _mm256_sign_epi32(abs_dq, coeff); | 
|  |  | 
|  | _mm256_storeu_si256((__m256i *)qcoeff, q); | 
|  | _mm256_storeu_si256((__m256i *)dqcoeff, dq); | 
|  |  | 
|  | *eob = get_max_lane_eob(iscan_ptr, *eob, nz_mask); | 
|  | } | 
|  |  | 
|  | static AOM_FORCE_INLINE void quantize(const __m256i *qp, | 
|  | const tran_low_t *coeff_ptr, | 
|  | const int16_t *iscan_ptr, | 
|  | tran_low_t *qcoeff, tran_low_t *dqcoeff, | 
|  | __m256i *eob) { | 
|  | const __m256i coeff = _mm256_loadu_si256((const __m256i *)coeff_ptr); | 
|  | const __m256i abs_coeff = _mm256_abs_epi32(coeff); | 
|  | const __m256i zbin_mask = _mm256_cmpgt_epi32(abs_coeff, qp[0]); | 
|  |  | 
|  | if (UNLIKELY(_mm256_movemask_epi8(zbin_mask) == 0)) { | 
|  | const __m256i zero = _mm256_setzero_si256(); | 
|  | _mm256_storeu_si256((__m256i *)qcoeff, zero); | 
|  | _mm256_storeu_si256((__m256i *)dqcoeff, zero); | 
|  | return; | 
|  | } | 
|  |  | 
|  | const __m256i tmp_rnd = | 
|  | _mm256_and_si256(_mm256_add_epi32(abs_coeff, qp[1]), zbin_mask); | 
|  | const __m256i tmp = mm256_mul_shift_epi32(&tmp_rnd, &qp[2]); | 
|  | const __m256i tmp2 = _mm256_add_epi32(tmp, tmp_rnd); | 
|  | const __m256i abs_q = mm256_mul_shift_epi32(&tmp2, &qp[4]); | 
|  | const __m256i abs_dq = _mm256_mullo_epi32(abs_q, qp[3]); | 
|  | const __m256i nz_mask = _mm256_cmpgt_epi32(abs_q, _mm256_setzero_si256()); | 
|  | const __m256i q = _mm256_sign_epi32(abs_q, coeff); | 
|  | const __m256i dq = _mm256_sign_epi32(abs_dq, coeff); | 
|  |  | 
|  | _mm256_storeu_si256((__m256i *)qcoeff, q); | 
|  | _mm256_storeu_si256((__m256i *)dqcoeff, dq); | 
|  |  | 
|  | *eob = get_max_lane_eob(iscan_ptr, *eob, nz_mask); | 
|  | } | 
|  |  | 
|  | void aom_highbd_quantize_b_avx2(const tran_low_t *coeff_ptr, intptr_t n_coeffs, | 
|  | const int16_t *zbin_ptr, | 
|  | const int16_t *round_ptr, | 
|  | const int16_t *quant_ptr, | 
|  | const int16_t *quant_shift_ptr, | 
|  | tran_low_t *qcoeff_ptr, tran_low_t *dqcoeff_ptr, | 
|  | const int16_t *dequant_ptr, uint16_t *eob_ptr, | 
|  | const int16_t *scan, const int16_t *iscan) { | 
|  | (void)scan; | 
|  | const int step = 8; | 
|  |  | 
|  | __m256i eob = _mm256_setzero_si256(); | 
|  | __m256i qp[5]; | 
|  |  | 
|  | init_qp(zbin_ptr, round_ptr, quant_ptr, dequant_ptr, quant_shift_ptr, qp, 0); | 
|  |  | 
|  | quantize(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob); | 
|  |  | 
|  | coeff_ptr += step; | 
|  | qcoeff_ptr += step; | 
|  | dqcoeff_ptr += step; | 
|  | iscan += step; | 
|  | n_coeffs -= step; | 
|  |  | 
|  | update_qp(qp); | 
|  |  | 
|  | while (n_coeffs > 0) { | 
|  | quantize(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob); | 
|  |  | 
|  | coeff_ptr += step; | 
|  | qcoeff_ptr += step; | 
|  | dqcoeff_ptr += step; | 
|  | iscan += step; | 
|  | n_coeffs -= step; | 
|  | } | 
|  |  | 
|  | *eob_ptr = get_max_eob(eob); | 
|  | } | 
|  |  | 
|  | void aom_highbd_quantize_b_32x32_avx2( | 
|  | const tran_low_t *coeff_ptr, intptr_t n_coeffs, const int16_t *zbin_ptr, | 
|  | const int16_t *round_ptr, const int16_t *quant_ptr, | 
|  | const int16_t *quant_shift_ptr, tran_low_t *qcoeff_ptr, | 
|  | tran_low_t *dqcoeff_ptr, const int16_t *dequant_ptr, uint16_t *eob_ptr, | 
|  | const int16_t *scan, const int16_t *iscan) { | 
|  | (void)scan; | 
|  | const unsigned int step = 8; | 
|  |  | 
|  | __m256i eob = _mm256_setzero_si256(); | 
|  | __m256i qp[5]; | 
|  | init_qp(zbin_ptr, round_ptr, quant_ptr, dequant_ptr, quant_shift_ptr, qp, 1); | 
|  |  | 
|  | quantize_logscale(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob, 1); | 
|  |  | 
|  | coeff_ptr += step; | 
|  | qcoeff_ptr += step; | 
|  | dqcoeff_ptr += step; | 
|  | iscan += step; | 
|  | n_coeffs -= step; | 
|  |  | 
|  | update_qp(qp); | 
|  |  | 
|  | while (n_coeffs > 0) { | 
|  | quantize_logscale(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob, 1); | 
|  |  | 
|  | coeff_ptr += step; | 
|  | qcoeff_ptr += step; | 
|  | dqcoeff_ptr += step; | 
|  | iscan += step; | 
|  | n_coeffs -= step; | 
|  | } | 
|  |  | 
|  | *eob_ptr = get_max_eob(eob); | 
|  | } | 
|  |  | 
|  | void aom_highbd_quantize_b_64x64_avx2( | 
|  | const tran_low_t *coeff_ptr, intptr_t n_coeffs, const int16_t *zbin_ptr, | 
|  | const int16_t *round_ptr, const int16_t *quant_ptr, | 
|  | const int16_t *quant_shift_ptr, tran_low_t *qcoeff_ptr, | 
|  | tran_low_t *dqcoeff_ptr, const int16_t *dequant_ptr, uint16_t *eob_ptr, | 
|  | const int16_t *scan, const int16_t *iscan) { | 
|  | (void)scan; | 
|  | const int step = 8; | 
|  |  | 
|  | __m256i eob = _mm256_setzero_si256(); | 
|  | __m256i qp[5]; | 
|  | init_qp(zbin_ptr, round_ptr, quant_ptr, dequant_ptr, quant_shift_ptr, qp, 2); | 
|  |  | 
|  | quantize_logscale(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob, 2); | 
|  |  | 
|  | coeff_ptr += step; | 
|  | qcoeff_ptr += step; | 
|  | dqcoeff_ptr += step; | 
|  | iscan += step; | 
|  | n_coeffs -= step; | 
|  |  | 
|  | update_qp(qp); | 
|  |  | 
|  | while (n_coeffs > 0) { | 
|  | quantize_logscale(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob, 2); | 
|  |  | 
|  | coeff_ptr += step; | 
|  | qcoeff_ptr += step; | 
|  | dqcoeff_ptr += step; | 
|  | iscan += step; | 
|  | n_coeffs -= step; | 
|  | } | 
|  |  | 
|  | *eob_ptr = get_max_eob(eob); | 
|  | } |