| /* |
| * Copyright (c) 2021, Alliance for Open Media. All rights reserved |
| * |
| * This source code is subject to the terms of the BSD 3-Clause Clear License |
| * and the Alliance for Open Media Patent License 1.0. If the BSD 3-Clause Clear |
| * License was not distributed with this source code in the LICENSE file, you |
| * can obtain it at aomedia.org/license/software-license/bsd-3-c-c/. If the |
| * Alliance for Open Media Patent License 1.0 was not distributed with this |
| * source code in the PATENTS file, you can obtain it at |
| * aomedia.org/license/patent-license/. |
| */ |
| |
| #include <immintrin.h> |
| |
| #include "config/aom_dsp_rtcd.h" |
| |
| #include "aom/aom_integer.h" |
| |
| static INLINE void update_qp(__m256i *qp) { |
| int i; |
| for (i = 0; i < 5; ++i) { |
| qp[i] = _mm256_permute2x128_si256(qp[i], qp[i], 0x11); |
| } |
| } |
| |
| static INLINE void init_qp(const int32_t *zbin_ptr, const int32_t *round_ptr, |
| const int32_t *quant_ptr, const int32_t *dequant_ptr, |
| const int32_t *quant_shift_ptr, __m256i *qp, |
| int log_scale) { |
| qp[0] = _mm256_loadu_si256((const __m256i *)zbin_ptr); |
| qp[1] = _mm256_loadu_si256((const __m256i *)round_ptr); |
| qp[2] = _mm256_loadu_si256((const __m256i *)quant_ptr); |
| qp[3] = _mm256_loadu_si256((const __m256i *)dequant_ptr); |
| qp[4] = _mm256_loadu_si256((const __m256i *)quant_shift_ptr); |
| |
| if (log_scale > 0) { |
| const __m256i rnd = _mm256_set1_epi32((int16_t)(1 << (log_scale - 1))); |
| qp[0] = _mm256_add_epi32(qp[0], rnd); |
| qp[0] = _mm256_srai_epi32(qp[0], log_scale); |
| |
| qp[1] = _mm256_add_epi32(qp[1], rnd); |
| qp[1] = _mm256_srai_epi32(qp[1], log_scale); |
| } |
| // Subtracting 1 here eliminates a _mm256_cmpeq_epi32() instruction when |
| // calculating the zbin mask. |
| qp[0] = _mm256_sub_epi32(qp[0], _mm256_set1_epi32(1)); |
| } |
| |
| // Note: |
| // *x is vector multiplied by *y which is 16 int32_t parallel multiplication |
| // and right shift 16. |
| static INLINE __m256i mm256_mul_shift_epi32(const __m256i *x, const __m256i *y, |
| int log_scale) { |
| __m256i prod_lo = _mm256_mul_epi32(*x, *y); |
| __m256i prod_hi = _mm256_srli_epi64(*x, 32); |
| const __m256i mult_hi = _mm256_srli_epi64(*y, 32); |
| prod_hi = _mm256_mul_epi32(prod_hi, mult_hi); |
| |
| prod_lo = _mm256_srli_epi64(prod_lo, 16 - log_scale); |
| prod_hi = _mm256_slli_epi64(prod_hi, 32 - (16 - log_scale)); |
| |
| return _mm256_blend_epi32(prod_lo, prod_hi, 0xAA); |
| } |
| |
| static AOM_FORCE_INLINE __m256i get_max_lane_eob(const int16_t *iscan_ptr, |
| __m256i eobmax, |
| __m256i nz_mask) { |
| const __m256i packed_nz_mask = _mm256_packs_epi32(nz_mask, nz_mask); |
| const __m256i packed_nz_mask_perm = |
| _mm256_permute4x64_epi64(packed_nz_mask, 0xD8); |
| const __m256i iscan = |
| _mm256_castsi128_si256(_mm_loadu_si128((const __m128i *)iscan_ptr)); |
| const __m256i iscan_plus1 = _mm256_sub_epi16(iscan, packed_nz_mask_perm); |
| const __m256i nz_iscan = _mm256_and_si256(iscan_plus1, packed_nz_mask_perm); |
| return _mm256_max_epi16(eobmax, nz_iscan); |
| } |
| |
| // Get the max eob from the lower 128 bits. |
| static AOM_FORCE_INLINE uint16_t get_max_eob(__m256i eob) { |
| __m256i eob_s; |
| eob_s = _mm256_shuffle_epi32(eob, 0xe); |
| eob = _mm256_max_epi16(eob, eob_s); |
| eob_s = _mm256_shufflelo_epi16(eob, 0xe); |
| eob = _mm256_max_epi16(eob, eob_s); |
| eob_s = _mm256_shufflelo_epi16(eob, 1); |
| eob = _mm256_max_epi16(eob, eob_s); |
| return (uint16_t)_mm256_extract_epi16(eob, 0); |
| } |
| |
| static INLINE void quantize(const __m256i *qp, const tran_low_t *coeff_ptr, |
| const int16_t *iscan_ptr, tran_low_t *qcoeff, |
| tran_low_t *dqcoeff, __m256i *eob, int log_scale) { |
| const __m256i coeff = _mm256_loadu_si256((const __m256i *)coeff_ptr); |
| const __m256i abs_coeff = _mm256_abs_epi32(coeff); |
| const __m256i zbin_mask = _mm256_cmpgt_epi32(abs_coeff, qp[0]); |
| const __m256i offset = _mm256_set1_epi32((1 << QUANT_TABLE_BITS) >> 1); |
| if (UNLIKELY(_mm256_movemask_epi8(zbin_mask) == 0)) { |
| const __m256i zero = _mm256_setzero_si256(); |
| _mm256_storeu_si256((__m256i *)qcoeff, zero); |
| _mm256_storeu_si256((__m256i *)dqcoeff, zero); |
| return; |
| } |
| |
| const __m256i tmp_rnd = |
| _mm256_and_si256(_mm256_add_epi32(abs_coeff, qp[1]), zbin_mask); |
| const __m256i tmp = mm256_mul_shift_epi32(&tmp_rnd, &qp[2], 0); |
| const __m256i tmp2 = _mm256_add_epi32(tmp, tmp_rnd); |
| const __m256i abs_q = mm256_mul_shift_epi32(&tmp2, &qp[4], log_scale); |
| __m256i abs_dq = _mm256_mullo_epi32(abs_q, qp[3]); |
| abs_dq = _mm256_add_epi32(abs_dq, offset); |
| abs_dq = _mm256_srli_epi32(abs_dq, QUANT_TABLE_BITS + log_scale); |
| const __m256i nz_mask = _mm256_cmpgt_epi32(abs_q, _mm256_setzero_si256()); |
| const __m256i q = _mm256_sign_epi32(abs_q, coeff); |
| const __m256i dq = _mm256_sign_epi32(abs_dq, coeff); |
| |
| _mm256_storeu_si256((__m256i *)qcoeff, q); |
| _mm256_storeu_si256((__m256i *)dqcoeff, dq); |
| |
| *eob = get_max_lane_eob(iscan_ptr, *eob, nz_mask); |
| } |
| |
| void aom_highbd_quantize_b_avx2( |
| const tran_low_t *coeff_ptr, intptr_t n_coeffs, const int32_t *zbin_ptr, |
| const int32_t *round_ptr, const int32_t *quant_ptr, |
| const int32_t *quant_shift_ptr, tran_low_t *qcoeff_ptr, |
| tran_low_t *dqcoeff_ptr, const int32_t *dequant_ptr, uint16_t *eob_ptr, |
| const int16_t *scan, const int16_t *iscan, const int log_scale) { |
| (void)scan; |
| const unsigned int step = 8; |
| |
| __m256i qp[5]; |
| init_qp(zbin_ptr, round_ptr, quant_ptr, dequant_ptr, quant_shift_ptr, qp, |
| log_scale); |
| |
| __m256i eob = _mm256_setzero_si256(); |
| quantize(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob, log_scale); |
| |
| coeff_ptr += step; |
| qcoeff_ptr += step; |
| dqcoeff_ptr += step; |
| iscan += step; |
| n_coeffs -= step; |
| |
| update_qp(qp); |
| |
| while (n_coeffs > 0) { |
| quantize(qp, coeff_ptr, iscan, qcoeff_ptr, dqcoeff_ptr, &eob, log_scale); |
| |
| coeff_ptr += step; |
| qcoeff_ptr += step; |
| dqcoeff_ptr += step; |
| iscan += step; |
| n_coeffs -= step; |
| } |
| |
| *eob_ptr = get_max_eob(eob); |
| } |