| /* |
| * Copyright (c) 2016, Alliance for Open Media. All rights reserved |
| * |
| * This source code is subject to the terms of the BSD 2 Clause License and |
| * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License |
| * was not distributed with this source code in the LICENSE file, you can |
| * obtain it at www.aomedia.org/license/software. If the Alliance for Open |
| * Media Patent License 1.0 was not distributed with this source code in the |
| * PATENTS file, you can obtain it at www.aomedia.org/license/patent. |
| */ |
| |
| #include <math.h> |
| |
| #include "config/aom_dsp_rtcd.h" |
| #include "config/av1_rtcd.h" |
| |
| #include "aom_ports/mem.h" |
| #include "av1/common/av1_inv_txfm1d_cfg.h" |
| #include "av1/common/av1_txfm.h" |
| #include "av1/common/blockd.h" |
| #include "av1/common/enums.h" |
| #include "av1/common/idct.h" |
| #if CONFIG_DSPL_RESIDUAL |
| #include "av1/common/resize.h" |
| #include "av1/common/scan.h" |
| #endif // CONFIG_DSPL_RESIDUAL |
| #if CONFIG_NN_RECON |
| #include "av1/common/cnn_tflite.h" |
| #include "av1/common/nn_recon.h" |
| #endif // CONFIG_NN_RECON |
| |
| int av1_get_tx_scale(const TX_SIZE tx_size) { |
| const int pels = tx_size_2d[tx_size]; |
| // Largest possible pels is 4096 (64x64). |
| return (pels > 256) + (pels > 1024); |
| } |
| |
| // NOTE: The implementation of all inverses need to be aware of the fact |
| // that input and output could be the same buffer. |
| |
| // idct |
| void av1_highbd_iwht4x4_add(const tran_low_t *input, uint8_t *dest, int stride, |
| int eob, int bd) { |
| if (eob > 1) |
| av1_highbd_iwht4x4_16_add(input, dest, stride, bd); |
| else |
| av1_highbd_iwht4x4_1_add(input, dest, stride, bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_4x4_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| assert(av1_ext_tx_used[txfm_param->tx_set_type][txfm_param->tx_type]); |
| int eob = txfm_param->eob; |
| int bd = txfm_param->bd; |
| int lossless = txfm_param->lossless; |
| const int32_t *src = cast_to_int32(input); |
| const TX_TYPE tx_type = txfm_param->tx_type; |
| if (lossless) { |
| assert(tx_type == DCT_DCT); |
| av1_highbd_iwht4x4_add(input, dest, stride, eob, bd); |
| return; |
| } |
| |
| av1_inv_txfm2d_add_4x4_c(src, CONVERT_TO_SHORTPTR(dest), stride, tx_type, |
| txfm_param->mode, bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_4x8_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| assert(av1_ext_tx_used[txfm_param->tx_set_type][txfm_param->tx_type]); |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_4x8_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_8x4_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| assert(av1_ext_tx_used[txfm_param->tx_set_type][txfm_param->tx_type]); |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_8x4_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_16x32_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_16x32_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_32x16_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_32x16_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_16x4_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_16x4_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_4x16_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_4x16_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_32x8_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_32x8_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_8x32_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_8x32_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_32x64_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_32x64_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_64x32_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_64x32_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_16x64_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_16x64_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_64x16_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_64x16_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_8x8_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| int bd = txfm_param->bd; |
| const TX_TYPE tx_type = txfm_param->tx_type; |
| const int32_t *src = cast_to_int32(input); |
| |
| av1_inv_txfm2d_add_8x8_c(src, CONVERT_TO_SHORTPTR(dest), stride, tx_type, |
| txfm_param->mode, bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_16x16_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| int bd = txfm_param->bd; |
| const TX_TYPE tx_type = txfm_param->tx_type; |
| const int32_t *src = cast_to_int32(input); |
| |
| av1_inv_txfm2d_add_16x16_c(src, CONVERT_TO_SHORTPTR(dest), stride, tx_type, |
| txfm_param->mode, bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_8x16_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_8x16_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_16x8_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_16x8_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_32x32_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int bd = txfm_param->bd; |
| const TX_TYPE tx_type = txfm_param->tx_type; |
| const int32_t *src = cast_to_int32(input); |
| |
| av1_inv_txfm2d_add_32x32_c(src, CONVERT_TO_SHORTPTR(dest), stride, tx_type, |
| txfm_param->mode, bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_64x64_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int bd = txfm_param->bd; |
| const TX_TYPE tx_type = txfm_param->tx_type; |
| const int32_t *src = cast_to_int32(input); |
| assert(tx_type == DCT_DCT); |
| av1_inv_txfm2d_add_64x64_c(src, CONVERT_TO_SHORTPTR(dest), stride, tx_type, |
| txfm_param->mode, bd); |
| } |
| |
| #if CONFIG_FLEX_PARTITION |
| void av1_highbd_inv_txfm_add_4x32_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_4x32_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_32x4_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_32x4_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_8x64_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_8x64_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_64x8_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_64x8_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_4x64_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_4x64_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| |
| void av1_highbd_inv_txfm_add_64x4_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| const int32_t *src = cast_to_int32(input); |
| av1_inv_txfm2d_add_64x4_c(src, CONVERT_TO_SHORTPTR(dest), stride, |
| txfm_param->tx_type, txfm_param->mode, |
| txfm_param->bd); |
| } |
| #endif // CONFIG_FLEX_PARTITION |
| |
| static void init_txfm_param(const MACROBLOCKD *xd, int plane, TX_SIZE tx_size, |
| TX_TYPE tx_type, int eob, int reduced_tx_set, |
| TxfmParam *txfm_param) { |
| (void)plane; |
| txfm_param->tx_type = tx_type; |
| txfm_param->tx_size = tx_size; |
| txfm_param->eob = eob; |
| txfm_param->lossless = xd->lossless[xd->mi[0]->segment_id]; |
| txfm_param->bd = xd->bd; |
| txfm_param->is_hbd = is_cur_buf_hbd(xd); |
| txfm_param->tx_set_type = av1_get_ext_tx_set_type( |
| txfm_param->tx_size, is_inter_block(xd->mi[0]), reduced_tx_set); |
| txfm_param->mode = get_mode_dep_txfm_mode(xd->mi[0]); |
| } |
| |
| void av1_highbd_inv_txfm_add_c(const tran_low_t *input, uint8_t *dest, |
| int stride, const TxfmParam *txfm_param) { |
| assert(av1_ext_tx_used[txfm_param->tx_set_type][txfm_param->tx_type]); |
| const TX_SIZE tx_size = txfm_param->tx_size; |
| switch (tx_size) { |
| case TX_32X32: |
| av1_highbd_inv_txfm_add_32x32_c(input, dest, stride, txfm_param); |
| break; |
| case TX_16X16: |
| av1_highbd_inv_txfm_add_16x16_c(input, dest, stride, txfm_param); |
| break; |
| case TX_8X8: |
| av1_highbd_inv_txfm_add_8x8_c(input, dest, stride, txfm_param); |
| break; |
| case TX_4X8: |
| av1_highbd_inv_txfm_add_4x8_c(input, dest, stride, txfm_param); |
| break; |
| case TX_8X4: |
| av1_highbd_inv_txfm_add_8x4_c(input, dest, stride, txfm_param); |
| break; |
| case TX_8X16: |
| av1_highbd_inv_txfm_add_8x16_c(input, dest, stride, txfm_param); |
| break; |
| case TX_16X8: |
| av1_highbd_inv_txfm_add_16x8_c(input, dest, stride, txfm_param); |
| break; |
| case TX_16X32: |
| av1_highbd_inv_txfm_add_16x32_c(input, dest, stride, txfm_param); |
| break; |
| case TX_32X16: |
| av1_highbd_inv_txfm_add_32x16_c(input, dest, stride, txfm_param); |
| break; |
| case TX_64X64: |
| av1_highbd_inv_txfm_add_64x64_c(input, dest, stride, txfm_param); |
| break; |
| case TX_32X64: |
| av1_highbd_inv_txfm_add_32x64_c(input, dest, stride, txfm_param); |
| break; |
| case TX_64X32: |
| av1_highbd_inv_txfm_add_64x32_c(input, dest, stride, txfm_param); |
| break; |
| case TX_16X64: |
| av1_highbd_inv_txfm_add_16x64_c(input, dest, stride, txfm_param); |
| break; |
| case TX_64X16: |
| av1_highbd_inv_txfm_add_64x16_c(input, dest, stride, txfm_param); |
| break; |
| case TX_4X4: |
| // this is like av1_short_idct4x4 but has a special case around eob<=1 |
| // which is significant (not just an optimization) for the lossless |
| // case. |
| av1_highbd_inv_txfm_add_4x4_c(input, dest, stride, txfm_param); |
| break; |
| case TX_16X4: |
| av1_highbd_inv_txfm_add_16x4_c(input, dest, stride, txfm_param); |
| break; |
| case TX_4X16: |
| av1_highbd_inv_txfm_add_4x16_c(input, dest, stride, txfm_param); |
| break; |
| case TX_8X32: |
| av1_highbd_inv_txfm_add_8x32_c(input, dest, stride, txfm_param); |
| break; |
| case TX_32X8: |
| av1_highbd_inv_txfm_add_32x8_c(input, dest, stride, txfm_param); |
| break; |
| #if CONFIG_FLEX_PARTITION |
| case TX_4X32: |
| av1_highbd_inv_txfm_add_4x32_c(input, dest, stride, txfm_param); |
| break; |
| case TX_32X4: |
| av1_highbd_inv_txfm_add_32x4_c(input, dest, stride, txfm_param); |
| break; |
| case TX_8X64: |
| av1_highbd_inv_txfm_add_8x64_c(input, dest, stride, txfm_param); |
| break; |
| case TX_64X8: |
| av1_highbd_inv_txfm_add_64x8_c(input, dest, stride, txfm_param); |
| break; |
| case TX_4X64: |
| av1_highbd_inv_txfm_add_4x64_c(input, dest, stride, txfm_param); |
| break; |
| case TX_64X4: |
| av1_highbd_inv_txfm_add_64x4_c(input, dest, stride, txfm_param); |
| break; |
| #endif // CONFIG_FLEX_PARTITION |
| default: assert(0 && "Invalid transform size"); break; |
| } |
| } |
| |
| void av1_inv_txfm_add_c(const tran_low_t *dqcoeff, uint8_t *dst, int stride, |
| const TxfmParam *txfm_param) { |
| const TX_SIZE tx_size = txfm_param->tx_size; |
| DECLARE_ALIGNED(32, uint16_t, tmp[MAX_TX_SQUARE]); |
| int tmp_stride = MAX_TX_SIZE; |
| int w = tx_size_wide[tx_size]; |
| int h = tx_size_high[tx_size]; |
| for (int r = 0; r < h; ++r) { |
| for (int c = 0; c < w; ++c) { |
| tmp[r * tmp_stride + c] = dst[r * stride + c]; |
| } |
| } |
| |
| #if CONFIG_MODE_DEP_INTRA_TX || CONFIG_MODE_DEP_INTER_TX || CONFIG_LGT |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #elif CONFIG_DST_32X32 && CONFIG_SUPERRES_TX64 && CONFIG_DST7_16X16 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32 || |
| tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #elif CONFIG_DST7_16X16 && CONFIG_SUPERRES_TX64 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #elif CONFIG_DST_32X32 && CONFIG_SUPERRES_TX64 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #elif CONFIG_DST_32X32 && CONFIG_DST7_16X16 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32 || |
| tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16) |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #elif CONFIG_DST7_16X16 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16) |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #elif CONFIG_DST_32X32 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32) |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #elif CONFIG_SUPERRES_TX64 |
| if (txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #else |
| av1_highbd_inv_txfm_add(dqcoeff, CONVERT_TO_BYTEPTR(tmp), tmp_stride, |
| txfm_param); |
| #endif // CONFIG_MODE_DEP_INTRA_TX || CONFIG_MODE_DEP_INTER_TX || CONFIG_LGT |
| |
| for (int r = 0; r < h; ++r) { |
| for (int c = 0; c < w; ++c) { |
| dst[r * stride + c] = (uint8_t)tmp[r * tmp_stride + c]; |
| } |
| } |
| } |
| |
| #if CONFIG_DSPL_RESIDUAL |
| /*! |
| * Given dequantized coefficients and transform parameters, this function |
| * inverts the transform and saves result in a signed 16-bit integer buffer dst. |
| * |
| * This function is necessitated by the reduced output range of |
| * av1_inverse_transform_block(), which produces unsigned output in [0-255], |
| * [0-1023], etc depending on bitdepth. This is because after |
| * av1_inverse_transform_block() inverts transform, it adds the residuals to the |
| * prediction buffer to obtain the reconstruction. The reconstruction is always |
| * in this bounded range. |
| * |
| * However the CONFIG_DSPL_RESIDUAL experiment requires access to the residuals |
| * themselves, not the reconstruction. The residuals can range form |
| * [-255, 255], [-1023, 1023], etc depending on bitdepth. |
| * |
| * To extract these residuals, this function forces a 12-bit depth transform |
| * with prediction initialized to 1023. For 8 and 10 bit depths, 1023 can then |
| * be subtracted from the output to compute residual. |
| * |
| * Currently this function only supports 8-bit depth (as that is the target |
| * of CONFIG_DSPL_RESIDUAL at this moment). Extension to 10-bit depth |
| * should be straightforward but has not been tested. However, extension to |
| * 12-bit depth will require using inverse functions further down the call stack |
| * rather than using av1_highbd_inv_txfm_add() since residuals will range in |
| * [-4095, 4095]. |
| */ |
| void av1_inverse_transform_block_diff(const MACROBLOCKD *xd, |
| const tran_low_t *dqcoeff, int plane, |
| TX_TYPE tx_type, TX_SIZE tx_size, |
| int16_t *dst, int stride, int eob, |
| int reduced_tx_set) { |
| if (!eob) return; |
| |
| // This method has not been tested for 10-bit depth and most certainly won't |
| // work with 12-bit depth |
| assert(xd->bd == 8); |
| |
| // We initialize the residual buffer to 1023 and then subtract 1023 after the |
| // inverse and add operation is performed by av1_highbd_inv_txfm_add() to get |
| // residuals |
| const uint16_t residual_init_val = 1023; |
| const uint8_t txw = tx_size_wide[tx_size], txh = tx_size_high[tx_size]; |
| |
| DECLARE_ALIGNED(32, uint16_t, residual16[MAX_SB_SQUARE]); |
| uint8_t *const residual = CONVERT_TO_BYTEPTR(residual16); |
| |
| for (int r = 0; r < txh; ++r) |
| for (int c = 0; c < txw; ++c) |
| residual16[r * MAX_TX_SIZE + c] = residual_init_val; |
| |
| TxfmParam txfm_param; |
| init_txfm_param(xd, plane, tx_size, tx_type, eob, reduced_tx_set, |
| &txfm_param); |
| txfm_param.bd = 12; |
| txfm_param.is_hbd = 1; |
| |
| av1_highbd_inv_txfm_add(dqcoeff, residual, MAX_TX_SIZE, &txfm_param); |
| |
| for (int r = 0; r < txh; ++r) |
| for (int c = 0; c < txw; ++c) |
| dst[r * stride + c] = |
| (int16_t)residual16[r * MAX_TX_SIZE + c] - residual_init_val; |
| } |
| |
| /*! |
| * This function inverts transform blocks which have been coded using |
| * downsampling in four steps: unpack coefficients, perform inverse transform, |
| * upsample and add to the prediction buffer dst. For packing logic, please see |
| * av1_xform_quant(). |
| */ |
| void av1_inverse_dspl_transform_block(const MACROBLOCKD *xd, |
| const tran_low_t *dqcoeff, int plane, |
| TX_TYPE tx_type, TX_SIZE tx_size, |
| uint8_t *dst, int stride, int eob, |
| int reduced_tx_set) { |
| const TX_SIZE new_tx_size = dspl_tx_size_map[tx_size]; |
| const uint8_t txw = tx_size_wide[tx_size], txh = tx_size_high[tx_size]; |
| const uint8_t dspl_txw = tx_size_wide[new_tx_size], |
| dspl_txh = tx_size_high[new_tx_size]; |
| |
| // Since we have to have enough buffer memory on the stack for MAX_TX_SIZE |
| // size transforms, we will use MAX_TX_SIZE as stride for simplicity. |
| const int buf_stride = MAX_TX_SIZE; |
| assert(buf_stride >= dspl_txw); |
| |
| // Buffers |
| DECLARE_ALIGNED(32, tran_low_t, scan_buf[MAX_TX_SQUARE]); |
| DECLARE_ALIGNED(32, tran_low_t, dqcoeff_buf[MAX_TX_SQUARE]); |
| DECLARE_ALIGNED(32, int16_t, buf_diff[MAX_TX_SQUARE]); |
| DECLARE_ALIGNED(32, int16_t, buf_up_diff[MAX_TX_SQUARE]); |
| |
| const int dspl_eob = eob; |
| TxfmParam txfm_param; |
| init_txfm_param(xd, plane, tx_size, tx_type, dspl_eob, reduced_tx_set, |
| &txfm_param); |
| |
| // Unpack coefficients |
| const SCAN_ORDER *const scan_order = get_scan(tx_size, tx_type); |
| const SCAN_ORDER *const dspl_scan_order = get_scan(new_tx_size, tx_type); |
| const int size = av1_get_max_eob(tx_size); |
| memset(scan_buf, 0, size * sizeof(tran_low_t)); |
| scan_array(dqcoeff, scan_buf, eob, scan_order); |
| memset(dqcoeff_buf, 0, txw * txh * sizeof(tran_low_t)); |
| iscan_array(scan_buf, dqcoeff_buf, dspl_eob, dspl_scan_order); |
| |
| // Invert and compute difference |
| av1_inverse_transform_block_diff(xd, dqcoeff_buf, plane, tx_type, new_tx_size, |
| buf_diff, buf_stride, dspl_eob, |
| reduced_tx_set); |
| |
| // Upsample |
| av1_signed_up2(buf_diff, dspl_txh, dspl_txw, buf_stride, buf_up_diff, |
| buf_stride, 1, 1, txfm_param.bd); |
| |
| // Add to output |
| for (int r = 0; r < txh; ++r) { |
| for (int c = 0; c < txw; ++c) { |
| const tran_low_t residue = (tran_low_t)buf_up_diff[r * buf_stride + c]; |
| const uint16_t out = |
| highbd_clip_pixel_add(dst[r * stride + c], residue, txfm_param.bd); |
| dst[r * stride + c] = (uint8_t)out; |
| } |
| } |
| } |
| #endif // CONFIG_DSPL_RESIDUAL |
| |
| void av1_inverse_transform_block(const MACROBLOCKD *xd, |
| const tran_low_t *dqcoeff, int plane, |
| TX_TYPE tx_type, TX_SIZE tx_size, uint8_t *dst, |
| int stride, int eob, int reduced_tx_set) { |
| if (!eob) return; |
| |
| assert(eob <= av1_get_max_eob(tx_size)); |
| |
| TxfmParam txfm_param; |
| init_txfm_param(xd, plane, tx_size, tx_type, eob, reduced_tx_set, |
| &txfm_param); |
| assert(av1_ext_tx_used[txfm_param.tx_set_type][txfm_param.tx_type]); |
| |
| if (txfm_param.is_hbd) { |
| #if CONFIG_MODE_DEP_INTRA_TX || CONFIG_MODE_DEP_INTER_TX || CONFIG_LGT |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 && CONFIG_SUPERRES_TX64 && CONFIG_DST_32X32 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 && CONFIG_SUPERRES_TX64 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 && CONFIG_DST_32X32 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32) |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST_32X32 && CONFIG_SUPERRES_TX64 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16) |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST_32X32 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32) |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_SUPERRES_TX64 |
| if (txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_highbd_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #else |
| av1_highbd_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #endif // CONFIG_MODE_DEP_INTRA_TX || CONFIG_MODE_DEP_INTER_TX || CONFIG_LGT |
| } else { |
| #if CONFIG_MODE_DEP_INTRA_TX || CONFIG_MODE_DEP_INTER_TX || CONFIG_LGT |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 && CONFIG_SUPERRES_TX64 && CONFIG_DST_32X32 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 && CONFIG_SUPERRES_TX64 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST_32X32 && CONFIG_SUPERRES_TX64 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32 || |
| txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 && CONFIG_DST_32X32 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16 || |
| tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32) |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST7_16X16 |
| if (tx_size_wide[tx_size] == 16 || tx_size_high[tx_size] == 16) |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DST_32X32 |
| if (tx_size_wide[tx_size] == 32 || tx_size_high[tx_size] == 32) |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_SUPERRES_TX64 |
| if (txsize_sqr_up_map[tx_size] == TX_64X64) |
| av1_inv_txfm_add_c(dqcoeff, dst, stride, &txfm_param); |
| else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #elif CONFIG_DSPL_RESIDUAL |
| DSPL_TYPE dspl_type = xd->mi[0]->dspl_type; |
| if (plane == 0 && dspl_type == DSPL_XY && xd->bd == 8) { |
| av1_inverse_dspl_transform_block(xd, dqcoeff, plane, tx_type, tx_size, |
| dst, stride, eob, reduced_tx_set); |
| } else { |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| } |
| #else |
| av1_inv_txfm_add(dqcoeff, dst, stride, &txfm_param); |
| #endif // CONFIG_MODE_DEP_INTRA_TX || CONFIG_MODE_DEP_INTER_TX || CONFIG_LGT |
| #if CONFIG_NN_RECON |
| if (xd->mi[0]->use_nn_recon && plane == 0) { |
| av1_cnn_recon(xd, dst, stride, tx_size); |
| } |
| #endif // CONFIG_NN_RECON |
| } |
| } |