| /* |
| * |
| * Copyright (c) 2021, Alliance for Open Media. All rights reserved |
| * |
| * This source code is subject to the terms of the BSD 3-Clause Clear License |
| * and the Alliance for Open Media Patent License 1.0. If the BSD 3-Clause Clear |
| * License was not distributed with this source code in the LICENSE file, you |
| * can obtain it at aomedia.org/license/software-license/bsd-3-c-c/. If the |
| * Alliance for Open Media Patent License 1.0 was not distributed with this |
| * source code in the PATENTS file, you can obtain it at |
| * aomedia.org/license/patent-license/. |
| */ |
| |
| #include "config/aom_config.h" |
| |
| #include "aom_mem/aom_mem.h" |
| |
| #include "av1/common/alloccommon.h" |
| #include "av1/common/av1_common_int.h" |
| #include "av1/common/blockd.h" |
| #include "av1/common/entropymode.h" |
| #include "av1/common/entropymv.h" |
| |
| int av1_get_MBs(int width, int height) { |
| const int aligned_width = ALIGN_POWER_OF_TWO(width, 3); |
| const int aligned_height = ALIGN_POWER_OF_TWO(height, 3); |
| const int mi_cols = aligned_width >> MI_SIZE_LOG2; |
| const int mi_rows = aligned_height >> MI_SIZE_LOG2; |
| |
| const int mb_cols = (mi_cols + 2) >> 2; |
| const int mb_rows = (mi_rows + 2) >> 2; |
| return mb_rows * mb_cols; |
| } |
| |
| void av1_free_ref_frame_buffers(BufferPool *pool) { |
| int i; |
| for (i = 0; i < FRAME_BUFFERS; ++i) { |
| if (pool->frame_bufs[i].ref_count > 0 && |
| pool->frame_bufs[i].raw_frame_buffer.data != NULL) { |
| pool->release_fb_cb(pool->cb_priv, &pool->frame_bufs[i].raw_frame_buffer); |
| pool->frame_bufs[i].raw_frame_buffer.data = NULL; |
| pool->frame_bufs[i].raw_frame_buffer.size = 0; |
| pool->frame_bufs[i].raw_frame_buffer.priv = NULL; |
| pool->frame_bufs[i].ref_count = 0; |
| } |
| aom_free(pool->frame_bufs[i].mvs); |
| pool->frame_bufs[i].mvs = NULL; |
| aom_free(pool->frame_bufs[i].seg_map); |
| pool->frame_bufs[i].seg_map = NULL; |
| aom_free_frame_buffer(&pool->frame_bufs[i].buf); |
| #if CONFIG_TEMP_LR |
| for (int p = 0; p < MAX_MB_PLANE; ++p) { |
| av1_free_restoration_struct(&pool->frame_bufs[i].rst_info[p]); |
| } |
| #endif // CONFIG_TEMP_LR |
| } |
| } |
| |
| // Assumes cm->rst_info[p].restoration_unit_size is already initialized |
| void av1_alloc_restoration_buffers(AV1_COMMON *cm) { |
| const int num_planes = av1_num_planes(cm); |
| for (int p = 0; p < num_planes; ++p) |
| av1_alloc_restoration_struct(cm, &cm->rst_info[p], p); |
| |
| if (cm->rst_tmpbuf == NULL) { |
| CHECK_MEM_ERROR(cm, cm->rst_tmpbuf, |
| (int32_t *)aom_memalign(16, RESTORATION_TMPBUF_SIZE)); |
| } |
| |
| if (cm->rlbs == NULL) { |
| CHECK_MEM_ERROR(cm, cm->rlbs, aom_malloc(sizeof(RestorationLineBuffers))); |
| } |
| |
| // For striped loop restoration, we divide each row of tiles into "stripes", |
| // of height 64 luma pixels but with an offset by RESTORATION_UNIT_OFFSET |
| // luma pixels to match the output from CDEF. We will need to store 2 * |
| // RESTORATION_CTX_VERT lines of data for each stripe, and also need to be |
| // able to quickly answer the question "Where is the <n>'th stripe for tile |
| // row <m>?" To make that efficient, we generate the rst_last_stripe array. |
| int num_stripes = 0; |
| for (int i = 0; i < cm->tiles.rows; ++i) { |
| TileInfo tile_info; |
| av1_tile_set_row(&tile_info, cm, i); |
| const int mi_h = tile_info.mi_row_end - tile_info.mi_row_start; |
| const int ext_h = RESTORATION_UNIT_OFFSET + (mi_h << MI_SIZE_LOG2); |
| const int tile_stripes = |
| (ext_h + RESTORATION_PROC_UNIT_SIZE - 1) / RESTORATION_PROC_UNIT_SIZE; |
| num_stripes += tile_stripes; |
| } |
| #if CONFIG_EXT_SUPERRES |
| // TODO(yuec, debargha): This is a temporary fix to handle mismatch between |
| // stripes in the coded domain vs. stripes in the upscaled domain. In AV1 |
| // they were the same and there was no issue. However in |
| // CONFIG_EXT_SUPERRES, the actual number of stripes in the upscaled domain |
| // can be twice as many as stripes in the coded domain. Hence this change. |
| // Going forward, wee need to rethink striped loop-restoration in the |
| // context of 2D superres, and implement a different strategy or remove |
| // striping altogether, based on consultation with hardware teams. |
| num_stripes <<= 1; |
| #endif // CONFIG_EXT_SUPERRES |
| |
| // Now we need to allocate enough space to store the line buffers for the |
| // stripes |
| const int frame_w = cm->superres_upscaled_width; |
| |
| for (int p = 0; p < num_planes; ++p) { |
| const int is_uv = p > 0; |
| const int ss_x = is_uv && cm->seq_params.subsampling_x; |
| const int plane_w = ((frame_w + ss_x) >> ss_x) + 2 * RESTORATION_EXTRA_HORZ; |
| const int stride = ALIGN_POWER_OF_TWO(plane_w, 5); |
| const int buf_size = num_stripes * stride * RESTORATION_CTX_VERT << 1; |
| RestorationStripeBoundaries *boundaries = &cm->rst_info[p].boundaries; |
| |
| if (buf_size != boundaries->stripe_boundary_size || |
| boundaries->stripe_boundary_above == NULL || |
| boundaries->stripe_boundary_below == NULL) { |
| aom_free(boundaries->stripe_boundary_above); |
| aom_free(boundaries->stripe_boundary_below); |
| |
| CHECK_MEM_ERROR(cm, boundaries->stripe_boundary_above, |
| aom_memalign(32, buf_size)); |
| CHECK_MEM_ERROR(cm, boundaries->stripe_boundary_below, |
| aom_memalign(32, buf_size)); |
| |
| boundaries->stripe_boundary_size = buf_size; |
| } |
| boundaries->stripe_boundary_stride = stride; |
| } |
| } |
| |
| void av1_free_restoration_buffers(AV1_COMMON *cm) { |
| int p; |
| for (p = 0; p < MAX_MB_PLANE; ++p) |
| av1_free_restoration_struct(&cm->rst_info[p]); |
| aom_free(cm->rst_tmpbuf); |
| cm->rst_tmpbuf = NULL; |
| aom_free(cm->rlbs); |
| cm->rlbs = NULL; |
| for (p = 0; p < MAX_MB_PLANE; ++p) { |
| RestorationStripeBoundaries *boundaries = &cm->rst_info[p].boundaries; |
| aom_free(boundaries->stripe_boundary_above); |
| aom_free(boundaries->stripe_boundary_below); |
| boundaries->stripe_boundary_above = NULL; |
| boundaries->stripe_boundary_below = NULL; |
| } |
| |
| aom_free_frame_buffer(&cm->rst_frame); |
| } |
| |
| void av1_free_above_context_buffers(CommonContexts *above_contexts) { |
| int i; |
| const int num_planes = above_contexts->num_planes; |
| |
| for (int tile_row = 0; tile_row < above_contexts->num_tile_rows; tile_row++) { |
| for (i = 0; i < num_planes; i++) { |
| aom_free(above_contexts->entropy[i][tile_row]); |
| above_contexts->entropy[i][tile_row] = NULL; |
| aom_free(above_contexts->partition[i][tile_row]); |
| above_contexts->partition[i][tile_row] = NULL; |
| } |
| aom_free(above_contexts->txfm[tile_row]); |
| above_contexts->txfm[tile_row] = NULL; |
| } |
| for (i = 0; i < num_planes; i++) { |
| aom_free(above_contexts->entropy[i]); |
| above_contexts->entropy[i] = NULL; |
| aom_free(above_contexts->partition[i]); |
| above_contexts->partition[i] = NULL; |
| } |
| aom_free(above_contexts->txfm); |
| above_contexts->txfm = NULL; |
| |
| above_contexts->num_tile_rows = 0; |
| above_contexts->num_mi_cols = 0; |
| above_contexts->num_planes = 0; |
| } |
| |
| void av1_free_sbi(CommonSBInfoParams *sbi_params) { |
| for (int i = 0; i < sbi_params->sbi_alloc_size; ++i) { |
| av1_free_ptree_recursive(sbi_params->sbi_grid_base[i].ptree_root[0]); |
| av1_free_ptree_recursive(sbi_params->sbi_grid_base[i].ptree_root[1]); |
| } |
| |
| aom_free(sbi_params->sbi_grid_base); |
| sbi_params->sbi_grid_base = NULL; |
| sbi_params->sbi_alloc_size = 0; |
| } |
| |
| void av1_free_context_buffers(AV1_COMMON *cm) { |
| cm->mi_params.free_mi(&cm->mi_params); |
| av1_free_sbi(&cm->sbi_params); |
| |
| av1_free_above_context_buffers(&cm->above_contexts); |
| |
| #if CONFIG_LPF_MASK |
| av1_free_loop_filter_mask(cm); |
| #endif |
| } |
| |
| int av1_alloc_above_context_buffers(CommonContexts *above_contexts, |
| int num_tile_rows, int num_mi_cols, |
| int num_planes) { |
| const int aligned_mi_cols = |
| ALIGN_POWER_OF_TWO(num_mi_cols, MAX_MIB_SIZE_LOG2); |
| |
| // Allocate above context buffers |
| above_contexts->num_tile_rows = num_tile_rows; |
| above_contexts->num_mi_cols = aligned_mi_cols; |
| above_contexts->num_planes = num_planes; |
| for (int plane_idx = 0; plane_idx < num_planes; plane_idx++) { |
| above_contexts->entropy[plane_idx] = (ENTROPY_CONTEXT **)aom_calloc( |
| num_tile_rows, sizeof(above_contexts->entropy[0])); |
| if (!above_contexts->entropy[plane_idx]) return 1; |
| above_contexts->partition[plane_idx] = (PARTITION_CONTEXT **)aom_calloc( |
| num_tile_rows, sizeof(above_contexts->partition[plane_idx])); |
| if (!above_contexts->partition[plane_idx]) return 1; |
| } |
| |
| above_contexts->txfm = |
| (TXFM_CONTEXT **)aom_calloc(num_tile_rows, sizeof(above_contexts->txfm)); |
| if (!above_contexts->txfm) return 1; |
| |
| for (int tile_row = 0; tile_row < num_tile_rows; tile_row++) { |
| for (int plane_idx = 0; plane_idx < num_planes; plane_idx++) { |
| above_contexts->entropy[plane_idx][tile_row] = |
| (ENTROPY_CONTEXT *)aom_calloc( |
| aligned_mi_cols, sizeof(*above_contexts->entropy[0][tile_row])); |
| if (!above_contexts->entropy[plane_idx][tile_row]) return 1; |
| above_contexts->partition[plane_idx][tile_row] = |
| (PARTITION_CONTEXT *)aom_calloc( |
| aligned_mi_cols, |
| sizeof(*above_contexts->partition[plane_idx][tile_row])); |
| if (!above_contexts->partition[plane_idx][tile_row]) return 1; |
| } |
| above_contexts->txfm[tile_row] = (TXFM_CONTEXT *)aom_calloc( |
| aligned_mi_cols, sizeof(*above_contexts->txfm[tile_row])); |
| if (!above_contexts->txfm[tile_row]) return 1; |
| } |
| |
| return 0; |
| } |
| |
| // Allocate the dynamically allocated arrays in 'mi_params' assuming |
| // 'mi_params->set_mb_mi()' was already called earlier to initialize the rest of |
| // the struct members. |
| static int alloc_mi(CommonModeInfoParams *mi_params |
| #if CONFIG_PC_WIENER |
| , |
| AV1_COMMON *cm |
| #endif // CONFIG_PC_WIENER |
| ) { |
| const int aligned_mi_rows = calc_mi_size(mi_params->mi_rows); |
| const int mi_grid_size = mi_params->mi_stride * aligned_mi_rows; |
| const int alloc_size_1d = mi_size_wide[mi_params->mi_alloc_bsize]; |
| const int alloc_mi_size = |
| mi_params->mi_alloc_stride * (aligned_mi_rows / alloc_size_1d); |
| |
| if (mi_params->mi_alloc_size < alloc_mi_size || |
| mi_params->mi_grid_size < mi_grid_size) { |
| mi_params->free_mi(mi_params); |
| |
| mi_params->mi_alloc = |
| aom_calloc(alloc_mi_size, sizeof(*mi_params->mi_alloc)); |
| if (!mi_params->mi_alloc) return 1; |
| mi_params->mi_alloc_size = alloc_mi_size; |
| |
| mi_params->mi_grid_base = (MB_MODE_INFO **)aom_calloc( |
| mi_grid_size, sizeof(*mi_params->mi_grid_base)); |
| if (!mi_params->mi_grid_base) return 1; |
| mi_params->mi_grid_size = mi_grid_size; |
| #if CONFIG_PC_WIENER |
| av1_alloc_txk_skip_array(mi_params, cm); |
| av1_alloc_class_id_array(mi_params, cm); |
| #endif // CONFIG_PC_WIENER |
| #if CONFIG_C071_SUBBLK_WARPMV |
| mi_params->mi_alloc_sub = |
| aom_calloc(alloc_mi_size, sizeof(*mi_params->mi_alloc_sub)); |
| if (!mi_params->mi_alloc_sub) return 1; |
| mi_params->submi_grid_base = (SUBMB_INFO **)aom_calloc( |
| mi_grid_size, sizeof(*mi_params->submi_grid_base)); |
| if (!mi_params->submi_grid_base) return 1; |
| #endif // CONFIG_C071_SUBBLK_WARPMV |
| |
| mi_params->blk_skip = |
| aom_calloc(mi_grid_size, sizeof(*mi_params->blk_skip)); |
| if (!mi_params->blk_skip) return 1; |
| mi_params->tx_type_map = |
| aom_calloc(mi_grid_size, sizeof(*mi_params->tx_type_map)); |
| if (!mi_params->tx_type_map) return 1; |
| #if CONFIG_CROSS_CHROMA_TX |
| mi_params->cctx_type_map = |
| aom_calloc(mi_grid_size, sizeof(*mi_params->cctx_type_map)); |
| if (!mi_params->cctx_type_map) return 1; |
| #endif // CONFIG_CROSS_CHROMA_TX |
| } |
| |
| return 0; |
| } |
| |
| static void set_sb_si(AV1_COMMON *cm) { |
| CommonSBInfoParams *const sbi_params = &cm->sbi_params; |
| const int mib_size_log2 = cm->seq_params.mib_size_log2; |
| sbi_params->sb_cols = |
| ALIGN_POWER_OF_TWO(cm->mi_params.mi_cols, mib_size_log2) >> mib_size_log2; |
| sbi_params->sb_rows = |
| ALIGN_POWER_OF_TWO(cm->mi_params.mi_rows, mib_size_log2) >> mib_size_log2; |
| sbi_params->sbi_stride = cm->mi_params.mi_stride >> mib_size_log2; |
| } |
| |
| static int alloc_sbi(CommonSBInfoParams *sbi_params) { |
| const int sbi_size = |
| sbi_params->sbi_stride * calc_mi_size(sbi_params->sb_rows); |
| |
| if (sbi_params->sbi_alloc_size < sbi_size) { |
| av1_free_sbi(sbi_params); |
| sbi_params->sbi_grid_base = aom_calloc(sbi_size, sizeof(SB_INFO)); |
| |
| if (!sbi_params->sbi_grid_base) return 1; |
| |
| sbi_params->sbi_alloc_size = sbi_size; |
| for (int i = 0; i < sbi_size; ++i) { |
| sbi_params->sbi_grid_base[i].ptree_root[0] = NULL; |
| sbi_params->sbi_grid_base[i].ptree_root[1] = NULL; |
| } |
| } |
| |
| return 0; |
| } |
| |
| int av1_duplicate_sbi(CommonSBInfoParams *to, const CommonSBInfoParams *from) { |
| av1_free_sbi(to); |
| memcpy(to, from, sizeof(*to)); |
| const int sbi_size = to->sbi_alloc_size; |
| |
| to->sbi_grid_base = aom_calloc(sbi_size, sizeof(SB_INFO)); |
| |
| if (!to->sbi_grid_base) return 1; |
| |
| for (int i = 0; i < sbi_size; ++i) { |
| to->sbi_grid_base[i].ptree_root[0] = NULL; |
| to->sbi_grid_base[i].ptree_root[1] = NULL; |
| } |
| for (int i = 0; i < sbi_size; ++i) { |
| to->sbi_grid_base[i].ptree_root[0] = av1_duplicate_ptree_recursive( |
| from->sbi_grid_base[i].ptree_root[0], NULL); |
| to->sbi_grid_base[i].ptree_root[1] = av1_duplicate_ptree_recursive( |
| from->sbi_grid_base[i].ptree_root[1], NULL); |
| } |
| return 0; |
| } |
| |
| int av1_alloc_context_buffers(AV1_COMMON *cm, int width, int height) { |
| CommonModeInfoParams *const mi_params = &cm->mi_params; |
| mi_params->set_mb_mi(mi_params, width, height); |
| #if CONFIG_PC_WIENER |
| if (alloc_mi(mi_params, cm)) goto fail; |
| #else |
| if (alloc_mi(mi_params)) goto fail; |
| #endif // CONFIG_PC_WIENER |
| CommonSBInfoParams *const sbi_params = &cm->sbi_params; |
| set_sb_si(cm); |
| if (alloc_sbi(sbi_params)) goto fail; |
| |
| return 0; |
| |
| fail: |
| // clear the mi_* values to force a realloc on resync |
| mi_params->set_mb_mi(mi_params, 0, 0); |
| av1_free_context_buffers(cm); |
| return 1; |
| } |
| |
| void av1_remove_common(AV1_COMMON *cm) { |
| av1_free_context_buffers(cm); |
| |
| aom_free(cm->fc); |
| cm->fc = NULL; |
| aom_free(cm->default_frame_context); |
| cm->default_frame_context = NULL; |
| } |
| |
| void av1_init_mi_buffers(CommonModeInfoParams *mi_params) { |
| mi_params->setup_mi(mi_params); |
| } |
| |
| #if CONFIG_LPF_MASK |
| int av1_alloc_loop_filter_mask(AV1_COMMON *cm) { |
| aom_free(cm->lf.lfm); |
| cm->lf.lfm = NULL; |
| |
| // Each lfm holds bit masks for all the 4x4 blocks in a max |
| // 64x64 (128x128 for ext_partitions) region. The stride |
| // and rows are rounded up / truncated to a multiple of 16 |
| // (32 for ext_partition). |
| cm->lf.lfm_stride = |
| (cm->mi_params.mi_cols + (MI_SIZE_64X64 - 1)) >> MIN_MIB_SIZE_LOG2; |
| cm->lf.lfm_num = |
| ((cm->mi_params.mi_rows + (MI_SIZE_64X64 - 1)) >> MIN_MIB_SIZE_LOG2) * |
| cm->lf.lfm_stride; |
| cm->lf.lfm = |
| (LoopFilterMask *)aom_calloc(cm->lf.lfm_num, sizeof(*cm->lf.lfm)); |
| if (!cm->lf.lfm) return 1; |
| |
| unsigned int i; |
| for (i = 0; i < cm->lf.lfm_num; ++i) av1_zero(cm->lf.lfm[i]); |
| |
| return 0; |
| } |
| |
| void av1_free_loop_filter_mask(AV1_COMMON *cm) { |
| if (cm->lf.lfm == NULL) return; |
| |
| aom_free(cm->lf.lfm); |
| cm->lf.lfm = NULL; |
| cm->lf.lfm_num = 0; |
| cm->lf.lfm_stride = 0; |
| } |
| #endif |
| |
| int av1_copy_mi_neq(const AV1_COMMON *cm, CommonModeInfoParams *to, |
| const CommonModeInfoParams *from) { |
| assert(from->mi_rows == to->mi_rows); |
| assert(from->mi_cols == to->mi_cols); |
| assert(from->mi_alloc_bsize == to->mi_alloc_bsize); |
| const int aligned_mi_rows = calc_mi_size(from->mi_rows); |
| const int aligned_mi_cols = calc_mi_size(from->mi_cols); |
| const int mi_alloc_size_1d = mi_size_wide[from->mi_alloc_bsize]; |
| const int mi_alloc_stride = |
| (aligned_mi_cols + mi_alloc_size_1d - 1) / mi_alloc_size_1d; |
| for (int i = 0; i < aligned_mi_rows; ++i) { |
| memcpy(&to->mi_alloc[to->mi_alloc_stride * i], |
| &from->mi_alloc[from->mi_alloc_stride * i], |
| mi_alloc_stride * sizeof(*to->mi_alloc)); |
| } |
| for (int i = 0; i < aligned_mi_rows; ++i) { |
| for (int j = 0; j < aligned_mi_cols; ++j) { |
| if (from->mi_grid_base[i * from->mi_stride + j] == NULL) { |
| to->mi_grid_base[i * to->mi_stride + j] = NULL; |
| } else { |
| to->mi_grid_base[i * to->mi_stride + j] = |
| to->mi_alloc + |
| (int)(from->mi_grid_base[i * from->mi_stride + j] - from->mi_alloc); |
| } |
| } |
| } |
| #if CONFIG_C071_SUBBLK_WARPMV |
| for (int i = 0; i < aligned_mi_rows; ++i) { |
| memcpy(&to->mi_alloc_sub[to->mi_alloc_stride * i], |
| &from->mi_alloc_sub[from->mi_alloc_stride * i], |
| mi_alloc_stride * sizeof(*to->mi_alloc_sub)); |
| } |
| for (int i = 0; i < aligned_mi_rows; ++i) { |
| for (int j = 0; j < aligned_mi_cols; ++j) { |
| if (from->submi_grid_base[i * from->mi_stride + j] == NULL) { |
| to->submi_grid_base[i * to->mi_stride + j] = NULL; |
| } else { |
| to->submi_grid_base[i * to->mi_stride + j] = |
| to->mi_alloc_sub + |
| (int)(from->submi_grid_base[i * from->mi_stride + j] - |
| from->mi_alloc_sub); |
| } |
| } |
| } |
| #endif // CONFIG_C071_SUBBLK_WARPMV |
| for (int i = 0; i < aligned_mi_rows; ++i) { |
| memcpy(&to->blk_skip[i * to->mi_stride], |
| &from->blk_skip[i * from->mi_stride], |
| aligned_mi_cols * sizeof(*to->blk_skip)); |
| } |
| for (int i = 0; i < aligned_mi_rows; ++i) { |
| memcpy(&to->tx_type_map[i * to->mi_stride], |
| &from->tx_type_map[i * from->mi_stride], |
| aligned_mi_cols * sizeof(*to->tx_type_map)); |
| } |
| #if CONFIG_CROSS_CHROMA_TX |
| for (int i = 0; i < aligned_mi_rows; ++i) { |
| memcpy(&to->cctx_type_map[i * to->mi_stride], |
| &from->cctx_type_map[i * from->mi_stride], |
| aligned_mi_cols * sizeof(*to->cctx_type_map)); |
| } |
| #endif // CONFIG_CROSS_CHROMA_TX |
| #if CONFIG_PC_WIENER |
| for (int plane = 0; plane < MAX_MB_PLANE; plane++) { |
| int w = from->mi_cols << MI_SIZE_LOG2; |
| int h = from->mi_rows << MI_SIZE_LOG2; |
| w = ((w + MAX_SB_SIZE - 1) >> MAX_SB_SIZE_LOG2) << MAX_SB_SIZE_LOG2; |
| h = ((h + MAX_SB_SIZE - 1) >> MAX_SB_SIZE_LOG2) << MAX_SB_SIZE_LOG2; |
| w >>= ((plane == 0) ? 0 : cm->seq_params.subsampling_x); |
| h >>= ((plane == 0) ? 0 : cm->seq_params.subsampling_y); |
| int stride = (w + MIN_TX_SIZE - 1) >> MIN_TX_SIZE_LOG2; |
| int rows = (h + MIN_TX_SIZE - 1) >> MIN_TX_SIZE_LOG2; |
| for (int i = 0; i < rows; ++i) { |
| memcpy(&to->tx_skip[plane][i * to->tx_skip_stride[plane]], |
| &from->tx_skip[plane][i * from->tx_skip_stride[plane]], |
| stride * sizeof(to->tx_skip[plane][0])); |
| memcpy(&to->wiener_class_id[plane][i * to->wiener_class_id_stride[plane]], |
| &from->wiener_class_id[plane] |
| [i * from->wiener_class_id_stride[plane]], |
| stride * sizeof(to->wiener_class_id[plane][0])); |
| } |
| } |
| #endif // CONFIG_PC_WIENER |
| return 0; |
| } |
| |
| int av1_copy_mi(CommonModeInfoParams *to, const CommonModeInfoParams *from) { |
| assert(to->mi_alloc_size == from->mi_alloc_size); |
| assert(to->mi_grid_size == from->mi_grid_size); |
| for (int plane = 0; plane < MAX_MB_PLANE; plane++) { |
| assert(to->tx_skip_buf_size[plane] == from->tx_skip_buf_size[plane]); |
| assert(to->wiener_class_id_buf_size[plane] == |
| from->wiener_class_id_buf_size[plane]); |
| } |
| |
| memcpy(to->mi_alloc, from->mi_alloc, |
| to->mi_alloc_size * sizeof(*to->mi_alloc)); |
| for (int i = 0; i < to->mi_grid_size; ++i) { |
| if (from->mi_grid_base[i] == NULL) { |
| to->mi_grid_base[i] = NULL; |
| } else { |
| to->mi_grid_base[i] = |
| to->mi_alloc + (int)(from->mi_grid_base[i] - from->mi_alloc); |
| } |
| } |
| #if CONFIG_C071_SUBBLK_WARPMV |
| memcpy(to->mi_alloc_sub, from->mi_alloc_sub, |
| to->mi_alloc_size * sizeof(*to->mi_alloc_sub)); |
| for (int i = 0; i < to->mi_grid_size; ++i) { |
| if (from->submi_grid_base[i] == NULL) { |
| to->submi_grid_base[i] = NULL; |
| } else { |
| to->submi_grid_base[i] = |
| to->mi_alloc_sub + |
| (int)(from->submi_grid_base[i] - from->mi_alloc_sub); |
| } |
| } |
| #endif // CONFIG_C071_SUBBLK_WARPMV |
| memcpy(to->blk_skip, from->blk_skip, |
| to->mi_grid_size * sizeof(*to->blk_skip)); |
| memcpy(to->tx_type_map, from->tx_type_map, |
| to->mi_grid_size * sizeof(*to->tx_type_map)); |
| #if CONFIG_CROSS_CHROMA_TX |
| memcpy(to->cctx_type_map, from->cctx_type_map, |
| to->mi_grid_size * sizeof(*to->cctx_type_map)); |
| #endif // CONFIG_CROSS_CHROMA_TX |
| #if CONFIG_PC_WIENER |
| for (int plane = 0; plane < MAX_MB_PLANE; plane++) { |
| memcpy(to->tx_skip[plane], from->tx_skip[plane], |
| to->tx_skip_buf_size[plane] * sizeof(*to->tx_skip[plane])); |
| memcpy(to->wiener_class_id[plane], from->wiener_class_id[plane], |
| to->wiener_class_id_buf_size[plane] * |
| sizeof(*to->wiener_class_id[plane])); |
| } |
| #endif // CONFIG_PC_WIENER |
| return 0; |
| } |
| |
| int av1_duplicate_mi(AV1_COMMON *cm, CommonModeInfoParams *to) { |
| const CommonModeInfoParams *const from = &cm->mi_params; |
| if (to->mi_alloc_size > 0) { |
| assert(to->free_mi); |
| to->free_mi(to); |
| } |
| to->set_mb_mi = from->set_mb_mi; |
| to->free_mi = from->free_mi; |
| to->setup_mi = from->setup_mi; |
| to->set_mb_mi(to, cm->width, cm->height); |
| |
| int ret = alloc_mi(to |
| #if CONFIG_PC_WIENER |
| , |
| cm |
| #endif // CONFIG_PC_WIENER |
| ); |
| if (ret) return 1; |
| |
| return av1_copy_mi_neq(cm, to, from); |
| } |
| |
| #if CONFIG_TEMP_LR |
| int av1_copy_rst_info(RestorationInfo *to, RestorationInfo *from) { |
| av1_free_restoration_struct(to); |
| memcpy(to, from, sizeof(*to)); |
| if (to->units_per_tile) { |
| to->unit_info = (RestorationUnitInfo *)aom_memalign( |
| 16, sizeof(*to->unit_info) * to->units_per_tile); |
| if (to->unit_info == NULL) return 1; |
| memcpy(to->unit_info, from->unit_info, |
| sizeof(*to->unit_info) * to->units_per_tile); |
| } |
| return 0; |
| } |
| #endif // CONFIG_TEMP_LR |