Revert "Update cdf and mode rate per superblock in RDO"
This reverts commit b855b8e80827997fc54c3aaa4de1d3dd6cf0ab15.
Reason for revert:
breaks the rate assumptions in:
AV1/DatarateTestLarge.BasicRateTargetingVBR
AV1/DatarateTestLarge.ChangingDropFrameThresh/2
Change-Id: I92c4709b6c1bd98a78fe5aa6f0167a534fed5997
diff --git a/av1/encoder/block.h b/av1/encoder/block.h
index f368456..30f14d4 100644
--- a/av1/encoder/block.h
+++ b/av1/encoder/block.h
@@ -148,8 +148,6 @@
int *nmvjointcost;
int nmv_vec_cost[NMV_CONTEXTS][MV_JOINTS];
- int nmvcost_array[NMV_CONTEXTS][2][MV_VALS];
- int nmvcost_hp_array[NMV_CONTEXTS][2][MV_VALS];
int *nmvcost[NMV_CONTEXTS][2];
int *nmvcost_hp[NMV_CONTEXTS][2];
int **mv_cost_stack[NMV_CONTEXTS];
@@ -195,7 +193,6 @@
#if CONFIG_EXT_INTER
int inter_compound_mode_cost[INTER_MODE_CONTEXTS][INTER_COMPOUND_MODES];
- int compound_type_cost[BLOCK_SIZES_ALL][COMPOUND_TYPES];
#if CONFIG_COMPOUND_SINGLEREF
int inter_singleref_comp_mode_cost[INTER_MODE_CONTEXTS]
[INTER_SINGLEREF_COMP_MODES];
diff --git a/av1/encoder/encodeframe.c b/av1/encoder/encodeframe.c
index ef651b0..f730fac 100644
--- a/av1/encoder/encodeframe.c
+++ b/av1/encoder/encodeframe.c
@@ -477,9 +477,6 @@
has_subpel_mv_component(xd->mi[0], xd, dir + 2))) {
const int ctx = av1_get_pred_context_switchable_interp(xd, dir);
++counts->switchable_interp[ctx][mbmi->interp_filter[dir]];
- update_cdf(xd->tile_ctx->switchable_interp_cdf[ctx],
- av1_switchable_interp_ind[mbmi->interp_filter[dir]],
- SWITCHABLE_FILTERS);
}
}
}
@@ -1553,7 +1550,6 @@
const MB_MODE_INFO *const mbmi = &mi->mbmi;
const MB_MODE_INFO_EXT *const mbmi_ext = x->mbmi_ext;
const BLOCK_SIZE bsize = mbmi->sb_type;
- FRAME_CONTEXT *fc = xd->tile_ctx;
#if CONFIG_DELTA_Q
// delta quant applies to both intra and inter
@@ -1598,10 +1594,6 @@
if (!supertx_enabled)
#endif
counts->intra_inter[av1_get_intra_inter_context(xd)][inter_block]++;
-#if CONFIG_NEW_MULTISYMBOL
- update_cdf(fc->intra_inter_cdf[av1_get_intra_inter_context(xd)],
- inter_block, 2);
-#endif
// If the segment reference feature is enabled we have only a single
// reference frame allowed for the segment so exclude it from
// the reference frame counts used to work out probabilities.
@@ -1619,21 +1611,12 @@
// This flag is also updated for 4x4 blocks
rdc->single_ref_used_flag = 1;
#if !SUB8X8_COMP_REF
- if (mbmi->sb_type != BLOCK_4X4) {
+ if (mbmi->sb_type != BLOCK_4X4)
counts->comp_inter[av1_get_reference_mode_context(cm, xd)]
[has_second_ref(mbmi)]++;
-#if CONFIG_NEW_MULTISYMBOL
- update_cdf(av1_get_reference_mode_cdf(cm, xd), has_second_ref(mbmi),
- 2);
-#endif
- }
#else
counts->comp_inter[av1_get_reference_mode_context(cm, xd)]
[has_second_ref(mbmi)]++;
-#if CONFIG_NEW_MULTISYMBOL
- update_cdf(av1_get_reference_mode_cdf(cm, xd), has_second_ref(mbmi),
- 2);
-#endif
#endif
}
@@ -1742,24 +1725,11 @@
const int bsize_group = size_group_lookup[bsize];
if (mbmi->ref_frame[1] == INTRA_FRAME) {
counts->interintra[bsize_group][1]++;
-#if CONFIG_NEW_MULTISYMBOL
- update_cdf(fc->interintra_cdf[bsize_group], 1, 2);
-#endif
counts->interintra_mode[bsize_group][mbmi->interintra_mode]++;
- update_cdf(fc->interintra_mode_cdf[bsize_group],
- mbmi->interintra_mode, INTERINTRA_MODES);
- if (is_interintra_wedge_used(bsize)) {
+ if (is_interintra_wedge_used(bsize))
counts->wedge_interintra[bsize][mbmi->use_wedge_interintra]++;
-#if CONFIG_NEW_MULTISYMBOL
- update_cdf(fc->wedge_interintra_cdf[bsize],
- mbmi->use_wedge_interintra, 2);
-#endif
- }
} else {
counts->interintra[bsize_group][0]++;
-#if CONFIG_NEW_MULTISYMBOL
- update_cdf(fc->interintra_cdf[bsize_group], 0, 2);
-#endif
}
}
#endif // CONFIG_INTERINTRA
@@ -1799,21 +1769,17 @@
{
if (motion_allowed == WARPED_CAUSAL) {
counts->motion_mode[mbmi->sb_type][mbmi->motion_mode]++;
- update_cdf(fc->motion_mode_cdf[mbmi->sb_type], mbmi->motion_mode,
- MOTION_MODES);
+ update_cdf(xd->tile_ctx->motion_mode_cdf[mbmi->sb_type],
+ mbmi->motion_mode, MOTION_MODES);
} else if (motion_allowed == OBMC_CAUSAL) {
counts->obmc[mbmi->sb_type][mbmi->motion_mode == OBMC_CAUSAL]++;
-#if CONFIG_NEW_MULTISYMBOL
- update_cdf(fc->obmc_cdf[mbmi->sb_type],
- mbmi->motion_mode == OBMC_CAUSAL, 2);
-#endif
}
}
#else
if (motion_allowed > SIMPLE_TRANSLATION) {
counts->motion_mode[mbmi->sb_type][mbmi->motion_mode]++;
- update_cdf(fc->motion_mode_cdf[mbmi->sb_type], mbmi->motion_mode,
- MOTION_MODES);
+ update_cdf(xd->tile_ctx->motion_mode_cdf[mbmi->sb_type],
+ mbmi->motion_mode, MOTION_MODES);
}
#endif // CONFIG_MOTION_VAR && CONFIG_WARPED_MOTION
@@ -1846,16 +1812,7 @@
&& mbmi->motion_mode == SIMPLE_TRANSLATION
#endif // CONFIG_MOTION_VAR || CONFIG_WARPED_MOTION
) {
-#if CONFIG_WEDGE && CONFIG_COMPOUND_SEGMENT
- if (is_interinter_compound_used(COMPOUND_WEDGE, bsize)) {
-#endif
- counts
- ->compound_interinter[bsize][mbmi->interinter_compound_type]++;
- update_cdf(fc->compound_type_cdf[bsize],
- mbmi->interinter_compound_type, COMPOUND_TYPES);
-#if CONFIG_WEDGE && CONFIG_COMPOUND_SEGMENT
- }
-#endif
+ counts->compound_interinter[bsize][mbmi->interinter_compound_type]++;
}
#endif // CONFIG_EXT_INTER
}
@@ -1869,8 +1826,6 @@
if (has_second_ref(mbmi)) {
mode_ctx = mbmi_ext->compound_mode_context[mbmi->ref_frame[0]];
++counts->inter_compound_mode[mode_ctx][INTER_COMPOUND_OFFSET(mode)];
- update_cdf(fc->inter_compound_mode_cdf[mode_ctx],
- INTER_COMPOUND_OFFSET(mode), INTER_COMPOUND_MODES);
#if CONFIG_COMPOUND_SINGLEREF
} else if (is_inter_singleref_comp_mode(mode)) {
mode_ctx = mbmi_ext->compound_mode_context[mbmi->ref_frame[0]];
@@ -4506,6 +4461,7 @@
// Initialize the left context for the new SB row
av1_zero_left_context(xd);
+ av1_fill_mode_rates(cm, x, xd->tile_ctx);
#if CONFIG_DELTA_Q
// Reset delta for every tile
@@ -4542,7 +4498,6 @@
av1_fill_token_costs_from_cdf(x->token_tail_costs,
x->e_mbd.tile_ctx->coef_tail_cdfs);
#endif
- av1_fill_mode_rates(cm, x, xd->tile_ctx);
if (sf->adaptive_pred_interp_filter) {
for (i = 0; i < leaf_nodes; ++i)
@@ -5706,18 +5661,21 @@
const MODE_INFO *mi, const MODE_INFO *above_mi,
const MODE_INFO *left_mi, const int intraonly,
const int mi_row, const int mi_col) {
- FRAME_CONTEXT *fc = xd->tile_ctx;
const MB_MODE_INFO *const mbmi = &mi->mbmi;
+#if CONFIG_ENTROPY_STATS
const PREDICTION_MODE y_mode = mbmi->mode;
const UV_PREDICTION_MODE uv_mode = mbmi->uv_mode;
+#else // CONFIG_ENTROPY_STATS
(void)counts;
(void)above_mi;
(void)left_mi;
(void)intraonly;
+#endif // CONFIG_ENTROPY_STATS
const BLOCK_SIZE bsize = mbmi->sb_type;
const int unify_bsize = CONFIG_CB4X4;
if (bsize < BLOCK_8X8 && !unify_bsize) {
+#if CONFIG_ENTROPY_STATS
int idx, idy;
const int num_4x4_w = num_4x4_blocks_wide_lookup[bsize];
const int num_4x4_h = num_4x4_blocks_high_lookup[bsize];
@@ -5726,36 +5684,24 @@
const int bidx = idy * 2 + idx;
const PREDICTION_MODE bmode = mi->bmi[bidx].as_mode;
if (intraonly) {
-#if CONFIG_ENTROPY_STATS
const PREDICTION_MODE a = av1_above_block_mode(mi, above_mi, bidx);
const PREDICTION_MODE l = av1_left_block_mode(mi, left_mi, bidx);
++counts->kf_y_mode[a][l][bmode];
-#endif // CONFIG_ENTROPY_STATS
- update_cdf(get_y_mode_cdf(fc, mi, above_mi, left_mi, bidx),
- av1_intra_mode_ind[bmode], INTRA_MODES);
} else {
-#if CONFIG_ENTROPY_STATS
++counts->y_mode[0][bmode];
-#endif // CONFIG_ENTROPY_STATS
- update_cdf(fc->y_mode_cdf[0], av1_intra_mode_ind[bmode], INTRA_MODES);
}
}
+#endif // CONFIG_ENTROPY_STATS
} else {
- if (intraonly) {
#if CONFIG_ENTROPY_STATS
+ if (intraonly) {
const PREDICTION_MODE above = av1_above_block_mode(mi, above_mi, 0);
const PREDICTION_MODE left = av1_left_block_mode(mi, left_mi, 0);
++counts->kf_y_mode[above][left][y_mode];
-#endif // CONFIG_ENTROPY_STATS
- update_cdf(get_y_mode_cdf(fc, mi, above_mi, left_mi, 0),
- av1_intra_mode_ind[y_mode], INTRA_MODES);
} else {
-#if CONFIG_ENTROPY_STATS
++counts->y_mode[size_group_lookup[bsize]][y_mode];
-#endif // CONFIG_ENTROPY_STATS
- update_cdf(fc->y_mode_cdf[size_group_lookup[bsize]],
- av1_intra_mode_ind[y_mode], INTRA_MODES);
}
+#endif // CONFIG_ENTROPY_STATS
#if CONFIG_FILTER_INTRA
if (mbmi->mode == DC_PRED && mbmi->palette_mode_info.palette_size[0] == 0) {
const int use_filter_intra_mode =
@@ -5797,7 +5743,6 @@
#if CONFIG_ENTROPY_STATS
++counts->uv_mode[y_mode][uv_mode];
#endif // CONFIG_ENTROPY_STATS
- update_cdf(fc->uv_mode_cdf[y_mode], av1_intra_mode_ind[uv_mode], INTRA_MODES);
}
#if CONFIG_VAR_TX
@@ -5944,7 +5889,6 @@
FRAME_COUNTS *counts) {
MB_MODE_INFO *mbmi = &xd->mi[0]->mbmi;
int is_inter = is_inter_block(mbmi);
- FRAME_CONTEXT *fc = xd->tile_ctx;
#if !CONFIG_TXK_SEL
TX_TYPE tx_type = mbmi->tx_type;
@@ -5965,14 +5909,9 @@
if (eset > 0) {
if (is_inter) {
++counts->inter_ext_tx[eset][txsize_sqr_map[tx_size]][tx_type];
- update_cdf(fc->inter_ext_tx_cdf[eset][txsize_sqr_map[tx_size]],
- av1_ext_tx_inter_ind[eset][tx_type], ext_tx_cnt_inter[eset]);
} else {
++counts->intra_ext_tx[eset][txsize_sqr_map[tx_size]][mbmi->mode]
[tx_type];
- update_cdf(
- fc->intra_ext_tx_cdf[eset][txsize_sqr_map[tx_size]][mbmi->mode],
- av1_ext_tx_intra_ind[eset][tx_type], ext_tx_cnt_intra[eset]);
}
}
}
@@ -5985,15 +5924,9 @@
!segfeature_active(&cm->seg, mbmi->segment_id, SEG_LVL_SKIP)) {
if (is_inter) {
++counts->inter_ext_tx[tx_size][tx_type];
- update_cdf(fc->inter_ext_tx_cdf[tx_size], av1_ext_tx_ind[tx_type],
- TX_TYPES);
} else {
++counts->intra_ext_tx[tx_size][intra_mode_to_tx_type_context[mbmi->mode]]
[tx_type];
- update_cdf(
- fc->intra_ext_tx_cdf[tx_size]
- [intra_mode_to_tx_type_context[mbmi->mode]],
- av1_ext_tx_ind[tx_type], TX_TYPES);
}
}
#endif // CONFIG_EXT_TX
diff --git a/av1/encoder/encodemv.c b/av1/encoder/encodemv.c
index 84bf921..fd61fe6 100644
--- a/av1/encoder/encodemv.c
+++ b/av1/encoder/encodemv.c
@@ -97,31 +97,22 @@
sign_cost[0] = av1_cost_zero(mvcomp->sign);
sign_cost[1] = av1_cost_one(mvcomp->sign);
- av1_cost_tokens_from_cdf(class_cost, mvcomp->class_cdf, NULL);
-#if CONFIG_NEW_MULTISYMBOL
- av1_cost_tokens_from_cdf(class0_cost, mvcomp->class0_cdf, NULL);
-#else
+ av1_cost_tokens(class_cost, mvcomp->classes, av1_mv_class_tree);
av1_cost_tokens(class0_cost, mvcomp->class0, av1_mv_class0_tree);
-#endif
for (i = 0; i < MV_OFFSET_BITS; ++i) {
bits_cost[i][0] = av1_cost_zero(mvcomp->bits[i]);
bits_cost[i][1] = av1_cost_one(mvcomp->bits[i]);
}
for (i = 0; i < CLASS0_SIZE; ++i)
- av1_cost_tokens_from_cdf(class0_fp_cost[i], mvcomp->class0_fp_cdf[i], NULL);
- av1_cost_tokens_from_cdf(fp_cost, mvcomp->fp_cdf, NULL);
+ av1_cost_tokens(class0_fp_cost[i], mvcomp->class0_fp[i], av1_mv_fp_tree);
+ av1_cost_tokens(fp_cost, mvcomp->fp, av1_mv_fp_tree);
if (precision > MV_SUBPEL_LOW_PRECISION) {
-#if CONFIG_NEW_MULTISYMBOL
- av1_cost_tokens_from_cdf(class0_hp_cost, mvcomp->class0_hp_cdf, NULL);
- av1_cost_tokens_from_cdf(hp_cost, mvcomp->hp_cdf, NULL);
-#else
class0_hp_cost[0] = av1_cost_zero(mvcomp->class0_hp);
class0_hp_cost[1] = av1_cost_one(mvcomp->class0_hp);
hp_cost[0] = av1_cost_zero(mvcomp->hp);
hp_cost[1] = av1_cost_one(mvcomp->hp);
-#endif
}
mvcost[0] = 0;
for (v = 1; v <= MV_MAX; ++v) {
@@ -226,64 +217,15 @@
void av1_build_nmv_cost_table(int *mvjoint, int *mvcost[2],
const nmv_context *ctx,
MvSubpelPrecision precision) {
- av1_cost_tokens_from_cdf(mvjoint, ctx->joint_cdf, NULL);
+ av1_cost_tokens(mvjoint, ctx->joints, av1_mv_joint_tree);
build_nmv_component_cost_table(mvcost[0], &ctx->comps[0], precision);
build_nmv_component_cost_table(mvcost[1], &ctx->comps[1], precision);
}
-static void inc_mv_component_cdf(int v, nmv_component *mvcomp,
- MvSubpelPrecision precision) {
- int s, z, c, o, d, f;
- assert(v != 0);
- s = v < 0;
- z = (s ? -v : v) - 1; /* magnitude - 1 */
-
- c = av1_get_mv_class(z, &o);
- update_cdf(mvcomp->class_cdf, c, MV_CLASSES);
- (void)precision;
-
- d = (o >> 3); /* int mv data */
- f = (o >> 1) & 3; /* fractional pel mv data */
-#if CONFIG_NEW_MULTISYMBOL
- int e = (o & 1); /* high precision mv data */
-#endif
-
- if (c == MV_CLASS_0) {
- update_cdf(mvcomp->class0_fp_cdf[d], f, MV_FP_SIZE);
-#if CONFIG_NEW_MULTISYMBOL
- if (precision > MV_SUBPEL_LOW_PRECISION)
- update_cdf(mvcomp->class0_hp_cdf, e, 2);
-#endif
- } else {
- update_cdf(mvcomp->fp_cdf, f, MV_FP_SIZE);
-#if CONFIG_NEW_MULTISYMBOL
- if (precision > MV_SUBPEL_LOW_PRECISION) update_cdf(mvcomp->hp_cdf, e, 2);
-#endif
- }
-}
-
-static void inc_mv_cdf(const MV *mv, nmv_context *const nmvc,
- MvSubpelPrecision precision) {
- const MV_JOINT_TYPE j = av1_get_mv_joint(mv);
- update_cdf(nmvc->joint_cdf, j, MV_JOINTS);
-
- if (mv_joint_vertical(j)) {
- int v = mv->row;
- nmv_component *mvcomp = &nmvc->comps[0];
- inc_mv_component_cdf(v, mvcomp, precision);
- }
-
- if (mv_joint_horizontal(j)) {
- int v = mv->col;
- nmv_component *mvcomp = &nmvc->comps[1];
- inc_mv_component_cdf(v, mvcomp, precision);
- }
-}
-
#if CONFIG_EXT_INTER
static void inc_mvs(const MB_MODE_INFO *mbmi, const MB_MODE_INFO_EXT *mbmi_ext,
const int_mv mvs[2], const int_mv pred_mvs[2],
- nmv_context_counts *nmv_counts, FRAME_CONTEXT *fc) {
+ nmv_context_counts *nmv_counts) {
int i;
PREDICTION_MODE mode = mbmi->mode;
@@ -299,7 +241,6 @@
nmv_context_counts *counts = &nmv_counts[nmv_ctx];
(void)pred_mvs;
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
}
} else if (mode == NEAREST_NEWMV || mode == NEAR_NEWMV) {
const MV *ref = &mbmi_ext->ref_mvs[mbmi->ref_frame[1]][0].as_mv;
@@ -311,7 +252,6 @@
mbmi_ext->ref_mv_stack[rf_type], 1, mbmi->ref_mv_idx);
nmv_context_counts *counts = &nmv_counts[nmv_ctx];
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
} else if (mode == NEW_NEARESTMV || mode == NEW_NEARMV) {
const MV *ref = &mbmi_ext->ref_mvs[mbmi->ref_frame[0]][0].as_mv;
const MV diff = { mvs[0].as_mv.row - ref->row,
@@ -322,7 +262,6 @@
mbmi_ext->ref_mv_stack[rf_type], 0, mbmi->ref_mv_idx);
nmv_context_counts *counts = &nmv_counts[nmv_ctx];
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
#if CONFIG_COMPOUND_SINGLEREF
} else {
assert( // mode == SR_NEAREST_NEWMV ||
@@ -339,19 +278,17 @@
diff.row = mvs[0].as_mv.row - ref->row;
diff.col = mvs[0].as_mv.col - ref->col;
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
}
diff.row = mvs[1].as_mv.row - ref->row;
diff.col = mvs[1].as_mv.col - ref->col;
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
#endif // CONFIG_COMPOUND_SINGLEREF
}
}
static void inc_mvs_sub8x8(const MODE_INFO *mi, int block, const int_mv mvs[2],
const MB_MODE_INFO_EXT *mbmi_ext,
- nmv_context_counts *nmv_counts, FRAME_CONTEXT *fc) {
+ nmv_context_counts *nmv_counts) {
int i;
PREDICTION_MODE mode = mi->bmi[block].as_mode;
const MB_MODE_INFO *mbmi = &mi->mbmi;
@@ -367,7 +304,6 @@
mbmi_ext->ref_mv_stack[rf_type], i, mbmi->ref_mv_idx);
nmv_context_counts *counts = &nmv_counts[nmv_ctx];
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
}
} else if (mode == NEAREST_NEWMV || mode == NEAR_NEWMV) {
const MV *ref = &mi->bmi[block].ref_mv[1].as_mv;
@@ -379,7 +315,6 @@
mbmi_ext->ref_mv_stack[rf_type], 1, mbmi->ref_mv_idx);
nmv_context_counts *counts = &nmv_counts[nmv_ctx];
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
} else if (mode == NEW_NEARESTMV || mode == NEW_NEARMV) {
const MV *ref = &mi->bmi[block].ref_mv[0].as_mv;
const MV diff = { mvs[0].as_mv.row - ref->row,
@@ -390,13 +325,12 @@
mbmi_ext->ref_mv_stack[rf_type], 0, mbmi->ref_mv_idx);
nmv_context_counts *counts = &nmv_counts[nmv_ctx];
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
}
}
#else // !CONFIG_EXT_INTER
static void inc_mvs(const MB_MODE_INFO *mbmi, const MB_MODE_INFO_EXT *mbmi_ext,
const int_mv mvs[2], const int_mv pred_mvs[2],
- nmv_context_counts *nmv_counts, FRAME_CONTEXT *fc) {
+ nmv_context_counts *nmv_counts) {
int i;
for (i = 0; i < 1 + has_second_ref(mbmi); ++i) {
@@ -409,7 +343,6 @@
const MV diff = { mvs[i].as_mv.row - ref->row,
mvs[i].as_mv.col - ref->col };
av1_inc_mv(&diff, counts, 1);
- inc_mv_cdf(&diff, &fc->nmvc[nmv_ctx], 1);
}
}
#endif // CONFIG_EXT_INTER
@@ -436,12 +369,11 @@
#if CONFIG_EXT_INTER
if (have_newmv_in_inter_mode(mi->bmi[i].as_mode))
- inc_mvs_sub8x8(mi, i, mi->bmi[i].as_mv, mbmi_ext, td->counts->mv,
- xd->tile_ctx);
+ inc_mvs_sub8x8(mi, i, mi->bmi[i].as_mv, mbmi_ext, td->counts->mv);
#else
if (mi->bmi[i].as_mode == NEWMV)
inc_mvs(mbmi, mbmi_ext, mi->bmi[i].as_mv, mi->bmi[i].pred_mv,
- td->counts->mv, xd->tile_ctx);
+ td->counts->mv);
#endif // CONFIG_EXT_INTER
}
}
@@ -451,7 +383,6 @@
#else
if (mbmi->mode == NEWMV)
#endif // CONFIG_EXT_INTER
- inc_mvs(mbmi, mbmi_ext, mbmi->mv, mbmi->pred_mv, td->counts->mv,
- xd->tile_ctx);
+ inc_mvs(mbmi, mbmi_ext, mbmi->mv, mbmi->pred_mv, td->counts->mv);
}
}
diff --git a/av1/encoder/encoder.c b/av1/encoder/encoder.c
index 2d3a741..972c9bf 100644
--- a/av1/encoder/encoder.c
+++ b/av1/encoder/encoder.c
@@ -249,6 +249,23 @@
}
}
+void av1_set_high_precision_mv(AV1_COMP *cpi, int allow_high_precision_mv) {
+ MACROBLOCK *const mb = &cpi->td.mb;
+ cpi->common.allow_high_precision_mv = allow_high_precision_mv;
+
+ if (cpi->common.allow_high_precision_mv) {
+ int i;
+ for (i = 0; i < NMV_CONTEXTS; ++i) {
+ mb->mv_cost_stack[i] = mb->nmvcost_hp[i];
+ }
+ } else {
+ int i;
+ for (i = 0; i < NMV_CONTEXTS; ++i) {
+ mb->mv_cost_stack[i] = mb->nmvcost[i];
+ }
+ }
+}
+
static BLOCK_SIZE select_sb_size(const AV1_COMP *const cpi) {
#if CONFIG_EXT_PARTITION
if (cpi->oxcf.superblock_size == AOM_SUPERBLOCK_SIZE_64X64)
@@ -523,11 +540,18 @@
static void save_coding_context(AV1_COMP *cpi) {
CODING_CONTEXT *const cc = &cpi->coding_context;
AV1_COMMON *cm = &cpi->common;
+ int i;
// Stores a snapshot of key state variables which can subsequently be
// restored with a call to av1_restore_coding_context. These functions are
// intended for use in a re-code loop in av1_compress_frame where the
// quantizer value is adjusted between loop iterations.
+ for (i = 0; i < NMV_CONTEXTS; ++i) {
+ av1_copy(cc->nmv_vec_cost[i], cpi->td.mb.nmv_vec_cost[i]);
+ av1_copy(cc->nmv_costs, cpi->nmv_costs);
+ av1_copy(cc->nmv_costs_hp, cpi->nmv_costs_hp);
+ }
+
av1_copy(cc->last_ref_lf_deltas, cm->lf.last_ref_deltas);
av1_copy(cc->last_mode_lf_deltas, cm->lf.last_mode_deltas);
@@ -537,9 +561,16 @@
static void restore_coding_context(AV1_COMP *cpi) {
CODING_CONTEXT *const cc = &cpi->coding_context;
AV1_COMMON *cm = &cpi->common;
+ int i;
// Restore key state variables to the snapshot state stored in the
// previous call to av1_save_coding_context.
+ for (i = 0; i < NMV_CONTEXTS; ++i) {
+ av1_copy(cpi->td.mb.nmv_vec_cost[i], cc->nmv_vec_cost[i]);
+ av1_copy(cpi->nmv_costs, cc->nmv_costs);
+ av1_copy(cpi->nmv_costs_hp, cc->nmv_costs_hp);
+ }
+
av1_copy(cm->lf.last_ref_deltas, cc->last_ref_lf_deltas);
av1_copy(cm->lf.last_mode_deltas, cc->last_mode_lf_deltas);
@@ -2296,7 +2327,7 @@
set_compound_tools(cm);
#endif // CONFIG_EXT_INTER
av1_reset_segment_features(cm);
- cm->allow_high_precision_mv = 0;
+ av1_set_high_precision_mv(cpi, 0);
set_rc_buffer_sizes(rc, &cpi->oxcf);
@@ -2426,6 +2457,11 @@
realloc_segmentation_maps(cpi);
+ for (i = 0; i < NMV_CONTEXTS; ++i) {
+ memset(cpi->nmv_costs, 0, sizeof(cpi->nmv_costs));
+ memset(cpi->nmv_costs_hp, 0, sizeof(cpi->nmv_costs_hp));
+ }
+
for (i = 0; i < (sizeof(cpi->mbgraph_stats) / sizeof(cpi->mbgraph_stats[0]));
i++) {
CHECK_MEM_ERROR(
@@ -2488,6 +2524,13 @@
cpi->first_time_stamp_ever = INT64_MAX;
+ for (i = 0; i < NMV_CONTEXTS; ++i) {
+ cpi->td.mb.nmvcost[i][0] = &cpi->nmv_costs[i][0][MV_MAX];
+ cpi->td.mb.nmvcost[i][1] = &cpi->nmv_costs[i][1][MV_MAX];
+ cpi->td.mb.nmvcost_hp[i][0] = &cpi->nmv_costs_hp[i][0][MV_MAX];
+ cpi->td.mb.nmvcost_hp[i][1] = &cpi->nmv_costs_hp[i][1][MV_MAX];
+ }
+
#ifdef OUTPUT_YUV_SKINMAP
yuv_skinmap_file = fopen("skinmap.yuv", "ab");
#endif
@@ -3875,8 +3918,9 @@
*q = av1_rc_pick_q_and_bounds(cpi, bottom_index, top_index);
#endif
- if (!frame_is_intra_only(cm))
- cm->allow_high_precision_mv = (*q) < HIGH_PRECISION_MV_QTHRESH;
+ if (!frame_is_intra_only(cm)) {
+ av1_set_high_precision_mv(cpi, (*q) < HIGH_PRECISION_MV_QTHRESH);
+ }
// Configure experimental use of segmentation for enhanced coding of
// static regions if indicated.
@@ -5689,7 +5733,7 @@
aom_usec_timer_start(&cmptimer);
- cm->allow_high_precision_mv = ALTREF_HIGH_PRECISION_MV;
+ av1_set_high_precision_mv(cpi, ALTREF_HIGH_PRECISION_MV);
// Is multi-arf enabled.
// Note that at the moment multi_arf is only configured for 2 pass VBR
diff --git a/av1/encoder/encoder.h b/av1/encoder/encoder.h
index 8239f30..cc7eeb6 100644
--- a/av1/encoder/encoder.h
+++ b/av1/encoder/encoder.h
@@ -58,6 +58,10 @@
#endif // CONFIG_SPEED_REFS
typedef struct {
+ int nmv_vec_cost[NMV_CONTEXTS][MV_JOINTS];
+ int nmv_costs[NMV_CONTEXTS][2][MV_VALS];
+ int nmv_costs_hp[NMV_CONTEXTS][2][MV_VALS];
+
// 0 = Intra, Last, GF, ARF
signed char last_ref_lf_deltas[TOTAL_REFS_PER_FRAME];
// 0 = ZERO_MV, MV
@@ -452,6 +456,9 @@
int gmparams_cost[TOTAL_REFS_PER_FRAME];
#endif // CONFIG_GLOBAL_MOTION
+ int nmv_costs[NMV_CONTEXTS][2][MV_VALS];
+ int nmv_costs_hp[NMV_CONTEXTS][2][MV_VALS];
+
int64_t last_time_stamp_seen;
int64_t last_end_time_stamp_seen;
int64_t first_time_stamp_ever;
@@ -745,6 +752,7 @@
void av1_update_reference_frames(AV1_COMP *cpi);
+void av1_set_high_precision_mv(AV1_COMP *cpi, int allow_high_precision_mv);
#if CONFIG_TEMPMV_SIGNALING
void av1_set_temporal_mv_prediction(AV1_COMP *cpi, int allow_tempmv_prediction);
#endif
diff --git a/av1/encoder/rd.c b/av1/encoder/rd.c
index 6dfeb68..8a2a63c 100644
--- a/av1/encoder/rd.c
+++ b/av1/encoder/rd.c
@@ -61,29 +61,6 @@
4, 4, 8, 8, 16, 16
};
-void av1_fill_nmv_costs(AV1_COMMON *const cm, MACROBLOCK *x,
- FRAME_CONTEXT *fc) {
- for (int i = 0; i < NMV_CONTEXTS; ++i) {
- memset(x->nmvcost_array, 0, sizeof(x->nmvcost_array));
- memset(x->nmvcost_hp_array, 0, sizeof(x->nmvcost_hp_array));
- x->nmvcost[i][0] = &x->nmvcost_array[i][0][MV_MAX];
- x->nmvcost[i][1] = &x->nmvcost_array[i][1][MV_MAX];
- x->nmvcost_hp[i][0] = &x->nmvcost_hp_array[i][0][MV_MAX];
- x->nmvcost_hp[i][1] = &x->nmvcost_hp_array[i][1][MV_MAX];
- if (cm->allow_high_precision_mv)
- x->mv_cost_stack[i] = x->nmvcost_hp[i];
- else
- x->mv_cost_stack[i] = x->nmvcost[i];
-
- av1_build_nmv_cost_table(
- x->nmv_vec_cost[i],
- cm->allow_high_precision_mv ? x->nmvcost_hp[i] : x->nmvcost[i],
- &fc->nmvc[i], cm->allow_high_precision_mv);
- }
- x->mvcost = x->mv_cost_stack[0];
- x->nmvjointcost = x->nmv_vec_cost[0];
-}
-
void av1_fill_mode_rates(AV1_COMMON *const cm, MACROBLOCK *x,
FRAME_CONTEXT *fc) {
int i, j;
@@ -223,15 +200,8 @@
av1_cost_tokens(x->switchable_restore_cost, fc->switchable_restore_prob,
av1_switchable_restore_tree);
#endif // CONFIG_LOOP_RESTORATION
-#if CONFIG_INTRABC
- av1_fill_nmv_costs(cm, x, fc);
-#endif
if (!frame_is_intra_only(cm)) {
-#if !CONFIG_INTRABC
- av1_fill_nmv_costs(cm, x, fc);
-#endif
-
for (i = 0; i < NEWMV_MODE_CONTEXTS; ++i) {
#if CONFIG_NEW_MULTISYMBOL
av1_cost_tokens_from_cdf(x->newmv_mode_cost[i], fc->newmv_cdf[i], NULL);
@@ -271,9 +241,6 @@
for (i = 0; i < INTER_MODE_CONTEXTS; ++i)
av1_cost_tokens_from_cdf(x->inter_compound_mode_cost[i],
fc->inter_compound_mode_cdf[i], NULL);
- for (i = 0; i < BLOCK_SIZES_ALL; ++i)
- av1_cost_tokens_from_cdf(x->compound_type_cost[i],
- fc->compound_type_cdf[i], NULL);
#if CONFIG_COMPOUND_SINGLEREF
for (i = 0; i < INTER_MODE_CONTEXTS; ++i)
av1_cost_tokens_from_cdf(x->inter_singleref_comp_mode_cost[i],
@@ -569,6 +536,7 @@
AV1_COMMON *const cm = &cpi->common;
MACROBLOCK *const x = &cpi->td.mb;
RD_OPT *const rd = &cpi->rd;
+ int nmv_ctx;
aom_clear_system_state();
@@ -578,7 +546,15 @@
set_block_thresholds(cm, rd);
- av1_fill_nmv_costs(cm, x, cm->fc);
+ for (nmv_ctx = 0; nmv_ctx < NMV_CONTEXTS; ++nmv_ctx) {
+ av1_build_nmv_cost_table(
+ x->nmv_vec_cost[nmv_ctx],
+ cm->allow_high_precision_mv ? x->nmvcost_hp[nmv_ctx]
+ : x->nmvcost[nmv_ctx],
+ &cm->fc->nmvc[nmv_ctx], cm->allow_high_precision_mv);
+ }
+ x->mvcost = x->mv_cost_stack[0];
+ x->nmvjointcost = x->nmv_vec_cost[0];
#if CONFIG_INTRABC
if (frame_is_intra_only(cm) && cm->allow_screen_content_tools &&
diff --git a/av1/encoder/rdopt.c b/av1/encoder/rdopt.c
index 69bc56c..180001a 100644
--- a/av1/encoder/rdopt.c
+++ b/av1/encoder/rdopt.c
@@ -9232,6 +9232,7 @@
int best_tmp_rate_mv = rate_mv;
int tmp_skip_txfm_sb;
int64_t tmp_skip_sse_sb;
+ int compound_type_cost[COMPOUND_TYPES];
DECLARE_ALIGNED(16, uint8_t, pred0[2 * MAX_SB_SQUARE]);
DECLARE_ALIGNED(16, uint8_t, pred1[2 * MAX_SB_SQUARE]);
uint8_t *preds0[1] = { pred0 };
@@ -9243,7 +9244,6 @@
masked_compound_used = masked_compound_used && cm->allow_masked_compound;
#endif // CONFIG_COMPOUND_SEGMENT || CONFIG_WEDGE
COMPOUND_TYPE cur_type;
- int best_compmode_interinter_cost = 0;
best_mv[0].as_int = cur_mv[0].as_int;
best_mv[1].as_int = cur_mv[1].as_int;
@@ -9265,6 +9265,8 @@
#endif // CONFIG_EXT_INTER && CONFIG_COMPOUND_SINGLEREF
if (masked_compound_used) {
+ av1_cost_tokens(compound_type_cost, cm->fc->compound_type_prob[bsize],
+ av1_compound_type_tree);
// get inter predictors to use for masked compound modes
av1_build_inter_predictors_for_planes_single_buf(
xd, bsize, 0, 0, mi_row, mi_col, 0, preds0, strides);
@@ -9286,7 +9288,7 @@
else
#endif // CONFIG_WEDGE && CONFIG_COMPOUND_SEGMENT
masked_type_cost +=
- x->compound_type_cost[bsize][mbmi->interinter_compound_type];
+ compound_type_cost[mbmi->interinter_compound_type];
}
rs2 = av1_cost_literal(get_interinter_compound_type_bits(
bsize, mbmi->interinter_compound_type)) +
@@ -9340,7 +9342,6 @@
#endif // CONFIG_COMPOUND_SEGMENT
best_compound_data.interinter_compound_type =
mbmi->interinter_compound_type;
- best_compmode_interinter_cost = rs2;
if (have_newmv_in_inter_mode(this_mode)) {
if (use_masked_motion_search(cur_type)) {
best_tmp_rate_mv = tmp_rate_mv;
@@ -9385,7 +9386,12 @@
pred_exists = 0;
- compmode_interinter_cost = best_compmode_interinter_cost;
+ compmode_interinter_cost =
+ av1_cost_literal(get_interinter_compound_type_bits(
+ bsize, mbmi->interinter_compound_type)) +
+ (masked_compound_used
+ ? compound_type_cost[mbmi->interinter_compound_type]
+ : 0);
}
#endif // CONFIG_WEDGE || CONFIG_COMPOUND_SEGMENT