/* * Copyright (c) 2016, Alliance for Open Media. All rights reserved * * This source code is subject to the terms of the BSD 2 Clause License and * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License * was not distributed with this source code in the LICENSE file, you can * obtain it at www.aomedia.org/license/software. If the Alliance for Open * Media Patent License 1.0 was not distributed with this source code in the * PATENTS file, you can obtain it at www.aomedia.org/license/patent. */ #include #include #include #include "config/av1_rtcd.h" #include "aom_dsp/aom_dsp_common.h" #include "aom_mem/aom_mem.h" #include "aom_ports/bitops.h" #include "aom_ports/mem.h" #include "aom_ports/system_state.h" #include "av1/common/common.h" #include "av1/common/entropy.h" #include "av1/common/entropymode.h" #include "av1/common/mvref_common.h" #include "av1/common/pred_common.h" #include "av1/common/quant_common.h" #include "av1/common/reconinter.h" #include "av1/common/reconintra.h" #include "av1/common/seg_common.h" #include "av1/encoder/av1_quantize.h" #include "av1/encoder/cost.h" #include "av1/encoder/encodemb.h" #include "av1/encoder/encodemv.h" #include "av1/encoder/encoder.h" #include "av1/encoder/encodetxb.h" #include "av1/encoder/mcomp.h" #include "av1/encoder/ratectrl.h" #include "av1/encoder/rd.h" #include "av1/encoder/tokenize.h" #define RD_THRESH_POW 1.25 // Factor to weigh the rate for switchable interp filters. #define SWITCHABLE_INTERP_RATE_FACTOR 1 // The baseline rd thresholds for breaking out of the rd loop for // certain modes are assumed to be based on 8x8 blocks. // This table is used to correct for block size. // The factors here are << 2 (2 = x0.5, 32 = x8 etc). static const uint8_t rd_thresh_block_size_factor[BLOCK_SIZES_ALL] = { 2, 3, 3, 4, 6, 6, 8, 12, 12, 16, 24, 24, 32, 48, 48, 64, 4, 4, 8, 8, 16, 16 }; static const int use_intra_ext_tx_for_txsize[EXT_TX_SETS_INTRA][EXT_TX_SIZES] = { { 1, 1, 1, 1 }, // unused { 1, 1, 0, 0 }, { 0, 0, 1, 0 }, }; static const int use_inter_ext_tx_for_txsize[EXT_TX_SETS_INTER][EXT_TX_SIZES] = { { 1, 1, 1, 1 }, // unused { 1, 1, 0, 0 }, { 0, 0, 1, 0 }, { 0, 0, 0, 1 }, }; static const int av1_ext_tx_set_idx_to_type[2][AOMMAX(EXT_TX_SETS_INTRA, EXT_TX_SETS_INTER)] = { { // Intra EXT_TX_SET_DCTONLY, EXT_TX_SET_DTT4_IDTX_1DDCT, EXT_TX_SET_DTT4_IDTX, }, { // Inter EXT_TX_SET_DCTONLY, EXT_TX_SET_ALL16, EXT_TX_SET_DTT9_IDTX_1DDCT, EXT_TX_SET_DCT_IDTX, }, }; void av1_fill_mode_rates(AV1_COMMON *const cm, MACROBLOCK *x, FRAME_CONTEXT *fc) { int i, j; for (i = 0; i < PARTITION_CONTEXTS; ++i) av1_cost_tokens_from_cdf(x->partition_cost[i], fc->partition_cdf[i], NULL); if (cm->skip_mode_flag) { for (i = 0; i < SKIP_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->skip_mode_cost[i], fc->skip_mode_cdfs[i], NULL); } } for (i = 0; i < SKIP_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->skip_cost[i], fc->skip_cdfs[i], NULL); } for (i = 0; i < KF_MODE_CONTEXTS; ++i) for (j = 0; j < KF_MODE_CONTEXTS; ++j) av1_cost_tokens_from_cdf(x->y_mode_costs[i][j], fc->kf_y_cdf[i][j], NULL); for (i = 0; i < BLOCK_SIZE_GROUPS; ++i) av1_cost_tokens_from_cdf(x->mbmode_cost[i], fc->y_mode_cdf[i], NULL); for (i = 0; i < CFL_ALLOWED_TYPES; ++i) for (j = 0; j < INTRA_MODES; ++j) av1_cost_tokens_from_cdf(x->intra_uv_mode_cost[i][j], fc->uv_mode_cdf[i][j], NULL); av1_cost_tokens_from_cdf(x->filter_intra_mode_cost, fc->filter_intra_mode_cdf, NULL); for (i = 0; i < BLOCK_SIZES_ALL; ++i) { if (av1_filter_intra_allowed_bsize(cm, i)) av1_cost_tokens_from_cdf(x->filter_intra_cost[i], fc->filter_intra_cdfs[i], NULL); } for (i = 0; i < SWITCHABLE_FILTER_CONTEXTS; ++i) av1_cost_tokens_from_cdf(x->switchable_interp_costs[i], fc->switchable_interp_cdf[i], NULL); for (i = 0; i < PALATTE_BSIZE_CTXS; ++i) { av1_cost_tokens_from_cdf(x->palette_y_size_cost[i], fc->palette_y_size_cdf[i], NULL); av1_cost_tokens_from_cdf(x->palette_uv_size_cost[i], fc->palette_uv_size_cdf[i], NULL); for (j = 0; j < PALETTE_Y_MODE_CONTEXTS; ++j) { av1_cost_tokens_from_cdf(x->palette_y_mode_cost[i][j], fc->palette_y_mode_cdf[i][j], NULL); } } for (i = 0; i < PALETTE_UV_MODE_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->palette_uv_mode_cost[i], fc->palette_uv_mode_cdf[i], NULL); } for (i = 0; i < PALETTE_SIZES; ++i) { for (j = 0; j < PALETTE_COLOR_INDEX_CONTEXTS; ++j) { av1_cost_tokens_from_cdf(x->palette_y_color_cost[i][j], fc->palette_y_color_index_cdf[i][j], NULL); av1_cost_tokens_from_cdf(x->palette_uv_color_cost[i][j], fc->palette_uv_color_index_cdf[i][j], NULL); } } int sign_cost[CFL_JOINT_SIGNS]; av1_cost_tokens_from_cdf(sign_cost, fc->cfl_sign_cdf, NULL); for (int joint_sign = 0; joint_sign < CFL_JOINT_SIGNS; joint_sign++) { int *cost_u = x->cfl_cost[joint_sign][CFL_PRED_U]; int *cost_v = x->cfl_cost[joint_sign][CFL_PRED_V]; if (CFL_SIGN_U(joint_sign) == CFL_SIGN_ZERO) { memset(cost_u, 0, CFL_ALPHABET_SIZE * sizeof(*cost_u)); } else { const aom_cdf_prob *cdf_u = fc->cfl_alpha_cdf[CFL_CONTEXT_U(joint_sign)]; av1_cost_tokens_from_cdf(cost_u, cdf_u, NULL); } if (CFL_SIGN_V(joint_sign) == CFL_SIGN_ZERO) { memset(cost_v, 0, CFL_ALPHABET_SIZE * sizeof(*cost_v)); } else { const aom_cdf_prob *cdf_v = fc->cfl_alpha_cdf[CFL_CONTEXT_V(joint_sign)]; av1_cost_tokens_from_cdf(cost_v, cdf_v, NULL); } for (int u = 0; u < CFL_ALPHABET_SIZE; u++) cost_u[u] += sign_cost[joint_sign]; } for (i = 0; i < MAX_TX_CATS; ++i) for (j = 0; j < TX_SIZE_CONTEXTS; ++j) av1_cost_tokens_from_cdf(x->tx_size_cost[i][j], fc->tx_size_cdf[i][j], NULL); for (i = 0; i < TXFM_PARTITION_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->txfm_partition_cost[i], fc->txfm_partition_cdf[i], NULL); } for (i = TX_4X4; i < EXT_TX_SIZES; ++i) { int s; for (s = 1; s < EXT_TX_SETS_INTER; ++s) { if (use_inter_ext_tx_for_txsize[s][i]) { av1_cost_tokens_from_cdf( x->inter_tx_type_costs[s][i], fc->inter_ext_tx_cdf[s][i], av1_ext_tx_inv[av1_ext_tx_set_idx_to_type[1][s]]); } } for (s = 1; s < EXT_TX_SETS_INTRA; ++s) { if (use_intra_ext_tx_for_txsize[s][i]) { for (j = 0; j < INTRA_MODES; ++j) { av1_cost_tokens_from_cdf( x->intra_tx_type_costs[s][i][j], fc->intra_ext_tx_cdf[s][i][j], av1_ext_tx_inv[av1_ext_tx_set_idx_to_type[0][s]]); } } } } for (i = 0; i < DIRECTIONAL_MODES; ++i) { av1_cost_tokens_from_cdf(x->angle_delta_cost[i], fc->angle_delta_cdf[i], NULL); } av1_cost_tokens_from_cdf(x->switchable_restore_cost, fc->switchable_restore_cdf, NULL); av1_cost_tokens_from_cdf(x->wiener_restore_cost, fc->wiener_restore_cdf, NULL); av1_cost_tokens_from_cdf(x->sgrproj_restore_cost, fc->sgrproj_restore_cdf, NULL); av1_cost_tokens_from_cdf(x->intrabc_cost, fc->intrabc_cdf, NULL); if (!frame_is_intra_only(cm)) { for (i = 0; i < COMP_INTER_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->comp_inter_cost[i], fc->comp_inter_cdf[i], NULL); } for (i = 0; i < REF_CONTEXTS; ++i) { for (j = 0; j < SINGLE_REFS - 1; ++j) { av1_cost_tokens_from_cdf(x->single_ref_cost[i][j], fc->single_ref_cdf[i][j], NULL); } } for (i = 0; i < COMP_REF_TYPE_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->comp_ref_type_cost[i], fc->comp_ref_type_cdf[i], NULL); } for (i = 0; i < UNI_COMP_REF_CONTEXTS; ++i) { for (j = 0; j < UNIDIR_COMP_REFS - 1; ++j) { av1_cost_tokens_from_cdf(x->uni_comp_ref_cost[i][j], fc->uni_comp_ref_cdf[i][j], NULL); } } for (i = 0; i < REF_CONTEXTS; ++i) { for (j = 0; j < FWD_REFS - 1; ++j) { av1_cost_tokens_from_cdf(x->comp_ref_cost[i][j], fc->comp_ref_cdf[i][j], NULL); } } for (i = 0; i < REF_CONTEXTS; ++i) { for (j = 0; j < BWD_REFS - 1; ++j) { av1_cost_tokens_from_cdf(x->comp_bwdref_cost[i][j], fc->comp_bwdref_cdf[i][j], NULL); } } for (i = 0; i < INTRA_INTER_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->intra_inter_cost[i], fc->intra_inter_cdf[i], NULL); } for (i = 0; i < NEWMV_MODE_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->newmv_mode_cost[i], fc->newmv_cdf[i], NULL); } for (i = 0; i < GLOBALMV_MODE_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->zeromv_mode_cost[i], fc->zeromv_cdf[i], NULL); } for (i = 0; i < REFMV_MODE_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->refmv_mode_cost[i], fc->refmv_cdf[i], NULL); } for (i = 0; i < DRL_MODE_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->drl_mode_cost0[i], fc->drl_cdf[i], NULL); } for (i = 0; i < INTER_MODE_CONTEXTS; ++i) av1_cost_tokens_from_cdf(x->inter_compound_mode_cost[i], fc->inter_compound_mode_cdf[i], NULL); for (i = 0; i < BLOCK_SIZES_ALL; ++i) av1_cost_tokens_from_cdf(x->compound_type_cost[i], fc->compound_type_cdf[i], NULL); for (i = 0; i < BLOCK_SIZES_ALL; ++i) { if (get_interinter_wedge_bits(i)) { av1_cost_tokens_from_cdf(x->wedge_idx_cost[i], fc->wedge_idx_cdf[i], NULL); } } for (i = 0; i < BLOCK_SIZE_GROUPS; ++i) { av1_cost_tokens_from_cdf(x->interintra_cost[i], fc->interintra_cdf[i], NULL); av1_cost_tokens_from_cdf(x->interintra_mode_cost[i], fc->interintra_mode_cdf[i], NULL); } for (i = 0; i < BLOCK_SIZES_ALL; ++i) { av1_cost_tokens_from_cdf(x->wedge_interintra_cost[i], fc->wedge_interintra_cdf[i], NULL); } for (i = BLOCK_8X8; i < BLOCK_SIZES_ALL; i++) { av1_cost_tokens_from_cdf(x->motion_mode_cost[i], fc->motion_mode_cdf[i], NULL); } for (i = BLOCK_8X8; i < BLOCK_SIZES_ALL; i++) { av1_cost_tokens_from_cdf(x->motion_mode_cost1[i], fc->obmc_cdf[i], NULL); } for (i = 0; i < COMP_INDEX_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->comp_idx_cost[i], fc->compound_index_cdf[i], NULL); } for (i = 0; i < COMP_GROUP_IDX_CONTEXTS; ++i) { av1_cost_tokens_from_cdf(x->comp_group_idx_cost[i], fc->comp_group_idx_cdf[i], NULL); } } } // Values are now correlated to quantizer. static int sad_per_bit16lut_8[QINDEX_RANGE]; static int sad_per_bit4lut_8[QINDEX_RANGE]; static int sad_per_bit16lut_10[QINDEX_RANGE]; static int sad_per_bit4lut_10[QINDEX_RANGE]; static int sad_per_bit16lut_12[QINDEX_RANGE]; static int sad_per_bit4lut_12[QINDEX_RANGE]; static void init_me_luts_bd(int *bit16lut, int *bit4lut, int range, aom_bit_depth_t bit_depth) { int i; // Initialize the sad lut tables using a formulaic calculation for now. // This is to make it easier to resolve the impact of experimental changes // to the quantizer tables. for (i = 0; i < range; i++) { const double q = av1_convert_qindex_to_q(i, bit_depth); bit16lut[i] = (int)(0.0418 * q + 2.4107); bit4lut[i] = (int)(0.063 * q + 2.742); } } void av1_init_me_luts(void) { init_me_luts_bd(sad_per_bit16lut_8, sad_per_bit4lut_8, QINDEX_RANGE, AOM_BITS_8); init_me_luts_bd(sad_per_bit16lut_10, sad_per_bit4lut_10, QINDEX_RANGE, AOM_BITS_10); init_me_luts_bd(sad_per_bit16lut_12, sad_per_bit4lut_12, QINDEX_RANGE, AOM_BITS_12); } static const int rd_boost_factor[16] = { 64, 32, 32, 32, 24, 16, 12, 12, 8, 8, 4, 4, 2, 2, 1, 0 }; static const int rd_frame_type_factor[FRAME_UPDATE_TYPES] = { 128, 144, 128, 128, 144, // TODO(zoeliu): To adjust further following factor values. 128, 128, 128, // TODO(weitinglin): We should investigate if the values should be the same // as the value used by OVERLAY frame 144, // INTNL_OVERLAY_UPDATE 128 // INTNL_ARF_UPDATE }; int av1_compute_rd_mult(const AV1_COMP *cpi, int qindex) { const int64_t q = av1_dc_quant_Q3(qindex, 0, cpi->common.bit_depth); int64_t rdmult = 0; switch (cpi->common.bit_depth) { case AOM_BITS_8: rdmult = 88 * q * q / 24; break; case AOM_BITS_10: rdmult = ROUND_POWER_OF_TWO(88 * q * q / 24, 4); break; case AOM_BITS_12: rdmult = ROUND_POWER_OF_TWO(88 * q * q / 24, 8); break; default: assert(0 && "bit_depth should be AOM_BITS_8, AOM_BITS_10 or AOM_BITS_12"); return -1; } if (cpi->oxcf.pass == 2 && (cpi->common.frame_type != KEY_FRAME)) { const GF_GROUP *const gf_group = &cpi->twopass.gf_group; const FRAME_UPDATE_TYPE frame_type = gf_group->update_type[gf_group->index]; const int boost_index = AOMMIN(15, (cpi->rc.gfu_boost / 100)); rdmult = (rdmult * rd_frame_type_factor[frame_type]) >> 7; rdmult += ((rdmult * rd_boost_factor[boost_index]) >> 7); } if (rdmult < 1) rdmult = 1; return (int)rdmult; } static int compute_rd_thresh_factor(int qindex, aom_bit_depth_t bit_depth) { double q; switch (bit_depth) { case AOM_BITS_8: q = av1_dc_quant_Q3(qindex, 0, AOM_BITS_8) / 4.0; break; case AOM_BITS_10: q = av1_dc_quant_Q3(qindex, 0, AOM_BITS_10) / 16.0; break; case AOM_BITS_12: q = av1_dc_quant_Q3(qindex, 0, AOM_BITS_12) / 64.0; break; default: assert(0 && "bit_depth should be AOM_BITS_8, AOM_BITS_10 or AOM_BITS_12"); return -1; } // TODO(debargha): Adjust the function below. return AOMMAX((int)(pow(q, RD_THRESH_POW) * 5.12), 8); } void av1_initialize_me_consts(const AV1_COMP *cpi, MACROBLOCK *x, int qindex) { switch (cpi->common.bit_depth) { case AOM_BITS_8: x->sadperbit16 = sad_per_bit16lut_8[qindex]; x->sadperbit4 = sad_per_bit4lut_8[qindex]; break; case AOM_BITS_10: x->sadperbit16 = sad_per_bit16lut_10[qindex]; x->sadperbit4 = sad_per_bit4lut_10[qindex]; break; case AOM_BITS_12: x->sadperbit16 = sad_per_bit16lut_12[qindex]; x->sadperbit4 = sad_per_bit4lut_12[qindex]; break; default: assert(0 && "bit_depth should be AOM_BITS_8, AOM_BITS_10 or AOM_BITS_12"); } } static void set_block_thresholds(const AV1_COMMON *cm, RD_OPT *rd) { int i, bsize, segment_id; for (segment_id = 0; segment_id < MAX_SEGMENTS; ++segment_id) { const int qindex = clamp(av1_get_qindex(&cm->seg, segment_id, cm->base_qindex) + cm->y_dc_delta_q, 0, MAXQ); const int q = compute_rd_thresh_factor(qindex, cm->bit_depth); for (bsize = 0; bsize < BLOCK_SIZES_ALL; ++bsize) { // Threshold here seems unnecessarily harsh but fine given actual // range of values used for cpi->sf.thresh_mult[]. const int t = q * rd_thresh_block_size_factor[bsize]; const int thresh_max = INT_MAX / t; for (i = 0; i < MAX_MODES; ++i) rd->threshes[segment_id][bsize][i] = rd->thresh_mult[i] < thresh_max ? rd->thresh_mult[i] * t / 4 : INT_MAX; } } } void av1_set_mvcost(MACROBLOCK *x, int ref, int ref_mv_idx) { (void)ref; (void)ref_mv_idx; x->mvcost = x->mv_cost_stack; x->nmvjointcost = x->nmv_vec_cost; } void av1_fill_coeff_costs(MACROBLOCK *x, FRAME_CONTEXT *fc, const int num_planes) { const int nplanes = AOMMIN(num_planes, PLANE_TYPES); for (int eob_multi_size = 0; eob_multi_size < 7; ++eob_multi_size) { for (int plane = 0; plane < nplanes; ++plane) { LV_MAP_EOB_COST *pcost = &x->eob_costs[eob_multi_size][plane]; for (int ctx = 0; ctx < 2; ++ctx) { aom_cdf_prob *pcdf; switch (eob_multi_size) { case 0: pcdf = fc->eob_flag_cdf16[plane][ctx]; break; case 1: pcdf = fc->eob_flag_cdf32[plane][ctx]; break; case 2: pcdf = fc->eob_flag_cdf64[plane][ctx]; break; case 3: pcdf = fc->eob_flag_cdf128[plane][ctx]; break; case 4: pcdf = fc->eob_flag_cdf256[plane][ctx]; break; case 5: pcdf = fc->eob_flag_cdf512[plane][ctx]; break; case 6: default: pcdf = fc->eob_flag_cdf1024[plane][ctx]; break; } av1_cost_tokens_from_cdf(pcost->eob_cost[ctx], pcdf, NULL); } } } for (int tx_size = 0; tx_size < TX_SIZES; ++tx_size) { for (int plane = 0; plane < nplanes; ++plane) { LV_MAP_COEFF_COST *pcost = &x->coeff_costs[tx_size][plane]; for (int ctx = 0; ctx < TXB_SKIP_CONTEXTS; ++ctx) av1_cost_tokens_from_cdf(pcost->txb_skip_cost[ctx], fc->txb_skip_cdf[tx_size][ctx], NULL); for (int ctx = 0; ctx < SIG_COEF_CONTEXTS_EOB; ++ctx) av1_cost_tokens_from_cdf(pcost->base_eob_cost[ctx], fc->coeff_base_eob_cdf[tx_size][plane][ctx], NULL); for (int ctx = 0; ctx < SIG_COEF_CONTEXTS; ++ctx) av1_cost_tokens_from_cdf(pcost->base_cost[ctx], fc->coeff_base_cdf[tx_size][plane][ctx], NULL); for (int ctx = 0; ctx < EOB_COEF_CONTEXTS; ++ctx) av1_cost_tokens_from_cdf(pcost->eob_extra_cost[ctx], fc->eob_extra_cdf[tx_size][plane][ctx], NULL); for (int ctx = 0; ctx < DC_SIGN_CONTEXTS; ++ctx) av1_cost_tokens_from_cdf(pcost->dc_sign_cost[ctx], fc->dc_sign_cdf[plane][ctx], NULL); for (int ctx = 0; ctx < LEVEL_CONTEXTS; ++ctx) { int br_rate[BR_CDF_SIZE]; int prev_cost = 0; int i, j; av1_cost_tokens_from_cdf(br_rate, fc->coeff_br_cdf[tx_size][plane][ctx], NULL); // printf("br_rate: "); // for(j = 0; j < BR_CDF_SIZE; j++) // printf("%4d ", br_rate[j]); // printf("\n"); for (i = 0; i < COEFF_BASE_RANGE; i += BR_CDF_SIZE - 1) { for (j = 0; j < BR_CDF_SIZE - 1; j++) { pcost->lps_cost[ctx][i + j] = prev_cost + br_rate[j]; } prev_cost += br_rate[j]; } pcost->lps_cost[ctx][i] = prev_cost; // printf("lps_cost: %d %d %2d : ", tx_size, plane, ctx); // for (i = 0; i <= COEFF_BASE_RANGE; i++) // printf("%5d ", pcost->lps_cost[ctx][i]); // printf("\n"); } } } } void av1_initialize_rd_consts(AV1_COMP *cpi) { AV1_COMMON *const cm = &cpi->common; MACROBLOCK *const x = &cpi->td.mb; RD_OPT *const rd = &cpi->rd; aom_clear_system_state(); rd->RDMULT = av1_compute_rd_mult(cpi, cm->base_qindex + cm->y_dc_delta_q); set_error_per_bit(x, rd->RDMULT); set_block_thresholds(cm, rd); if (cm->cur_frame_force_integer_mv) { av1_build_nmv_cost_table(x->nmv_vec_cost, x->nmvcost, &cm->fc->nmvc, MV_SUBPEL_NONE); } else { av1_build_nmv_cost_table( x->nmv_vec_cost, cm->allow_high_precision_mv ? x->nmvcost_hp : x->nmvcost, &cm->fc->nmvc, cm->allow_high_precision_mv); } x->mvcost = x->mv_cost_stack; x->nmvjointcost = x->nmv_vec_cost; if (frame_is_intra_only(cm) && cm->allow_screen_content_tools && cpi->oxcf.pass != 1) { int *dvcost[2] = { &cpi->dv_cost[0][MV_MAX], &cpi->dv_cost[1][MV_MAX] }; av1_build_nmv_cost_table(cpi->dv_joint_cost, dvcost, &cm->fc->ndvc, MV_SUBPEL_NONE); } if (cpi->oxcf.pass != 1) { for (int i = 0; i < TRANS_TYPES; ++i) // IDENTITY: 1 bit // TRANSLATION: 3 bits // ROTZOOM: 2 bits // AFFINE: 3 bits cpi->gmtype_cost[i] = (1 + (i > 0 ? (i == ROTZOOM ? 1 : 2) : 0)) << AV1_PROB_COST_SHIFT; } } static void model_rd_norm(int xsq_q10, int *r_q10, int *d_q10) { // NOTE: The tables below must be of the same size. // The functions described below are sampled at the four most significant // bits of x^2 + 8 / 256. // Normalized rate: // This table models the rate for a Laplacian source with given variance // when quantized with a uniform quantizer with given stepsize. The // closed form expression is: // Rn(x) = H(sqrt(r)) + sqrt(r)*[1 + H(r)/(1 - r)], // where r = exp(-sqrt(2) * x) and x = qpstep / sqrt(variance), // and H(x) is the binary entropy function. static const int rate_tab_q10[] = { 65536, 6086, 5574, 5275, 5063, 4899, 4764, 4651, 4553, 4389, 4255, 4142, 4044, 3958, 3881, 3811, 3748, 3635, 3538, 3453, 3376, 3307, 3244, 3186, 3133, 3037, 2952, 2877, 2809, 2747, 2690, 2638, 2589, 2501, 2423, 2353, 2290, 2232, 2179, 2130, 2084, 2001, 1928, 1862, 1802, 1748, 1698, 1651, 1608, 1530, 1460, 1398, 1342, 1290, 1243, 1199, 1159, 1086, 1021, 963, 911, 864, 821, 781, 745, 680, 623, 574, 530, 490, 455, 424, 395, 345, 304, 269, 239, 213, 190, 171, 154, 126, 104, 87, 73, 61, 52, 44, 38, 28, 21, 16, 12, 10, 8, 6, 5, 3, 2, 1, 1, 1, 0, 0, }; // Normalized distortion: // This table models the normalized distortion for a Laplacian source // with given variance when quantized with a uniform quantizer // with given stepsize. The closed form expression is: // Dn(x) = 1 - 1/sqrt(2) * x / sinh(x/sqrt(2)) // where x = qpstep / sqrt(variance). // Note the actual distortion is Dn * variance. static const int dist_tab_q10[] = { 0, 0, 1, 1, 1, 2, 2, 2, 3, 3, 4, 5, 5, 6, 7, 7, 8, 9, 11, 12, 13, 15, 16, 17, 18, 21, 24, 26, 29, 31, 34, 36, 39, 44, 49, 54, 59, 64, 69, 73, 78, 88, 97, 106, 115, 124, 133, 142, 151, 167, 184, 200, 215, 231, 245, 260, 274, 301, 327, 351, 375, 397, 418, 439, 458, 495, 528, 559, 587, 613, 637, 659, 680, 717, 749, 777, 801, 823, 842, 859, 874, 899, 919, 936, 949, 960, 969, 977, 983, 994, 1001, 1006, 1010, 1013, 1015, 1017, 1018, 1020, 1022, 1022, 1023, 1023, 1023, 1024, }; static const int xsq_iq_q10[] = { 0, 4, 8, 12, 16, 20, 24, 28, 32, 40, 48, 56, 64, 72, 80, 88, 96, 112, 128, 144, 160, 176, 192, 208, 224, 256, 288, 320, 352, 384, 416, 448, 480, 544, 608, 672, 736, 800, 864, 928, 992, 1120, 1248, 1376, 1504, 1632, 1760, 1888, 2016, 2272, 2528, 2784, 3040, 3296, 3552, 3808, 4064, 4576, 5088, 5600, 6112, 6624, 7136, 7648, 8160, 9184, 10208, 11232, 12256, 13280, 14304, 15328, 16352, 18400, 20448, 22496, 24544, 26592, 28640, 30688, 32736, 36832, 40928, 45024, 49120, 53216, 57312, 61408, 65504, 73696, 81888, 90080, 98272, 106464, 114656, 122848, 131040, 147424, 163808, 180192, 196576, 212960, 229344, 245728, }; const int tmp = (xsq_q10 >> 2) + 8; const int k = get_msb(tmp) - 3; const int xq = (k << 3) + ((tmp >> k) & 0x7); const int one_q10 = 1 << 10; const int a_q10 = ((xsq_q10 - xsq_iq_q10[xq]) << 10) >> (2 + k); const int b_q10 = one_q10 - a_q10; *r_q10 = (rate_tab_q10[xq] * b_q10 + rate_tab_q10[xq + 1] * a_q10) >> 10; *d_q10 = (dist_tab_q10[xq] * b_q10 + dist_tab_q10[xq + 1] * a_q10) >> 10; } void av1_model_rd_from_var_lapndz(int64_t var, unsigned int n_log2, unsigned int qstep, int *rate, int64_t *dist) { // This function models the rate and distortion for a Laplacian // source with given variance when quantized with a uniform quantizer // with given stepsize. The closed form expressions are in: // Hang and Chen, "Source Model for transform video coder and its // application - Part I: Fundamental Theory", IEEE Trans. Circ. // Sys. for Video Tech., April 1997. if (var == 0) { *rate = 0; *dist = 0; } else { int d_q10, r_q10; static const uint32_t MAX_XSQ_Q10 = 245727; const uint64_t xsq_q10_64 = (((uint64_t)qstep * qstep << (n_log2 + 10)) + (var >> 1)) / var; const int xsq_q10 = (int)AOMMIN(xsq_q10_64, MAX_XSQ_Q10); model_rd_norm(xsq_q10, &r_q10, &d_q10); *rate = ROUND_POWER_OF_TWO(r_q10 << n_log2, 10 - AV1_PROB_COST_SHIFT); *dist = (var * (int64_t)d_q10 + 512) >> 10; } } static void get_entropy_contexts_plane(BLOCK_SIZE plane_bsize, const struct macroblockd_plane *pd, ENTROPY_CONTEXT t_above[MAX_MIB_SIZE], ENTROPY_CONTEXT t_left[MAX_MIB_SIZE]) { const int num_4x4_w = block_size_wide[plane_bsize] >> tx_size_wide_log2[0]; const int num_4x4_h = block_size_high[plane_bsize] >> tx_size_high_log2[0]; const ENTROPY_CONTEXT *const above = pd->above_context; const ENTROPY_CONTEXT *const left = pd->left_context; memcpy(t_above, above, sizeof(ENTROPY_CONTEXT) * num_4x4_w); memcpy(t_left, left, sizeof(ENTROPY_CONTEXT) * num_4x4_h); } void av1_get_entropy_contexts(BLOCK_SIZE bsize, const struct macroblockd_plane *pd, ENTROPY_CONTEXT t_above[MAX_MIB_SIZE], ENTROPY_CONTEXT t_left[MAX_MIB_SIZE]) { const BLOCK_SIZE plane_bsize = get_plane_block_size(bsize, pd->subsampling_x, pd->subsampling_y); get_entropy_contexts_plane(plane_bsize, pd, t_above, t_left); } void av1_mv_pred(const AV1_COMP *cpi, MACROBLOCK *x, uint8_t *ref_y_buffer, int ref_y_stride, int ref_frame, BLOCK_SIZE block_size) { int i; int zero_seen = 0; int best_sad = INT_MAX; int this_sad = INT_MAX; int max_mv = 0; uint8_t *src_y_ptr = x->plane[0].src.buf; uint8_t *ref_y_ptr; MV pred_mv[MAX_MV_REF_CANDIDATES + 1]; int num_mv_refs = 0; const MV_REFERENCE_FRAME ref_frames[2] = { ref_frame, NONE_FRAME }; const int_mv ref_mv = av1_get_ref_mv_from_stack(0, ref_frames, 0, x->mbmi_ext); const int_mv ref_mv1 = av1_get_ref_mv_from_stack(0, ref_frames, 1, x->mbmi_ext); pred_mv[num_mv_refs++] = ref_mv.as_mv; if (ref_mv.as_int != ref_mv1.as_int) { pred_mv[num_mv_refs++] = ref_mv1.as_mv; } if (cpi->sf.adaptive_motion_search && block_size < x->max_partition_size) pred_mv[num_mv_refs++] = x->pred_mv[ref_frame]; assert(num_mv_refs <= (int)(sizeof(pred_mv) / sizeof(pred_mv[0]))); // Get the sad for each candidate reference mv. for (i = 0; i < num_mv_refs; ++i) { const MV *this_mv = &pred_mv[i]; int fp_row, fp_col; fp_row = (this_mv->row + 3 + (this_mv->row >= 0)) >> 3; fp_col = (this_mv->col + 3 + (this_mv->col >= 0)) >> 3; max_mv = AOMMAX(max_mv, AOMMAX(abs(this_mv->row), abs(this_mv->col)) >> 3); if (fp_row == 0 && fp_col == 0 && zero_seen) continue; zero_seen |= (fp_row == 0 && fp_col == 0); ref_y_ptr = &ref_y_buffer[ref_y_stride * fp_row + fp_col]; // Find sad for current vector. this_sad = cpi->fn_ptr[block_size].sdf(src_y_ptr, x->plane[0].src.stride, ref_y_ptr, ref_y_stride); // Note if it is the best so far. if (this_sad < best_sad) { best_sad = this_sad; } } // Note the index of the mv that worked best in the reference list. x->max_mv_context[ref_frame] = max_mv; x->pred_mv_sad[ref_frame] = best_sad; } void av1_setup_pred_block(const MACROBLOCKD *xd, struct buf_2d dst[MAX_MB_PLANE], const YV12_BUFFER_CONFIG *src, int mi_row, int mi_col, const struct scale_factors *scale, const struct scale_factors *scale_uv, const int num_planes) { int i; dst[0].buf = src->y_buffer; dst[0].stride = src->y_stride; dst[1].buf = src->u_buffer; dst[2].buf = src->v_buffer; dst[1].stride = dst[2].stride = src->uv_stride; for (i = 0; i < num_planes; ++i) { setup_pred_plane(dst + i, xd->mi[0]->sb_type, dst[i].buf, i ? src->uv_crop_width : src->y_crop_width, i ? src->uv_crop_height : src->y_crop_height, dst[i].stride, mi_row, mi_col, i ? scale_uv : scale, xd->plane[i].subsampling_x, xd->plane[i].subsampling_y); } } int av1_raster_block_offset(BLOCK_SIZE plane_bsize, int raster_block, int stride) { const int bw = mi_size_wide_log2[plane_bsize]; const int y = 4 * (raster_block >> bw); const int x = 4 * (raster_block & ((1 << bw) - 1)); return y * stride + x; } int16_t *av1_raster_block_offset_int16(BLOCK_SIZE plane_bsize, int raster_block, int16_t *base) { const int stride = block_size_wide[plane_bsize]; return base + av1_raster_block_offset(plane_bsize, raster_block, stride); } YV12_BUFFER_CONFIG *av1_get_scaled_ref_frame(const AV1_COMP *cpi, int ref_frame) { const AV1_COMMON *const cm = &cpi->common; const int scaled_idx = cpi->scaled_ref_idx[ref_frame - 1]; const int ref_idx = get_ref_frame_buf_idx(cpi, ref_frame); return (scaled_idx != ref_idx && scaled_idx != INVALID_IDX) ? &cm->buffer_pool->frame_bufs[scaled_idx].buf : NULL; } int av1_get_switchable_rate(const AV1_COMMON *const cm, MACROBLOCK *x, const MACROBLOCKD *xd) { if (cm->interp_filter == SWITCHABLE) { const MB_MODE_INFO *const mbmi = xd->mi[0]; int inter_filter_cost = 0; int dir; for (dir = 0; dir < 2; ++dir) { const int ctx = av1_get_pred_context_switchable_interp(xd, dir); const InterpFilter filter = av1_extract_interp_filter(mbmi->interp_filters, dir); inter_filter_cost += x->switchable_interp_costs[ctx][filter]; } return SWITCHABLE_INTERP_RATE_FACTOR * inter_filter_cost; } else { return 0; } } void av1_set_rd_speed_thresholds(AV1_COMP *cpi) { int i; RD_OPT *const rd = &cpi->rd; SPEED_FEATURES *const sf = &cpi->sf; // Set baseline threshold values. for (i = 0; i < MAX_MODES; ++i) rd->thresh_mult[i] = cpi->oxcf.mode == 0; if (sf->adaptive_rd_thresh) { rd->thresh_mult[THR_NEARESTMV] = 300; rd->thresh_mult[THR_NEARESTL2] = 300; rd->thresh_mult[THR_NEARESTL3] = 300; rd->thresh_mult[THR_NEARESTB] = 300; rd->thresh_mult[THR_NEARESTA2] = 300; rd->thresh_mult[THR_NEARESTA] = 300; rd->thresh_mult[THR_NEARESTG] = 300; } else { rd->thresh_mult[THR_NEARESTMV] = 0; rd->thresh_mult[THR_NEARESTL2] = 0; rd->thresh_mult[THR_NEARESTL3] = 0; rd->thresh_mult[THR_NEARESTB] = 0; rd->thresh_mult[THR_NEARESTA2] = 0; rd->thresh_mult[THR_NEARESTA] = 0; rd->thresh_mult[THR_NEARESTG] = 0; } rd->thresh_mult[THR_DC] += 1000; rd->thresh_mult[THR_NEWMV] += 1000; rd->thresh_mult[THR_NEWL2] += 1000; rd->thresh_mult[THR_NEWL3] += 1000; rd->thresh_mult[THR_NEWB] += 1000; rd->thresh_mult[THR_NEWA2] = 1000; rd->thresh_mult[THR_NEWA] += 1000; rd->thresh_mult[THR_NEWG] += 1000; rd->thresh_mult[THR_NEARMV] += 1000; rd->thresh_mult[THR_NEARL2] += 1000; rd->thresh_mult[THR_NEARL3] += 1000; rd->thresh_mult[THR_NEARB] += 1000; rd->thresh_mult[THR_NEARA2] = 1000; rd->thresh_mult[THR_NEARA] += 1000; rd->thresh_mult[THR_NEARG] += 1000; rd->thresh_mult[THR_GLOBALMV] += 2000; rd->thresh_mult[THR_GLOBALL2] += 2000; rd->thresh_mult[THR_GLOBALL3] += 2000; rd->thresh_mult[THR_GLOBALB] += 2000; rd->thresh_mult[THR_GLOBALA2] = 2000; rd->thresh_mult[THR_GLOBALG] += 2000; rd->thresh_mult[THR_GLOBALA] += 2000; rd->thresh_mult[THR_PAETH] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTLA] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTL2A] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTL3A] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTGA] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTLB] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTL2B] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTL3B] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTGB] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTLA2] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTL2A2] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTL3A2] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTGA2] += 1000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTLL2] += 2000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTLL3] += 2000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTLG] += 2000; rd->thresh_mult[THR_COMP_NEAREST_NEARESTBA] += 2000; rd->thresh_mult[THR_COMP_NEAR_NEARLA] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWLA] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTLA] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWLA] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARLA] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWLA] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALLA] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARL2A] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWL2A] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTL2A] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWL2A] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARL2A] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWL2A] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALL2A] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARL3A] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWL3A] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTL3A] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWL3A] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARL3A] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWL3A] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALL3A] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARGA] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWGA] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTGA] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWGA] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARGA] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWGA] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALGA] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARLB] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWLB] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTLB] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWLB] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARLB] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWLB] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALLB] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARL2B] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWL2B] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTL2B] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWL2B] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARL2B] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWL2B] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALL2B] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARL3B] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWL3B] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTL3B] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWL3B] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARL3B] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWL3B] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALL3B] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARGB] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWGB] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTGB] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWGB] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARGB] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWGB] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALGB] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARLA2] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWLA2] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTLA2] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWLA2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARLA2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWLA2] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALLA2] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARL2A2] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWL2A2] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTL2A2] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWL2A2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARL2A2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWL2A2] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALL2A2] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARL3A2] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWL3A2] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTL3A2] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWL3A2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARL3A2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWL3A2] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALL3A2] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARGA2] += 1200; rd->thresh_mult[THR_COMP_NEAREST_NEWGA2] += 1500; rd->thresh_mult[THR_COMP_NEW_NEARESTGA2] += 1500; rd->thresh_mult[THR_COMP_NEAR_NEWGA2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEARGA2] += 1700; rd->thresh_mult[THR_COMP_NEW_NEWGA2] += 2000; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALGA2] += 2500; rd->thresh_mult[THR_H_PRED] += 2000; rd->thresh_mult[THR_V_PRED] += 2000; rd->thresh_mult[THR_D135_PRED] += 2500; rd->thresh_mult[THR_D203_PRED] += 2500; rd->thresh_mult[THR_D157_PRED] += 2500; rd->thresh_mult[THR_D67_PRED] += 2500; rd->thresh_mult[THR_D113_PRED] += 2500; rd->thresh_mult[THR_D45_PRED] += 2500; rd->thresh_mult[THR_COMP_NEAR_NEARLL2] += 1600; rd->thresh_mult[THR_COMP_NEAREST_NEWLL2] += 2000; rd->thresh_mult[THR_COMP_NEW_NEARESTLL2] += 2000; rd->thresh_mult[THR_COMP_NEAR_NEWLL2] += 2200; rd->thresh_mult[THR_COMP_NEW_NEARLL2] += 2200; rd->thresh_mult[THR_COMP_NEW_NEWLL2] += 2400; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALLL2] += 3200; rd->thresh_mult[THR_COMP_NEAR_NEARLL3] += 1600; rd->thresh_mult[THR_COMP_NEAREST_NEWLL3] += 2000; rd->thresh_mult[THR_COMP_NEW_NEARESTLL3] += 2000; rd->thresh_mult[THR_COMP_NEAR_NEWLL3] += 2200; rd->thresh_mult[THR_COMP_NEW_NEARLL3] += 2200; rd->thresh_mult[THR_COMP_NEW_NEWLL3] += 2400; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALLL3] += 3200; rd->thresh_mult[THR_COMP_NEAR_NEARLG] += 1600; rd->thresh_mult[THR_COMP_NEAREST_NEWLG] += 2000; rd->thresh_mult[THR_COMP_NEW_NEARESTLG] += 2000; rd->thresh_mult[THR_COMP_NEAR_NEWLG] += 2200; rd->thresh_mult[THR_COMP_NEW_NEARLG] += 2200; rd->thresh_mult[THR_COMP_NEW_NEWLG] += 2400; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALLG] += 3200; rd->thresh_mult[THR_COMP_NEAR_NEARBA] += 1600; rd->thresh_mult[THR_COMP_NEAREST_NEWBA] += 2000; rd->thresh_mult[THR_COMP_NEW_NEARESTBA] += 2000; rd->thresh_mult[THR_COMP_NEAR_NEWBA] += 2200; rd->thresh_mult[THR_COMP_NEW_NEARBA] += 2200; rd->thresh_mult[THR_COMP_NEW_NEWBA] += 2400; rd->thresh_mult[THR_COMP_GLOBAL_GLOBALBA] += 3200; } void av1_set_rd_speed_thresholds_sub8x8(AV1_COMP *cpi) { static const int thresh_mult[MAX_REFS] = { 2500, 2500, 2500, 2500, 2500, 2500, 2500, 4500, 4500, 4500, 4500, 4500, 4500, 4500, 4500, 4500, 4500, 4500, 4500, 2500 }; RD_OPT *const rd = &cpi->rd; memcpy(rd->thresh_mult_sub8x8, thresh_mult, sizeof(thresh_mult)); } void av1_update_rd_thresh_fact(const AV1_COMMON *const cm, int (*factor_buf)[MAX_MODES], int rd_thresh, int bsize, int best_mode_index) { if (rd_thresh > 0) { const int top_mode = MAX_MODES; int mode; for (mode = 0; mode < top_mode; ++mode) { const BLOCK_SIZE min_size = AOMMAX(bsize - 1, BLOCK_4X4); const BLOCK_SIZE max_size = AOMMIN(bsize + 2, (int)cm->seq_params.sb_size); BLOCK_SIZE bs; for (bs = min_size; bs <= max_size; ++bs) { int *const fact = &factor_buf[bs][mode]; if (mode == best_mode_index) { *fact -= (*fact >> 4); } else { *fact = AOMMIN(*fact + RD_THRESH_INC, rd_thresh * RD_THRESH_MAX_FACT); } } } } } int av1_get_intra_cost_penalty(int qindex, int qdelta, aom_bit_depth_t bit_depth) { const int q = av1_dc_quant_Q3(qindex, qdelta, bit_depth); switch (bit_depth) { case AOM_BITS_8: return 20 * q; case AOM_BITS_10: return 5 * q; case AOM_BITS_12: return ROUND_POWER_OF_TWO(5 * q, 2); default: assert(0 && "bit_depth should be AOM_BITS_8, AOM_BITS_10 or AOM_BITS_12"); return -1; } }