/* * Copyright (c) 2014 The WebM project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "vp9/encoder/vp9_encodeframe.h" #include "vp9/encoder/vp9_encoder.h" #include "vp9/encoder/vp9_ethread.h" static void accumulate_frame_counts(VP9_COMMON *cm, ThreadData *td) { int i, j, k, l, m; for (i = 0; i < BLOCK_SIZE_GROUPS; i++) for (j = 0; j < INTRA_MODES; j++) cm->counts.y_mode[i][j] += td->counts->y_mode[i][j]; for (i = 0; i < INTRA_MODES; i++) for (j = 0; j < INTRA_MODES; j++) cm->counts.uv_mode[i][j] += td->counts->uv_mode[i][j]; for (i = 0; i < PARTITION_CONTEXTS; i++) for (j = 0; j < PARTITION_TYPES; j++) cm->counts.partition[i][j] += td->counts->partition[i][j]; for (i = 0; i < TX_SIZES; i++) for (j = 0; j < PLANE_TYPES; j++) for (k = 0; k < REF_TYPES; k++) for (l = 0; l < COEF_BANDS; l++) for (m = 0; m < COEFF_CONTEXTS; m++) cm->counts.eob_branch[i][j][k][l][m] += td->counts->eob_branch[i][j][k][l][m]; // cm->counts.coef is only updated at frame level, so not need // to accumulate it here. // for (n = 0; n < UNCONSTRAINED_NODES + 1; n++) // cm->counts.coef[i][j][k][l][m][n] += // td->counts->coef[i][j][k][l][m][n]; for (i = 0; i < SWITCHABLE_FILTER_CONTEXTS; i++) for (j = 0; j < SWITCHABLE_FILTERS; j++) cm->counts.switchable_interp[i][j] += td->counts->switchable_interp[i][j]; for (i = 0; i < INTER_MODE_CONTEXTS; i++) for (j = 0; j < INTER_MODES; j++) cm->counts.inter_mode[i][j] += td->counts->inter_mode[i][j]; for (i = 0; i < INTRA_INTER_CONTEXTS; i++) for (j = 0; j < 2; j++) cm->counts.intra_inter[i][j] += td->counts->intra_inter[i][j]; for (i = 0; i < COMP_INTER_CONTEXTS; i++) for (j = 0; j < 2; j++) cm->counts.comp_inter[i][j] += td->counts->comp_inter[i][j]; for (i = 0; i < REF_CONTEXTS; i++) for (j = 0; j < 2; j++) for (k = 0; k < 2; k++) cm->counts.single_ref[i][j][k] += td->counts->single_ref[i][j][k]; for (i = 0; i < REF_CONTEXTS; i++) for (j = 0; j < 2; j++) cm->counts.comp_ref[i][j] += td->counts->comp_ref[i][j]; for (i = 0; i < TX_SIZE_CONTEXTS; i++) { for (j = 0; j < TX_SIZES; j++) cm->counts.tx.p32x32[i][j] += td->counts->tx.p32x32[i][j]; for (j = 0; j < TX_SIZES - 1; j++) cm->counts.tx.p16x16[i][j] += td->counts->tx.p16x16[i][j]; for (j = 0; j < TX_SIZES - 2; j++) cm->counts.tx.p8x8[i][j] += td->counts->tx.p8x8[i][j]; } for (i = 0; i < SKIP_CONTEXTS; i++) for (j = 0; j < 2; j++) cm->counts.skip[i][j] += td->counts->skip[i][j]; for (i = 0; i < MV_JOINTS; i++) cm->counts.mv.joints[i] += td->counts->mv.joints[i]; for (k = 0; k < 2; k++) { nmv_component_counts *comps = &cm->counts.mv.comps[k]; nmv_component_counts *comps_t = &td->counts->mv.comps[k]; for (i = 0; i < 2; i++) { comps->sign[i] += comps_t->sign[i]; comps->class0_hp[i] += comps_t->class0_hp[i]; comps->hp[i] += comps_t->hp[i]; } for (i = 0; i < MV_CLASSES; i++) comps->classes[i] += comps_t->classes[i]; for (i = 0; i < CLASS0_SIZE; i++) { comps->class0[i] += comps_t->class0[i]; for (j = 0; j < MV_FP_SIZE; j++) comps->class0_fp[i][j] += comps_t->class0_fp[i][j]; } for (i = 0; i < MV_OFFSET_BITS; i++) for (j = 0; j < 2; j++) comps->bits[i][j] += comps_t->bits[i][j]; for (i = 0; i < MV_FP_SIZE; i++) comps->fp[i] += comps_t->fp[i]; } } static void accumulate_rd_opt(ThreadData *td, ThreadData *td_t) { int i, j, k, l, m, n; for (i = 0; i < REFERENCE_MODES; i++) td->rd_counts.comp_pred_diff[i] += td_t->rd_counts.comp_pred_diff[i]; for (i = 0; i < SWITCHABLE_FILTER_CONTEXTS; i++) td->rd_counts.filter_diff[i] += td_t->rd_counts.filter_diff[i]; for (i = 0; i < TX_MODES; i++) td->rd_counts.tx_select_diff[i] += td_t->rd_counts.tx_select_diff[i]; for (i = 0; i < TX_SIZES; i++) for (j = 0; j < PLANE_TYPES; j++) for (k = 0; k < REF_TYPES; k++) for (l = 0; l < COEF_BANDS; l++) for (m = 0; m < COEFF_CONTEXTS; m++) for (n = 0; n < ENTROPY_TOKENS; n++) td->rd_counts.coef_counts[i][j][k][l][m][n] += td_t->rd_counts.coef_counts[i][j][k][l][m][n]; } static int enc_worker_hook(EncWorkerData *const thread_data, void *unused) { VP9_COMP *const cpi = thread_data->cpi; const VP9_COMMON *const cm = &cpi->common; const int tile_cols = 1 << cm->log2_tile_cols; const int tile_rows = 1 << cm->log2_tile_rows; int t; (void) unused; for (t = thread_data->start; t < tile_rows * tile_cols; t += cpi->num_workers) { int tile_row = t / tile_cols; int tile_col = t % tile_cols; vp9_encode_tile(cpi, thread_data->td, tile_row, tile_col); } return 0; } void vp9_encode_tiles_mt(VP9_COMP *cpi) { VP9_COMMON *const cm = &cpi->common; const int tile_cols = 1 << cm->log2_tile_cols; const VP9WorkerInterface *const winterface = vp9_get_worker_interface(); const int num_workers = MIN(cpi->oxcf.max_threads, tile_cols); int i; vp9_init_tile_data(cpi); // Only run once to create threads and allocate thread data. if (cpi->num_workers == 0) { CHECK_MEM_ERROR(cm, cpi->workers, vpx_malloc(num_workers * sizeof(*cpi->workers))); for (i = 0; i < num_workers; i++) { VP9Worker *const worker = &cpi->workers[i]; EncWorkerData *thread_data; ++cpi->num_workers; winterface->init(worker); CHECK_MEM_ERROR(cm, worker->data1, (EncWorkerData*)vpx_calloc(1, sizeof(EncWorkerData))); thread_data = (EncWorkerData*)worker->data1; if (i < num_workers - 1) { thread_data->cpi = cpi; // Allocate thread data. CHECK_MEM_ERROR(cm, thread_data->td, vpx_calloc(1, sizeof(*thread_data->td))); // Set up pc_tree. thread_data->td->leaf_tree = NULL; thread_data->td->pc_tree = NULL; vp9_setup_pc_tree(cm, thread_data->td); // Allocate frame counters in thread data. CHECK_MEM_ERROR(cm, thread_data->td->counts, vpx_calloc(1, sizeof(*thread_data->td->counts))); // Create threads if (!winterface->reset(worker)) vpx_internal_error(&cm->error, VPX_CODEC_ERROR, "Tile encoder thread creation failed"); } else { // Main thread acts as a worker and uses the thread data in cpi. thread_data->cpi = cpi; thread_data->td = &cpi->td; } // data2 is unused. worker->data2 = NULL; winterface->sync(worker); worker->hook = (VP9WorkerHook)enc_worker_hook; } } for (i = 0; i < num_workers; i++) { VP9Worker *const worker = &cpi->workers[i]; EncWorkerData *const thread_data = (EncWorkerData*)worker->data1; // Before encoding a frame, copy the thread data from cpi. thread_data->td->mb = cpi->td.mb; thread_data->td->rd_counts = cpi->td.rd_counts; vpx_memcpy(thread_data->td->counts, &cpi->common.counts, sizeof(cpi->common.counts)); // Handle use_nonrd_pick_mode case. if (cpi->sf.use_nonrd_pick_mode) { MACROBLOCK *const x = &thread_data->td->mb; MACROBLOCKD *const xd = &x->e_mbd; struct macroblock_plane *const p = x->plane; struct macroblockd_plane *const pd = xd->plane; PICK_MODE_CONTEXT *ctx = &thread_data->td->pc_root->none; int j; for (j = 0; j < MAX_MB_PLANE; ++j) { p[j].coeff = ctx->coeff_pbuf[j][0]; p[j].qcoeff = ctx->qcoeff_pbuf[j][0]; pd[j].dqcoeff = ctx->dqcoeff_pbuf[j][0]; p[j].eobs = ctx->eobs_pbuf[j][0]; } } } // Encode a frame for (i = 0; i < num_workers; i++) { VP9Worker *const worker = &cpi->workers[i]; EncWorkerData *const thread_data = (EncWorkerData*)worker->data1; // Set the starting tile for each thread. thread_data->start = i; if (i == num_workers - 1) winterface->execute(worker); else winterface->launch(worker); } // Encoding ends. for (i = 0; i < num_workers; i++) { VP9Worker *const worker = &cpi->workers[i]; winterface->sync(worker); } for (i = 0; i < num_workers; i++) { VP9Worker *const worker = &cpi->workers[i]; EncWorkerData *const thread_data = (EncWorkerData*)worker->data1; // Accumulate counters. if (i < num_workers - 1) { accumulate_frame_counts(&cpi->common, thread_data->td); accumulate_rd_opt(&cpi->td, thread_data->td); } } }