/* * Copyright (c) 2010 The VP8 project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "mcomp.h" #include "vpx_mem/vpx_mem.h" #include #include #include #ifdef ENTROPY_STATS static int mv_ref_ct [31] [4] [2]; static int mv_mode_cts [4] [2]; #endif static int mv_bits_sadcost[256]; void vp8cx_init_mv_bits_sadcost() { int i; for (i = 0; i < 256; i++) { mv_bits_sadcost[i] = (int)sqrt(i * 16); } } int vp8_mv_bit_cost(MV *mv, MV *ref, int *mvcost[2], int Weight) { // MV costing is based on the distribution of vectors in the previous frame and as such will tend to // over state the cost of vectors. In addition coding a new vector can have a knock on effect on the // cost of subsequent vectors and the quality of prediction from NEAR and NEAREST for subsequent blocks. // The "Weight" parameter allows, to a limited extent, for some account to be taken of these factors. return ((mvcost[0][(mv->row - ref->row) >> 1] + mvcost[1][(mv->col - ref->col) >> 1]) * Weight) >> 7; } int vp8_mv_err_cost(MV *mv, MV *ref, int *mvcost[2], int error_per_bit) { //int i; //return ((mvcost[0][(mv->row - ref->row)>>1] + mvcost[1][(mv->col - ref->col)>>1] + 128) * error_per_bit) >> 8; //return ( (vp8_mv_bit_cost(mv, ref, mvcost, 100) + 128) * error_per_bit) >> 8; //i = (vp8_mv_bit_cost(mv, ref, mvcost, 100) * error_per_bit + 128) >> 8; return ((mvcost[0][(mv->row - ref->row) >> 1] + mvcost[1][(mv->col - ref->col) >> 1]) * error_per_bit + 128) >> 8; //return (vp8_mv_bit_cost(mv, ref, mvcost, 128) * error_per_bit + 128) >> 8; } static int mv_bits(MV *mv, MV *ref, int *mvcost[2]) { // get the estimated number of bits for a motion vector, to be used for costing in SAD based // motion estimation return ((mvcost[0][(mv->row - ref->row) >> 1] + mvcost[1][(mv->col - ref->col)>> 1]) + 128) >> 8; } void vp8_init_dsmotion_compensation(MACROBLOCK *x, int stride) { int Len; int search_site_count = 0; // Generate offsets for 4 search sites per step. Len = MAX_FIRST_STEP; x->ss[search_site_count].mv.col = 0; x->ss[search_site_count].mv.row = 0; x->ss[search_site_count].offset = 0; search_site_count++; while (Len > 0) { // Compute offsets for search sites. x->ss[search_site_count].mv.col = 0; x->ss[search_site_count].mv.row = -Len; x->ss[search_site_count].offset = -Len * stride; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = 0; x->ss[search_site_count].mv.row = Len; x->ss[search_site_count].offset = Len * stride; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = -Len; x->ss[search_site_count].mv.row = 0; x->ss[search_site_count].offset = -Len; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = Len; x->ss[search_site_count].mv.row = 0; x->ss[search_site_count].offset = Len; search_site_count++; // Contract. Len /= 2; } x->ss_count = search_site_count; x->searches_per_step = 4; } void vp8_init3smotion_compensation(MACROBLOCK *x, int stride) { int Len; int search_site_count = 0; // Generate offsets for 8 search sites per step. Len = MAX_FIRST_STEP; x->ss[search_site_count].mv.col = 0; x->ss[search_site_count].mv.row = 0; x->ss[search_site_count].offset = 0; search_site_count++; while (Len > 0) { // Compute offsets for search sites. x->ss[search_site_count].mv.col = 0; x->ss[search_site_count].mv.row = -Len; x->ss[search_site_count].offset = -Len * stride; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = 0; x->ss[search_site_count].mv.row = Len; x->ss[search_site_count].offset = Len * stride; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = -Len; x->ss[search_site_count].mv.row = 0; x->ss[search_site_count].offset = -Len; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = Len; x->ss[search_site_count].mv.row = 0; x->ss[search_site_count].offset = Len; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = -Len; x->ss[search_site_count].mv.row = -Len; x->ss[search_site_count].offset = -Len * stride - Len; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = Len; x->ss[search_site_count].mv.row = -Len; x->ss[search_site_count].offset = -Len * stride + Len; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = -Len; x->ss[search_site_count].mv.row = Len; x->ss[search_site_count].offset = Len * stride - Len; search_site_count++; // Compute offsets for search sites. x->ss[search_site_count].mv.col = Len; x->ss[search_site_count].mv.row = Len; x->ss[search_site_count].offset = Len * stride + Len; search_site_count++; // Contract. Len /= 2; } x->ss_count = search_site_count; x->searches_per_step = 8; } #define MVC(r,c) (((mvcost[0][(r)-rr] + mvcost[1][(c) - rc]) * error_per_bit + 128 )>>8 ) // estimated cost of a motion vector (r,c) #define PRE(r,c) (*(d->base_pre) + d->pre + ((r)>>2) * d->pre_stride + ((c)>>2)) // pointer to predictor base of a motionvector #define SP(x) (((x)&3)<<1) // convert motion vector component to offset for svf calc #define DIST(r,c) svf( PRE(r,c), d->pre_stride, SP(c),SP(r), z,b->src_stride,&sse) // returns subpixel variance error function. #define IFMVCV(r,c,s,e) if ( c >= minc && c <= maxc && r >= minr && r <= maxr) s else e; #define ERR(r,c) (MVC(r,c)+DIST(r,c)) // returns distortion + motion vector cost #define CHECK_BETTER(v,r,c) IFMVCV(r,c,{if((v = ERR(r,c)) < besterr) { besterr = v; br=r; bc=c; }}, v=INT_MAX;)// checks if (r,c) has better score than previous best #define MIN(x,y) (((x)<(y))?(x):(y)) #define MAX(x,y) (((x)>(y))?(x):(y)) //#define CHECK_BETTER(v,r,c) if((v = ERR(r,c)) < besterr) { besterr = v; br=r; bc=c; } int vp8_find_best_sub_pixel_step_iteratively(MACROBLOCK *x, BLOCK *b, BLOCKD *d, MV *bestmv, MV *ref_mv, int error_per_bit, vp8_subpixvariance_fn_t svf, vp8_variance_fn_t vf, int *mvcost[2]) { unsigned char *y = *(d->base_pre) + d->pre + (bestmv->row) * d->pre_stride + bestmv->col; unsigned char *z = (*(b->base_src) + b->src); int rr = ref_mv->row >> 1, rc = ref_mv->col >> 1; int br = bestmv->row << 2, bc = bestmv->col << 2; int tr = br, tc = bc; unsigned int besterr = INT_MAX; unsigned int left, right, up, down, diag; unsigned int sse; unsigned int whichdir; unsigned int halfiters = 4; unsigned int quarteriters = 4; int minc = MAX(x->mv_col_min << 2, (ref_mv->col >> 1) - ((1 << mvlong_width) - 1)); int maxc = MIN(x->mv_col_max << 2, (ref_mv->col >> 1) + ((1 << mvlong_width) - 1)); int minr = MAX(x->mv_row_min << 2, (ref_mv->row >> 1) - ((1 << mvlong_width) - 1)); int maxr = MIN(x->mv_row_max << 2, (ref_mv->row >> 1) + ((1 << mvlong_width) - 1)); // central mv bestmv->row <<= 3; bestmv->col <<= 3; // calculate central point error besterr = vf(y, d->pre_stride, z, b->src_stride, &sse); besterr += vp8_mv_err_cost(bestmv, ref_mv, mvcost, error_per_bit); // TODO: Each subsequent iteration checks at least one point in common with the last iteration could be 2 ( if diag selected) while (--halfiters) { // 1/2 pel CHECK_BETTER(left, tr, tc - 2); CHECK_BETTER(right, tr, tc + 2); CHECK_BETTER(up, tr - 2, tc); CHECK_BETTER(down, tr + 2, tc); whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); switch (whichdir) { case 0: CHECK_BETTER(diag, tr - 2, tc - 2); break; case 1: CHECK_BETTER(diag, tr - 2, tc + 2); break; case 2: CHECK_BETTER(diag, tr + 2, tc - 2); break; case 3: CHECK_BETTER(diag, tr + 2, tc + 2); break; } // no reason to check the same one again. if (tr == br && tc == bc) break; tr = br; tc = bc; } // TODO: Each subsequent iteration checks at least one point in common with the last iteration could be 2 ( if diag selected) // 1/4 pel while (--quarteriters) { CHECK_BETTER(left, tr, tc - 1); CHECK_BETTER(right, tr, tc + 1); CHECK_BETTER(up, tr - 1, tc); CHECK_BETTER(down, tr + 1, tc); whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); switch (whichdir) { case 0: CHECK_BETTER(diag, tr - 1, tc - 1); break; case 1: CHECK_BETTER(diag, tr - 1, tc + 1); break; case 2: CHECK_BETTER(diag, tr + 1, tc - 1); break; case 3: CHECK_BETTER(diag, tr + 1, tc + 1); break; } // no reason to check the same one again. if (tr == br && tc == bc) break; tr = br; tc = bc; } bestmv->row = br << 1; bestmv->col = bc << 1; if ((abs(bestmv->col - ref_mv->col) > MAX_FULL_PEL_VAL) || (abs(bestmv->row - ref_mv->row) > MAX_FULL_PEL_VAL)) return INT_MAX; return besterr; } #undef MVC #undef PRE #undef SP #undef DIST #undef ERR #undef CHECK_BETTER #undef MIN #undef MAX int vp8_find_best_sub_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d, MV *bestmv, MV *ref_mv, int error_per_bit, vp8_subpixvariance_fn_t svf, vp8_variance_fn_t vf, int *mvcost[2]) { int bestmse = INT_MAX; MV startmv; //MV this_mv; MV this_mv; unsigned char *y = *(d->base_pre) + d->pre + (bestmv->row) * d->pre_stride + bestmv->col; unsigned char *z = (*(b->base_src) + b->src); int left, right, up, down, diag; unsigned int sse; int whichdir ; // Trap uncodable vectors if ((abs((bestmv->col << 3) - ref_mv->col) > MAX_FULL_PEL_VAL) || (abs((bestmv->row << 3) - ref_mv->row) > MAX_FULL_PEL_VAL)) { bestmv->row <<= 3; bestmv->col <<= 3; return INT_MAX; } // central mv bestmv->row <<= 3; bestmv->col <<= 3; startmv = *bestmv; // calculate central point error bestmse = vf(y, d->pre_stride, z, b->src_stride, &sse); bestmse += vp8_mv_err_cost(bestmv, ref_mv, mvcost, error_per_bit); // go left then right and check error this_mv.row = startmv.row; this_mv.col = ((startmv.col - 8) | 4); left = svf(y - 1, d->pre_stride, 4, 0, z, b->src_stride, &sse); left += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (left < bestmse) { *bestmv = this_mv; bestmse = left; } this_mv.col += 8; right = svf(y, d->pre_stride, 4, 0, z, b->src_stride, &sse); right += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (right < bestmse) { *bestmv = this_mv; bestmse = right; } // go up then down and check error this_mv.col = startmv.col; this_mv.row = ((startmv.row - 8) | 4); up = svf(y - d->pre_stride, d->pre_stride, 0, 4, z, b->src_stride, &sse); up += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (up < bestmse) { *bestmv = this_mv; bestmse = up; } this_mv.row += 8; down = svf(y, d->pre_stride, 0, 4, z, b->src_stride, &sse); down += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (down < bestmse) { *bestmv = this_mv; bestmse = down; } // now check 1 more diagonal whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); // whichdir must be 0-4. Therefore, one of the cases below // must run through. However, because there is no default // and diag is not set elsewhere, we get a compile warning diag = 0; //for(whichdir =0;whichdir<4;whichdir++) //{ this_mv = startmv; switch (whichdir) { case 0: this_mv.col = (this_mv.col - 8) | 4; this_mv.row = (this_mv.row - 8) | 4; diag = svf(y - 1 - d->pre_stride, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; case 1: this_mv.col += 4; this_mv.row = (this_mv.row - 8) | 4; diag = svf(y - d->pre_stride, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; case 2: this_mv.col = (this_mv.col - 8) | 4; this_mv.row += 4; diag = svf(y - 1, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; case 3: this_mv.col += 4; this_mv.row += 4; diag = svf(y, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; } diag += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (diag < bestmse) { *bestmv = this_mv; bestmse = diag; } // } // time to check quarter pels. if (bestmv->row < startmv.row) y -= d->pre_stride; if (bestmv->col < startmv.col) y--; startmv = *bestmv; // go left then right and check error this_mv.row = startmv.row; if (startmv.col & 7) { this_mv.col = startmv.col - 2; left = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); } else { this_mv.col = (startmv.col - 8) | 6; left = svf(y - 1, d->pre_stride, 6, this_mv.row & 7, z, b->src_stride, &sse); } left += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (left < bestmse) { *bestmv = this_mv; bestmse = left; } this_mv.col += 4; right = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); right += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (right < bestmse) { *bestmv = this_mv; bestmse = right; } // go up then down and check error this_mv.col = startmv.col; if (startmv.row & 7) { this_mv.row = startmv.row - 2; up = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); } else { this_mv.row = (startmv.row - 8) | 6; up = svf(y - d->pre_stride, d->pre_stride, this_mv.col & 7, 6, z, b->src_stride, &sse); } up += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (up < bestmse) { *bestmv = this_mv; bestmse = up; } this_mv.row += 4; down = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); down += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (down < bestmse) { *bestmv = this_mv; bestmse = down; } // now check 1 more diagonal whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); // for(whichdir=0;whichdir<4;whichdir++) // { this_mv = startmv; switch (whichdir) { case 0: if (startmv.row & 7) { this_mv.row -= 2; if (startmv.col & 7) { this_mv.col -= 2; diag = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); } else { this_mv.col = (startmv.col - 8) | 6; diag = svf(y - 1, d->pre_stride, 6, this_mv.row & 7, z, b->src_stride, &sse);; } } else { this_mv.row = (startmv.row - 8) | 6; if (startmv.col & 7) { this_mv.col -= 2; diag = svf(y - d->pre_stride, d->pre_stride, this_mv.col & 7, 6, z, b->src_stride, &sse); } else { this_mv.col = (startmv.col - 8) | 6; diag = svf(y - d->pre_stride - 1, d->pre_stride, 6, 6, z, b->src_stride, &sse); } } break; case 1: this_mv.col += 2; if (startmv.row & 7) { this_mv.row -= 2; diag = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); } else { this_mv.row = (startmv.row - 8) | 6; diag = svf(y - d->pre_stride, d->pre_stride, this_mv.col & 7, 6, z, b->src_stride, &sse); } break; case 2: this_mv.row += 2; if (startmv.col & 7) { this_mv.col -= 2; diag = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); } else { this_mv.col = (startmv.col - 8) | 6; diag = svf(y - 1, d->pre_stride, 6, this_mv.row & 7, z, b->src_stride, &sse);; } break; case 3: this_mv.col += 2; this_mv.row += 2; diag = svf(y, d->pre_stride, this_mv.col & 7, this_mv.row & 7, z, b->src_stride, &sse); break; } diag += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (diag < bestmse) { *bestmv = this_mv; bestmse = diag; } // } return bestmse; } int vp8_find_best_half_pixel_step(MACROBLOCK *mb, BLOCK *b, BLOCKD *d, MV *bestmv, MV *ref_mv, int error_per_bit, vp8_subpixvariance_fn_t svf, vp8_variance_fn_t vf, int *mvcost[2]) { int bestmse = INT_MAX; MV startmv; //MV this_mv; MV this_mv; unsigned char *y = *(d->base_pre) + d->pre + (bestmv->row) * d->pre_stride + bestmv->col; unsigned char *z = (*(b->base_src) + b->src); int left, right, up, down, diag; unsigned int sse; // Trap uncodable vectors if ((abs((bestmv->col << 3) - ref_mv->col) > MAX_FULL_PEL_VAL) || (abs((bestmv->row << 3) - ref_mv->row) > MAX_FULL_PEL_VAL)) { bestmv->row <<= 3; bestmv->col <<= 3; return INT_MAX; } // central mv bestmv->row <<= 3; bestmv->col <<= 3; startmv = *bestmv; // calculate central point error bestmse = vf(y, d->pre_stride, z, b->src_stride, &sse); bestmse += vp8_mv_err_cost(bestmv, ref_mv, mvcost, error_per_bit); // go left then right and check error this_mv.row = startmv.row; this_mv.col = ((startmv.col - 8) | 4); left = svf(y - 1, d->pre_stride, 4, 0, z, b->src_stride, &sse); left += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (left < bestmse) { *bestmv = this_mv; bestmse = left; } this_mv.col += 8; right = svf(y, d->pre_stride, 4, 0, z, b->src_stride, &sse); right += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (right < bestmse) { *bestmv = this_mv; bestmse = right; } // go up then down and check error this_mv.col = startmv.col; this_mv.row = ((startmv.row - 8) | 4); up = svf(y - d->pre_stride, d->pre_stride, 0, 4, z, b->src_stride, &sse); up += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (up < bestmse) { *bestmv = this_mv; bestmse = up; } this_mv.row += 8; down = svf(y, d->pre_stride, 0, 4, z, b->src_stride, &sse); down += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (down < bestmse) { *bestmv = this_mv; bestmse = down; } // somewhat strangely not doing all the diagonals for half pel is slower than doing them. #if 0 // now check 1 more diagonal - whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); this_mv = startmv; switch (whichdir) { case 0: this_mv.col = (this_mv.col - 8) | 4; this_mv.row = (this_mv.row - 8) | 4; diag = svf(y - 1 - d->pre_stride, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; case 1: this_mv.col += 4; this_mv.row = (this_mv.row - 8) | 4; diag = svf(y - d->pre_stride, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; case 2: this_mv.col = (this_mv.col - 8) | 4; this_mv.row += 4; diag = svf(y - 1, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; case 3: this_mv.col += 4; this_mv.row += 4; diag = svf(y, d->pre_stride, 4, 4, z, b->src_stride, &sse); break; } diag += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (diag < bestmse) { *bestmv = this_mv; bestmse = diag; } #else this_mv.col = (this_mv.col - 8) | 4; this_mv.row = (this_mv.row - 8) | 4; diag = svf(y - 1 - d->pre_stride, d->pre_stride, 4, 4, z, b->src_stride, &sse); diag += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (diag < bestmse) { *bestmv = this_mv; bestmse = diag; } this_mv.col += 8; diag = svf(y - d->pre_stride, d->pre_stride, 4, 4, z, b->src_stride, &sse); diag += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (diag < bestmse) { *bestmv = this_mv; bestmse = diag; } this_mv.col = (this_mv.col - 8) | 4; this_mv.row = startmv.row + 4; diag = svf(y - 1, d->pre_stride, 4, 4, z, b->src_stride, &sse); diag += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (diag < bestmse) { *bestmv = this_mv; bestmse = diag; } this_mv.col += 8; diag = svf(y, d->pre_stride, 4, 4, z, b->src_stride, &sse); diag += vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); if (diag < bestmse) { *bestmv = this_mv; bestmse = diag; } #endif return bestmse; } #define MVC(r,c) (((mvsadcost[0][((r)<<2)-rr] + mvsadcost[1][((c)<<2) - rc]) * error_per_bit + 128 )>>8 ) // estimated cost of a motion vector (r,c) #define PRE(r,c) (*(d->base_pre) + d->pre + (r) * d->pre_stride + (c)) // pointer to predictor base of a motionvector #define DIST(r,c,v) sf( src,src_stride,PRE(r,c),d->pre_stride, v) // returns sad error score. #define ERR(r,c,v) (MVC(r,c)+DIST(r,c,v)) // returns distortion + motion vector cost #define CHECK_BETTER(v,r,c) if ((v = ERR(r,c,besterr)) < besterr) { besterr = v; br=r; bc=c; } // checks if (r,c) has better score than previous best int vp8_hex_search ( MACROBLOCK *x, BLOCK *b, BLOCKD *d, MV *ref_mv, MV *best_mv, int search_param, int error_per_bit, int *num00, vp8_variance_fn_t vf, vp8_sad_fn_t sf, int *mvsadcost[2], int *mvcost[2] ) { MV hex[6] = { { -2, 0}, { -1, -2}, { -1, 2}, {2, 0}, {1, 2}, {1, -2} } ; MV neighbors[8] = { { -1, -1}, { -1, 0}, { -1, 1}, {0, -1}, {0, 1}, {1, -1}, {1, 0}, {1, 1} } ; int i, j; unsigned char *src = (*(b->base_src) + b->src); int src_stride = b->src_stride; int rr = ref_mv->row, rc = ref_mv->col, br = rr, bc = rc, tr, tc; unsigned int besterr, thiserr = 0x7fffffff; if (rc < x->mv_col_min) bc = x->mv_col_min; if (rc > x->mv_col_max) bc = x->mv_col_max; if (rr < x->mv_row_min) br = x->mv_row_min; if (rr > x->mv_row_max) br = x->mv_row_max; rr >>= 1; rc >>= 1; br >>= 3; bc >>= 3; besterr = ERR(br, bc, thiserr); // hex search jbb changed to 127 to avoid max 256 problem steping by 2. for (j = 0; j < 127; j++) { tr = br; tc = bc; for (i = 0; i < 6; i++) { int nr = tr + hex[i].row, nc = tc + hex[i].col; if (nc < x->mv_col_min) continue; if (nc > x->mv_col_max) continue; if (nr < x->mv_row_min) continue; if (nr > x->mv_row_max) continue; CHECK_BETTER(thiserr, nr, nc); } if (tr == br && tc == bc) break; } // check 8 1 away neighbors tr = br; tc = bc; for (i = 0; i < 8; i++) { int nr = tr + neighbors[i].row, nc = tc + neighbors[i].col; if (nc < x->mv_col_min) continue; if (nc > x->mv_col_max) continue; if (nr < x->mv_row_min) continue; if (nr > x->mv_row_max) continue; CHECK_BETTER(thiserr, nr, nc); } best_mv->row = br; best_mv->col = bc; return vf(src, src_stride, PRE(br, bc), d->pre_stride, &thiserr) + MVC(br, bc) ; } #undef MVC #undef PRE #undef SP #undef DIST #undef ERR #undef CHECK_BETTER int vp8_diamond_search_sad ( MACROBLOCK *x, BLOCK *b, BLOCKD *d, MV *ref_mv, MV *best_mv, int search_param, int error_per_bit, int *num00, vp8_variance_fn_ptr_t *fn_ptr, int *mvsadcost[2], int *mvcost[2] ) { int i, j, step; unsigned char *what = (*(b->base_src) + b->src); int what_stride = b->src_stride; unsigned char *in_what; int in_what_stride = d->pre_stride; unsigned char *best_address; int tot_steps; MV this_mv; int bestsad = INT_MAX; int best_site = 0; int last_site = 0; int ref_row = ref_mv->row >> 3; int ref_col = ref_mv->col >> 3; int this_row_offset; int this_col_offset; search_site *ss; unsigned char *check_here; int thissad; // Work out the start point for the search in_what = (unsigned char *)(*(d->base_pre) + d->pre + (ref_row * (d->pre_stride)) + ref_col); best_address = in_what; // We need to check that the starting point for the search (as indicated by ref_mv) is within the buffer limits if ((ref_col > x->mv_col_min) && (ref_col < x->mv_col_max) && (ref_row > x->mv_row_min) && (ref_row < x->mv_row_max)) { // Check the starting position bestsad = fn_ptr->sdf(what, what_stride, in_what, in_what_stride, 0x7fffffff) + vp8_mv_err_cost(ref_mv, ref_mv, mvsadcost, error_per_bit); } // search_param determines the length of the initial step and hence the number of iterations // 0 = initial step (MAX_FIRST_STEP) pel : 1 = (MAX_FIRST_STEP/2) pel, 2 = (MAX_FIRST_STEP/4) pel... etc. ss = &x->ss[search_param * x->searches_per_step]; tot_steps = (x->ss_count / x->searches_per_step) - search_param; i = 1; best_mv->row = ref_row; best_mv->col = ref_col; *num00 = 0; for (step = 0; step < tot_steps ; step++) { for (j = 0 ; j < x->searches_per_step ; j++) { // Trap illegal vectors this_row_offset = best_mv->row + ss[i].mv.row; this_col_offset = best_mv->col + ss[i].mv.col; if ((this_col_offset > x->mv_col_min) && (this_col_offset < x->mv_col_max) && (this_row_offset > x->mv_row_min) && (this_row_offset < x->mv_row_max)) { check_here = ss[i].offset + best_address; thissad = fn_ptr->sdf(what, what_stride, check_here , in_what_stride, bestsad); if (thissad < bestsad) { this_mv.row = this_row_offset << 3; this_mv.col = this_col_offset << 3; thissad += vp8_mv_err_cost(&this_mv, ref_mv, mvsadcost, error_per_bit); if (thissad < bestsad) { bestsad = thissad; best_site = i; } } } i++; } if (best_site != last_site) { best_mv->row += ss[best_site].mv.row; best_mv->col += ss[best_site].mv.col; best_address += ss[best_site].offset; last_site = best_site; } else if (best_address == in_what) (*num00)++; } this_mv.row = best_mv->row << 3; this_mv.col = best_mv->col << 3; if (bestsad == INT_MAX) return INT_MAX; return fn_ptr->vf(what, what_stride, best_address, in_what_stride, (unsigned int *)(&thissad)) + vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); } int vp8_diamond_search_sadx4 ( MACROBLOCK *x, BLOCK *b, BLOCKD *d, MV *ref_mv, MV *best_mv, int search_param, int error_per_bit, int *num00, vp8_variance_fn_ptr_t *fn_ptr, int *mvsadcost[2], int *mvcost[2] ) { int i, j, step; unsigned char *what = (*(b->base_src) + b->src); int what_stride = b->src_stride; unsigned char *in_what; int in_what_stride = d->pre_stride; unsigned char *best_address; int tot_steps; MV this_mv; int bestsad = INT_MAX; int best_site = 0; int last_site = 0; int ref_row = ref_mv->row >> 3; int ref_col = ref_mv->col >> 3; int this_row_offset; int this_col_offset; search_site *ss; unsigned char *check_here; unsigned int thissad; // Work out the start point for the search in_what = (unsigned char *)(*(d->base_pre) + d->pre + (ref_row * (d->pre_stride)) + ref_col); best_address = in_what; // We need to check that the starting point for the search (as indicated by ref_mv) is within the buffer limits if ((ref_col > x->mv_col_min) && (ref_col < x->mv_col_max) && (ref_row > x->mv_row_min) && (ref_row < x->mv_row_max)) { // Check the starting position bestsad = fn_ptr->sdf(what, what_stride, in_what, in_what_stride, 0x7fffffff) + vp8_mv_err_cost(ref_mv, ref_mv, mvsadcost, error_per_bit); } // search_param determines the length of the initial step and hence the number of iterations // 0 = initial step (MAX_FIRST_STEP) pel : 1 = (MAX_FIRST_STEP/2) pel, 2 = (MAX_FIRST_STEP/4) pel... etc. ss = &x->ss[search_param * x->searches_per_step]; tot_steps = (x->ss_count / x->searches_per_step) - search_param; i = 1; best_mv->row = ref_row; best_mv->col = ref_col; *num00 = 0; for (step = 0; step < tot_steps ; step++) { int check_row_min, check_col_min, check_row_max, check_col_max; check_row_min = x->mv_row_min - best_mv->row; check_row_max = x->mv_row_max - best_mv->row; check_col_min = x->mv_col_min - best_mv->col; check_col_max = x->mv_col_max - best_mv->col; for (j = 0 ; j < x->searches_per_step ; j += 4) { unsigned char *block_offset[4]; unsigned int valid_block[4]; int all_in = 1, t; for (t = 0; t < 4; t++) { valid_block [t] = (ss[t+i].mv.col > check_col_min); valid_block [t] &= (ss[t+i].mv.col < check_col_max); valid_block [t] &= (ss[t+i].mv.row > check_row_min); valid_block [t] &= (ss[t+i].mv.row < check_row_max); all_in &= valid_block[t]; block_offset[t] = ss[i+t].offset + best_address; } if (all_in) { unsigned int sad_array[4]; fn_ptr->sdx4df(what, what_stride, block_offset, in_what_stride, sad_array); for (t = 0; t < 4; t++, i++) { thissad = sad_array[t]; if (thissad < bestsad) { this_mv.row = (best_mv->row + ss[i].mv.row) << 3; this_mv.col = (best_mv->col + ss[i].mv.col) << 3; thissad += vp8_mv_err_cost(&this_mv, ref_mv, mvsadcost, error_per_bit); if (thissad < bestsad) { bestsad = thissad; best_site = i; } } } } else { int t; for (t = 0; t < 4; i++, t++) { // Trap illegal vectors if (valid_block[t]) { check_here = block_offset[t]; thissad = fn_ptr->sdf(what, what_stride, check_here , in_what_stride, bestsad); if (thissad < bestsad) { this_row_offset = best_mv->row + ss[i].mv.row; this_col_offset = best_mv->col + ss[i].mv.col; this_mv.row = this_row_offset << 3; this_mv.col = this_col_offset << 3; thissad += vp8_mv_err_cost(&this_mv, ref_mv, mvsadcost, error_per_bit); if (thissad < bestsad) { bestsad = thissad; best_site = i; } } } } } } if (best_site != last_site) { best_mv->row += ss[best_site].mv.row; best_mv->col += ss[best_site].mv.col; best_address += ss[best_site].offset; last_site = best_site; } else if (best_address == in_what) (*num00)++; } this_mv.row = best_mv->row << 3; this_mv.col = best_mv->col << 3; if (bestsad == INT_MAX) return INT_MAX; return fn_ptr->vf(what, what_stride, best_address, in_what_stride, (unsigned int *)(&thissad)) + vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); } int vp8_full_search_sad(MACROBLOCK *x, BLOCK *b, BLOCKD *d, MV *ref_mv, int error_per_bit, int distance, vp8_variance_fn_ptr_t *fn_ptr, int *mvcost[2], int *mvsadcost[2]) { unsigned char *what = (*(b->base_src) + b->src); int what_stride = b->src_stride; unsigned char *in_what; int in_what_stride = d->pre_stride; int mv_stride = d->pre_stride; unsigned char *bestaddress; MV *best_mv = &d->bmi.mv.as_mv; MV this_mv; int bestsad = INT_MAX; int r, c; unsigned char *check_here; int thissad; int ref_row = ref_mv->row >> 3; int ref_col = ref_mv->col >> 3; int row_min = ref_row - distance; int row_max = ref_row + distance; int col_min = ref_col - distance; int col_max = ref_col + distance; // Work out the mid point for the search in_what = *(d->base_pre) + d->pre; bestaddress = in_what + (ref_row * d->pre_stride) + ref_col; best_mv->row = ref_row; best_mv->col = ref_col; // We need to check that the starting point for the search (as indicated by ref_mv) is within the buffer limits if ((ref_col > x->mv_col_min) && (ref_col < x->mv_col_max) && (ref_row > x->mv_row_min) && (ref_row < x->mv_row_max)) { // Baseline value at the centre //bestsad = fn_ptr->sf( what,what_stride,bestaddress,in_what_stride) + (int)sqrt(vp8_mv_err_cost(ref_mv,ref_mv, mvcost,error_per_bit*14)); bestsad = fn_ptr->sdf(what, what_stride, bestaddress, in_what_stride, 0x7fffffff) + vp8_mv_err_cost(ref_mv, ref_mv, mvsadcost, error_per_bit); } // Apply further limits to prevent us looking using vectors that stretch beyiond the UMV border if (col_min < x->mv_col_min) col_min = x->mv_col_min; if (col_max > x->mv_col_max) col_max = x->mv_col_max; if (row_min < x->mv_row_min) row_min = x->mv_row_min; if (row_max > x->mv_row_max) row_max = x->mv_row_max; for (r = row_min; r < row_max ; r++) { this_mv.row = r << 3; check_here = r * mv_stride + in_what + col_min; for (c = col_min; c < col_max; c++) { thissad = fn_ptr->sdf(what, what_stride, check_here , in_what_stride, bestsad); this_mv.col = c << 3; //thissad += (int)sqrt(vp8_mv_err_cost(&this_mv,ref_mv, mvcost,error_per_bit*14)); //thissad += error_per_bit * mv_bits_sadcost[mv_bits(&this_mv, ref_mv, mvcost)]; thissad += vp8_mv_err_cost(&this_mv, ref_mv, mvsadcost, error_per_bit); //mv_bits(error_per_bit, &this_mv, ref_mv, mvsadcost); if (thissad < bestsad) { bestsad = thissad; best_mv->row = r; best_mv->col = c; bestaddress = check_here; } check_here++; } } this_mv.row = best_mv->row << 3; this_mv.col = best_mv->col << 3; if (bestsad < INT_MAX) return fn_ptr->vf(what, what_stride, bestaddress, in_what_stride, (unsigned int *)(&thissad)) + vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); else return INT_MAX; } int vp8_full_search_sadx3(MACROBLOCK *x, BLOCK *b, BLOCKD *d, MV *ref_mv, int error_per_bit, int distance, vp8_variance_fn_ptr_t *fn_ptr, int *mvcost[2], int *mvsadcost[2]) { unsigned char *what = (*(b->base_src) + b->src); int what_stride = b->src_stride; unsigned char *in_what; int in_what_stride = d->pre_stride; int mv_stride = d->pre_stride; unsigned char *bestaddress; MV *best_mv = &d->bmi.mv.as_mv; MV this_mv; int bestsad = INT_MAX; int r, c; unsigned char *check_here; unsigned int thissad; int ref_row = ref_mv->row >> 3; int ref_col = ref_mv->col >> 3; int row_min = ref_row - distance; int row_max = ref_row + distance; int col_min = ref_col - distance; int col_max = ref_col + distance; unsigned int sad_array[3]; // Work out the mid point for the search in_what = *(d->base_pre) + d->pre; bestaddress = in_what + (ref_row * d->pre_stride) + ref_col; best_mv->row = ref_row; best_mv->col = ref_col; // We need to check that the starting point for the search (as indicated by ref_mv) is within the buffer limits if ((ref_col > x->mv_col_min) && (ref_col < x->mv_col_max) && (ref_row > x->mv_row_min) && (ref_row < x->mv_row_max)) { // Baseline value at the centre bestsad = fn_ptr->sdf(what, what_stride, bestaddress, in_what_stride, 0x7fffffff) + vp8_mv_err_cost(ref_mv, ref_mv, mvsadcost, error_per_bit); } // Apply further limits to prevent us looking using vectors that stretch beyiond the UMV border if (col_min < x->mv_col_min) col_min = x->mv_col_min; if (col_max > x->mv_col_max) col_max = x->mv_col_max; if (row_min < x->mv_row_min) row_min = x->mv_row_min; if (row_max > x->mv_row_max) row_max = x->mv_row_max; for (r = row_min; r < row_max ; r++) { this_mv.row = r << 3; check_here = r * mv_stride + in_what + col_min; c = col_min; while ((c + 3) < col_max) { int i; fn_ptr->sdx3f(what, what_stride, check_here , in_what_stride, sad_array); for (i = 0; i < 3; i++) { thissad = sad_array[i]; if (thissad < bestsad) { this_mv.col = c << 3; thissad += vp8_mv_err_cost(&this_mv, ref_mv, mvsadcost, error_per_bit); if (thissad < bestsad) { bestsad = thissad; best_mv->row = r; best_mv->col = c; bestaddress = check_here; } } check_here++; c++; } } while (c < col_max) { thissad = fn_ptr->sdf(what, what_stride, check_here , in_what_stride, bestsad); if (thissad < bestsad) { this_mv.col = c << 3; thissad += vp8_mv_err_cost(&this_mv, ref_mv, mvsadcost, error_per_bit); if (thissad < bestsad) { bestsad = thissad; best_mv->row = r; best_mv->col = c; bestaddress = check_here; } } check_here ++; c ++; } } this_mv.row = best_mv->row << 3; this_mv.col = best_mv->col << 3; if (bestsad < INT_MAX) return fn_ptr->vf(what, what_stride, bestaddress, in_what_stride, (unsigned int *)(&thissad)) + vp8_mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); else return INT_MAX; } #ifdef ENTROPY_STATS void print_mode_context(void) { FILE *f = fopen("modecont.c", "w"); int i, j; fprintf(f, "#include \"entropy.h\"\n"); fprintf(f, "const int vp8_mode_contexts[6][4] =\n"); fprintf(f, "{\n"); for (j = 0; j < 6; j++) { fprintf(f, " { // %d \n", j); fprintf(f, " "); for (i = 0; i < 4; i++) { int overal_prob; int this_prob; int count; // = mv_ref_ct[j][i][0]+mv_ref_ct[j][i][1]; // Overall probs count = mv_mode_cts[i][0] + mv_mode_cts[i][1]; if (count) overal_prob = 256 * mv_mode_cts[i][0] / count; else overal_prob = 128; if (overal_prob == 0) overal_prob = 1; // context probs count = mv_ref_ct[j][i][0] + mv_ref_ct[j][i][1]; if (count) this_prob = 256 * mv_ref_ct[j][i][0] / count; else this_prob = 128; if (this_prob == 0) this_prob = 1; fprintf(f, "%5d, ", this_prob); //fprintf(f,"%5d, %5d, %8d,", this_prob, overal_prob, (this_prob << 10)/overal_prob); //fprintf(f,"%8d, ", (this_prob << 10)/overal_prob); } fprintf(f, " },\n"); } fprintf(f, "};\n"); fclose(f); } /* MV ref count ENTROPY_STATS stats code */ #ifdef ENTROPY_STATS void init_mv_ref_counts() { vpx_memset(mv_ref_ct, 0, sizeof(mv_ref_ct)); vpx_memset(mv_mode_cts, 0, sizeof(mv_mode_cts)); } void accum_mv_refs(MB_PREDICTION_MODE m, const int ct[4]) { if (m == ZEROMV) { ++mv_ref_ct [ct[0]] [0] [0]; ++mv_mode_cts[0][0]; } else { ++mv_ref_ct [ct[0]] [0] [1]; ++mv_mode_cts[0][1]; if (m == NEARESTMV) { ++mv_ref_ct [ct[1]] [1] [0]; ++mv_mode_cts[1][0]; } else { ++mv_ref_ct [ct[1]] [1] [1]; ++mv_mode_cts[1][1]; if (m == NEARMV) { ++mv_ref_ct [ct[2]] [2] [0]; ++mv_mode_cts[2][0]; } else { ++mv_ref_ct [ct[2]] [2] [1]; ++mv_mode_cts[2][1]; if (m == NEWMV) { ++mv_ref_ct [ct[3]] [3] [0]; ++mv_mode_cts[3][0]; } else { ++mv_ref_ct [ct[3]] [3] [1]; ++mv_mode_cts[3][1]; } } } } } #endif/* END MV ref count ENTROPY_STATS stats code */ #endif