From aac982e89c9bc883316025036f074eb6b9d45298 Mon Sep 17 00:00:00 2001 From: Debargha Mukherjee Date: Tue, 11 Sep 2012 14:52:20 -0700 Subject: Revert "Hybrid transform cleanups" This reverts commit c5af82b7ed6d89a4a12003b8d1c6b4befcadce92 There was a mismatch bug introduced by this patch. Hence reverting until it is fixed. --- vp8/encoder/encodeframe.c | 3 ++ vp8/encoder/tokenize.c | 91 ++++++++--------------------------------------- 2 files changed, 17 insertions(+), 77 deletions(-) (limited to 'vp8/encoder') diff --git a/vp8/encoder/encodeframe.c b/vp8/encoder/encodeframe.c index 91a1512db..4670c314d 100644 --- a/vp8/encoder/encodeframe.c +++ b/vp8/encoder/encodeframe.c @@ -1834,6 +1834,7 @@ void vp8cx_encode_intra_macro_block(VP8_COMP *cpi, else #endif if (cpi->common.txfm_mode == ALLOW_8X8 + && mbmi->mode != I8X8_PRED && mbmi->mode != B_PRED) { mbmi->txfm_size = TX_8X8; cpi->t8x8_count++; @@ -1930,6 +1931,7 @@ void vp8cx_encode_inter_macroblock (VP8_COMP *cpi, MACROBLOCK *x, } else #endif if (cpi->common.txfm_mode == ALLOW_8X8 + && mbmi->mode != I8X8_PRED && mbmi->mode != B_PRED && mbmi->mode != SPLITMV) { mbmi->txfm_size = TX_8X8; @@ -2110,6 +2112,7 @@ void vp8cx_encode_inter_superblock(VP8_COMP *cpi, MACROBLOCK *x, TOKENEXTRA **t, /* test code: set transform size based on mode selection */ if (cpi->common.txfm_mode == ALLOW_8X8 + && x->e_mbd.mode_info_context->mbmi.mode != I8X8_PRED && x->e_mbd.mode_info_context->mbmi.mode != B_PRED && x->e_mbd.mode_info_context->mbmi.mode != SPLITMV) { x->e_mbd.mode_info_context->mbmi.txfm_size = TX_8X8; diff --git a/vp8/encoder/tokenize.c b/vp8/encoder/tokenize.c index 5b1f8ffb4..d0c315283 100644 --- a/vp8/encoder/tokenize.c +++ b/vp8/encoder/tokenize.c @@ -42,8 +42,6 @@ void vp8_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run); void vp8_stuff_mb_8x8(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run); -void vp8_stuff_mb_8x8_4x4uv(VP8_COMP *cpi, - MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run); #if CONFIG_TX16X16 || CONFIG_HYBRIDTRANSFORM16X16 void vp8_stuff_mb_16x16(VP8_COMP *cpi, MACROBLOCKD *xd, TOKENEXTRA **t, int dry_run); @@ -766,10 +764,6 @@ int mb_is_skippable_8x8(MACROBLOCKD *xd) { return (mby_is_skippable_8x8(xd) & mbuv_is_skippable_8x8(xd)); } -int mb_is_skippable_8x8_4x4uv(MACROBLOCKD *xd) { - return (mby_is_skippable_8x8(xd) & mbuv_is_skippable(xd)); -} - #if CONFIG_TX16X16 || CONFIG_HYBRIDTRANSFORM16X16 int mby_is_skippable_16x16(MACROBLOCKD *xd) { int skip = 1; @@ -828,10 +822,7 @@ void vp8_tokenize_mb(VP8_COMP *cpi, break; #endif case TX_8X8: - if (xd->mode_info_context->mbmi.mode == I8X8_PRED) - xd->mode_info_context->mbmi.mb_skip_coeff = mb_is_skippable_8x8_4x4uv(xd); - else - xd->mode_info_context->mbmi.mb_skip_coeff = mb_is_skippable_8x8(xd); + xd->mode_info_context->mbmi.mb_skip_coeff = mb_is_skippable_8x8(xd); break; default: xd->mode_info_context->mbmi.mb_skip_coeff = mb_is_skippable(xd, has_y2_block); @@ -847,12 +838,9 @@ void vp8_tokenize_mb(VP8_COMP *cpi, vp8_stuff_mb_16x16(cpi, xd, t, dry_run); else #endif - if (tx_type == TX_8X8) { - if (xd->mode_info_context->mbmi.mode == I8X8_PRED) - vp8_stuff_mb_8x8_4x4uv(cpi, xd, t, dry_run); - else - vp8_stuff_mb_8x8(cpi, xd, t, dry_run); - } else + if (tx_type == TX_8X8) + vp8_stuff_mb_8x8(cpi, xd, t, dry_run); + else vp8_stuff_mb(cpi, xd, t, dry_run); } else { vp8_fix_contexts(xd); @@ -907,10 +895,6 @@ void vp8_tokenize_mb(VP8_COMP *cpi, if (tx_type == TX_8X8) { ENTROPY_CONTEXT *A = (ENTROPY_CONTEXT *)xd->above_context; ENTROPY_CONTEXT *L = (ENTROPY_CONTEXT *)xd->left_context; -#if CONFIG_HYBRIDTRANSFORM8X8 - if (xd->mode_info_context->mbmi.mode == I8X8_PRED) - plane_type = PLANE_TYPE_Y_WITH_DC; -#endif for (b = 0; b < 16; b += 4) { tokenize1st_order_b_8x8(xd, xd->block + b, t, plane_type, xd->frame_type, @@ -920,34 +904,21 @@ void vp8_tokenize_mb(VP8_COMP *cpi, *(A + vp8_block2above_8x8[b] + 1) = *(A + vp8_block2above_8x8[b]); *(L + vp8_block2left_8x8[b] + 1) = *(L + vp8_block2left_8x8[b]); } -#if CONFIG_HYBRIDTRANSFORM8X8 - if (xd->mode_info_context->mbmi.mode == I8X8_PRED) { - tokenize1st_order_chroma(xd, t, PLANE_TYPE_UV, cpi, dry_run); - } else -#endif - { - for (b = 16; b < 24; b += 4) { - tokenize1st_order_b_8x8(xd, - xd->block + b, t, 2, xd->frame_type, - A + vp8_block2above_8x8[b], - L + vp8_block2left_8x8[b], - cpi, dry_run); - *(A + vp8_block2above_8x8[b] + 1) = *(A + vp8_block2above_8x8[b]); - *(L + vp8_block2left_8x8[b] + 1) = *(L + vp8_block2left_8x8[b]); - } + for (b = 16; b < 24; b += 4) { + tokenize1st_order_b_8x8(xd, + xd->block + b, t, 2, xd->frame_type, + A + vp8_block2above_8x8[b], + L + vp8_block2left_8x8[b], + cpi, dry_run); + *(A + vp8_block2above_8x8[b] + 1) = *(A + vp8_block2above_8x8[b]); + *(L + vp8_block2left_8x8[b] + 1) = *(L + vp8_block2left_8x8[b]); } } else { #if CONFIG_HYBRIDTRANSFORM if(active_ht) { tokenize1st_order_ht(xd, t, plane_type, cpi, dry_run); - } else -#endif - { - tokenize1st_order_b(xd, t, plane_type, cpi, dry_run); - } - } + } else { - /* #if CONFIG_HYBRIDTRANSFORM8X8 if (xd->mode_info_context->mbmi.mode == I8X8_PRED) { ENTROPY_CONTEXT *A = (ENTROPY_CONTEXT *)xd->above_context; @@ -975,7 +946,6 @@ void vp8_tokenize_mb(VP8_COMP *cpi, tokenize1st_order_b(xd, t, plane_type, cpi, dry_run); #endif } - */ if (dry_run) *t = t_backup; } @@ -1371,6 +1341,7 @@ void vp8_stuff_mb_8x8(VP8_COMP *cpi, *t = t_backup; } + #if CONFIG_TX16X16 || CONFIG_HYBRIDTRANSFORM16X16 static __inline void stuff1st_order_b_16x16(const BLOCKD *const b, @@ -1522,40 +1493,6 @@ void vp8_stuff_mb(VP8_COMP *cpi, MACROBLOCKD *xd, if (dry_run) *t = t_backup; } - -void vp8_stuff_mb_8x8_4x4uv(VP8_COMP *cpi, - MACROBLOCKD *xd, - TOKENEXTRA **t, - int dry_run) { - ENTROPY_CONTEXT *A = (ENTROPY_CONTEXT *)xd->above_context; - ENTROPY_CONTEXT *L = (ENTROPY_CONTEXT *)xd->left_context; - int plane_type; - int b; - TOKENEXTRA *t_backup = *t; - - stuff2nd_order_b_8x8(xd->block + 24, t, 1, xd->frame_type, - A + vp8_block2above_8x8[24], - L + vp8_block2left_8x8[24], cpi, dry_run); - plane_type = 0; - - for (b = 0; b < 16; b += 4) { - stuff1st_order_b_8x8(xd->block + b, t, plane_type, xd->frame_type, - A + vp8_block2above_8x8[b], - L + vp8_block2left_8x8[b], - cpi, dry_run); - *(A + vp8_block2above_8x8[b] + 1) = *(A + vp8_block2above_8x8[b]); - *(L + vp8_block2left_8x8[b] + 1) = *(L + vp8_block2left_8x8[b]); - } - for (b = 16; b < 24; b++) - stuff1st_order_buv(t, - A + vp8_block2above[b], - L + vp8_block2left[b], - cpi, dry_run); - - if (dry_run) - *t = t_backup; -} - void vp8_fix_contexts(MACROBLOCKD *xd) { /* Clear entropy contexts for Y2 blocks */ if ((xd->mode_info_context->mbmi.mode != B_PRED -- cgit v1.2.3