diff options
author | John Koleszar <jkoleszar@google.com> | 2012-12-05 08:41:35 -0800 |
---|---|---|
committer | John Koleszar <jkoleszar@google.com> | 2012-12-05 08:41:35 -0800 |
commit | 5d91a1e0ae6bbe36005a70cdc3f91b8fbbc51e92 (patch) | |
tree | cb46d7c6f24b1dd891c7518e5d6af70197000929 /vp9 | |
parent | d138262ac0e97f571fe16de46cbcc92ba8782151 (diff) | |
parent | 4a4d2aa55c970815375c795c570d29645b431b53 (diff) | |
download | libvpx-5d91a1e0ae6bbe36005a70cdc3f91b8fbbc51e92.tar libvpx-5d91a1e0ae6bbe36005a70cdc3f91b8fbbc51e92.tar.gz libvpx-5d91a1e0ae6bbe36005a70cdc3f91b8fbbc51e92.tar.bz2 libvpx-5d91a1e0ae6bbe36005a70cdc3f91b8fbbc51e92.zip |
Merge remote-tracking branch 'origin/vp9-preview' into experimental
Diffstat (limited to 'vp9')
-rw-r--r-- | vp9/common/x86/vp9_asm_stubs.c | 23 | ||||
-rw-r--r-- | vp9/encoder/vp9_encodeframe.c | 4 | ||||
-rw-r--r-- | vp9/encoder/vp9_mbgraph.c | 2 | ||||
-rw-r--r-- | vp9/encoder/x86/vp9_variance_mmx.c | 26 | ||||
-rw-r--r-- | vp9/encoder/x86/vp9_variance_sse2.c | 2 | ||||
-rw-r--r-- | vp9/vp9cx.mk | 1 |
6 files changed, 27 insertions, 31 deletions
diff --git a/vp9/common/x86/vp9_asm_stubs.c b/vp9/common/x86/vp9_asm_stubs.c index f804af698..e53ede066 100644 --- a/vp9/common/x86/vp9_asm_stubs.c +++ b/vp9/common/x86/vp9_asm_stubs.c @@ -97,6 +97,29 @@ extern void vp9_filter_block1d8_v6_only_sse2(unsigned char *src_ptr, extern prototype_subpixel_predict(vp9_bilinear_predict8x8_mmx); +/////////////////////////////////////////////////////////////////////////// +// the mmx function that does the bilinear filtering and var calculation // +// int one pass // +/////////////////////////////////////////////////////////////////////////// +DECLARE_ALIGNED(16, const short, vp9_bilinear_filters_mmx[16][8]) = { + { 128, 128, 128, 128, 0, 0, 0, 0 }, + { 120, 120, 120, 120, 8, 8, 8, 8 }, + { 112, 112, 112, 112, 16, 16, 16, 16 }, + { 104, 104, 104, 104, 24, 24, 24, 24 }, + { 96, 96, 96, 96, 32, 32, 32, 32 }, + { 88, 88, 88, 88, 40, 40, 40, 40 }, + { 80, 80, 80, 80, 48, 48, 48, 48 }, + { 72, 72, 72, 72, 56, 56, 56, 56 }, + { 64, 64, 64, 64, 64, 64, 64, 64 }, + { 56, 56, 56, 56, 72, 72, 72, 72 }, + { 48, 48, 48, 48, 80, 80, 80, 80 }, + { 40, 40, 40, 40, 88, 88, 88, 88 }, + { 32, 32, 32, 32, 96, 96, 96, 96 }, + { 24, 24, 24, 24, 104, 104, 104, 104 }, + { 16, 16, 16, 16, 112, 112, 112, 112 }, + { 8, 8, 8, 8, 120, 120, 120, 120 } +}; + #if HAVE_MMX void vp9_sixtap_predict4x4_mmx(unsigned char *src_ptr, int src_pixels_per_line, diff --git a/vp9/encoder/vp9_encodeframe.c b/vp9/encoder/vp9_encodeframe.c index b7f8cf49f..d2008a775 100644 --- a/vp9/encoder/vp9_encodeframe.c +++ b/vp9/encoder/vp9_encodeframe.c @@ -463,7 +463,7 @@ static void update_state(VP9_COMP *cpi, MACROBLOCK *x, if (mb_mode == B_PRED) { for (i = 0; i < 16; i++) { xd->block[i].bmi.as_mode = xd->mode_info_context->bmi[i].as_mode; - assert(xd->block[i].bmi.as_mode.first < MB_MODE_COUNT); + assert(xd->block[i].bmi.as_mode.first < B_MODE_COUNT); } } else if (mb_mode == I8X8_PRED) { for (i = 0; i < 16; i++) { @@ -931,7 +931,7 @@ static void pick_sb_modes (VP9_COMP *cpi, xd->mode_info_context->mbmi.segment_id = 0; x->active_ptr = cpi->active_map + map_index; - + cpi->update_context = 0; // TODO Do we need this now?? /* Find best coding mode & reconstruct the MB so it is available diff --git a/vp9/encoder/vp9_mbgraph.c b/vp9/encoder/vp9_mbgraph.c index 059572e24..8511bc572 100644 --- a/vp9/encoder/vp9_mbgraph.c +++ b/vp9/encoder/vp9_mbgraph.c @@ -367,8 +367,6 @@ static void separate_arf_mbs(VP9_COMP *cpi) { CHECK_MEM_ERROR(arf_not_zz, vpx_calloc(cm->mb_rows * cm->mb_cols * sizeof(*arf_not_zz), 1)); - vpx_memset(arf_not_zz, 0, sizeof(arf_not_zz)); - // We are not interested in results beyond the alt ref itself. if (n_frames > cpi->frames_till_gf_update_due) n_frames = cpi->frames_till_gf_update_due; diff --git a/vp9/encoder/x86/vp9_variance_mmx.c b/vp9/encoder/x86/vp9_variance_mmx.c index fa69130a1..bad1cfa74 100644 --- a/vp9/encoder/x86/vp9_variance_mmx.c +++ b/vp9/encoder/x86/vp9_variance_mmx.c @@ -191,31 +191,7 @@ unsigned int vp9_variance8x16_mmx( } - - - -/////////////////////////////////////////////////////////////////////////// -// the mmx function that does the bilinear filtering and var calculation // -// int one pass // -/////////////////////////////////////////////////////////////////////////// -DECLARE_ALIGNED(16, const short, vp9_bilinear_filters_mmx[16][8]) = { - { 128, 128, 128, 128, 0, 0, 0, 0 }, - { 120, 120, 120, 120, 8, 8, 8, 8 }, - { 112, 112, 112, 112, 16, 16, 16, 16 }, - { 104, 104, 104, 104, 24, 24, 24, 24 }, - { 96, 96, 96, 96, 32, 32, 32, 32 }, - { 88, 88, 88, 88, 40, 40, 40, 40 }, - { 80, 80, 80, 80, 48, 48, 48, 48 }, - { 72, 72, 72, 72, 56, 56, 56, 56 }, - { 64, 64, 64, 64, 64, 64, 64, 64 }, - { 56, 56, 56, 56, 72, 72, 72, 72 }, - { 48, 48, 48, 48, 80, 80, 80, 80 }, - { 40, 40, 40, 40, 88, 88, 88, 88 }, - { 32, 32, 32, 32, 96, 96, 96, 96 }, - { 24, 24, 24, 24, 104, 104, 104, 104 }, - { 16, 16, 16, 16, 112, 112, 112, 112 }, - { 8, 8, 8, 8, 120, 120, 120, 120 } -}; +DECLARE_ALIGNED(16, extern const short, vp9_bilinear_filters_mmx[16][8]); unsigned int vp9_sub_pixel_variance4x4_mmx ( diff --git a/vp9/encoder/x86/vp9_variance_sse2.c b/vp9/encoder/x86/vp9_variance_sse2.c index 280efcd7a..36fae6e8c 100644 --- a/vp9/encoder/x86/vp9_variance_sse2.c +++ b/vp9/encoder/x86/vp9_variance_sse2.c @@ -137,7 +137,7 @@ void vp9_half_vert_variance16x_h_sse2 unsigned int *sumsquared ); -DECLARE_ALIGNED(16, extern short, vp9_bilinear_filters_mmx[16][8]); +DECLARE_ALIGNED(16, extern const short, vp9_bilinear_filters_mmx[16][8]); unsigned int vp9_variance4x4_wmt( const unsigned char *src_ptr, diff --git a/vp9/vp9cx.mk b/vp9/vp9cx.mk index 4ab0a9696..020a5a2d6 100644 --- a/vp9/vp9cx.mk +++ b/vp9/vp9cx.mk @@ -82,7 +82,6 @@ VP9_CX_SRCS-yes += encoder/vp9_mbgraph.h VP9_CX_SRCS-$(ARCH_X86)$(ARCH_X86_64) += encoder/x86/vp9_mcomp_x86.h VP9_CX_SRCS-$(ARCH_X86)$(ARCH_X86_64) += encoder/x86/vp9_quantize_x86.h -VP9_CX_SRCS-$(ARCH_X86)$(ARCH_X86_64) += encoder/x86/vp9_temporal_filter_x86.h VP9_CX_SRCS-$(ARCH_X86)$(ARCH_X86_64) += encoder/x86/vp9_x86_csystemdependent.c VP9_CX_SRCS-$(HAVE_MMX) += encoder/x86/vp9_variance_mmx.c VP9_CX_SRCS-$(HAVE_MMX) += encoder/x86/vp9_variance_impl_mmx.asm |