summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorJohn Koleszar <jkoleszar@google.com>2012-11-07 14:32:26 -0800
committerJohn Koleszar <jkoleszar@google.com>2012-11-07 14:32:26 -0800
commit8959c8b11d11ad44c7bf0cc49e54974e161d374c (patch)
tree9fc43cd73c50c016c927eca525f8ef573e86fa52
parent2c08c28191d65bb5a54c78567c2f54960b203f6d (diff)
parent5338d983d655664614fffc29142331a928ce018f (diff)
downloadlibvpx-8959c8b11d11ad44c7bf0cc49e54974e161d374c.tar
libvpx-8959c8b11d11ad44c7bf0cc49e54974e161d374c.tar.gz
libvpx-8959c8b11d11ad44c7bf0cc49e54974e161d374c.tar.bz2
libvpx-8959c8b11d11ad44c7bf0cc49e54974e161d374c.zip
Merge with upstream experimental changes (2)
Include upstream changes (variance fixes) into the merged code base. Change-Id: I4182654c1411c1b15cd23235d3822702613abce1
-rw-r--r--test/test.mk4
-rw-r--r--test/variance_test.cc123
-rw-r--r--vp9/common/findnearmv.c4
-rw-r--r--vp9/encoder/variance_c.c13
-rw-r--r--vp9/encoder/x86/variance_mmx.c20
-rw-r--r--vp9/encoder/x86/variance_sse2.c26
-rw-r--r--vp9/encoder/x86/variance_ssse3.c4
7 files changed, 160 insertions, 34 deletions
diff --git a/test/test.mk b/test/test.mk
index 5e9565c33..4fb464e64 100644
--- a/test/test.mk
+++ b/test/test.mk
@@ -1,5 +1,6 @@
-LIBVPX_TEST_SRCS-yes += acm_random.h
LIBVPX_TEST_SRCS-yes += test.mk
+LIBVPX_TEST_SRCS-yes += acm_random.h
+
LIBVPX_TEST_SRCS-yes += test_libvpx.cc
LIBVPX_TEST_SRCS-yes += util.h
LIBVPX_TEST_SRCS-yes += video_source.h
@@ -64,6 +65,7 @@ LIBVPX_TEST_SRCS-$(CONFIG_VP9_ENCODER) += fdct4x4_test.cc
LIBVPX_TEST_SRCS-$(CONFIG_VP9_ENCODER) += fdct8x8_test.cc
#LIBVPX_TEST_SRCS-$(CONFIG_VP9_ENCODER) += dct16x16_test.cc
LIBVPX_TEST_SRCS-yes += idct8x8_test.cc
+LIBVPX_TEST_SRCS-yes += variance_test.cc
endif # VP9
diff --git a/test/variance_test.cc b/test/variance_test.cc
new file mode 100644
index 000000000..f838c05e2
--- /dev/null
+++ b/test/variance_test.cc
@@ -0,0 +1,123 @@
+/*
+ * Copyright (c) 2012 The WebM project authors. All Rights Reserved.
+ *
+ * Use of this source code is governed by a BSD-style license
+ * that can be found in the LICENSE file in the root of the source
+ * tree. An additional intellectual property rights grant can be found
+ * in the file PATENTS. All contributing project authors may
+ * be found in the AUTHORS file in the root of the source tree.
+ */
+#include <stdlib.h>
+#include <new>
+
+#include "third_party/googletest/src/include/gtest/gtest.h"
+
+#include "vpx_config.h"
+extern "C" {
+#include "vp9/encoder/variance.h"
+#include "vpx/vpx_integer.h"
+#include "vpx_rtcd.h"
+}
+
+namespace {
+
+using ::std::tr1::get;
+using ::std::tr1::make_tuple;
+using ::std::tr1::tuple;
+
+class VP9VarianceTest :
+ public ::testing::TestWithParam<tuple<int, int, vp9_variance_fn_t> > {
+ public:
+ virtual void SetUp() {
+ const tuple<int, int, vp9_variance_fn_t>& params = GetParam();
+ width_ = get<0>(params);
+ height_ = get<1>(params);
+ variance_ = get<2>(params);
+
+ block_size_ = width_ * height_;
+ src_ = new uint8_t[block_size_];
+ ref_ = new uint8_t[block_size_];
+ ASSERT_TRUE(src_ != NULL);
+ ASSERT_TRUE(ref_ != NULL);
+ }
+
+ virtual void TearDown() {
+ delete[] src_;
+ delete[] ref_;
+ }
+
+ protected:
+ uint8_t* src_;
+ uint8_t* ref_;
+ int width_;
+ int height_;
+ int block_size_;
+ vp9_variance_fn_t variance_;
+};
+
+TEST_P(VP9VarianceTest, Zero) {
+ for (int i = 0; i <= 255; ++i) {
+ memset(src_, i, block_size_);
+ for (int j = 0; j <= 255; ++j) {
+ memset(ref_, j, block_size_);
+ unsigned int sse;
+ const unsigned int var = variance_(src_, width_, ref_, width_, &sse);
+ EXPECT_EQ(0u, var) << "src values: " << i << "ref values: " << j;
+ }
+ }
+}
+
+TEST_P(VP9VarianceTest, OneQuarter) {
+ memset(src_, 255, block_size_);
+ const int half = block_size_ / 2;
+ memset(ref_, 255, half);
+ memset(ref_ + half, 0, half);
+ unsigned int sse;
+ const unsigned int var = variance_(src_, width_, ref_, width_, &sse);
+ const unsigned int expected = block_size_ * 255 * 255 / 4;
+ EXPECT_EQ(expected, var);
+}
+
+const vp9_variance_fn_t variance4x4_c = vp9_variance4x4_c;
+const vp9_variance_fn_t variance8x8_c = vp9_variance8x8_c;
+const vp9_variance_fn_t variance8x16_c = vp9_variance8x16_c;
+const vp9_variance_fn_t variance16x8_c = vp9_variance16x8_c;
+const vp9_variance_fn_t variance16x16_c = vp9_variance16x16_c;
+INSTANTIATE_TEST_CASE_P(
+ C, VP9VarianceTest,
+ ::testing::Values(make_tuple(4, 4, variance4x4_c),
+ make_tuple(8, 8, variance8x8_c),
+ make_tuple(8, 16, variance8x16_c),
+ make_tuple(16, 8, variance16x8_c),
+ make_tuple(16, 16, variance16x16_c)));
+
+#if HAVE_MMX
+const vp9_variance_fn_t variance4x4_mmx = vp9_variance4x4_mmx;
+const vp9_variance_fn_t variance8x8_mmx = vp9_variance8x8_mmx;
+const vp9_variance_fn_t variance8x16_mmx = vp9_variance8x16_mmx;
+const vp9_variance_fn_t variance16x8_mmx = vp9_variance16x8_mmx;
+const vp9_variance_fn_t variance16x16_mmx = vp9_variance16x16_mmx;
+INSTANTIATE_TEST_CASE_P(
+ MMX, VP9VarianceTest,
+ ::testing::Values(make_tuple(4, 4, variance4x4_mmx),
+ make_tuple(8, 8, variance8x8_mmx),
+ make_tuple(8, 16, variance8x16_mmx),
+ make_tuple(16, 8, variance16x8_mmx),
+ make_tuple(16, 16, variance16x16_mmx)));
+#endif
+
+#if HAVE_SSE2
+const vp9_variance_fn_t variance4x4_wmt = vp9_variance4x4_wmt;
+const vp9_variance_fn_t variance8x8_wmt = vp9_variance8x8_wmt;
+const vp9_variance_fn_t variance8x16_wmt = vp9_variance8x16_wmt;
+const vp9_variance_fn_t variance16x8_wmt = vp9_variance16x8_wmt;
+const vp9_variance_fn_t variance16x16_wmt = vp9_variance16x16_wmt;
+INSTANTIATE_TEST_CASE_P(
+ SSE2, VP9VarianceTest,
+ ::testing::Values(make_tuple(4, 4, variance4x4_wmt),
+ make_tuple(8, 8, variance8x8_wmt),
+ make_tuple(8, 16, variance8x16_wmt),
+ make_tuple(16, 8, variance16x8_wmt),
+ make_tuple(16, 16, variance16x16_wmt)));
+#endif
+} // namespace
diff --git a/vp9/common/findnearmv.c b/vp9/common/findnearmv.c
index cc560cf3a..4f3d15a82 100644
--- a/vp9/common/findnearmv.c
+++ b/vp9/common/findnearmv.c
@@ -197,7 +197,7 @@ unsigned int vp9_variance2x16_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 2, 16, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 5));
+ return (var - (((unsigned int)avg * avg) >> 5));
}
unsigned int vp9_variance16x2_c(const unsigned char *src_ptr,
@@ -210,7 +210,7 @@ unsigned int vp9_variance16x2_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 2, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 5));
+ return (var - (((unsigned int)avg * avg) >> 5));
}
unsigned int vp9_sub_pixel_variance16x2_c(const unsigned char *src_ptr,
diff --git a/vp9/encoder/variance_c.c b/vp9/encoder/variance_c.c
index 62de70301..8ac956e95 100644
--- a/vp9/encoder/variance_c.c
+++ b/vp9/encoder/variance_c.c
@@ -37,7 +37,8 @@ unsigned int vp9_variance32x32_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 32, 32, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 10));
+ // TODO(rbultje): in extreme cases these products will rollover.
+ return (var - (((unsigned int)avg * avg) >> 10));
}
#endif
@@ -51,7 +52,7 @@ unsigned int vp9_variance16x16_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 16, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 8));
+ return (var - (((unsigned int)avg * avg) >> 8));
}
unsigned int vp9_variance8x16_c(const unsigned char *src_ptr,
@@ -64,7 +65,7 @@ unsigned int vp9_variance8x16_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 16, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 7));
+ return (var - (((unsigned int)avg * avg) >> 7));
}
unsigned int vp9_variance16x8_c(const unsigned char *src_ptr,
@@ -77,7 +78,7 @@ unsigned int vp9_variance16x8_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 8, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 7));
+ return (var - (((unsigned int)avg * avg) >> 7));
}
@@ -91,7 +92,7 @@ unsigned int vp9_variance8x8_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 8, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 6));
+ return (var - (((unsigned int)avg * avg) >> 6));
}
unsigned int vp9_variance4x4_c(const unsigned char *src_ptr,
@@ -104,7 +105,7 @@ unsigned int vp9_variance4x4_c(const unsigned char *src_ptr,
variance(src_ptr, source_stride, ref_ptr, recon_stride, 4, 4, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 4));
+ return (var - (((unsigned int)avg * avg) >> 4));
}
diff --git a/vp9/encoder/x86/variance_mmx.c b/vp9/encoder/x86/variance_mmx.c
index 3fedc6b5c..3b5f2a1b9 100644
--- a/vp9/encoder/x86/variance_mmx.c
+++ b/vp9/encoder/x86/variance_mmx.c
@@ -89,7 +89,7 @@ unsigned int vp9_variance4x4_mmx(
vp9_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 4));
+ return (var - (((unsigned int)avg * avg) >> 4));
}
@@ -105,7 +105,7 @@ unsigned int vp9_variance8x8_mmx(
vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 6));
+ return (var - (((unsigned int)avg * avg) >> 6));
}
@@ -148,7 +148,7 @@ unsigned int vp9_variance16x16_mmx(
var = sse0 + sse1 + sse2 + sse3;
avg = sum0 + sum1 + sum2 + sum3;
*sse = var;
- return (var - ((avg * avg) >> 8));
+ return (var - (((unsigned int)avg * avg) >> 8));
}
unsigned int vp9_variance16x8_mmx(
@@ -166,7 +166,7 @@ unsigned int vp9_variance16x8_mmx(
var = sse0 + sse1;
avg = sum0 + sum1;
*sse = var;
- return (var - ((avg * avg) >> 7));
+ return (var - (((unsigned int)avg * avg) >> 7));
}
@@ -187,7 +187,7 @@ unsigned int vp9_variance8x16_mmx(
avg = sum0 + sum1;
*sse = var;
- return (var - ((avg * avg) >> 7));
+ return (var - (((unsigned int)avg * avg) >> 7));
}
@@ -237,7 +237,7 @@ unsigned int vp9_sub_pixel_variance4x4_mmx
&xsum, &xxsum
);
*sse = xxsum;
- return (xxsum - ((xsum * xsum) >> 4));
+ return (xxsum - (((unsigned int)xsum * xsum) >> 4));
}
@@ -261,7 +261,7 @@ unsigned int vp9_sub_pixel_variance8x8_mmx
&xsum, &xxsum
);
*sse = xxsum;
- return (xxsum - ((xsum * xsum) >> 6));
+ return (xxsum - (((unsigned int)xsum * xsum) >> 6));
}
unsigned int vp9_sub_pixel_variance16x16_mmx
@@ -296,7 +296,7 @@ unsigned int vp9_sub_pixel_variance16x16_mmx
xxsum0 += xxsum1;
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 8));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 8));
}
@@ -347,7 +347,7 @@ unsigned int vp9_sub_pixel_variance16x8_mmx
xxsum0 += xxsum1;
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 7));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 7));
}
unsigned int vp9_sub_pixel_variance8x16_mmx
@@ -369,7 +369,7 @@ unsigned int vp9_sub_pixel_variance8x16_mmx
&xsum, &xxsum
);
*sse = xxsum;
- return (xxsum - ((xsum * xsum) >> 7));
+ return (xxsum - (((unsigned int)xsum * xsum) >> 7));
}
diff --git a/vp9/encoder/x86/variance_sse2.c b/vp9/encoder/x86/variance_sse2.c
index e63c53a53..9f897da51 100644
--- a/vp9/encoder/x86/variance_sse2.c
+++ b/vp9/encoder/x86/variance_sse2.c
@@ -150,7 +150,7 @@ unsigned int vp9_variance4x4_wmt(
vp9_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 4));
+ return (var - (((unsigned int)avg * avg) >> 4));
}
@@ -166,7 +166,7 @@ unsigned int vp9_variance8x8_wmt
vp9_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg);
*sse = var;
- return (var - ((avg * avg) >> 6));
+ return (var - (((unsigned int)avg * avg) >> 6));
}
@@ -184,7 +184,7 @@ unsigned int vp9_variance16x16_wmt
vp9_get16x16var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, &sum0);
*sse = sse0;
- return (sse0 - ((sum0 * sum0) >> 8));
+ return (sse0 - (((unsigned int)sum0 * sum0) >> 8));
}
unsigned int vp9_mse16x16_wmt(
const unsigned char *src_ptr,
@@ -218,7 +218,7 @@ unsigned int vp9_variance16x8_wmt
var = sse0 + sse1;
avg = sum0 + sum1;
*sse = var;
- return (var - ((avg * avg) >> 7));
+ return (var - (((unsigned int)avg * avg) >> 7));
}
@@ -238,7 +238,7 @@ unsigned int vp9_variance8x16_wmt
var = sse0 + sse1;
avg = sum0 + sum1;
*sse = var;
- return (var - ((avg * avg) >> 7));
+ return (var - (((unsigned int)avg * avg) >> 7));
}
@@ -261,7 +261,7 @@ unsigned int vp9_sub_pixel_variance4x4_wmt
&xsum, &xxsum
);
*sse = xxsum;
- return (xxsum - ((xsum * xsum) >> 4));
+ return (xxsum - (((unsigned int)xsum * xsum) >> 4));
}
@@ -302,7 +302,7 @@ unsigned int vp9_sub_pixel_variance8x8_wmt
}
*sse = xxsum;
- return (xxsum - ((xsum * xsum) >> 6));
+ return (xxsum - (((unsigned int)xsum * xsum) >> 6));
}
unsigned int vp9_sub_pixel_variance16x16_wmt
@@ -355,7 +355,7 @@ unsigned int vp9_sub_pixel_variance16x16_wmt
}
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 8));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 8));
}
unsigned int vp9_sub_pixel_mse16x16_wmt(
@@ -417,7 +417,7 @@ unsigned int vp9_sub_pixel_variance16x8_wmt
}
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 7));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 7));
}
unsigned int vp9_sub_pixel_variance8x16_wmt
@@ -457,7 +457,7 @@ unsigned int vp9_sub_pixel_variance8x16_wmt
}
*sse = xxsum;
- return (xxsum - ((xsum * xsum) >> 7));
+ return (xxsum - (((unsigned int)xsum * xsum) >> 7));
}
@@ -476,7 +476,7 @@ unsigned int vp9_variance_halfpixvar16x16_h_wmt(
&xsum0, &xxsum0);
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 8));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 8));
}
@@ -494,7 +494,7 @@ unsigned int vp9_variance_halfpixvar16x16_v_wmt(
&xsum0, &xxsum0);
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 8));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 8));
}
@@ -513,5 +513,5 @@ unsigned int vp9_variance_halfpixvar16x16_hv_wmt(
&xsum0, &xxsum0);
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 8));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 8));
}
diff --git a/vp9/encoder/x86/variance_ssse3.c b/vp9/encoder/x86/variance_ssse3.c
index f37abb881..81b06d117 100644
--- a/vp9/encoder/x86/variance_ssse3.c
+++ b/vp9/encoder/x86/variance_ssse3.c
@@ -106,7 +106,7 @@ unsigned int vp9_sub_pixel_variance16x16_ssse3
}
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 8));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 8));
}
unsigned int vp9_sub_pixel_variance16x8_ssse3
@@ -147,5 +147,5 @@ unsigned int vp9_sub_pixel_variance16x8_ssse3
}
*sse = xxsum0;
- return (xxsum0 - ((xsum0 * xsum0) >> 7));
+ return (xxsum0 - (((unsigned int)xsum0 * xsum0) >> 7));
}