summaryrefslogtreecommitdiff
path: root/vp8/encoder/x86/denoising_sse2.c
diff options
context:
space:
mode:
authorChristian Duvivier <cduvivier@google.com>2012-05-21 07:54:20 -0700
committerJim Bankoski <jimbankoski@google.com>2012-05-21 07:54:20 -0700
commit38ddb426d097349c0926977068f97655d2d77f26 (patch)
treec26516a481194c0610d4975f6f15ead11b1cff2d /vp8/encoder/x86/denoising_sse2.c
parent51fe9168d5fb570abc93b3d15dc9a4e5c320cd95 (diff)
downloadlibvpx-38ddb426d097349c0926977068f97655d2d77f26.tar
libvpx-38ddb426d097349c0926977068f97655d2d77f26.tar.gz
libvpx-38ddb426d097349c0926977068f97655d2d77f26.tar.bz2
libvpx-38ddb426d097349c0926977068f97655d2d77f26.zip
Inline Intrinsic optimized Denoiser
Faster version of denoiser, cut cost by 1.7x for C path, by 3.3x for SSE2 path. Change-Id: I154786308550763bc0e3497e5fa5bfd1ce651beb
Diffstat (limited to 'vp8/encoder/x86/denoising_sse2.c')
-rw-r--r--vp8/encoder/x86/denoising_sse2.c128
1 files changed, 128 insertions, 0 deletions
diff --git a/vp8/encoder/x86/denoising_sse2.c b/vp8/encoder/x86/denoising_sse2.c
new file mode 100644
index 000000000..9bf938275
--- /dev/null
+++ b/vp8/encoder/x86/denoising_sse2.c
@@ -0,0 +1,128 @@
+/*
+ * Copyright (c) 2012 The WebM project authors. All Rights Reserved.
+ *
+ * Use of this source code is governed by a BSD-style license
+ * that can be found in the LICENSE file in the root of the source
+ * tree. An additional intellectual property rights grant can be found
+ * in the file PATENTS. All contributing project authors may
+ * be found in the AUTHORS file in the root of the source tree.
+ */
+
+#include "vp8/encoder/denoising.h"
+
+#include "vp8/common/reconinter.h"
+#include "vpx/vpx_integer.h"
+#include "vpx_mem/vpx_mem.h"
+#include "vpx_rtcd.h"
+
+#include <emmintrin.h>
+
+void vp8_denoiser_filter_sse2(YV12_BUFFER_CONFIG *mc_running_avg,
+ YV12_BUFFER_CONFIG *running_avg,
+ MACROBLOCK *signal, unsigned int motion_magnitude,
+ int y_offset, int uv_offset)
+{
+ unsigned char *sig = signal->thismb;
+ int sig_stride = 16;
+ unsigned char *mc_running_avg_y = mc_running_avg->y_buffer + y_offset;
+ int mc_avg_y_stride = mc_running_avg->y_stride;
+ unsigned char *running_avg_y = running_avg->y_buffer + y_offset;
+ int avg_y_stride = running_avg->y_stride;
+ const union coeff_pair *LUT = vp8_get_filter_coeff_LUT(motion_magnitude);
+ int r, c;
+
+ for (r = 0; r < 16; ++r)
+ {
+ __m128i filter_coefficient_00, filter_coefficient_04;
+ __m128i filter_coefficient_08, filter_coefficient_12;
+ __m128i v_sig0, v_sig1;
+ __m128i v_mc_running_avg_y0, v_mc_running_avg_y1;
+ __m128i state0, state1, state2, state3;
+ __m128i res0, res1, res2, res3;
+ __m128i v_running_avg_y;
+ __m128i diff0, diff1, diff0sq, diff1sq, diff_sq;
+ const __m128i kNOISE_DIFF2_THRESHOLD =
+ _mm_set1_epi8(NOISE_DIFF2_THRESHOLD);
+ __m128i take_running, p0, p1, p2;
+ const __m128i k_zero = _mm_set1_epi16(0);
+ const __m128i k_128 = _mm_set1_epi32(128);
+
+ // Calculate absolute differences
+ DECLARE_ALIGNED_ARRAY(16,unsigned char,abs_diff,16);
+ DECLARE_ALIGNED_ARRAY(16,uint32_t,filter_coefficient,16);
+ __m128i v_sig = _mm_loadu_si128((__m128i *)(&sig[0]));
+ __m128i v_mc_running_avg_y = _mm_loadu_si128(
+ (__m128i *)(&mc_running_avg_y[0]));
+ __m128i a_minus_b = _mm_subs_epu8(v_sig, v_mc_running_avg_y);
+ __m128i b_minus_a = _mm_subs_epu8(v_mc_running_avg_y, v_sig);
+ __m128i v_abs_diff = _mm_adds_epu8(a_minus_b, b_minus_a);
+ _mm_store_si128((__m128i *)(&abs_diff[0]), v_abs_diff);
+
+ // Use LUT to get filter coefficients (two 16b value; f and 256-f)
+ for (c = 0; c < 16; ++c)
+ {
+ filter_coefficient[c] = LUT[abs_diff[c]].as_int;
+ }
+
+ // Filtering...
+ // load filter coefficients (two 16b value; f and 256-f)
+ filter_coefficient_00 = _mm_load_si128(
+ (__m128i *)(&filter_coefficient[ 0]));
+ filter_coefficient_04 = _mm_load_si128(
+ (__m128i *)(&filter_coefficient[ 4]));
+ filter_coefficient_08 = _mm_load_si128(
+ (__m128i *)(&filter_coefficient[ 8]));
+ filter_coefficient_12 = _mm_load_si128(
+ (__m128i *)(&filter_coefficient[12]));
+
+ // expand sig from 8b to 16b
+ v_sig0 = _mm_unpacklo_epi8(v_sig, k_zero);
+ v_sig1 = _mm_unpackhi_epi8(v_sig, k_zero);
+ // expand mc_running_avg_y from 8b to 16b
+ v_mc_running_avg_y0 = _mm_unpacklo_epi8(v_mc_running_avg_y, k_zero);
+ v_mc_running_avg_y1 = _mm_unpackhi_epi8(v_mc_running_avg_y, k_zero);
+ // interleave sig and mc_running_avg_y for upcoming multiply-add
+ state0 = _mm_unpacklo_epi16(v_mc_running_avg_y0, v_sig0);
+ state1 = _mm_unpackhi_epi16(v_mc_running_avg_y0, v_sig0);
+ state2 = _mm_unpacklo_epi16(v_mc_running_avg_y1, v_sig1);
+ state3 = _mm_unpackhi_epi16(v_mc_running_avg_y1, v_sig1);
+ // blend values
+ res0 = _mm_madd_epi16(filter_coefficient_00, state0);
+ res1 = _mm_madd_epi16(filter_coefficient_04, state1);
+ res2 = _mm_madd_epi16(filter_coefficient_08, state2);
+ res3 = _mm_madd_epi16(filter_coefficient_12, state3);
+ res0 = _mm_add_epi32(res0, k_128);
+ res1 = _mm_add_epi32(res1, k_128);
+ res2 = _mm_add_epi32(res2, k_128);
+ res3 = _mm_add_epi32(res3, k_128);
+ res0 = _mm_srai_epi32(res0, 8);
+ res1 = _mm_srai_epi32(res1, 8);
+ res2 = _mm_srai_epi32(res2, 8);
+ res3 = _mm_srai_epi32(res3, 8);
+ // combine the 32b results into a single 8b vector
+ res0 = _mm_packs_epi32(res0, res1);
+ res2 = _mm_packs_epi32(res2, res3);
+ v_running_avg_y = _mm_packus_epi16(res0, res2);
+
+ // Depending on the magnitude of the difference between the signal and
+ // filtered version, either replace the signal by the filtered one or
+ // update the filter state with the signal when the change in a pixel
+ // isn't classified as noise.
+ diff0 = _mm_sub_epi16(v_sig0, res0);
+ diff1 = _mm_sub_epi16(v_sig1, res2);
+ diff0sq = _mm_mullo_epi16(diff0, diff0);
+ diff1sq = _mm_mullo_epi16(diff1, diff1);
+ diff_sq = _mm_packus_epi16(diff0sq, diff1sq);
+ take_running = _mm_cmplt_epi8(diff_sq, kNOISE_DIFF2_THRESHOLD);
+ p0 = _mm_and_si128(take_running, v_running_avg_y);
+ p1 = _mm_andnot_si128(take_running, v_sig);
+ p2 = _mm_or_si128(p0, p1);
+ _mm_storeu_si128((__m128i *)(&running_avg_y[0]), p2);
+ _mm_storeu_si128((__m128i *)(&sig[0]), p2);
+
+ // Update pointers for next iteration.
+ sig += sig_stride;
+ mc_running_avg_y += mc_avg_y_stride;
+ running_avg_y += avg_y_stride;
+ }
+}