diff options
author | Johann <johannkoenig@google.com> | 2013-04-02 11:20:55 -0700 |
---|---|---|
committer | Johann <johannkoenig@google.com> | 2013-04-02 11:20:55 -0700 |
commit | 6c147b9d937fa57494546e69b550ba5f6fe406fc (patch) | |
tree | 5ffd50f2da321317194f20462d08d7dfb918a1f8 /vp9/common/x86/vp9_sadmxn_sse2.c | |
parent | 416bd7529c848fef0170d219691a70f1a81f7a47 (diff) | |
download | libvpx-6c147b9d937fa57494546e69b550ba5f6fe406fc.tar libvpx-6c147b9d937fa57494546e69b550ba5f6fe406fc.tar.gz libvpx-6c147b9d937fa57494546e69b550ba5f6fe406fc.tar.bz2 libvpx-6c147b9d937fa57494546e69b550ba5f6fe406fc.zip |
vp9_sadmxn_x86 only contains SSE2 functions
Rename the file and clean up includes. In the future we would like to
pattern match the files which need additional compiler flags.
Change-Id: I2c76256467f392a78dd4ccc71e6e0a580e158e56
Diffstat (limited to 'vp9/common/x86/vp9_sadmxn_sse2.c')
-rw-r--r-- | vp9/common/x86/vp9_sadmxn_sse2.c | 95 |
1 files changed, 95 insertions, 0 deletions
diff --git a/vp9/common/x86/vp9_sadmxn_sse2.c b/vp9/common/x86/vp9_sadmxn_sse2.c new file mode 100644 index 000000000..ed873a5ba --- /dev/null +++ b/vp9/common/x86/vp9_sadmxn_sse2.c @@ -0,0 +1,95 @@ +/* + * Copyright (c) 2012 The WebM project authors. All Rights Reserved. + * + * Use of this source code is governed by a BSD-style license + * that can be found in the LICENSE file in the root of the source + * tree. An additional intellectual property rights grant can be found + * in the file PATENTS. All contributing project authors may + * be found in the AUTHORS file in the root of the source tree. + */ + +#include <emmintrin.h> /* SSE2 */ +#include "vpx/vpx_integer.h" +#include "vpx_ports/emmintrin_compat.h" + +unsigned int vp9_sad16x3_sse2( + const unsigned char *src_ptr, + int src_stride, + const unsigned char *ref_ptr, + int ref_stride) { + __m128i s0, s1, s2; + __m128i r0, r1, r2; + __m128i sad; + + s0 = _mm_loadu_si128((const __m128i *)(src_ptr + 0 * src_stride)); + s1 = _mm_loadu_si128((const __m128i *)(src_ptr + 1 * src_stride)); + s2 = _mm_loadu_si128((const __m128i *)(src_ptr + 2 * src_stride)); + + r0 = _mm_loadu_si128((const __m128i *)(ref_ptr + 0 * ref_stride)); + r1 = _mm_loadu_si128((const __m128i *)(ref_ptr + 1 * ref_stride)); + r2 = _mm_loadu_si128((const __m128i *)(ref_ptr + 2 * ref_stride)); + + sad = _mm_sad_epu8(s0, r0); + sad = _mm_add_epi16(sad, _mm_sad_epu8(s1, r1)); + sad = _mm_add_epi16(sad, _mm_sad_epu8(s2, r2)); + sad = _mm_add_epi16(sad, _mm_srli_si128(sad, 8)); + + return _mm_cvtsi128_si32(sad); +} + +unsigned int vp9_sad3x16_sse2( + const unsigned char *src_ptr, + int src_stride, + const unsigned char *ref_ptr, + int ref_stride) { + int r; + __m128i s0, s1, s2, s3; + __m128i r0, r1, r2, r3; + __m128i sad = _mm_setzero_si128(); + __m128i mask; + const int offset = (uintptr_t)src_ptr & 3; + + /* In current use case, the offset is 1 if CONFIG_SUBPELREFMV is off. + * Here, for offset=1, we adjust src_ptr to be 4-byte aligned. Then, movd + * takes much less time. + */ + if (offset == 1) + src_ptr -= 1; + + /* mask = 0xffffffffffff0000ffffffffffff0000 */ + mask = _mm_cmpeq_epi32(sad, sad); + mask = _mm_slli_epi64(mask, 16); + + for (r = 0; r < 16; r += 4) { + s0 = _mm_cvtsi32_si128 (*(const int *)(src_ptr + 0 * src_stride)); + s1 = _mm_cvtsi32_si128 (*(const int *)(src_ptr + 1 * src_stride)); + s2 = _mm_cvtsi32_si128 (*(const int *)(src_ptr + 2 * src_stride)); + s3 = _mm_cvtsi32_si128 (*(const int *)(src_ptr + 3 * src_stride)); + r0 = _mm_cvtsi32_si128 (*(const int *)(ref_ptr + 0 * ref_stride)); + r1 = _mm_cvtsi32_si128 (*(const int *)(ref_ptr + 1 * ref_stride)); + r2 = _mm_cvtsi32_si128 (*(const int *)(ref_ptr + 2 * ref_stride)); + r3 = _mm_cvtsi32_si128 (*(const int *)(ref_ptr + 3 * ref_stride)); + + s0 = _mm_unpacklo_epi8(s0, s1); + r0 = _mm_unpacklo_epi8(r0, r1); + s2 = _mm_unpacklo_epi8(s2, s3); + r2 = _mm_unpacklo_epi8(r2, r3); + s0 = _mm_unpacklo_epi64(s0, s2); + r0 = _mm_unpacklo_epi64(r0, r2); + + // throw out extra byte + if (offset == 1) + s0 = _mm_and_si128(s0, mask); + else + s0 = _mm_slli_epi64(s0, 16); + r0 = _mm_slli_epi64(r0, 16); + + sad = _mm_add_epi16(sad, _mm_sad_epu8(s0, r0)); + + src_ptr += src_stride*4; + ref_ptr += ref_stride*4; + } + + sad = _mm_add_epi16(sad, _mm_srli_si128(sad, 8)); + return _mm_cvtsi128_si32(sad); +} |