diff options
author | James Zern <jzern@google.com> | 2014-02-20 16:03:23 -0800 |
---|---|---|
committer | Gerrit Code Review <gerrit@gerrit.golo.chromium.org> | 2014-02-20 16:03:23 -0800 |
commit | e36cfc91a42bc78bce9b90986c93e8f2d1176c22 (patch) | |
tree | c9bfa6044b878b059cb2ed390441a3b05b6d6730 /vp9 | |
parent | e2f614be53f075301572057686956a2757831737 (diff) | |
parent | d12b39daab0a9dd1e3b9f0a4f04d2e456f0746f5 (diff) | |
download | libvpx-e36cfc91a42bc78bce9b90986c93e8f2d1176c22.tar libvpx-e36cfc91a42bc78bce9b90986c93e8f2d1176c22.tar.gz libvpx-e36cfc91a42bc78bce9b90986c93e8f2d1176c22.tar.bz2 libvpx-e36cfc91a42bc78bce9b90986c93e8f2d1176c22.zip |
Merge "vp9_subpel_variance_impl_intrin_avx2.c: make some tables static"
Diffstat (limited to 'vp9')
-rw-r--r-- | vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c | 53 |
1 files changed, 27 insertions, 26 deletions
diff --git a/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c b/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c index a8f98e94b..b8bfa8900 100644 --- a/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c +++ b/vp9/encoder/x86/vp9_subpel_variance_impl_intrin_avx2.c @@ -12,7 +12,7 @@ #include "vpx_ports/mem.h" #include "vp9/encoder/vp9_variance.h" -DECLARE_ALIGNED(32, const unsigned char, vp9_bilinear_filters_avx2[512])= { +DECLARE_ALIGNED(32, static const uint8_t, bilinear_filters_avx2[512]) = { 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 16, 0, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, @@ -44,7 +44,8 @@ DECLARE_ALIGNED(32, const unsigned char, vp9_bilinear_filters_avx2[512])= { 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 2, 14, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, - 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15}; + 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15, 1, 15 +}; unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src, int src_stride, @@ -137,12 +138,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src, int64_t y_offset64; y_offset64 = y_offset; y_offset64 <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2 + y_offset64)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + y_offset64)); #else y_offset <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2 + y_offset)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + y_offset)); #endif pw8 = _mm256_set1_epi16(8); for (i = 0; i < height ; i++) { @@ -283,12 +284,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src, int64_t y_offset64; y_offset64 = y_offset; y_offset64 <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2+y_offset64)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + y_offset64)); #else y_offset <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2 + y_offset)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + y_offset)); #endif pw8 = _mm256_set1_epi16(8); // load source and another source starting from the next @@ -354,12 +355,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src, int64_t x_offset64; x_offset64 = x_offset; x_offset64 <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2+x_offset64)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + x_offset64)); #else x_offset <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2 + x_offset)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + x_offset)); #endif pw8 = _mm256_set1_epi16(8); for (i = 0; i < height ; i++) { @@ -413,12 +414,12 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src, int64_t x_offset64; x_offset64 = x_offset; x_offset64 <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2+x_offset64)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + x_offset64)); #else x_offset <<= 5; - filter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2 + x_offset)); + filter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + x_offset)); #endif pw8 = _mm256_set1_epi16(8); // load source and another source starting from the next @@ -508,17 +509,17 @@ unsigned int vp9_sub_pixel_variance32xh_avx2(const uint8_t *src, x_offset64 <<= 5; y_offset64 = y_offset; y_offset64 <<= 5; - xfilter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2+x_offset64)); - yfilter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2+y_offset64)); + xfilter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + x_offset64)); + yfilter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + y_offset64)); #else x_offset <<= 5; - xfilter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2 + x_offset)); + xfilter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + x_offset)); y_offset <<= 5; - yfilter = _mm256_load_si256((__m256i const *) - (vp9_bilinear_filters_avx2 + y_offset)); + yfilter = _mm256_load_si256( + (__m256i const *)(bilinear_filters_avx2 + y_offset)); #endif pw8 = _mm256_set1_epi16(8); // load source and another source starting from the next |