/* * Copyright (c) 2014 The WebM project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #ifndef VP9_ENCODER_X86_VP9_DCT_SSE2_H_ #define VP9_ENCODER_X86_VP9_DCT_SSE2_H_ #ifdef __cplusplus extern "C" { #endif #define pair_set_epi32(a, b) \ _mm_set_epi32((int)(b), (int)(a), (int)(b), (int)(a)) void vp9_fdct4x4_sse2(const int16_t *input, tran_low_t *output, int stride); void vp9_fdct8x8_sse2(const int16_t *input, tran_low_t *output, int stride); void vp9_fdct16x16_sse2(const int16_t *input, tran_low_t *output, int stride); void vp9_highbd_fdct4x4_sse2(const int16_t *input, tran_low_t *output, int stride); void vp9_highbd_fdct8x8_sse2(const int16_t *input, tran_low_t *output, int stride); void vp9_highbd_fdct16x16_sse2(const int16_t *input, tran_low_t *output, int stride); static INLINE __m128i k_madd_epi32(__m128i a, __m128i b) { __m128i buf0, buf1; buf0 = _mm_mul_epu32(a, b); a = _mm_srli_epi64(a, 32); b = _mm_srli_epi64(b, 32); buf1 = _mm_mul_epu32(a, b); return _mm_add_epi64(buf0, buf1); } static INLINE __m128i k_packs_epi64(__m128i a, __m128i b) { __m128i buf0 = _mm_shuffle_epi32(a, _MM_SHUFFLE(0, 0, 2, 0)); __m128i buf1 = _mm_shuffle_epi32(b, _MM_SHUFFLE(0, 0, 2, 0)); return _mm_unpacklo_epi64(buf0, buf1); } static INLINE int check_epi16_overflow_x2(__m128i reg0, __m128i reg1) { const __m128i max_overflow = _mm_set1_epi16(0x7fff); const __m128i min_overflow = _mm_set1_epi16(0x8000); __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(reg0, max_overflow), _mm_cmpeq_epi16(reg0, min_overflow)); __m128i cmp1 = _mm_or_si128(_mm_cmpeq_epi16(reg1, max_overflow), _mm_cmpeq_epi16(reg1, min_overflow)); cmp0 = _mm_or_si128(cmp0, cmp1); return _mm_movemask_epi8(cmp0); } static INLINE int check_epi16_overflow_x4(__m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3) { const __m128i max_overflow = _mm_set1_epi16(0x7fff); const __m128i min_overflow = _mm_set1_epi16(0x8000); __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(reg0, max_overflow), _mm_cmpeq_epi16(reg0, min_overflow)); __m128i cmp1 = _mm_or_si128(_mm_cmpeq_epi16(reg1, max_overflow), _mm_cmpeq_epi16(reg1, min_overflow)); __m128i cmp2 = _mm_or_si128(_mm_cmpeq_epi16(reg2, max_overflow), _mm_cmpeq_epi16(reg2, min_overflow)); __m128i cmp3 = _mm_or_si128(_mm_cmpeq_epi16(reg3, max_overflow), _mm_cmpeq_epi16(reg3, min_overflow)); cmp0 = _mm_or_si128(_mm_or_si128(cmp0, cmp1), _mm_or_si128(cmp2, cmp3)); return _mm_movemask_epi8(cmp0); } static INLINE int check_epi16_overflow_x8(__m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, __m128i reg4, __m128i reg5, __m128i reg6, __m128i reg7) { int res0, res1; res0 = check_epi16_overflow_x4(reg0, reg1, reg2, reg3); res1 = check_epi16_overflow_x4(reg4, reg5, reg6, reg7); return res0 + res1; } static INLINE int check_epi16_overflow_x12(__m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, __m128i reg4, __m128i reg5, __m128i reg6, __m128i reg7, __m128i reg8, __m128i reg9, __m128i reg10, __m128i reg11) { int res0, res1; res0 = check_epi16_overflow_x4(reg0, reg1, reg2, reg3); res1 = check_epi16_overflow_x4(reg4, reg5, reg6, reg7); if (!res0) res0 = check_epi16_overflow_x4(reg8, reg9, reg10, reg11); return res0 + res1; } static INLINE int check_epi16_overflow_x16(__m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, __m128i reg4, __m128i reg5, __m128i reg6, __m128i reg7, __m128i reg8, __m128i reg9, __m128i reg10, __m128i reg11, __m128i reg12, __m128i reg13, __m128i reg14, __m128i reg15) { int res0, res1; res0 = check_epi16_overflow_x4(reg0, reg1, reg2, reg3); res1 = check_epi16_overflow_x4(reg4, reg5, reg6, reg7); if (!res0) { res0 = check_epi16_overflow_x4(reg8, reg9, reg10, reg11); if (!res1) res1 = check_epi16_overflow_x4(reg12, reg13, reg14, reg15); } return res0 + res1; } static INLINE int check_epi16_overflow_x32(__m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, __m128i reg4, __m128i reg5, __m128i reg6, __m128i reg7, __m128i reg8, __m128i reg9, __m128i reg10, __m128i reg11, __m128i reg12, __m128i reg13, __m128i reg14, __m128i reg15, __m128i reg16, __m128i reg17, __m128i reg18, __m128i reg19, __m128i reg20, __m128i reg21, __m128i reg22, __m128i reg23, __m128i reg24, __m128i reg25, __m128i reg26, __m128i reg27, __m128i reg28, __m128i reg29, __m128i reg30, __m128i reg31) { int res0, res1; res0 = check_epi16_overflow_x4(reg0, reg1, reg2, reg3); res1 = check_epi16_overflow_x4(reg4, reg5, reg6, reg7); if (!res0) { res0 = check_epi16_overflow_x4(reg8, reg9, reg10, reg11); if (!res1) { res1 = check_epi16_overflow_x4(reg12, reg13, reg14, reg15); if (!res0) { res0 = check_epi16_overflow_x4(reg16, reg17, reg18, reg19); if (!res1) { res1 = check_epi16_overflow_x4(reg20, reg21, reg22, reg23); if (!res0) { res0 = check_epi16_overflow_x4(reg24, reg25, reg26, reg27); if (!res1) res1 = check_epi16_overflow_x4(reg28, reg29, reg30, reg31); } } } } } return res0 + res1; } static INLINE int k_check_epi32_overflow_4(__m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, const __m128i* zero) { __m128i minus_one = _mm_set1_epi32(-1); // Check for overflows __m128i reg0_shifted = _mm_slli_epi64(reg0, 1); __m128i reg1_shifted = _mm_slli_epi64(reg1, 1); __m128i reg2_shifted = _mm_slli_epi64(reg2, 1); __m128i reg3_shifted = _mm_slli_epi64(reg3, 1); __m128i reg0_top_dwords = _mm_shuffle_epi32( reg0_shifted, _MM_SHUFFLE(0, 0, 3, 1)); __m128i reg1_top_dwords = _mm_shuffle_epi32( reg1_shifted, _MM_SHUFFLE(0, 0, 3, 1)); __m128i reg2_top_dwords = _mm_shuffle_epi32( reg2_shifted, _MM_SHUFFLE(0, 0, 3, 1)); __m128i reg3_top_dwords = _mm_shuffle_epi32( reg3_shifted, _MM_SHUFFLE(0, 0, 3, 1)); __m128i top_dwords_01 = _mm_unpacklo_epi64(reg0_top_dwords, reg1_top_dwords); __m128i top_dwords_23 = _mm_unpacklo_epi64(reg2_top_dwords, reg3_top_dwords); __m128i valid_positve_01 = _mm_cmpeq_epi32(top_dwords_01, *zero); __m128i valid_positve_23 = _mm_cmpeq_epi32(top_dwords_23, *zero); __m128i valid_negative_01 = _mm_cmpeq_epi32(top_dwords_01, minus_one); __m128i valid_negative_23 = _mm_cmpeq_epi32(top_dwords_23, minus_one); int overflow_01 = _mm_movemask_epi8( _mm_cmpeq_epi32(valid_positve_01, valid_negative_01)); int overflow_23 = _mm_movemask_epi8( _mm_cmpeq_epi32(valid_positve_23, valid_negative_23)); return (overflow_01 + overflow_23); } static INLINE int k_check_epi32_overflow_8(__m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, __m128i reg4, __m128i reg5, __m128i reg6, __m128i reg7, const __m128i* zero) { int overflow = k_check_epi32_overflow_4(reg0, reg1, reg2, reg3, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg4, reg5, reg6, reg7, zero); } return overflow; } static INLINE int k_check_epi32_overflow_16( __m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, __m128i reg4, __m128i reg5, __m128i reg6, __m128i reg7, __m128i reg8, __m128i reg9, __m128i reg10, __m128i reg11, __m128i reg12, __m128i reg13, __m128i reg14, __m128i reg15, const __m128i* zero) { int overflow = k_check_epi32_overflow_4(reg0, reg1, reg2, reg3, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg4, reg5, reg6, reg7, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg8, reg9, reg10, reg11, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg12, reg13, reg14, reg15, zero); } } } return overflow; } static INLINE int k_check_epi32_overflow_32( __m128i reg0, __m128i reg1, __m128i reg2, __m128i reg3, __m128i reg4, __m128i reg5, __m128i reg6, __m128i reg7, __m128i reg8, __m128i reg9, __m128i reg10, __m128i reg11, __m128i reg12, __m128i reg13, __m128i reg14, __m128i reg15, __m128i reg16, __m128i reg17, __m128i reg18, __m128i reg19, __m128i reg20, __m128i reg21, __m128i reg22, __m128i reg23, __m128i reg24, __m128i reg25, __m128i reg26, __m128i reg27, __m128i reg28, __m128i reg29, __m128i reg30, __m128i reg31, const __m128i* zero) { int overflow = k_check_epi32_overflow_4(reg0, reg1, reg2, reg3, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg4, reg5, reg6, reg7, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg8, reg9, reg10, reg11, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg12, reg13, reg14, reg15, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg16, reg17, reg18, reg19, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg20, reg21, reg22, reg23, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg24, reg25, reg26, reg27, zero); if (!overflow) { overflow = k_check_epi32_overflow_4(reg28, reg29, reg30, reg31, zero); } } } } } } } return overflow; } static INLINE void store_output(const __m128i output, tran_low_t* dst_ptr) { #if CONFIG_VP9_HIGHBITDEPTH const __m128i zero = _mm_setzero_si128(); const __m128i sign_bits = _mm_cmplt_epi16(output, zero); __m128i out0 = _mm_unpacklo_epi16(output, sign_bits); __m128i out1 = _mm_unpackhi_epi16(output, sign_bits); _mm_store_si128((__m128i *)(dst_ptr), out0); _mm_store_si128((__m128i *)(dst_ptr + 4), out1); #else _mm_store_si128((__m128i *)(dst_ptr), output); #endif // CONFIG_VP9_HIGHBITDEPTH } static INLINE void storeu_output(const __m128i output, tran_low_t* dst_ptr) { #if CONFIG_VP9_HIGHBITDEPTH const __m128i zero = _mm_setzero_si128(); const __m128i sign_bits = _mm_cmplt_epi16(output, zero); __m128i out0 = _mm_unpacklo_epi16(output, sign_bits); __m128i out1 = _mm_unpackhi_epi16(output, sign_bits); _mm_storeu_si128((__m128i *)(dst_ptr), out0); _mm_storeu_si128((__m128i *)(dst_ptr + 4), out1); #else _mm_storeu_si128((__m128i *)(dst_ptr), output); #endif // CONFIG_VP9_HIGHBITDEPTH } static INLINE __m128i mult_round_shift(const __m128i in0, const __m128i in1, const __m128i multiplier, const __m128i rounding, const int shift) { const __m128i u0 = _mm_madd_epi16(in0, multiplier); const __m128i u1 = _mm_madd_epi16(in1, multiplier); const __m128i v0 = _mm_add_epi32(u0, rounding); const __m128i v1 = _mm_add_epi32(u1, rounding); const __m128i w0 = _mm_srai_epi32(v0, shift); const __m128i w1 = _mm_srai_epi32(v1, shift); return _mm_packs_epi32(w0, w1); } static INLINE void transpose_and_output8x8( const __m128i in00, const __m128i in01, const __m128i in02, const __m128i in03, const __m128i in04, const __m128i in05, const __m128i in06, const __m128i in07, const int pass, int16_t* out0_ptr, tran_low_t* out1_ptr) { // 00 01 02 03 04 05 06 07 // 10 11 12 13 14 15 16 17 // 20 21 22 23 24 25 26 27 // 30 31 32 33 34 35 36 37 // 40 41 42 43 44 45 46 47 // 50 51 52 53 54 55 56 57 // 60 61 62 63 64 65 66 67 // 70 71 72 73 74 75 76 77 const __m128i tr0_0 = _mm_unpacklo_epi16(in00, in01); const __m128i tr0_1 = _mm_unpacklo_epi16(in02, in03); const __m128i tr0_2 = _mm_unpackhi_epi16(in00, in01); const __m128i tr0_3 = _mm_unpackhi_epi16(in02, in03); const __m128i tr0_4 = _mm_unpacklo_epi16(in04, in05); const __m128i tr0_5 = _mm_unpacklo_epi16(in06, in07); const __m128i tr0_6 = _mm_unpackhi_epi16(in04, in05); const __m128i tr0_7 = _mm_unpackhi_epi16(in06, in07); // 00 10 01 11 02 12 03 13 // 20 30 21 31 22 32 23 33 // 04 14 05 15 06 16 07 17 // 24 34 25 35 26 36 27 37 // 40 50 41 51 42 52 43 53 // 60 70 61 71 62 72 63 73 // 54 54 55 55 56 56 57 57 // 64 74 65 75 66 76 67 77 const __m128i tr1_0 = _mm_unpacklo_epi32(tr0_0, tr0_1); const __m128i tr1_1 = _mm_unpacklo_epi32(tr0_2, tr0_3); const __m128i tr1_2 = _mm_unpackhi_epi32(tr0_0, tr0_1); const __m128i tr1_3 = _mm_unpackhi_epi32(tr0_2, tr0_3); const __m128i tr1_4 = _mm_unpacklo_epi32(tr0_4, tr0_5); const __m128i tr1_5 = _mm_unpacklo_epi32(tr0_6, tr0_7); const __m128i tr1_6 = _mm_unpackhi_epi32(tr0_4, tr0_5); const __m128i tr1_7 = _mm_unpackhi_epi32(tr0_6, tr0_7); // 00 10 20 30 01 11 21 31 // 40 50 60 70 41 51 61 71 // 02 12 22 32 03 13 23 33 // 42 52 62 72 43 53 63 73 // 04 14 24 34 05 15 21 36 // 44 54 64 74 45 55 61 76 // 06 16 26 36 07 17 27 37 // 46 56 66 76 47 57 67 77 const __m128i tr2_0 = _mm_unpacklo_epi64(tr1_0, tr1_4); const __m128i tr2_1 = _mm_unpackhi_epi64(tr1_0, tr1_4); const __m128i tr2_2 = _mm_unpacklo_epi64(tr1_2, tr1_6); const __m128i tr2_3 = _mm_unpackhi_epi64(tr1_2, tr1_6); const __m128i tr2_4 = _mm_unpacklo_epi64(tr1_1, tr1_5); const __m128i tr2_5 = _mm_unpackhi_epi64(tr1_1, tr1_5); const __m128i tr2_6 = _mm_unpacklo_epi64(tr1_3, tr1_7); const __m128i tr2_7 = _mm_unpackhi_epi64(tr1_3, tr1_7); // 00 10 20 30 40 50 60 70 // 01 11 21 31 41 51 61 71 // 02 12 22 32 42 52 62 72 // 03 13 23 33 43 53 63 73 // 04 14 24 34 44 54 64 74 // 05 15 25 35 45 55 65 75 // 06 16 26 36 46 56 66 76 // 07 17 27 37 47 57 67 77 if (pass == 0) { _mm_storeu_si128((__m128i*)(out0_ptr + 0 * 16), tr2_0); _mm_storeu_si128((__m128i*)(out0_ptr + 1 * 16), tr2_1); _mm_storeu_si128((__m128i*)(out0_ptr + 2 * 16), tr2_2); _mm_storeu_si128((__m128i*)(out0_ptr + 3 * 16), tr2_3); _mm_storeu_si128((__m128i*)(out0_ptr + 4 * 16), tr2_4); _mm_storeu_si128((__m128i*)(out0_ptr + 5 * 16), tr2_5); _mm_storeu_si128((__m128i*)(out0_ptr + 6 * 16), tr2_6); _mm_storeu_si128((__m128i*)(out0_ptr + 7 * 16), tr2_7); } else { storeu_output(tr2_0, (out1_ptr + 0 * 16)); storeu_output(tr2_1, (out1_ptr + 1 * 16)); storeu_output(tr2_2, (out1_ptr + 2 * 16)); storeu_output(tr2_3, (out1_ptr + 3 * 16)); storeu_output(tr2_4, (out1_ptr + 4 * 16)); storeu_output(tr2_5, (out1_ptr + 5 * 16)); storeu_output(tr2_6, (out1_ptr + 6 * 16)); storeu_output(tr2_7, (out1_ptr + 7 * 16)); } } #ifdef __cplusplus } // extern "C" #endif #endif // VP9_ENCODER_X86_VP9_DCT_SSE2_H_