diff options
author | Jingning Han <jingning@google.com> | 2017-01-12 15:06:30 -0800 |
---|---|---|
committer | Jingning Han <jingning@google.com> | 2017-01-12 15:13:18 -0800 |
commit | f65170ea842dc8031bc431e9c5d37230820d3eed (patch) | |
tree | 356d4be1e0982dfbfaedc01816d411d1cd96d6f4 /vpx_dsp | |
parent | 9a780fa7db79b709787a9ca56fc324a118158da7 (diff) | |
download | libvpx-f65170ea842dc8031bc431e9c5d37230820d3eed.tar libvpx-f65170ea842dc8031bc431e9c5d37230820d3eed.tar.gz libvpx-f65170ea842dc8031bc431e9c5d37230820d3eed.tar.bz2 libvpx-f65170ea842dc8031bc431e9c5d37230820d3eed.zip |
Rework 8x8 transpose SSSE3 for inverse 2D-DCT
Use same transpose process as inv_txfm_sse2 does.
Change-Id: Ic4827825bd174cba57a0a80e19bf458a648e7d94
Diffstat (limited to 'vpx_dsp')
-rw-r--r-- | vpx_dsp/x86/inv_txfm_ssse3_x86_64.asm | 83 |
1 files changed, 46 insertions, 37 deletions
diff --git a/vpx_dsp/x86/inv_txfm_ssse3_x86_64.asm b/vpx_dsp/x86/inv_txfm_ssse3_x86_64.asm index dee64e3ad..a0dbd60e6 100644 --- a/vpx_dsp/x86/inv_txfm_ssse3_x86_64.asm +++ b/vpx_dsp/x86/inv_txfm_ssse3_x86_64.asm @@ -10,10 +10,6 @@ %include "third_party/x86inc/x86inc.asm" -; This file provides SSSE3 version of the inverse transformation. Part -; of the functions are originally derived from the ffmpeg project. -; Note that the current version applies to x86 64-bit only. - SECTION_RODATA pw_11585x2: times 8 dw 23170 @@ -141,31 +137,44 @@ SECTION .text packssdw m%2, m%6 %endmacro -; matrix transpose -%macro INTERLEAVE_2X 4 - punpckh%1 m%4, m%2, m%3 - punpckl%1 m%2, m%3 - SWAP %3, %4 -%endmacro - -%macro TRANSPOSE8X8 9 - INTERLEAVE_2X wd, %1, %2, %9 - INTERLEAVE_2X wd, %3, %4, %9 - INTERLEAVE_2X wd, %5, %6, %9 - INTERLEAVE_2X wd, %7, %8, %9 - - INTERLEAVE_2X dq, %1, %3, %9 - INTERLEAVE_2X dq, %2, %4, %9 - INTERLEAVE_2X dq, %5, %7, %9 - INTERLEAVE_2X dq, %6, %8, %9 - - INTERLEAVE_2X qdq, %1, %5, %9 - INTERLEAVE_2X qdq, %3, %7, %9 - INTERLEAVE_2X qdq, %2, %6, %9 - INTERLEAVE_2X qdq, %4, %8, %9 - - SWAP %2, %5 - SWAP %4, %7 +; 8x8 transpose. This follows same operations as SSE2 does. +%macro TRANSPOSE8X8 10 + ; stage 1 + punpcklwd m%9, m%1, m%2 + punpcklwd m%10, m%3, m%4 + punpckhwd m%1, m%2 + punpckhwd m%3, m%4 + + punpcklwd m%2, m%5, m%6 + punpcklwd m%4, m%7, m%8 + punpckhwd m%5, m%6 + punpckhwd m%7, m%8 + + ; stage 2 + punpckldq m%6, m%9, m%10 + punpckldq m%8, m%1, m%3 + punpckhdq m%9, m%10 + punpckhdq m%1, m%3 + + punpckldq m%10, m%2, m%4 + punpckldq m%3, m%5, m%7 + punpckhdq m%2, m%4 + punpckhdq m%5, m%7 + + ; stage 3 + punpckhqdq m%4, m%9, m%2 ; out3 + punpcklqdq m%9, m%2 ; out2 + punpcklqdq m%7, m%1, m%5 ; out6 + punpckhqdq m%1, m%5 ; out7 + + punpckhqdq m%2, m%6, m%10 ; out1 + punpcklqdq m%6, m%10 ; out0 + punpcklqdq m%5, m%8, m%3 ; out4 + punpckhqdq m%8, m%3 ; out5 + + SWAP %6, %1 + SWAP %3, %9 + SWAP %8, %6 %endmacro %macro IDCT8_1D 0 @@ -247,9 +256,9 @@ cglobal idct8x8_64_add, 3, 5, 13, input, output, stride mova m6, [inputq + 96] mova m7, [inputq + 112] %endif - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 IDCT8_1D - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 IDCT8_1D pxor m12, m12 @@ -821,7 +830,7 @@ idct32x32_34_transpose: mova m7, [r3 + 16 * 28] %endif - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 IDCT32X32_34 16*0, 16*32, 16*64, 16*96 lea stp, [stp + 16 * 8] @@ -843,7 +852,7 @@ idct32x32_34_transpose_2: mova m6, [r3 + 16 * 6] mova m7, [r3 + 16 * 7] - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 IDCT32X32_34 16*0, 16*8, 16*16, 16*24 @@ -1250,7 +1259,7 @@ idct32x32_135_transpose: mova m6, [r3 + 16 * 24] mova m7, [r3 + 16 * 28] %endif - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 mova [r4 + 0], m0 mova [r4 + 16 * 1], m1 @@ -1299,7 +1308,7 @@ idct32x32_135_transpose_2: mova m6, [r3 + 16 * 6] mova m7, [r3 + 16 * 7] - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 mova [r4 + 0], m0 mova [r4 + 16 * 1], m1 @@ -1715,7 +1724,7 @@ idct32x32_1024_transpose: mova m7, [r3 + 16 * 28] %endif - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 mova [r4 + 0], m0 mova [r4 + 16 * 1], m1 @@ -1764,7 +1773,7 @@ idct32x32_1024_transpose_2: mova m6, [r3 + 16 * 6] mova m7, [r3 + 16 * 7] - TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9 + TRANSPOSE8X8 0, 1, 2, 3, 4, 5, 6, 7, 9, 10 mova [r4 + 0], m0 mova [r4 + 16 * 1], m1 |