/* * Copyright (c) 2015 The WebM project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include #include #include #include "vpx_dsp/ppc/bitdepth_conversion_vsx.h" #include "vpx_dsp/ppc/types_vsx.h" #include "./vpx_dsp_rtcd.h" #include "vpx_dsp/inv_txfm.h" static int16x8_t cospi1_v = { 16364, 16364, 16364, 16364, 16364, 16364, 16364, 16364 }; static int16x8_t cospi2_v = { 16305, 16305, 16305, 16305, 16305, 16305, 16305, 16305 }; static int16x8_t cospi3_v = { 16207, 16207, 16207, 16207, 16207, 16207, 16207, 16207 }; static int16x8_t cospi4_v = { 16069, 16069, 16069, 16069, 16069, 16069, 16069, 16069 }; static int16x8_t cospi4m_v = { -16069, -16069, -16069, -16069, -16069, -16069, -16069, -16069 }; static int16x8_t cospi5_v = { 15893, 15893, 15893, 15893, 15893, 15893, 15893, 15893 }; static int16x8_t cospi6_v = { 15679, 15679, 15679, 15679, 15679, 15679, 15679, 15679 }; static int16x8_t cospi7_v = { 15426, 15426, 15426, 15426, 15426, 15426, 15426, 15426 }; static int16x8_t cospi8_v = { 15137, 15137, 15137, 15137, 15137, 15137, 15137, 15137 }; static int16x8_t cospi8m_v = { -15137, -15137, -15137, -15137, -15137, -15137, -15137, -15137 }; static int16x8_t cospi9_v = { 14811, 14811, 14811, 14811, 14811, 14811, 14811, 14811 }; static int16x8_t cospi10_v = { 14449, 14449, 14449, 14449, 14449, 14449, 14449, 14449 }; static int16x8_t cospi11_v = { 14053, 14053, 14053, 14053, 14053, 14053, 14053, 14053 }; static int16x8_t cospi12_v = { 13623, 13623, 13623, 13623, 13623, 13623, 13623, 13623 }; static int16x8_t cospi13_v = { 13160, 13160, 13160, 13160, 13160, 13160, 13160, 13160 }; static int16x8_t cospi14_v = { 12665, 12665, 12665, 12665, 12665, 12665, 12665, 12665 }; static int16x8_t cospi15_v = { 12140, 12140, 12140, 12140, 12140, 12140, 12140, 12140 }; static int16x8_t cospi16_v = { 11585, 11585, 11585, 11585, 11585, 11585, 11585, 11585 }; static int16x8_t cospi17_v = { 11003, 11003, 11003, 11003, 11003, 11003, 11003, 11003 }; static int16x8_t cospi18_v = { 10394, 10394, 10394, 10394, 10394, 10394, 10394, 10394 }; static int16x8_t cospi19_v = { 9760, 9760, 9760, 9760, 9760, 9760, 9760, 9760 }; static int16x8_t cospi20_v = { 9102, 9102, 9102, 9102, 9102, 9102, 9102, 9102 }; static int16x8_t cospi20m_v = { -9102, -9102, -9102, -9102, -9102, -9102, -9102, -9102 }; static int16x8_t cospi21_v = { 8423, 8423, 8423, 8423, 8423, 8423, 8423, 8423 }; static int16x8_t cospi22_v = { 7723, 7723, 7723, 7723, 7723, 7723, 7723, 7723 }; static int16x8_t cospi23_v = { 7005, 7005, 7005, 7005, 7005, 7005, 7005, 7005 }; static int16x8_t cospi24_v = { 6270, 6270, 6270, 6270, 6270, 6270, 6270, 6270 }; static int16x8_t cospi24_mv = { -6270, -6270, -6270, -6270, -6270, -6270, -6270, -6270 }; static int16x8_t cospi25_v = { 5520, 5520, 5520, 5520, 5520, 5520, 5520, 5520 }; static int16x8_t cospi26_v = { 4756, 4756, 4756, 4756, 4756, 4756, 4756, 4756 }; static int16x8_t cospi27_v = { 3981, 3981, 3981, 3981, 3981, 3981, 3981, 3981 }; static int16x8_t cospi28_v = { 3196, 3196, 3196, 3196, 3196, 3196, 3196, 3196 }; static int16x8_t cospi29_v = { 2404, 2404, 2404, 2404, 2404, 2404, 2404, 2404 }; static int16x8_t cospi30_v = { 1606, 1606, 1606, 1606, 1606, 1606, 1606, 1606 }; static int16x8_t cospi31_v = { 804, 804, 804, 804, 804, 804, 804, 804 }; #define ROUND_SHIFT_INIT \ const int32x4_t shift = vec_sl(vec_splat_s32(1), vec_splat_u32(13)); \ const uint32x4_t shift14 = vec_splat_u32(14); #define DCT_CONST_ROUND_SHIFT(vec) vec = vec_sra(vec_add(vec, shift), shift14); #define PIXEL_ADD_INIT \ int16x8_t add8 = vec_splat_s16(8); \ uint16x8_t shift4 = vec_splat_u16(4); #define PIXEL_ADD4(out, in) out = vec_sra(vec_add(in, add8), shift4); #define IDCT4(in0, in1, out0, out1) \ t0 = vec_add(in0, in1); \ t1 = vec_sub(in0, in1); \ tmp16_0 = vec_mergeh(t0, t1); \ temp1 = vec_sra(vec_add(vec_mule(tmp16_0, cospi16_v), shift), shift14); \ temp2 = vec_sra(vec_add(vec_mulo(tmp16_0, cospi16_v), shift), shift14); \ \ tmp16_0 = vec_mergel(in0, in1); \ temp3 = vec_sub(vec_mule(tmp16_0, cospi24_v), vec_mulo(tmp16_0, cospi8_v)); \ DCT_CONST_ROUND_SHIFT(temp3); \ temp4 = vec_add(vec_mule(tmp16_0, cospi8_v), vec_mulo(tmp16_0, cospi24_v)); \ DCT_CONST_ROUND_SHIFT(temp4); \ \ step0 = vec_packs(temp1, temp2); \ step1 = vec_packs(temp4, temp3); \ out0 = vec_add(step0, step1); \ out1 = vec_sub(step0, step1); \ out1 = vec_perm(out1, out1, mask0); void vpx_idct4x4_16_add_vsx(const tran_low_t *input, uint8_t *dest, int stride) { int i, j; int32x4_t temp1, temp2, temp3, temp4; int16x8_t step0, step1, tmp16_0, tmp16_1, t_out0, t_out1; uint8x16_t mask0 = { 0x8, 0x9, 0xA, 0xB, 0xC, 0xD, 0xE, 0xF, 0x0, 0x1, 0x2, 0x3, 0x4, 0x5, 0x6, 0x7 }; uint8x16_t mask1 = { 0x0, 0x1, 0x2, 0x3, 0x4, 0x5, 0x6, 0x7, 0x10, 0x11, 0x12, 0x13, 0x14, 0x15, 0x16, 0x17 }; int16x8_t v0 = load_tran_low(0, input); int16x8_t v1 = load_tran_low(8 * sizeof(*input), input); int16x8_t t0 = vec_mergeh(v0, v1); int16x8_t t1 = vec_mergel(v0, v1); uint8x16_t dest0 = vec_vsx_ld(0, dest); uint8x16_t dest1 = vec_vsx_ld(stride, dest); uint8x16_t dest2 = vec_vsx_ld(2 * stride, dest); uint8x16_t dest3 = vec_vsx_ld(3 * stride, dest); uint8x16_t zerov = vec_splat_u8(0); int16x8_t d_u0 = (int16x8_t)vec_mergeh(dest0, zerov); int16x8_t d_u1 = (int16x8_t)vec_mergeh(dest1, zerov); int16x8_t d_u2 = (int16x8_t)vec_mergeh(dest2, zerov); int16x8_t d_u3 = (int16x8_t)vec_mergeh(dest3, zerov); uint8x16_t output_v; uint8_t tmp_dest[16]; ROUND_SHIFT_INIT PIXEL_ADD_INIT; v0 = vec_mergeh(t0, t1); v1 = vec_mergel(t0, t1); IDCT4(v0, v1, t_out0, t_out1); // transpose t0 = vec_mergeh(t_out0, t_out1); t1 = vec_mergel(t_out0, t_out1); v0 = vec_mergeh(t0, t1); v1 = vec_mergel(t0, t1); IDCT4(v0, v1, t_out0, t_out1); PIXEL_ADD4(v0, t_out0); PIXEL_ADD4(v1, t_out1); tmp16_0 = vec_add(vec_perm(d_u0, d_u1, mask1), v0); tmp16_1 = vec_add(vec_perm(d_u2, d_u3, mask1), v1); output_v = vec_packsu(tmp16_0, tmp16_1); vec_vsx_st(output_v, 0, tmp_dest); for (i = 0; i < 4; i++) for (j = 0; j < 4; j++) dest[j * stride + i] = tmp_dest[j * 4 + i]; } #define TRANSPOSE8x8(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, \ out3, out4, out5, out6, out7) \ out0 = vec_mergeh(in0, in1); \ out1 = vec_mergel(in0, in1); \ out2 = vec_mergeh(in2, in3); \ out3 = vec_mergel(in2, in3); \ out4 = vec_mergeh(in4, in5); \ out5 = vec_mergel(in4, in5); \ out6 = vec_mergeh(in6, in7); \ out7 = vec_mergel(in6, in7); \ in0 = (int16x8_t)vec_mergeh((int32x4_t)out0, (int32x4_t)out2); \ in1 = (int16x8_t)vec_mergel((int32x4_t)out0, (int32x4_t)out2); \ in2 = (int16x8_t)vec_mergeh((int32x4_t)out1, (int32x4_t)out3); \ in3 = (int16x8_t)vec_mergel((int32x4_t)out1, (int32x4_t)out3); \ in4 = (int16x8_t)vec_mergeh((int32x4_t)out4, (int32x4_t)out6); \ in5 = (int16x8_t)vec_mergel((int32x4_t)out4, (int32x4_t)out6); \ in6 = (int16x8_t)vec_mergeh((int32x4_t)out5, (int32x4_t)out7); \ in7 = (int16x8_t)vec_mergel((int32x4_t)out5, (int32x4_t)out7); \ out0 = vec_perm(in0, in4, tr8_mask0); \ out1 = vec_perm(in0, in4, tr8_mask1); \ out2 = vec_perm(in1, in5, tr8_mask0); \ out3 = vec_perm(in1, in5, tr8_mask1); \ out4 = vec_perm(in2, in6, tr8_mask0); \ out5 = vec_perm(in2, in6, tr8_mask1); \ out6 = vec_perm(in3, in7, tr8_mask0); \ out7 = vec_perm(in3, in7, tr8_mask1); /* for the: temp1 = step[x] * cospi_q - step[y] * cospi_z * temp2 = step[x] * cospi_z + step[y] * cospi_q */ #define STEP8_0(inpt0, inpt1, outpt0, outpt1, cospi0, cospi1) \ tmp16_0 = vec_mergeh(inpt0, inpt1); \ tmp16_1 = vec_mergel(inpt0, inpt1); \ temp10 = vec_sub(vec_mule(tmp16_0, cospi0), vec_mulo(tmp16_0, cospi1)); \ temp11 = vec_sub(vec_mule(tmp16_1, cospi0), vec_mulo(tmp16_1, cospi1)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt0 = vec_packs(temp10, temp11); \ temp10 = vec_add(vec_mule(tmp16_0, cospi1), vec_mulo(tmp16_0, cospi0)); \ temp11 = vec_add(vec_mule(tmp16_1, cospi1), vec_mulo(tmp16_1, cospi0)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt1 = vec_packs(temp10, temp11); #define STEP8_1(inpt0, inpt1, outpt0, outpt1, cospi) \ tmp16_2 = vec_sub(inpt0, inpt1); \ tmp16_3 = vec_add(inpt0, inpt1); \ tmp16_0 = vec_mergeh(tmp16_2, tmp16_3); \ tmp16_1 = vec_mergel(tmp16_2, tmp16_3); \ temp10 = vec_mule(tmp16_0, cospi); \ temp11 = vec_mule(tmp16_1, cospi); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt0 = vec_packs(temp10, temp11); \ temp10 = vec_mulo(tmp16_0, cospi); \ temp11 = vec_mulo(tmp16_1, cospi); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt1 = vec_packs(temp10, temp11); #define IDCT8(in0, in1, in2, in3, in4, in5, in6, in7) \ /* stage 1 */ \ step0 = in0; \ step2 = in4; \ step1 = in2; \ step3 = in6; \ \ STEP8_0(in1, in7, step4, step7, cospi28_v, cospi4_v); \ STEP8_0(in5, in3, step5, step6, cospi12_v, cospi20_v); \ \ /* stage 2 */ \ STEP8_1(step0, step2, in1, in0, cospi16_v); \ STEP8_0(step1, step3, in2, in3, cospi24_v, cospi8_v); \ in4 = vec_add(step4, step5); \ in5 = vec_sub(step4, step5); \ in6 = vec_sub(step7, step6); \ in7 = vec_add(step6, step7); \ \ /* stage 3 */ \ step0 = vec_add(in0, in3); \ step1 = vec_add(in1, in2); \ step2 = vec_sub(in1, in2); \ step3 = vec_sub(in0, in3); \ step4 = in4; \ STEP8_1(in6, in5, step5, step6, cospi16_v); \ step7 = in7; \ \ /* stage 4 */ \ in0 = vec_add(step0, step7); \ in1 = vec_add(step1, step6); \ in2 = vec_add(step2, step5); \ in3 = vec_add(step3, step4); \ in4 = vec_sub(step3, step4); \ in5 = vec_sub(step2, step5); \ in6 = vec_sub(step1, step6); \ in7 = vec_sub(step0, step7); #define PIXEL_ADD(in, out, add, shiftx) \ out = vec_add(vec_sra(vec_add(in, add), shiftx), out); static uint8x16_t tr8_mask0 = { 0x0, 0x1, 0x2, 0x3, 0x4, 0x5, 0x6, 0x7, 0x10, 0x11, 0x12, 0x13, 0x14, 0x15, 0x16, 0x17 }; static uint8x16_t tr8_mask1 = { 0x8, 0x9, 0xA, 0xB, 0xC, 0xD, 0xE, 0xF, 0x18, 0x19, 0x1A, 0x1B, 0x1C, 0x1D, 0x1E, 0x1F }; void vpx_idct8x8_64_add_vsx(const tran_low_t *input, uint8_t *dest, int stride) { int32x4_t temp10, temp11; int16x8_t step0, step1, step2, step3, step4, step5, step6, step7; int16x8_t tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, tmp16_0, tmp16_1, tmp16_2, tmp16_3; int16x8_t src0 = load_tran_low(0, input); int16x8_t src1 = load_tran_low(8 * sizeof(*input), input); int16x8_t src2 = load_tran_low(16 * sizeof(*input), input); int16x8_t src3 = load_tran_low(24 * sizeof(*input), input); int16x8_t src4 = load_tran_low(32 * sizeof(*input), input); int16x8_t src5 = load_tran_low(40 * sizeof(*input), input); int16x8_t src6 = load_tran_low(48 * sizeof(*input), input); int16x8_t src7 = load_tran_low(56 * sizeof(*input), input); uint8x16_t dest0 = vec_vsx_ld(0, dest); uint8x16_t dest1 = vec_vsx_ld(stride, dest); uint8x16_t dest2 = vec_vsx_ld(2 * stride, dest); uint8x16_t dest3 = vec_vsx_ld(3 * stride, dest); uint8x16_t dest4 = vec_vsx_ld(4 * stride, dest); uint8x16_t dest5 = vec_vsx_ld(5 * stride, dest); uint8x16_t dest6 = vec_vsx_ld(6 * stride, dest); uint8x16_t dest7 = vec_vsx_ld(7 * stride, dest); uint8x16_t zerov = vec_splat_u8(0); int16x8_t d_u0 = (int16x8_t)vec_mergeh(dest0, zerov); int16x8_t d_u1 = (int16x8_t)vec_mergeh(dest1, zerov); int16x8_t d_u2 = (int16x8_t)vec_mergeh(dest2, zerov); int16x8_t d_u3 = (int16x8_t)vec_mergeh(dest3, zerov); int16x8_t d_u4 = (int16x8_t)vec_mergeh(dest4, zerov); int16x8_t d_u5 = (int16x8_t)vec_mergeh(dest5, zerov); int16x8_t d_u6 = (int16x8_t)vec_mergeh(dest6, zerov); int16x8_t d_u7 = (int16x8_t)vec_mergeh(dest7, zerov); int16x8_t add = vec_sl(vec_splat_s16(8), vec_splat_u16(1)); uint16x8_t shift5 = vec_splat_u16(5); uint8x16_t output0, output1, output2, output3; ROUND_SHIFT_INIT; TRANSPOSE8x8(src0, src1, src2, src3, src4, src5, src6, src7, tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7); IDCT8(tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7); TRANSPOSE8x8(tmp0, tmp1, tmp2, tmp3, tmp4, tmp5, tmp6, tmp7, src0, src1, src2, src3, src4, src5, src6, src7); IDCT8(src0, src1, src2, src3, src4, src5, src6, src7); PIXEL_ADD(src0, d_u0, add, shift5); PIXEL_ADD(src1, d_u1, add, shift5); PIXEL_ADD(src2, d_u2, add, shift5); PIXEL_ADD(src3, d_u3, add, shift5); PIXEL_ADD(src4, d_u4, add, shift5); PIXEL_ADD(src5, d_u5, add, shift5); PIXEL_ADD(src6, d_u6, add, shift5); PIXEL_ADD(src7, d_u7, add, shift5); output0 = vec_packsu(d_u0, d_u1); output1 = vec_packsu(d_u2, d_u3); output2 = vec_packsu(d_u4, d_u5); output3 = vec_packsu(d_u6, d_u7); vec_vsx_st(xxpermdi(output0, dest0, 1), 0, dest); vec_vsx_st(xxpermdi(output0, dest1, 3), stride, dest); vec_vsx_st(xxpermdi(output1, dest2, 1), 2 * stride, dest); vec_vsx_st(xxpermdi(output1, dest3, 3), 3 * stride, dest); vec_vsx_st(xxpermdi(output2, dest4, 1), 4 * stride, dest); vec_vsx_st(xxpermdi(output2, dest5, 3), 5 * stride, dest); vec_vsx_st(xxpermdi(output3, dest6, 1), 6 * stride, dest); vec_vsx_st(xxpermdi(output3, dest7, 3), 7 * stride, dest); } #define LOAD_INPUT16(load, source, offset, step, in0, in1, in2, in3, in4, in5, \ in6, in7, in8, in9, inA, inB, inC, inD, inE, inF) \ in0 = load(offset, source); \ in1 = load((step) + (offset), source); \ in2 = load(2 * (step) + (offset), source); \ in3 = load(3 * (step) + (offset), source); \ in4 = load(4 * (step) + (offset), source); \ in5 = load(5 * (step) + (offset), source); \ in6 = load(6 * (step) + (offset), source); \ in7 = load(7 * (step) + (offset), source); \ in8 = load(8 * (step) + (offset), source); \ in9 = load(9 * (step) + (offset), source); \ inA = load(10 * (step) + (offset), source); \ inB = load(11 * (step) + (offset), source); \ inC = load(12 * (step) + (offset), source); \ inD = load(13 * (step) + (offset), source); \ inE = load(14 * (step) + (offset), source); \ inF = load(15 * (step) + (offset), source); #define STEP16_1(inpt0, inpt1, outpt0, outpt1, cospi) \ tmp16_0 = vec_mergeh(inpt0, inpt1); \ tmp16_1 = vec_mergel(inpt0, inpt1); \ temp10 = vec_mule(tmp16_0, cospi); \ temp11 = vec_mule(tmp16_1, cospi); \ temp20 = vec_mulo(tmp16_0, cospi); \ temp21 = vec_mulo(tmp16_1, cospi); \ temp30 = vec_sub(temp10, temp20); \ temp10 = vec_add(temp10, temp20); \ temp20 = vec_sub(temp11, temp21); \ temp21 = vec_add(temp11, temp21); \ DCT_CONST_ROUND_SHIFT(temp30); \ DCT_CONST_ROUND_SHIFT(temp20); \ outpt0 = vec_packs(temp30, temp20); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp21); \ outpt1 = vec_packs(temp10, temp21); #define IDCT16(in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, inA, inB, \ inC, inD, inE, inF, out0, out1, out2, out3, out4, out5, out6, \ out7, out8, out9, outA, outB, outC, outD, outE, outF) \ /* stage 1 */ \ /* out0 = in0; */ \ out1 = in8; \ out2 = in4; \ out3 = inC; \ out4 = in2; \ out5 = inA; \ out6 = in6; \ out7 = inE; \ out8 = in1; \ out9 = in9; \ outA = in5; \ outB = inD; \ outC = in3; \ outD = inB; \ outE = in7; \ outF = inF; \ \ /* stage 2 */ \ /* in0 = out0; */ \ in1 = out1; \ in2 = out2; \ in3 = out3; \ in4 = out4; \ in5 = out5; \ in6 = out6; \ in7 = out7; \ \ STEP8_0(out8, outF, in8, inF, cospi30_v, cospi2_v); \ STEP8_0(out9, outE, in9, inE, cospi14_v, cospi18_v); \ STEP8_0(outA, outD, inA, inD, cospi22_v, cospi10_v); \ STEP8_0(outB, outC, inB, inC, cospi6_v, cospi26_v); \ \ /* stage 3 */ \ out0 = in0; \ out1 = in1; \ out2 = in2; \ out3 = in3; \ \ STEP8_0(in4, in7, out4, out7, cospi28_v, cospi4_v); \ STEP8_0(in5, in6, out5, out6, cospi12_v, cospi20_v); \ \ out8 = vec_add(in8, in9); \ out9 = vec_sub(in8, in9); \ outA = vec_sub(inB, inA); \ outB = vec_add(inA, inB); \ outC = vec_add(inC, inD); \ outD = vec_sub(inC, inD); \ outE = vec_sub(inF, inE); \ outF = vec_add(inE, inF); \ \ /* stage 4 */ \ STEP16_1(out0, out1, in1, in0, cospi16_v); \ STEP8_0(out2, out3, in2, in3, cospi24_v, cospi8_v); \ in4 = vec_add(out4, out5); \ in5 = vec_sub(out4, out5); \ in6 = vec_sub(out7, out6); \ in7 = vec_add(out6, out7); \ \ in8 = out8; \ inF = outF; \ tmp16_0 = vec_mergeh(out9, outE); \ tmp16_1 = vec_mergel(out9, outE); \ temp10 = vec_sub(vec_mulo(tmp16_0, cospi24_v), vec_mule(tmp16_0, cospi8_v)); \ temp11 = vec_sub(vec_mulo(tmp16_1, cospi24_v), vec_mule(tmp16_1, cospi8_v)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ in9 = vec_packs(temp10, temp11); \ temp10 = vec_add(vec_mule(tmp16_0, cospi24_v), vec_mulo(tmp16_0, cospi8_v)); \ temp11 = vec_add(vec_mule(tmp16_1, cospi24_v), vec_mulo(tmp16_1, cospi8_v)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ inE = vec_packs(temp10, temp11); \ \ tmp16_0 = vec_mergeh(outA, outD); \ tmp16_1 = vec_mergel(outA, outD); \ temp10 = \ vec_sub(vec_mule(tmp16_0, cospi24_mv), vec_mulo(tmp16_0, cospi8_v)); \ temp11 = \ vec_sub(vec_mule(tmp16_1, cospi24_mv), vec_mulo(tmp16_1, cospi8_v)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ inA = vec_packs(temp10, temp11); \ temp10 = vec_sub(vec_mulo(tmp16_0, cospi24_v), vec_mule(tmp16_0, cospi8_v)); \ temp11 = vec_sub(vec_mulo(tmp16_1, cospi24_v), vec_mule(tmp16_1, cospi8_v)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ inD = vec_packs(temp10, temp11); \ \ inB = outB; \ inC = outC; \ \ /* stage 5 */ \ out0 = vec_add(in0, in3); \ out1 = vec_add(in1, in2); \ out2 = vec_sub(in1, in2); \ out3 = vec_sub(in0, in3); \ out4 = in4; \ STEP16_1(in6, in5, out5, out6, cospi16_v); \ out7 = in7; \ \ out8 = vec_add(in8, inB); \ out9 = vec_add(in9, inA); \ outA = vec_sub(in9, inA); \ outB = vec_sub(in8, inB); \ outC = vec_sub(inF, inC); \ outD = vec_sub(inE, inD); \ outE = vec_add(inD, inE); \ outF = vec_add(inC, inF); \ \ /* stage 6 */ \ in0 = vec_add(out0, out7); \ in1 = vec_add(out1, out6); \ in2 = vec_add(out2, out5); \ in3 = vec_add(out3, out4); \ in4 = vec_sub(out3, out4); \ in5 = vec_sub(out2, out5); \ in6 = vec_sub(out1, out6); \ in7 = vec_sub(out0, out7); \ in8 = out8; \ in9 = out9; \ STEP16_1(outD, outA, inA, inD, cospi16_v); \ STEP16_1(outC, outB, inB, inC, cospi16_v); \ inE = outE; \ inF = outF; \ \ /* stage 7 */ \ out0 = vec_add(in0, inF); \ out1 = vec_add(in1, inE); \ out2 = vec_add(in2, inD); \ out3 = vec_add(in3, inC); \ out4 = vec_add(in4, inB); \ out5 = vec_add(in5, inA); \ out6 = vec_add(in6, in9); \ out7 = vec_add(in7, in8); \ out8 = vec_sub(in7, in8); \ out9 = vec_sub(in6, in9); \ outA = vec_sub(in5, inA); \ outB = vec_sub(in4, inB); \ outC = vec_sub(in3, inC); \ outD = vec_sub(in2, inD); \ outE = vec_sub(in1, inE); \ outF = vec_sub(in0, inF); #define PIXEL_ADD_STORE16(in0, in1, dst, offset) \ d_uh = (int16x8_t)vec_mergeh(dst, zerov); \ d_ul = (int16x8_t)vec_mergel(dst, zerov); \ PIXEL_ADD(in0, d_uh, add, shift6); \ PIXEL_ADD(in1, d_ul, add, shift6); \ vec_vsx_st(vec_packsu(d_uh, d_ul), offset, dest); void vpx_idct16x16_256_add_vsx(const tran_low_t *input, uint8_t *dest, int stride) { int32x4_t temp10, temp11, temp20, temp21, temp30; int16x8_t src00, src01, src02, src03, src04, src05, src06, src07, src10, src11, src12, src13, src14, src15, src16, src17; int16x8_t src20, src21, src22, src23, src24, src25, src26, src27, src30, src31, src32, src33, src34, src35, src36, src37; int16x8_t tmp00, tmp01, tmp02, tmp03, tmp04, tmp05, tmp06, tmp07, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, tmp16, tmp17, tmp16_0, tmp16_1; int16x8_t tmp20, tmp21, tmp22, tmp23, tmp24, tmp25, tmp26, tmp27, tmp30, tmp31, tmp32, tmp33, tmp34, tmp35, tmp36, tmp37; uint8x16_t dest0, dest1, dest2, dest3, dest4, dest5, dest6, dest7, dest8, dest9, destA, destB, destC, destD, destE, destF; int16x8_t d_uh, d_ul; int16x8_t add = vec_sl(vec_splat_s16(8), vec_splat_u16(2)); uint16x8_t shift6 = vec_splat_u16(6); uint8x16_t zerov = vec_splat_u8(0); ROUND_SHIFT_INIT; // transform rows // load and transform the upper half of 16x16 matrix LOAD_INPUT16(load_tran_low, input, 0, 8 * sizeof(*input), src00, src10, src01, src11, src02, src12, src03, src13, src04, src14, src05, src15, src06, src16, src07, src17); TRANSPOSE8x8(src00, src01, src02, src03, src04, src05, src06, src07, tmp00, tmp01, tmp02, tmp03, tmp04, tmp05, tmp06, tmp07); TRANSPOSE8x8(src10, src11, src12, src13, src14, src15, src16, src17, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, tmp16, tmp17); IDCT16(tmp00, tmp01, tmp02, tmp03, tmp04, tmp05, tmp06, tmp07, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, tmp16, tmp17, src00, src01, src02, src03, src04, src05, src06, src07, src10, src11, src12, src13, src14, src15, src16, src17); TRANSPOSE8x8(src00, src01, src02, src03, src04, src05, src06, src07, tmp00, tmp01, tmp02, tmp03, tmp04, tmp05, tmp06, tmp07); TRANSPOSE8x8(src10, src11, src12, src13, src14, src15, src16, src17, tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, tmp16, tmp17); // load and transform the lower half of 16x16 matrix LOAD_INPUT16(load_tran_low, input, 8 * 8 * 2 * sizeof(*input), 8 * sizeof(*input), src20, src30, src21, src31, src22, src32, src23, src33, src24, src34, src25, src35, src26, src36, src27, src37); TRANSPOSE8x8(src20, src21, src22, src23, src24, src25, src26, src27, tmp20, tmp21, tmp22, tmp23, tmp24, tmp25, tmp26, tmp27); TRANSPOSE8x8(src30, src31, src32, src33, src34, src35, src36, src37, tmp30, tmp31, tmp32, tmp33, tmp34, tmp35, tmp36, tmp37); IDCT16(tmp20, tmp21, tmp22, tmp23, tmp24, tmp25, tmp26, tmp27, tmp30, tmp31, tmp32, tmp33, tmp34, tmp35, tmp36, tmp37, src20, src21, src22, src23, src24, src25, src26, src27, src30, src31, src32, src33, src34, src35, src36, src37); TRANSPOSE8x8(src20, src21, src22, src23, src24, src25, src26, src27, tmp20, tmp21, tmp22, tmp23, tmp24, tmp25, tmp26, tmp27); TRANSPOSE8x8(src30, src31, src32, src33, src34, src35, src36, src37, tmp30, tmp31, tmp32, tmp33, tmp34, tmp35, tmp36, tmp37); // transform columns // left half first IDCT16(tmp00, tmp01, tmp02, tmp03, tmp04, tmp05, tmp06, tmp07, tmp20, tmp21, tmp22, tmp23, tmp24, tmp25, tmp26, tmp27, src00, src01, src02, src03, src04, src05, src06, src07, src20, src21, src22, src23, src24, src25, src26, src27); // right half IDCT16(tmp10, tmp11, tmp12, tmp13, tmp14, tmp15, tmp16, tmp17, tmp30, tmp31, tmp32, tmp33, tmp34, tmp35, tmp36, tmp37, src10, src11, src12, src13, src14, src15, src16, src17, src30, src31, src32, src33, src34, src35, src36, src37); // load dest LOAD_INPUT16(vec_vsx_ld, dest, 0, stride, dest0, dest1, dest2, dest3, dest4, dest5, dest6, dest7, dest8, dest9, destA, destB, destC, destD, destE, destF); PIXEL_ADD_STORE16(src00, src10, dest0, 0); PIXEL_ADD_STORE16(src01, src11, dest1, stride); PIXEL_ADD_STORE16(src02, src12, dest2, 2 * stride); PIXEL_ADD_STORE16(src03, src13, dest3, 3 * stride); PIXEL_ADD_STORE16(src04, src14, dest4, 4 * stride); PIXEL_ADD_STORE16(src05, src15, dest5, 5 * stride); PIXEL_ADD_STORE16(src06, src16, dest6, 6 * stride); PIXEL_ADD_STORE16(src07, src17, dest7, 7 * stride); PIXEL_ADD_STORE16(src20, src30, dest8, 8 * stride); PIXEL_ADD_STORE16(src21, src31, dest9, 9 * stride); PIXEL_ADD_STORE16(src22, src32, destA, 10 * stride); PIXEL_ADD_STORE16(src23, src33, destB, 11 * stride); PIXEL_ADD_STORE16(src24, src34, destC, 12 * stride); PIXEL_ADD_STORE16(src25, src35, destD, 13 * stride); PIXEL_ADD_STORE16(src26, src36, destE, 14 * stride); PIXEL_ADD_STORE16(src27, src37, destF, 15 * stride); } #define LOAD_8x32(load, in00, in01, in02, in03, in10, in11, in12, in13, in20, \ in21, in22, in23, in30, in31, in32, in33, in40, in41, in42, \ in43, in50, in51, in52, in53, in60, in61, in62, in63, in70, \ in71, in72, in73, offset) \ /* load the first row from the 8x32 block*/ \ in00 = load(offset, input); \ in01 = load(offset + 16, input); \ in02 = load(offset + 2 * 16, input); \ in03 = load(offset + 3 * 16, input); \ \ in10 = load(offset + 4 * 16, input); \ in11 = load(offset + 5 * 16, input); \ in12 = load(offset + 6 * 16, input); \ in13 = load(offset + 7 * 16, input); \ \ in20 = load(offset + 8 * 16, input); \ in21 = load(offset + 9 * 16, input); \ in22 = load(offset + 10 * 16, input); \ in23 = load(offset + 11 * 16, input); \ \ in30 = load(offset + 12 * 16, input); \ in31 = load(offset + 13 * 16, input); \ in32 = load(offset + 14 * 16, input); \ in33 = load(offset + 15 * 16, input); \ \ in40 = load(offset + 16 * 16, input); \ in41 = load(offset + 17 * 16, input); \ in42 = load(offset + 18 * 16, input); \ in43 = load(offset + 19 * 16, input); \ \ in50 = load(offset + 20 * 16, input); \ in51 = load(offset + 21 * 16, input); \ in52 = load(offset + 22 * 16, input); \ in53 = load(offset + 23 * 16, input); \ \ in60 = load(offset + 24 * 16, input); \ in61 = load(offset + 25 * 16, input); \ in62 = load(offset + 26 * 16, input); \ in63 = load(offset + 27 * 16, input); \ \ /* load the last row from the 8x32 block*/ \ in70 = load(offset + 28 * 16, input); \ in71 = load(offset + 29 * 16, input); \ in72 = load(offset + 30 * 16, input); \ in73 = load(offset + 31 * 16, input); /* for the: temp1 = -step[x] * cospi_q + step[y] * cospi_z * temp2 = step[x] * cospi_z + step[y] * cospi_q */ #define STEP32(inpt0, inpt1, outpt0, outpt1, cospi0, cospi1) \ tmp16_0 = vec_mergeh(inpt0, inpt1); \ tmp16_1 = vec_mergel(inpt0, inpt1); \ temp10 = vec_sub(vec_mulo(tmp16_0, cospi1), vec_mule(tmp16_0, cospi0)); \ temp11 = vec_sub(vec_mulo(tmp16_1, cospi1), vec_mule(tmp16_1, cospi0)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt0 = vec_packs(temp10, temp11); \ temp10 = vec_add(vec_mule(tmp16_0, cospi1), vec_mulo(tmp16_0, cospi0)); \ temp11 = vec_add(vec_mule(tmp16_1, cospi1), vec_mulo(tmp16_1, cospi0)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt1 = vec_packs(temp10, temp11); /* for the: temp1 = -step[x] * cospi_q - step[y] * cospi_z * temp2 = -step[x] * cospi_z + step[y] * cospi_q */ #define STEP32_1(inpt0, inpt1, outpt0, outpt1, cospi0, cospi1, cospi1m) \ tmp16_0 = vec_mergeh(inpt0, inpt1); \ tmp16_1 = vec_mergel(inpt0, inpt1); \ temp10 = vec_sub(vec_mulo(tmp16_0, cospi1m), vec_mule(tmp16_0, cospi0)); \ temp11 = vec_sub(vec_mulo(tmp16_1, cospi1m), vec_mule(tmp16_1, cospi0)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt0 = vec_packs(temp10, temp11); \ temp10 = vec_sub(vec_mulo(tmp16_0, cospi0), vec_mule(tmp16_0, cospi1)); \ temp11 = vec_sub(vec_mulo(tmp16_1, cospi0), vec_mule(tmp16_1, cospi1)); \ DCT_CONST_ROUND_SHIFT(temp10); \ DCT_CONST_ROUND_SHIFT(temp11); \ outpt1 = vec_packs(temp10, temp11); #define IDCT32(in0, in1, in2, in3, out) \ \ /* stage 1 */ \ /* out[0][0] = in[0][0]; */ \ out[0][1] = in2[0]; \ out[0][2] = in1[0]; \ out[0][3] = in3[0]; \ out[0][4] = in0[4]; \ out[0][5] = in2[4]; \ out[0][6] = in1[4]; \ out[0][7] = in3[4]; \ out[1][0] = in0[2]; \ out[1][1] = in2[2]; \ out[1][2] = in1[2]; \ out[1][3] = in3[2]; \ out[1][4] = in0[6]; \ out[1][5] = in2[6]; \ out[1][6] = in1[6]; \ out[1][7] = in3[6]; \ \ STEP8_0(in0[1], in3[7], out[2][0], out[3][7], cospi31_v, cospi1_v); \ STEP8_0(in2[1], in1[7], out[2][1], out[3][6], cospi15_v, cospi17_v); \ STEP8_0(in1[1], in2[7], out[2][2], out[3][5], cospi23_v, cospi9_v); \ STEP8_0(in3[1], in0[7], out[2][3], out[3][4], cospi7_v, cospi25_v); \ STEP8_0(in0[5], in3[3], out[2][4], out[3][3], cospi27_v, cospi5_v); \ STEP8_0(in2[5], in1[3], out[2][5], out[3][2], cospi11_v, cospi21_v); \ STEP8_0(in1[5], in2[3], out[2][6], out[3][1], cospi19_v, cospi13_v); \ STEP8_0(in3[5], in0[3], out[2][7], out[3][0], cospi3_v, cospi29_v); \ \ /* stage 2 */ \ /* in0[0] = out[0][0]; */ \ in0[1] = out[0][1]; \ in0[2] = out[0][2]; \ in0[3] = out[0][3]; \ in0[4] = out[0][4]; \ in0[5] = out[0][5]; \ in0[6] = out[0][6]; \ in0[7] = out[0][7]; \ \ STEP8_0(out[1][0], out[1][7], in1[0], in1[7], cospi30_v, cospi2_v); \ STEP8_0(out[1][1], out[1][6], in1[1], in1[6], cospi14_v, cospi18_v); \ STEP8_0(out[1][2], out[1][5], in1[2], in1[5], cospi22_v, cospi10_v); \ STEP8_0(out[1][3], out[1][4], in1[3], in1[4], cospi6_v, cospi26_v); \ \ in2[0] = vec_add(out[2][0], out[2][1]); \ in2[1] = vec_sub(out[2][0], out[2][1]); \ in2[2] = vec_sub(out[2][3], out[2][2]); \ in2[3] = vec_add(out[2][3], out[2][2]); \ in2[4] = vec_add(out[2][4], out[2][5]); \ in2[5] = vec_sub(out[2][4], out[2][5]); \ in2[6] = vec_sub(out[2][7], out[2][6]); \ in2[7] = vec_add(out[2][7], out[2][6]); \ in3[0] = vec_add(out[3][0], out[3][1]); \ in3[1] = vec_sub(out[3][0], out[3][1]); \ in3[2] = vec_sub(out[3][3], out[3][2]); \ in3[3] = vec_add(out[3][3], out[3][2]); \ in3[4] = vec_add(out[3][4], out[3][5]); \ in3[5] = vec_sub(out[3][4], out[3][5]); \ in3[6] = vec_sub(out[3][7], out[3][6]); \ in3[7] = vec_add(out[3][6], out[3][7]); \ \ /* stage 3 */ \ out[0][0] = in0[0]; \ out[0][1] = in0[1]; \ out[0][2] = in0[2]; \ out[0][3] = in0[3]; \ \ STEP8_0(in0[4], in0[7], out[0][4], out[0][7], cospi28_v, cospi4_v); \ STEP8_0(in0[5], in0[6], out[0][5], out[0][6], cospi12_v, cospi20_v); \ \ out[1][0] = vec_add(in1[0], in1[1]); \ out[1][1] = vec_sub(in1[0], in1[1]); \ out[1][2] = vec_sub(in1[3], in1[2]); \ out[1][3] = vec_add(in1[2], in1[3]); \ out[1][4] = vec_add(in1[4], in1[5]); \ out[1][5] = vec_sub(in1[4], in1[5]); \ out[1][6] = vec_sub(in1[7], in1[6]); \ out[1][7] = vec_add(in1[6], in1[7]); \ \ out[2][0] = in2[0]; \ out[3][7] = in3[7]; \ STEP32(in2[1], in3[6], out[2][1], out[3][6], cospi4_v, cospi28_v); \ STEP32_1(in2[2], in3[5], out[2][2], out[3][5], cospi28_v, cospi4_v, \ cospi4m_v); \ out[2][3] = in2[3]; \ out[2][4] = in2[4]; \ STEP32(in2[5], in3[2], out[2][5], out[3][2], cospi20_v, cospi12_v); \ STEP32_1(in2[6], in3[1], out[2][6], out[3][1], cospi12_v, cospi20_v, \ cospi20m_v); \ out[2][7] = in2[7]; \ out[3][0] = in3[0]; \ out[3][3] = in3[3]; \ out[3][4] = in3[4]; \ \ /* stage 4 */ \ STEP16_1(out[0][0], out[0][1], in0[1], in0[0], cospi16_v); \ STEP8_0(out[0][2], out[0][3], in0[2], in0[3], cospi24_v, cospi8_v); \ in0[4] = vec_add(out[0][4], out[0][5]); \ in0[5] = vec_sub(out[0][4], out[0][5]); \ in0[6] = vec_sub(out[0][7], out[0][6]); \ in0[7] = vec_add(out[0][7], out[0][6]); \ \ in1[0] = out[1][0]; \ in1[7] = out[1][7]; \ STEP32(out[1][1], out[1][6], in1[1], in1[6], cospi8_v, cospi24_v); \ STEP32_1(out[1][2], out[1][5], in1[2], in1[5], cospi24_v, cospi8_v, \ cospi8m_v); \ in1[3] = out[1][3]; \ in1[4] = out[1][4]; \ \ in2[0] = vec_add(out[2][0], out[2][3]); \ in2[1] = vec_add(out[2][1], out[2][2]); \ in2[2] = vec_sub(out[2][1], out[2][2]); \ in2[3] = vec_sub(out[2][0], out[2][3]); \ in2[4] = vec_sub(out[2][7], out[2][4]); \ in2[5] = vec_sub(out[2][6], out[2][5]); \ in2[6] = vec_add(out[2][5], out[2][6]); \ in2[7] = vec_add(out[2][4], out[2][7]); \ \ in3[0] = vec_add(out[3][0], out[3][3]); \ in3[1] = vec_add(out[3][1], out[3][2]); \ in3[2] = vec_sub(out[3][1], out[3][2]); \ in3[3] = vec_sub(out[3][0], out[3][3]); \ in3[4] = vec_sub(out[3][7], out[3][4]); \ in3[5] = vec_sub(out[3][6], out[3][5]); \ in3[6] = vec_add(out[3][5], out[3][6]); \ in3[7] = vec_add(out[3][4], out[3][7]); \ \ /* stage 5 */ \ out[0][0] = vec_add(in0[0], in0[3]); \ out[0][1] = vec_add(in0[1], in0[2]); \ out[0][2] = vec_sub(in0[1], in0[2]); \ out[0][3] = vec_sub(in0[0], in0[3]); \ out[0][4] = in0[4]; \ STEP16_1(in0[6], in0[5], out[0][5], out[0][6], cospi16_v); \ out[0][7] = in0[7]; \ \ out[1][0] = vec_add(in1[0], in1[3]); \ out[1][1] = vec_add(in1[1], in1[2]); \ out[1][2] = vec_sub(in1[1], in1[2]); \ out[1][3] = vec_sub(in1[0], in1[3]); \ out[1][4] = vec_sub(in1[7], in1[4]); \ out[1][5] = vec_sub(in1[6], in1[5]); \ out[1][6] = vec_add(in1[5], in1[6]); \ out[1][7] = vec_add(in1[4], in1[7]); \ \ out[2][0] = in2[0]; \ out[2][1] = in2[1]; \ STEP32(in2[2], in3[5], out[2][2], out[3][5], cospi8_v, cospi24_v); \ STEP32(in2[3], in3[4], out[2][3], out[3][4], cospi8_v, cospi24_v); \ STEP32_1(in2[4], in3[3], out[2][4], out[3][3], cospi24_v, cospi8_v, \ cospi8m_v); \ STEP32_1(in2[5], in3[2], out[2][5], out[3][2], cospi24_v, cospi8_v, \ cospi8m_v); \ out[2][6] = in2[6]; \ out[2][7] = in2[7]; \ out[3][0] = in3[0]; \ out[3][1] = in3[1]; \ out[3][6] = in3[6]; \ out[3][7] = in3[7]; \ \ /* stage 6 */ \ in0[0] = vec_add(out[0][0], out[0][7]); \ in0[1] = vec_add(out[0][1], out[0][6]); \ in0[2] = vec_add(out[0][2], out[0][5]); \ in0[3] = vec_add(out[0][3], out[0][4]); \ in0[4] = vec_sub(out[0][3], out[0][4]); \ in0[5] = vec_sub(out[0][2], out[0][5]); \ in0[6] = vec_sub(out[0][1], out[0][6]); \ in0[7] = vec_sub(out[0][0], out[0][7]); \ in1[0] = out[1][0]; \ in1[1] = out[1][1]; \ STEP16_1(out[1][5], out[1][2], in1[2], in1[5], cospi16_v); \ STEP16_1(out[1][4], out[1][3], in1[3], in1[4], cospi16_v); \ in1[6] = out[1][6]; \ in1[7] = out[1][7]; \ \ in2[0] = vec_add(out[2][0], out[2][7]); \ in2[1] = vec_add(out[2][1], out[2][6]); \ in2[2] = vec_add(out[2][2], out[2][5]); \ in2[3] = vec_add(out[2][3], out[2][4]); \ in2[4] = vec_sub(out[2][3], out[2][4]); \ in2[5] = vec_sub(out[2][2], out[2][5]); \ in2[6] = vec_sub(out[2][1], out[2][6]); \ in2[7] = vec_sub(out[2][0], out[2][7]); \ \ in3[0] = vec_sub(out[3][7], out[3][0]); \ in3[1] = vec_sub(out[3][6], out[3][1]); \ in3[2] = vec_sub(out[3][5], out[3][2]); \ in3[3] = vec_sub(out[3][4], out[3][3]); \ in3[4] = vec_add(out[3][4], out[3][3]); \ in3[5] = vec_add(out[3][5], out[3][2]); \ in3[6] = vec_add(out[3][6], out[3][1]); \ in3[7] = vec_add(out[3][7], out[3][0]); \ \ /* stage 7 */ \ out[0][0] = vec_add(in0[0], in1[7]); \ out[0][1] = vec_add(in0[1], in1[6]); \ out[0][2] = vec_add(in0[2], in1[5]); \ out[0][3] = vec_add(in0[3], in1[4]); \ out[0][4] = vec_add(in0[4], in1[3]); \ out[0][5] = vec_add(in0[5], in1[2]); \ out[0][6] = vec_add(in0[6], in1[1]); \ out[0][7] = vec_add(in0[7], in1[0]); \ out[1][0] = vec_sub(in0[7], in1[0]); \ out[1][1] = vec_sub(in0[6], in1[1]); \ out[1][2] = vec_sub(in0[5], in1[2]); \ out[1][3] = vec_sub(in0[4], in1[3]); \ out[1][4] = vec_sub(in0[3], in1[4]); \ out[1][5] = vec_sub(in0[2], in1[5]); \ out[1][6] = vec_sub(in0[1], in1[6]); \ out[1][7] = vec_sub(in0[0], in1[7]); \ \ out[2][0] = in2[0]; \ out[2][1] = in2[1]; \ out[2][2] = in2[2]; \ out[2][3] = in2[3]; \ STEP16_1(in3[3], in2[4], out[2][4], out[3][3], cospi16_v); \ STEP16_1(in3[2], in2[5], out[2][5], out[3][2], cospi16_v); \ STEP16_1(in3[1], in2[6], out[2][6], out[3][1], cospi16_v); \ STEP16_1(in3[0], in2[7], out[2][7], out[3][0], cospi16_v); \ out[3][4] = in3[4]; \ out[3][5] = in3[5]; \ out[3][6] = in3[6]; \ out[3][7] = in3[7]; \ \ /* final */ \ in0[0] = vec_add(out[0][0], out[3][7]); \ in0[1] = vec_add(out[0][1], out[3][6]); \ in0[2] = vec_add(out[0][2], out[3][5]); \ in0[3] = vec_add(out[0][3], out[3][4]); \ in0[4] = vec_add(out[0][4], out[3][3]); \ in0[5] = vec_add(out[0][5], out[3][2]); \ in0[6] = vec_add(out[0][6], out[3][1]); \ in0[7] = vec_add(out[0][7], out[3][0]); \ in1[0] = vec_add(out[1][0], out[2][7]); \ in1[1] = vec_add(out[1][1], out[2][6]); \ in1[2] = vec_add(out[1][2], out[2][5]); \ in1[3] = vec_add(out[1][3], out[2][4]); \ in1[4] = vec_add(out[1][4], out[2][3]); \ in1[5] = vec_add(out[1][5], out[2][2]); \ in1[6] = vec_add(out[1][6], out[2][1]); \ in1[7] = vec_add(out[1][7], out[2][0]); \ in2[0] = vec_sub(out[1][7], out[2][0]); \ in2[1] = vec_sub(out[1][6], out[2][1]); \ in2[2] = vec_sub(out[1][5], out[2][2]); \ in2[3] = vec_sub(out[1][4], out[2][3]); \ in2[4] = vec_sub(out[1][3], out[2][4]); \ in2[5] = vec_sub(out[1][2], out[2][5]); \ in2[6] = vec_sub(out[1][1], out[2][6]); \ in2[7] = vec_sub(out[1][0], out[2][7]); \ in3[0] = vec_sub(out[0][7], out[3][0]); \ in3[1] = vec_sub(out[0][6], out[3][1]); \ in3[2] = vec_sub(out[0][5], out[3][2]); \ in3[3] = vec_sub(out[0][4], out[3][3]); \ in3[4] = vec_sub(out[0][3], out[3][4]); \ in3[5] = vec_sub(out[0][2], out[3][5]); \ in3[6] = vec_sub(out[0][1], out[3][6]); \ in3[7] = vec_sub(out[0][0], out[3][7]); // NOT A FULL TRANSPOSE! Transposes just each 8x8 block in each row, // does not transpose rows #define TRANSPOSE_8x32(in, out) \ /* transpose 4 of 8x8 blocks */ \ TRANSPOSE8x8(in[0][0], in[0][1], in[0][2], in[0][3], in[0][4], in[0][5], \ in[0][6], in[0][7], out[0][0], out[0][1], out[0][2], out[0][3], \ out[0][4], out[0][5], out[0][6], out[0][7]); \ TRANSPOSE8x8(in[1][0], in[1][1], in[1][2], in[1][3], in[1][4], in[1][5], \ in[1][6], in[1][7], out[1][0], out[1][1], out[1][2], out[1][3], \ out[1][4], out[1][5], out[1][6], out[1][7]); \ TRANSPOSE8x8(in[2][0], in[2][1], in[2][2], in[2][3], in[2][4], in[2][5], \ in[2][6], in[2][7], out[2][0], out[2][1], out[2][2], out[2][3], \ out[2][4], out[2][5], out[2][6], out[2][7]); \ TRANSPOSE8x8(in[3][0], in[3][1], in[3][2], in[3][3], in[3][4], in[3][5], \ in[3][6], in[3][7], out[3][0], out[3][1], out[3][2], out[3][3], \ out[3][4], out[3][5], out[3][6], out[3][7]); #define PIXEL_ADD_STORE32(in0, in1, in2, in3, step) \ dst = vec_vsx_ld((step)*stride, dest); \ d_uh = (int16x8_t)vec_mergeh(dst, zerov); \ d_ul = (int16x8_t)vec_mergel(dst, zerov); \ PIXEL_ADD(in0, d_uh, add, shift6); \ PIXEL_ADD(in1, d_ul, add, shift6); \ vec_vsx_st(vec_packsu(d_uh, d_ul), (step)*stride, dest); \ dst = vec_vsx_ld((step)*stride + 16, dest); \ d_uh = (int16x8_t)vec_mergeh(dst, zerov); \ d_ul = (int16x8_t)vec_mergel(dst, zerov); \ PIXEL_ADD(in2, d_uh, add, shift6); \ PIXEL_ADD(in3, d_ul, add, shift6); \ vec_vsx_st(vec_packsu(d_uh, d_ul), (step)*stride + 16, dest); #define ADD_STORE_BLOCK(in, offset) \ PIXEL_ADD_STORE32(in[0][0], in[1][0], in[2][0], in[3][0], offset + 0); \ PIXEL_ADD_STORE32(in[0][1], in[1][1], in[2][1], in[3][1], offset + 1); \ PIXEL_ADD_STORE32(in[0][2], in[1][2], in[2][2], in[3][2], offset + 2); \ PIXEL_ADD_STORE32(in[0][3], in[1][3], in[2][3], in[3][3], offset + 3); \ PIXEL_ADD_STORE32(in[0][4], in[1][4], in[2][4], in[3][4], offset + 4); \ PIXEL_ADD_STORE32(in[0][5], in[1][5], in[2][5], in[3][5], offset + 5); \ PIXEL_ADD_STORE32(in[0][6], in[1][6], in[2][6], in[3][6], offset + 6); \ PIXEL_ADD_STORE32(in[0][7], in[1][7], in[2][7], in[3][7], offset + 7); void vpx_idct32x32_1024_add_vsx(const tran_low_t *input, uint8_t *dest, int stride) { int16x8_t src0[4][8], src1[4][8], src2[4][8], src3[4][8], tmp[4][8]; int16x8_t tmp16_0, tmp16_1; int32x4_t temp10, temp11, temp20, temp21, temp30; uint8x16_t dst; int16x8_t d_uh, d_ul; int16x8_t add = vec_sl(vec_splat_s16(8), vec_splat_u16(2)); uint16x8_t shift6 = vec_splat_u16(6); uint8x16_t zerov = vec_splat_u8(0); ROUND_SHIFT_INIT; LOAD_8x32(load_tran_low, src0[0][0], src0[1][0], src0[2][0], src0[3][0], src0[0][1], src0[1][1], src0[2][1], src0[3][1], src0[0][2], src0[1][2], src0[2][2], src0[3][2], src0[0][3], src0[1][3], src0[2][3], src0[3][3], src0[0][4], src0[1][4], src0[2][4], src0[3][4], src0[0][5], src0[1][5], src0[2][5], src0[3][5], src0[0][6], src0[1][6], src0[2][6], src0[3][6], src0[0][7], src0[1][7], src0[2][7], src0[3][7], 0); // Rows // transpose the first row of 8x8 blocks TRANSPOSE_8x32(src0, tmp); // transform the 32x8 column IDCT32(tmp[0], tmp[1], tmp[2], tmp[3], src0); TRANSPOSE_8x32(tmp, src0); LOAD_8x32(load_tran_low, src1[0][0], src1[1][0], src1[2][0], src1[3][0], src1[0][1], src1[1][1], src1[2][1], src1[3][1], src1[0][2], src1[1][2], src1[2][2], src1[3][2], src1[0][3], src1[1][3], src1[2][3], src1[3][3], src1[0][4], src1[1][4], src1[2][4], src1[3][4], src1[0][5], src1[1][5], src1[2][5], src1[3][5], src1[0][6], src1[1][6], src1[2][6], src1[3][6], src1[0][7], src1[1][7], src1[2][7], src1[3][7], 512); TRANSPOSE_8x32(src1, tmp); IDCT32(tmp[0], tmp[1], tmp[2], tmp[3], src1); TRANSPOSE_8x32(tmp, src1); LOAD_8x32(load_tran_low, src2[0][0], src2[1][0], src2[2][0], src2[3][0], src2[0][1], src2[1][1], src2[2][1], src2[3][1], src2[0][2], src2[1][2], src2[2][2], src2[3][2], src2[0][3], src2[1][3], src2[2][3], src2[3][3], src2[0][4], src2[1][4], src2[2][4], src2[3][4], src2[0][5], src2[1][5], src2[2][5], src2[3][5], src2[0][6], src2[1][6], src2[2][6], src2[3][6], src2[0][7], src2[1][7], src2[2][7], src2[3][7], 1024); TRANSPOSE_8x32(src2, tmp); IDCT32(tmp[0], tmp[1], tmp[2], tmp[3], src2); TRANSPOSE_8x32(tmp, src2); LOAD_8x32(load_tran_low, src3[0][0], src3[1][0], src3[2][0], src3[3][0], src3[0][1], src3[1][1], src3[2][1], src3[3][1], src3[0][2], src3[1][2], src3[2][2], src3[3][2], src3[0][3], src3[1][3], src3[2][3], src3[3][3], src3[0][4], src3[1][4], src3[2][4], src3[3][4], src3[0][5], src3[1][5], src3[2][5], src3[3][5], src3[0][6], src3[1][6], src3[2][6], src3[3][6], src3[0][7], src3[1][7], src3[2][7], src3[3][7], 1536); TRANSPOSE_8x32(src3, tmp); IDCT32(tmp[0], tmp[1], tmp[2], tmp[3], src3); TRANSPOSE_8x32(tmp, src3); // Columns IDCT32(src0[0], src1[0], src2[0], src3[0], tmp); IDCT32(src0[1], src1[1], src2[1], src3[1], tmp); IDCT32(src0[2], src1[2], src2[2], src3[2], tmp); IDCT32(src0[3], src1[3], src2[3], src3[3], tmp); ADD_STORE_BLOCK(src0, 0); ADD_STORE_BLOCK(src1, 8); ADD_STORE_BLOCK(src2, 16); ADD_STORE_BLOCK(src3, 24); }