blob: 7a23174be2818b3bf9bca3cf925d6ae8022e77bb [file] [log] [blame]
/*
* Copyright (c) 2018, Alliance for Open Media. All rights reserved
*
* This source code is subject to the terms of the BSD 2 Clause License and
* the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License
* was not distributed with this source code in the LICENSE file, you can
* obtain it at www.aomedia.org/license/software. If the Alliance for Open
* Media Patent License 1.0 was not distributed with this source code in the
* PATENTS file, you can obtain it at www.aomedia.org/license/patent.
*/
#include <arm_neon.h>
#include "config/aom_config.h"
#include "config/aom_dsp_rtcd.h"
#include "config/av1_rtcd.h"
#include "av1/common/av1_inv_txfm1d.h"
#include "av1/common/av1_inv_txfm1d_cfg.h"
#include "av1/common/av1_txfm.h"
#include "av1/common/enums.h"
#include "av1/common/idct.h"
#include "av1/common/arm/av1_inv_txfm_neon.h"
#include "av1/common/arm/transpose_neon.h"
// 1D itx types
typedef enum ATTRIBUTE_PACKED {
IDCT_1D,
IADST_1D,
IFLIPADST_1D = IADST_1D,
IIDENTITY_1D,
ITX_TYPES_1D,
} ITX_TYPE_1D;
static const ITX_TYPE_1D vitx_1d_tab[TX_TYPES] = {
IDCT_1D, IADST_1D, IDCT_1D, IADST_1D,
IFLIPADST_1D, IDCT_1D, IFLIPADST_1D, IADST_1D,
IFLIPADST_1D, IIDENTITY_1D, IDCT_1D, IIDENTITY_1D,
IADST_1D, IIDENTITY_1D, IFLIPADST_1D, IIDENTITY_1D,
};
static const ITX_TYPE_1D hitx_1d_tab[TX_TYPES] = {
IDCT_1D, IDCT_1D, IADST_1D, IADST_1D,
IDCT_1D, IFLIPADST_1D, IFLIPADST_1D, IFLIPADST_1D,
IADST_1D, IIDENTITY_1D, IIDENTITY_1D, IDCT_1D,
IIDENTITY_1D, IADST_1D, IIDENTITY_1D, IFLIPADST_1D,
};
// 1D functions
static const transform_1d_neon lowbd_txfm_all_1d_arr[TX_SIZES][ITX_TYPES_1D] = {
{ av1_idct4_new, av1_iadst4_new, av1_iidentity4_c },
{ av1_idct8_new, av1_iadst8_new, av1_iidentity8_c },
{ av1_idct16_new, av1_iadst16_new, av1_iidentity16_c },
{ av1_idct32_new, NULL, NULL },
{ av1_idct64_new, NULL, NULL },
};
static INLINE void lowbd_add_flip_buffer_8xn_neon(int16x8_t *in,
uint8_t *output, int stride,
int flipud,
const int height) {
int j = flipud ? (height - 1) : 0;
const int step = flipud ? -1 : 1;
int16x8_t temp_output;
for (int i = 0; i < height; ++i, j += step) {
temp_output = vreinterpretq_s16_u16(vmovl_u8(vld1_u8(output)));
temp_output = vaddq_s16(temp_output, in[j]);
vst1_u8(output, vqmovun_s16(temp_output));
output += stride;
}
}
static INLINE uint8x16_t lowbd_get_recon_16x16_neon(const uint8x16_t pred,
int16x8_t res0,
int16x8_t res1) {
int16x8_t temp_output[2];
uint8x16_t temp_output_8q;
temp_output[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pred)));
temp_output[0] = vaddq_s16(temp_output[0], res0);
temp_output[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pred)));
temp_output[1] = vaddq_s16(temp_output[1], res1);
temp_output_8q =
vcombine_u8(vqmovun_s16(temp_output[0]), vqmovun_s16(temp_output[1]));
return temp_output_8q;
}
static INLINE void lowbd_add_flip_buffer_16xn_neon(int16x8_t *in,
uint8_t *output, int stride,
int flipud, int height) {
uint8x16_t temp_output_8q;
int j = flipud ? (height - 1) : 0;
const int step = flipud ? -1 : 1;
for (int i = 0; i < height; ++i, j += step) {
temp_output_8q = vld1q_u8(output + i * stride);
temp_output_8q =
lowbd_get_recon_16x16_neon(temp_output_8q, in[j], in[j + height]);
vst1q_u8((output + i * stride), temp_output_8q);
}
}
static INLINE void lowbd_inv_txfm2d_memset_neon(int16x8_t *a, int size,
int value) {
for (int i = 0; i < size; i++) {
a[i] = vdupq_n_s16((int16_t)value);
}
}
static INLINE void btf_16_lane_0_1_neon(const int16x8_t in0,
const int16x8_t in1, const int16x4_t c,
int16x8_t *t0, int16x8_t *t1) {
int32x4_t s0[2], s1[2];
int16x4_t v0[2], v1[2];
s0[0] = vmull_lane_s16(vget_low_s16(in0), c, 0);
s0[1] = vmull_lane_s16(vget_high_s16(in0), c, 0);
s1[0] = vmull_lane_s16(vget_low_s16(in0), c, 1);
s1[1] = vmull_lane_s16(vget_high_s16(in0), c, 1);
s0[0] = vmlal_lane_s16(s0[0], vget_low_s16(in1), c, 1);
s0[1] = vmlal_lane_s16(s0[1], vget_high_s16(in1), c, 1);
s1[0] = vmlsl_lane_s16(s1[0], vget_low_s16(in1), c, 0);
s1[1] = vmlsl_lane_s16(s1[1], vget_high_s16(in1), c, 0);
v0[0] = vrshrn_n_s32(s0[0], INV_COS_BIT);
v0[1] = vrshrn_n_s32(s0[1], INV_COS_BIT);
v1[0] = vrshrn_n_s32(s1[0], INV_COS_BIT);
v1[1] = vrshrn_n_s32(s1[1], INV_COS_BIT);
*t0 = vcombine_s16(v0[0], v0[1]);
*t1 = vcombine_s16(v1[0], v1[1]);
}
static INLINE void btf_16_lane_1_0_neon(const int16x8_t in0,
const int16x8_t in1, const int16x4_t c,
int16x8_t *t0, int16x8_t *t1) {
int32x4_t s0[2], s1[2];
int16x4_t v0[2], v1[2];
s0[0] = vmull_lane_s16(vget_low_s16(in0), c, 1);
s0[1] = vmull_lane_s16(vget_high_s16(in0), c, 1);
s1[0] = vmull_lane_s16(vget_low_s16(in0), c, 0);
s1[1] = vmull_lane_s16(vget_high_s16(in0), c, 0);
s0[0] = vmlal_lane_s16(s0[0], vget_low_s16(in1), c, 0);
s0[1] = vmlal_lane_s16(s0[1], vget_high_s16(in1), c, 0);
s1[0] = vmlsl_lane_s16(s1[0], vget_low_s16(in1), c, 1);
s1[1] = vmlsl_lane_s16(s1[1], vget_high_s16(in1), c, 1);
v0[0] = vrshrn_n_s32(s0[0], INV_COS_BIT);
v0[1] = vrshrn_n_s32(s0[1], INV_COS_BIT);
v1[0] = vrshrn_n_s32(s1[0], INV_COS_BIT);
v1[1] = vrshrn_n_s32(s1[1], INV_COS_BIT);
*t0 = vcombine_s16(v0[0], v0[1]);
*t1 = vcombine_s16(v1[0], v1[1]);
}
static INLINE void btf_16_lane_2_3_neon(const int16x8_t in0,
const int16x8_t in1, const int16x4_t c,
int16x8_t *t0, int16x8_t *t1) {
int32x4_t s0[2], s1[2];
int16x4_t v0[2], v1[2];
s0[0] = vmull_lane_s16(vget_low_s16(in0), c, 2);
s0[1] = vmull_lane_s16(vget_high_s16(in0), c, 2);
s1[0] = vmull_lane_s16(vget_low_s16(in0), c, 3);
s1[1] = vmull_lane_s16(vget_high_s16(in0), c, 3);
s0[0] = vmlal_lane_s16(s0[0], vget_low_s16(in1), c, 3);
s0[1] = vmlal_lane_s16(s0[1], vget_high_s16(in1), c, 3);
s1[0] = vmlsl_lane_s16(s1[0], vget_low_s16(in1), c, 2);
s1[1] = vmlsl_lane_s16(s1[1], vget_high_s16(in1), c, 2);
v0[0] = vrshrn_n_s32(s0[0], INV_COS_BIT);
v0[1] = vrshrn_n_s32(s0[1], INV_COS_BIT);
v1[0] = vrshrn_n_s32(s1[0], INV_COS_BIT);
v1[1] = vrshrn_n_s32(s1[1], INV_COS_BIT);
*t0 = vcombine_s16(v0[0], v0[1]);
*t1 = vcombine_s16(v1[0], v1[1]);
}
static INLINE void btf_16_neon(const int16x8_t in0, int16_t coef1,
int16_t coef2, int16x8_t *t0, int16x8_t *t1) {
int32x4_t s0_l, s0_h, s1_l, s1_h;
int16x4_t v0[2], v1[2];
s0_l = vmull_n_s16(vget_low_s16(in0), coef1);
s0_h = vmull_n_s16(vget_high_s16(in0), coef1);
s1_l = vmull_n_s16(vget_low_s16(in0), coef2);
s1_h = vmull_n_s16(vget_high_s16(in0), coef2);
v0[0] = vrshrn_n_s32(s0_l, INV_COS_BIT);
v0[1] = vrshrn_n_s32(s0_h, INV_COS_BIT);
v1[0] = vrshrn_n_s32(s1_l, INV_COS_BIT);
v1[1] = vrshrn_n_s32(s1_h, INV_COS_BIT);
*t0 = vcombine_s16(v0[0], v0[1]);
*t1 = vcombine_s16(v1[0], v1[1]);
}
static INLINE void btf_16_lane_3_2_neon(const int16x8_t in0,
const int16x8_t in1, const int16x4_t c,
int16x8_t *t0, int16x8_t *t1) {
int32x4_t s0[2], s1[2];
int16x4_t v0[2], v1[2];
s0[0] = vmull_lane_s16(vget_low_s16(in0), c, 3);
s0[1] = vmull_lane_s16(vget_high_s16(in0), c, 3);
s1[0] = vmull_lane_s16(vget_low_s16(in0), c, 2);
s1[1] = vmull_lane_s16(vget_high_s16(in0), c, 2);
s0[0] = vmlal_lane_s16(s0[0], vget_low_s16(in1), c, 2);
s0[1] = vmlal_lane_s16(s0[1], vget_high_s16(in1), c, 2);
s1[0] = vmlsl_lane_s16(s1[0], vget_low_s16(in1), c, 3);
s1[1] = vmlsl_lane_s16(s1[1], vget_high_s16(in1), c, 3);
v0[0] = vrshrn_n_s32(s0[0], INV_COS_BIT);
v0[1] = vrshrn_n_s32(s0[1], INV_COS_BIT);
v1[0] = vrshrn_n_s32(s1[0], INV_COS_BIT);
v1[1] = vrshrn_n_s32(s1[1], INV_COS_BIT);
*t0 = vcombine_s16(v0[0], v0[1]);
*t1 = vcombine_s16(v1[0], v1[1]);
}
static INLINE void btf_16_half_neon(int16x8_t *const x, const int16x4_t c) {
int32x4_t t0[2], t1[2];
int16x4_t v0[2], v1[2];
// Don't add/sub before multiply, which will overflow in iadst8.
const int32x4_t x0_lo = vmull_lane_s16(vget_low_s16(x[0]), c, 0);
const int32x4_t x0_hi = vmull_lane_s16(vget_high_s16(x[0]), c, 0);
const int32x4_t x1_lo = vmull_lane_s16(vget_low_s16(x[1]), c, 0);
const int32x4_t x1_hi = vmull_lane_s16(vget_high_s16(x[1]), c, 0);
t0[0] = vaddq_s32(x0_lo, x1_lo);
t0[1] = vaddq_s32(x0_hi, x1_hi);
t1[0] = vsubq_s32(x0_lo, x1_lo);
t1[1] = vsubq_s32(x0_hi, x1_hi);
v0[0] = vrshrn_n_s32(t0[0], INV_COS_BIT);
v0[1] = vrshrn_n_s32(t0[1], INV_COS_BIT);
v1[0] = vrshrn_n_s32(t1[0], INV_COS_BIT);
v1[1] = vrshrn_n_s32(t1[1], INV_COS_BIT);
x[0] = vcombine_s16(v0[0], v0[1]);
x[1] = vcombine_s16(v1[0], v1[1]);
}
static INLINE int16x4_t create_s16x4_neon(int16_t *const c0, int16_t *const c1,
int16_t *const c2,
int16_t *const c3) {
int16x4_t val = vdup_n_s16((int16_t)0);
val = vld1_lane_s16(c0, val, 0);
val = vld1_lane_s16(c1, val, 1);
val = vld1_lane_s16(c2, val, 2);
val = vld1_lane_s16(c3, val, 3);
return val;
}
static INLINE void iadst8_new_neon(int16x8_t *const in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 4), (int16_t *)(cospi + 60),
(int16_t *)(cospi + 20), (int16_t *)(cospi + 44));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 36), (int16_t *)(cospi + 28),
(int16_t *)(cospi + 52), (int16_t *)(cospi + 12));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
int16x8_t x[8];
int16x8_t s0, s1, s2, s3, s4, s5, s6, s7;
// Stage 1
x[0] = in[7];
x[1] = in[0];
x[2] = in[5];
x[3] = in[2];
x[4] = in[3];
x[5] = in[4];
x[6] = in[1];
x[7] = in[6];
// Stage 2
btf_16_lane_0_1_neon(x[0], x[1], c0, &s0, &s1);
btf_16_lane_2_3_neon(x[2], x[3], c0, &s2, &s3);
btf_16_lane_0_1_neon(x[4], x[5], c1, &s4, &s5);
btf_16_lane_2_3_neon(x[6], x[7], c1, &s6, &s7);
// Stage 3
x[0] = vqaddq_s16(s0, s4);
x[1] = vqaddq_s16(s1, s5);
x[2] = vqaddq_s16(s2, s6);
x[3] = vqaddq_s16(s3, s7);
x[4] = vqsubq_s16(s0, s4);
x[5] = vqsubq_s16(s1, s5);
x[6] = vqsubq_s16(s2, s6);
x[7] = vqsubq_s16(s3, s7);
// Stage 4
s0 = x[0];
s1 = x[1];
s2 = x[2];
s3 = x[3];
btf_16_lane_2_3_neon(x[4], x[5], c2, &s4, &s5);
btf_16_lane_3_2_neon(x[7], x[6], c2, &s7, &s6);
// Stage 5
x[0] = vqaddq_s16(s0, s2);
x[1] = vqaddq_s16(s1, s3);
x[2] = vqsubq_s16(s0, s2);
x[3] = vqsubq_s16(s1, s3);
x[4] = vqaddq_s16(s4, s6);
x[5] = vqaddq_s16(s5, s7);
x[6] = vqsubq_s16(s4, s6);
x[7] = vqsubq_s16(s5, s7);
// stage 6
btf_16_half_neon(x + 2, c2);
btf_16_half_neon(x + 6, c2);
// Stage 7
out[0] = x[0];
out[1] = vnegq_s16(x[4]);
out[2] = x[6];
out[3] = vnegq_s16(x[2]);
out[4] = x[3];
out[5] = vnegq_s16(x[7]);
out[6] = x[5];
out[7] = vnegq_s16(x[1]);
}
static INLINE void iadst8_low1_new_neon(int16x8_t *const in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
int16x8_t x[8];
int16x8_t s0, s1, s4, s5;
// Stage 1
x[1] = in[0];
// Stage 2
btf_16_neon(x[1], cospi[60], -cospi[4], &s0, &s1);
// Stage 3
x[0] = s0;
x[1] = s1;
x[4] = s0;
x[5] = s1;
// Stage 4
s0 = x[0];
s1 = x[1];
btf_16_lane_2_3_neon(x[4], x[5], c2, &s4, &s5);
// Stage 5
x[0] = s0;
x[1] = s1;
x[2] = s0;
x[3] = s1;
x[4] = s4;
x[5] = s5;
x[6] = s4;
x[7] = s5;
// stage 6
btf_16_half_neon(x + 2, c2);
btf_16_half_neon(x + 6, c2);
// Stage 7
out[0] = x[0];
out[1] = vnegq_s16(x[4]);
out[2] = x[6];
out[3] = vnegq_s16(x[2]);
out[4] = x[3];
out[5] = vnegq_s16(x[7]);
out[6] = x[5];
out[7] = vnegq_s16(x[1]);
}
static INLINE void idct8_new_neon(int16x8_t *in, int16x8_t *out, int8_t cos_bit,
int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1[8], step2[8];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 2
btf_16_lane_0_1_neon(in[1], in[7], c0, &step1[7], &step1[4]);
btf_16_lane_2_3_neon(in[5], in[3], c0, &step1[6], &step1[5]);
// stage 3
btf_16_lane_0_1_neon(in[0], in[4], c2, &step2[0], &step2[1]);
btf_16_lane_2_3_neon(in[2], in[6], c2, &step2[3], &step2[2]);
step2[4] = vqaddq_s16(step1[4], step1[5]);
step2[5] = vqsubq_s16(step1[4], step1[5]);
step2[6] = vqsubq_s16(step1[7], step1[6]);
step2[7] = vqaddq_s16(step1[7], step1[6]);
// stage 4
step1[0] = vqaddq_s16(step2[0], step2[3]);
step1[1] = vqaddq_s16(step2[1], step2[2]);
step1[2] = vqsubq_s16(step2[1], step2[2]);
step1[3] = vqsubq_s16(step2[0], step2[3]);
btf_16_lane_0_1_neon(step2[6], step2[5], c2, &step1[6], &step1[5]);
// stage 5
out[0] = vqaddq_s16(step1[0], step2[7]);
out[1] = vqaddq_s16(step1[1], step1[6]);
out[2] = vqaddq_s16(step1[2], step1[5]);
out[3] = vqaddq_s16(step1[3], step2[4]);
out[4] = vqsubq_s16(step1[3], step2[4]);
out[5] = vqsubq_s16(step1[2], step1[5]);
out[6] = vqsubq_s16(step1[1], step1[6]);
out[7] = vqsubq_s16(step1[0], step2[7]);
}
static INLINE void idct8_low1_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1;
int32x4_t t32[2];
// stage 1
// stage 2
// stage 3
t32[0] = vmull_n_s16(vget_low_s16(in[0]), (int16_t)cospi[32]);
t32[1] = vmull_n_s16(vget_high_s16(in[0]), (int16_t)cospi[32]);
step1 = vcombine_s16(vrshrn_n_s32(t32[0], INV_COS_BIT),
vrshrn_n_s32(t32[1], INV_COS_BIT));
// stage 4
// stage 5
out[0] = step1;
out[1] = step1;
out[2] = step1;
out[3] = step1;
out[4] = step1;
out[5] = step1;
out[6] = step1;
out[7] = step1;
}
void av1_round_shift_array_16_neon(int16x8_t *arr, int size, int bit) {
assert(!(size % 4));
if (!bit) return;
const int16x8_t dup_bits_n_16x8 = vdupq_n_s16((int16_t)(-bit));
for (int i = 0; i < size; i++) {
arr[i] = vrshlq_s16(arr[i], dup_bits_n_16x8);
}
}
static INLINE void flip_buf_ud_neon(int16x8_t *input, int size) {
int16x8_t temp[8];
for (int i = 0; i < size; ++i) {
temp[i] = input[size - 1 - i];
}
for (int i = 0; i < size; ++i) {
input[i] = temp[i];
}
}
static INLINE void load_buffer_32bit_to_16bit_neon(const int32_t *input,
int16x8_t *const a,
int out_size) {
for (int i = 0; i < 8; ++i) {
a[i] = vcombine_s16(vmovn_s32(vld1q_s32(input)),
vmovn_s32(vld1q_s32(input + 4)));
input += out_size;
}
}
static INLINE void identity8_new_neon(int16x8_t *input, int16x8_t *output,
int8_t cos_bit, int bit) {
(void)bit;
(void)cos_bit;
output[0] = vmulq_n_s16(input[0], (int16_t)2);
output[1] = vmulq_n_s16(input[1], (int16_t)2);
output[2] = vmulq_n_s16(input[2], (int16_t)2);
output[3] = vmulq_n_s16(input[3], (int16_t)2);
output[4] = vmulq_n_s16(input[4], (int16_t)2);
output[5] = vmulq_n_s16(input[5], (int16_t)2);
output[6] = vmulq_n_s16(input[6], (int16_t)2);
output[7] = vmulq_n_s16(input[7], (int16_t)2);
}
static INLINE void round_shift_for_rect(int16x8_t *input, int16x8_t *output,
int size) {
int32x4_t out_low, out_high;
int16x4_t low, high;
for (int z = 0; z < size; ++z) {
out_low = vmull_n_s16(vget_low_s16(input[z]), (int16_t)NewInvSqrt2);
out_high = vmull_n_s16(vget_high_s16(input[z]), (int16_t)NewInvSqrt2);
low = vqrshrn_n_s32(out_low, (int32_t)NewSqrt2Bits);
high = vqrshrn_n_s32(out_high, (int32_t)NewSqrt2Bits);
output[z] = vcombine_s16(low, high);
}
}
static INLINE void identity16_new_neon(int16x8_t *input, int16x8_t *output,
int8_t cos_bit, int bit) {
(void)bit;
(void)cos_bit;
int32x4_t out_low, out_high;
int16x4_t low, high;
int16_t scale = (int16_t)(2 * NewSqrt2);
for (int z = 0; z < 16; ++z) {
out_low = vmull_n_s16(vget_low_s16(input[z]), scale);
out_high = vmull_n_s16(vget_high_s16(input[z]), scale);
low = vqrshrn_n_s32(out_low, (int32_t)NewSqrt2Bits);
high = vqrshrn_n_s32(out_high, (int32_t)NewSqrt2Bits);
output[z] = vcombine_s16(low, high);
}
}
static INLINE void identity32_new_neon(int16x8_t *input, int16x8_t *output,
int8_t cos_bit, int bit) {
(void)bit;
(void)cos_bit;
for (int z = 0; z < 32; ++z) {
output[z] = vmulq_n_s16(input[z], (int16_t)4);
}
}
static INLINE void idct16_low1_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1;
int32x4_t t32[2];
// stage 4
t32[0] = vmull_n_s16(vget_low_s16(in[0]), cospi[32]);
t32[1] = vmull_n_s16(vget_high_s16(in[0]), cospi[32]);
step1 = vcombine_s16(vrshrn_n_s32(t32[0], INV_COS_BIT),
vrshrn_n_s32(t32[1], INV_COS_BIT));
// stage 6
// stage 7
out[0] = step1;
out[1] = step1;
out[2] = step1;
out[3] = step1;
out[4] = step1;
out[5] = step1;
out[6] = step1;
out[7] = step1;
out[8] = step1;
out[9] = step1;
out[10] = step1;
out[11] = step1;
out[12] = step1;
out[13] = step1;
out[14] = step1;
out[15] = step1;
}
static INLINE void idct16_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1[16], step2[16];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 4), (int16_t *)(cospi + 60),
(int16_t *)(cospi + 36), (int16_t *)(cospi + 28));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 20), (int16_t *)(cospi + 44),
(int16_t *)(cospi + 52), (int16_t *)(cospi + 12));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 2
btf_16_lane_0_1_neon(in[1], in[15], c0, &step2[15], &step2[8]);
btf_16_lane_2_3_neon(in[9], in[7], c0, &step2[14], &step2[9]);
btf_16_lane_0_1_neon(in[5], in[11], c1, &step2[13], &step2[10]);
btf_16_lane_2_3_neon(in[13], in[3], c1, &step2[12], &step2[11]);
step2[0] = in[0];
step2[1] = in[8];
step2[2] = in[4];
step2[3] = in[12];
step2[4] = in[2];
step2[5] = in[10];
step2[6] = in[6];
step2[7] = in[14];
// stage 3
btf_16_lane_0_1_neon(step2[4], step2[7], c2, &step1[7], &step1[4]);
btf_16_lane_2_3_neon(step2[5], step2[6], c2, &step1[6], &step1[5]);
step1[0] = step2[0];
step1[1] = step2[1];
step1[2] = step2[2];
step1[3] = step2[3];
step1[8] = vqaddq_s16(step2[8], step2[9]);
step1[9] = vqsubq_s16(step2[8], step2[9]);
step1[10] = vqsubq_s16(step2[11], step2[10]);
step1[11] = vqaddq_s16(step2[11], step2[10]);
step1[12] = vqaddq_s16(step2[12], step2[13]);
step1[13] = vqsubq_s16(step2[12], step2[13]);
step1[14] = vqsubq_s16(step2[15], step2[14]);
step1[15] = vqaddq_s16(step2[15], step2[14]);
// stage 4
btf_16_lane_0_1_neon(step1[0], step1[1], c3, &step2[0], &step2[1]);
btf_16_lane_2_3_neon(step1[2], step1[3], c3, &step2[3], &step2[2]);
btf_16_lane_2_3_neon(step1[14], step1[9], c3, &step2[14], &step2[9]);
btf_16_lane_3_2_neon(vnegq_s16(step1[10]), vnegq_s16(step1[13]), c3,
&step2[10], &step2[13]);
step2[4] = vqaddq_s16(step1[4], step1[5]);
step2[5] = vqsubq_s16(step1[4], step1[5]);
step2[6] = vqsubq_s16(step1[7], step1[6]);
step2[7] = vqaddq_s16(step1[7], step1[6]);
step2[8] = step1[8];
step2[11] = step1[11];
step2[12] = step1[12];
step2[15] = step1[15];
// stage 5
btf_16_lane_0_1_neon(step2[6], step2[5], c3, &step1[6], &step1[5]);
step1[0] = vqaddq_s16(step2[0], step2[3]);
step1[1] = vqaddq_s16(step2[1], step2[2]);
step1[2] = vqsubq_s16(step2[1], step2[2]);
step1[3] = vqsubq_s16(step2[0], step2[3]);
step1[4] = step2[4];
step1[7] = step2[7];
step1[8] = vqaddq_s16(step2[8], step2[11]);
step1[9] = vqaddq_s16(step2[9], step2[10]);
step1[10] = vqsubq_s16(step2[9], step2[10]);
step1[11] = vqsubq_s16(step2[8], step2[11]);
step1[12] = vqsubq_s16(step2[15], step2[12]);
step1[13] = vqsubq_s16(step2[14], step2[13]);
step1[14] = vqaddq_s16(step2[14], step2[13]);
step1[15] = vqaddq_s16(step2[15], step2[12]);
// stage 6
btf_16_lane_0_1_neon(step1[13], step1[10], c3, &step2[13], &step2[10]);
btf_16_lane_0_1_neon(step1[12], step1[11], c3, &step2[12], &step2[11]);
step2[0] = vqaddq_s16(step1[0], step1[7]);
step2[1] = vqaddq_s16(step1[1], step1[6]);
step2[2] = vqaddq_s16(step1[2], step1[5]);
step2[3] = vqaddq_s16(step1[3], step1[4]);
step2[4] = vqsubq_s16(step1[3], step1[4]);
step2[5] = vqsubq_s16(step1[2], step1[5]);
step2[6] = vqsubq_s16(step1[1], step1[6]);
step2[7] = vqsubq_s16(step1[0], step1[7]);
step2[8] = step1[8];
step2[9] = step1[9];
step2[14] = step1[14];
step2[15] = step1[15];
// stage 7
out[0] = vqaddq_s16(step2[0], step2[15]);
out[1] = vqaddq_s16(step2[1], step2[14]);
out[2] = vqaddq_s16(step2[2], step2[13]);
out[3] = vqaddq_s16(step2[3], step2[12]);
out[4] = vqaddq_s16(step2[4], step2[11]);
out[5] = vqaddq_s16(step2[5], step2[10]);
out[6] = vqaddq_s16(step2[6], step2[9]);
out[7] = vqaddq_s16(step2[7], step2[8]);
out[8] = vqsubq_s16(step2[7], step2[8]);
out[9] = vqsubq_s16(step2[6], step2[9]);
out[10] = vqsubq_s16(step2[5], step2[10]);
out[11] = vqsubq_s16(step2[4], step2[11]);
out[12] = vqsubq_s16(step2[3], step2[12]);
out[13] = vqsubq_s16(step2[2], step2[13]);
out[14] = vqsubq_s16(step2[1], step2[14]);
out[15] = vqsubq_s16(step2[0], step2[15]);
}
static INLINE void idct16_low8_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1[16], step2[16];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 1
// stage 2
step2[0] = in[0];
step2[2] = in[4];
step2[4] = in[2];
step2[6] = in[6];
btf_16_neon(in[1], cospi[60], cospi[4], &step2[8], &step2[15]);
btf_16_neon(in[7], -cospi[36], cospi[28], &step2[9], &step2[14]);
btf_16_neon(in[5], cospi[44], cospi[20], &step2[10], &step2[13]);
btf_16_neon(in[3], -cospi[52], cospi[12], &step2[11], &step2[12]);
// stage 3
btf_16_neon(step2[4], cospi[56], cospi[8], &step1[4], &step1[7]);
btf_16_neon(step2[6], -cospi[40], cospi[24], &step1[5], &step1[6]);
step1[0] = step2[0];
step1[2] = step2[2];
step1[8] = vqaddq_s16(step2[8], step2[9]);
step1[9] = vqsubq_s16(step2[8], step2[9]);
step1[10] = vqsubq_s16(step2[11], step2[10]);
step1[11] = vqaddq_s16(step2[11], step2[10]);
step1[12] = vqaddq_s16(step2[12], step2[13]);
step1[13] = vqsubq_s16(step2[12], step2[13]);
step1[14] = vqsubq_s16(step2[15], step2[14]);
step1[15] = vqaddq_s16(step2[15], step2[14]);
// stage 4
btf_16_neon(step1[0], cospi[32], cospi[32], &step2[0], &step2[1]);
btf_16_neon(step1[2], cospi[48], cospi[16], &step2[2], &step2[3]);
btf_16_lane_2_3_neon(step1[14], step1[9], c0, &step2[14], &step2[9]);
btf_16_lane_3_2_neon(vnegq_s16(step1[10]), vnegq_s16(step1[13]), c0,
&step2[10], &step2[13]);
step2[4] = vqaddq_s16(step1[4], step1[5]);
step2[5] = vqsubq_s16(step1[4], step1[5]);
step2[6] = vqsubq_s16(step1[7], step1[6]);
step2[7] = vqaddq_s16(step1[7], step1[6]);
step2[8] = step1[8];
step2[11] = step1[11];
step2[12] = step1[12];
step2[15] = step1[15];
// stage 5
btf_16_lane_0_1_neon(step2[6], step2[5], c0, &step1[6], &step1[5]);
step1[0] = vqaddq_s16(step2[0], step2[3]);
step1[1] = vqaddq_s16(step2[1], step2[2]);
step1[2] = vqsubq_s16(step2[1], step2[2]);
step1[3] = vqsubq_s16(step2[0], step2[3]);
step1[4] = step2[4];
step1[7] = step2[7];
step1[8] = vqaddq_s16(step2[8], step2[11]);
step1[9] = vqaddq_s16(step2[9], step2[10]);
step1[10] = vqsubq_s16(step2[9], step2[10]);
step1[11] = vqsubq_s16(step2[8], step2[11]);
step1[12] = vqsubq_s16(step2[15], step2[12]);
step1[13] = vqsubq_s16(step2[14], step2[13]);
step1[14] = vqaddq_s16(step2[14], step2[13]);
step1[15] = vqaddq_s16(step2[15], step2[12]);
// stage 6
btf_16_lane_0_1_neon(step1[13], step1[10], c0, &step2[13], &step2[10]);
btf_16_lane_0_1_neon(step1[12], step1[11], c0, &step2[12], &step2[11]);
step2[0] = vqaddq_s16(step1[0], step1[7]);
step2[1] = vqaddq_s16(step1[1], step1[6]);
step2[2] = vqaddq_s16(step1[2], step1[5]);
step2[3] = vqaddq_s16(step1[3], step1[4]);
step2[4] = vqsubq_s16(step1[3], step1[4]);
step2[5] = vqsubq_s16(step1[2], step1[5]);
step2[6] = vqsubq_s16(step1[1], step1[6]);
step2[7] = vqsubq_s16(step1[0], step1[7]);
step2[8] = step1[8];
step2[9] = step1[9];
step2[14] = step1[14];
step2[15] = step1[15];
// stage 7
out[0] = vqaddq_s16(step2[0], step2[15]);
out[1] = vqaddq_s16(step2[1], step2[14]);
out[2] = vqaddq_s16(step2[2], step2[13]);
out[3] = vqaddq_s16(step2[3], step2[12]);
out[4] = vqaddq_s16(step2[4], step2[11]);
out[5] = vqaddq_s16(step2[5], step2[10]);
out[6] = vqaddq_s16(step2[6], step2[9]);
out[7] = vqaddq_s16(step2[7], step2[8]);
out[8] = vqsubq_s16(step2[7], step2[8]);
out[9] = vqsubq_s16(step2[6], step2[9]);
out[10] = vqsubq_s16(step2[5], step2[10]);
out[11] = vqsubq_s16(step2[4], step2[11]);
out[12] = vqsubq_s16(step2[3], step2[12]);
out[13] = vqsubq_s16(step2[2], step2[13]);
out[14] = vqsubq_s16(step2[1], step2[14]);
out[15] = vqsubq_s16(step2[0], step2[15]);
}
static INLINE void iadst16_new_neon(int16x8_t *const in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 2), (int16_t *)(cospi + 62),
(int16_t *)(cospi + 10), (int16_t *)(cospi + 54));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 18), (int16_t *)(cospi + 46),
(int16_t *)(cospi + 26), (int16_t *)(cospi + 38));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 34), (int16_t *)(cospi + 30),
(int16_t *)(cospi + 42), (int16_t *)(cospi + 22));
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 50), (int16_t *)(cospi + 14),
(int16_t *)(cospi + 58), (int16_t *)(cospi + 6));
const int16x4_t c4 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
int16x8_t x[16];
int16x8_t t[14];
int16x8_t s0, s1, s2, s3, s4, s5, s6, s7;
int16x8_t s8, s9, s10, s11, s12, s13, s14, s15;
// Stage 1
x[0] = in[15];
x[1] = in[0];
x[2] = in[13];
x[3] = in[2];
x[4] = in[11];
x[5] = in[4];
x[6] = in[9];
x[7] = in[6];
x[8] = in[7];
x[9] = in[8];
x[10] = in[5];
x[11] = in[10];
x[12] = in[3];
x[13] = in[12];
x[14] = in[1];
x[15] = in[14];
// Stage 2
btf_16_lane_0_1_neon(x[0], x[1], c0, &s0, &s1);
btf_16_lane_2_3_neon(x[2], x[3], c0, &s2, &s3);
btf_16_lane_0_1_neon(x[4], x[5], c1, &s4, &s5);
btf_16_lane_2_3_neon(x[6], x[7], c1, &s6, &s7);
btf_16_lane_0_1_neon(x[8], x[9], c2, &s8, &s9);
btf_16_lane_2_3_neon(x[10], x[11], c2, &s10, &s11);
btf_16_lane_0_1_neon(x[12], x[13], c3, &s12, &s13);
btf_16_lane_2_3_neon(x[14], x[15], c3, &s14, &s15);
// Stage 3
x[0] = vqaddq_s16(s0, s8);
x[1] = vqaddq_s16(s1, s9);
x[2] = vqaddq_s16(s2, s10);
x[3] = vqaddq_s16(s3, s11);
x[4] = vqaddq_s16(s4, s12);
x[5] = vqaddq_s16(s5, s13);
x[6] = vqaddq_s16(s6, s14);
x[7] = vqaddq_s16(s7, s15);
x[8] = vqsubq_s16(s0, s8);
x[9] = vqsubq_s16(s1, s9);
x[10] = vqsubq_s16(s2, s10);
x[11] = vqsubq_s16(s3, s11);
x[12] = vqsubq_s16(s4, s12);
x[13] = vqsubq_s16(s5, s13);
x[14] = vqsubq_s16(s6, s14);
x[15] = vqsubq_s16(s7, s15);
// Stage 4
t[0] = x[0];
t[1] = x[1];
t[2] = x[2];
t[3] = x[3];
t[4] = x[4];
t[5] = x[5];
t[6] = x[6];
t[7] = x[7];
btf_16_lane_0_1_neon(x[8], x[9], c4, &s8, &s9);
btf_16_lane_2_3_neon(x[10], x[11], c4, &s10, &s11);
btf_16_lane_1_0_neon(x[13], x[12], c4, &s13, &s12);
btf_16_lane_3_2_neon(x[15], x[14], c4, &s15, &s14);
// Stage 5
x[0] = vqaddq_s16(t[0], t[4]);
x[1] = vqaddq_s16(t[1], t[5]);
x[2] = vqaddq_s16(t[2], t[6]);
x[3] = vqaddq_s16(t[3], t[7]);
x[4] = vqsubq_s16(t[0], t[4]);
x[5] = vqsubq_s16(t[1], t[5]);
x[6] = vqsubq_s16(t[2], t[6]);
x[7] = vqsubq_s16(t[3], t[7]);
x[8] = vqaddq_s16(s8, s12);
x[9] = vqaddq_s16(s9, s13);
x[10] = vqaddq_s16(s10, s14);
x[11] = vqaddq_s16(s11, s15);
x[12] = vqsubq_s16(s8, s12);
x[13] = vqsubq_s16(s9, s13);
x[14] = vqsubq_s16(s10, s14);
x[15] = vqsubq_s16(s11, s15);
// stage 6
t[0] = x[0];
t[1] = x[1];
t[2] = x[2];
t[3] = x[3];
btf_16_lane_2_3_neon(x[4], x[5], c, &s4, &s5);
btf_16_lane_3_2_neon(x[7], x[6], c, &s7, &s6);
t[8] = x[8];
t[9] = x[9];
t[10] = x[10];
t[11] = x[11];
btf_16_lane_2_3_neon(x[12], x[13], c, &s12, &s13);
btf_16_lane_3_2_neon(x[15], x[14], c, &s15, &s14);
// Stage 7
x[0] = vqaddq_s16(t[0], t[2]);
x[1] = vqaddq_s16(t[1], t[3]);
x[2] = vqsubq_s16(t[0], t[2]);
x[3] = vqsubq_s16(t[1], t[3]);
x[4] = vqaddq_s16(s4, s6);
x[5] = vqaddq_s16(s5, s7);
x[6] = vqsubq_s16(s4, s6);
x[7] = vqsubq_s16(s5, s7);
x[8] = vqaddq_s16(t[8], t[10]);
x[9] = vqaddq_s16(t[9], t[11]);
x[10] = vqsubq_s16(t[8], t[10]);
x[11] = vqsubq_s16(t[9], t[11]);
x[12] = vqaddq_s16(s12, s14);
x[13] = vqaddq_s16(s13, s15);
x[14] = vqsubq_s16(s12, s14);
x[15] = vqsubq_s16(s13, s15);
// Stage 8
btf_16_half_neon(x + 2, c);
btf_16_half_neon(x + 6, c);
btf_16_half_neon(x + 10, c);
btf_16_half_neon(x + 14, c);
// Stage 9
out[0] = x[0];
out[1] = vnegq_s16(x[8]);
out[2] = x[12];
out[3] = vnegq_s16(x[4]);
out[4] = x[6];
out[5] = vnegq_s16(x[14]);
out[6] = x[10];
out[7] = vnegq_s16(x[2]);
out[8] = x[3];
out[9] = vnegq_s16(x[11]);
out[10] = x[15];
out[11] = vnegq_s16(x[7]);
out[12] = x[5];
out[13] = vnegq_s16(x[13]);
out[14] = x[9];
out[15] = vnegq_s16(x[1]);
}
static INLINE void iadst16_low1_new_neon(int16x8_t *const in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
const int16x4_t c4 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
int16x8_t x[16];
int16x8_t t[10];
int16x8_t s0, s1, s4, s5;
int16x8_t s8, s9, s12, s13;
// Stage 1
x[1] = in[0];
// Stage 2
btf_16_neon(x[1], cospi[62], -cospi[2], &s0, &s1);
// Stage 3
x[0] = s0;
x[1] = s1;
x[8] = s0;
x[9] = s1;
// Stage 4
t[0] = x[0];
t[1] = x[1];
btf_16_lane_0_1_neon(x[8], x[9], c4, &s8, &s9);
// Stage 5
x[0] = t[0];
x[1] = t[1];
x[4] = t[0];
x[5] = t[1];
x[8] = s8;
x[9] = s9;
x[12] = s8;
x[13] = s9;
// stage 6
t[0] = x[0];
t[1] = x[1];
btf_16_lane_2_3_neon(x[4], x[5], c, &s4, &s5);
t[8] = x[8];
t[9] = x[9];
btf_16_lane_2_3_neon(x[12], x[13], c, &s12, &s13);
// Stage 7
x[0] = t[0];
x[1] = t[1];
x[2] = t[0];
x[3] = t[1];
x[4] = s4;
x[5] = s5;
x[6] = s4;
x[7] = s5;
x[8] = t[8];
x[9] = t[9];
x[10] = t[8];
x[11] = t[9];
x[12] = s12;
x[13] = s13;
x[14] = s12;
x[15] = s13;
// Stage 8
btf_16_half_neon(x + 2, c);
btf_16_half_neon(x + 6, c);
btf_16_half_neon(x + 10, c);
btf_16_half_neon(x + 14, c);
// Stage 9
out[0] = x[0];
out[1] = vnegq_s16(x[8]);
out[2] = x[12];
out[3] = vnegq_s16(x[4]);
out[4] = x[6];
out[5] = vnegq_s16(x[14]);
out[6] = x[10];
out[7] = vnegq_s16(x[2]);
out[8] = x[3];
out[9] = vnegq_s16(x[11]);
out[10] = x[15];
out[11] = vnegq_s16(x[7]);
out[12] = x[5];
out[13] = vnegq_s16(x[13]);
out[14] = x[9];
out[15] = vnegq_s16(x[1]);
}
static INLINE void iadst16_low8_new_neon(int16x8_t *const in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
const int16x4_t c4 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
int16x8_t x[16];
int16x8_t t[14];
int16x8_t s0, s1, s2, s3, s4, s5, s6, s7;
int16x8_t s8, s9, s10, s11, s12, s13, s14, s15;
// Stage 1
x[1] = in[0];
x[3] = in[2];
x[5] = in[4];
x[7] = in[6];
x[8] = in[7];
x[10] = in[5];
x[12] = in[3];
x[14] = in[1];
// Stage 2
btf_16_neon(x[1], cospi[62], -cospi[2], &s0, &s1);
btf_16_neon(x[3], cospi[54], -cospi[10], &s2, &s3);
btf_16_neon(x[5], cospi[46], -cospi[18], &s4, &s5);
btf_16_neon(x[7], cospi[38], -cospi[26], &s6, &s7);
btf_16_neon(x[8], cospi[34], cospi[30], &s8, &s9);
btf_16_neon(x[10], cospi[42], cospi[22], &s10, &s11);
btf_16_neon(x[12], cospi[50], cospi[14], &s12, &s13);
btf_16_neon(x[14], cospi[58], cospi[6], &s14, &s15);
// Stage 3
x[0] = vqaddq_s16(s0, s8);
x[1] = vqaddq_s16(s1, s9);
x[2] = vqaddq_s16(s2, s10);
x[3] = vqaddq_s16(s3, s11);
x[4] = vqaddq_s16(s4, s12);
x[5] = vqaddq_s16(s5, s13);
x[6] = vqaddq_s16(s6, s14);
x[7] = vqaddq_s16(s7, s15);
x[8] = vqsubq_s16(s0, s8);
x[9] = vqsubq_s16(s1, s9);
x[10] = vqsubq_s16(s2, s10);
x[11] = vqsubq_s16(s3, s11);
x[12] = vqsubq_s16(s4, s12);
x[13] = vqsubq_s16(s5, s13);
x[14] = vqsubq_s16(s6, s14);
x[15] = vqsubq_s16(s7, s15);
// Stage 4
t[0] = x[0];
t[1] = x[1];
t[2] = x[2];
t[3] = x[3];
t[4] = x[4];
t[5] = x[5];
t[6] = x[6];
t[7] = x[7];
btf_16_lane_0_1_neon(x[8], x[9], c4, &s8, &s9);
btf_16_lane_2_3_neon(x[10], x[11], c4, &s10, &s11);
btf_16_lane_1_0_neon(x[13], x[12], c4, &s13, &s12);
btf_16_lane_3_2_neon(x[15], x[14], c4, &s15, &s14);
// Stage 5
x[0] = vqaddq_s16(t[0], t[4]);
x[1] = vqaddq_s16(t[1], t[5]);
x[2] = vqaddq_s16(t[2], t[6]);
x[3] = vqaddq_s16(t[3], t[7]);
x[4] = vqsubq_s16(t[0], t[4]);
x[5] = vqsubq_s16(t[1], t[5]);
x[6] = vqsubq_s16(t[2], t[6]);
x[7] = vqsubq_s16(t[3], t[7]);
x[8] = vqaddq_s16(s8, s12);
x[9] = vqaddq_s16(s9, s13);
x[10] = vqaddq_s16(s10, s14);
x[11] = vqaddq_s16(s11, s15);
x[12] = vqsubq_s16(s8, s12);
x[13] = vqsubq_s16(s9, s13);
x[14] = vqsubq_s16(s10, s14);
x[15] = vqsubq_s16(s11, s15);
// stage 6
t[0] = x[0];
t[1] = x[1];
t[2] = x[2];
t[3] = x[3];
btf_16_lane_2_3_neon(x[4], x[5], c, &s4, &s5);
btf_16_lane_3_2_neon(x[7], x[6], c, &s7, &s6);
t[8] = x[8];
t[9] = x[9];
t[10] = x[10];
t[11] = x[11];
btf_16_lane_2_3_neon(x[12], x[13], c, &s12, &s13);
btf_16_lane_3_2_neon(x[15], x[14], c, &s15, &s14);
// Stage 7
x[0] = vqaddq_s16(t[0], t[2]);
x[1] = vqaddq_s16(t[1], t[3]);
x[2] = vqsubq_s16(t[0], t[2]);
x[3] = vqsubq_s16(t[1], t[3]);
x[4] = vqaddq_s16(s4, s6);
x[5] = vqaddq_s16(s5, s7);
x[6] = vqsubq_s16(s4, s6);
x[7] = vqsubq_s16(s5, s7);
x[8] = vqaddq_s16(t[8], t[10]);
x[9] = vqaddq_s16(t[9], t[11]);
x[10] = vqsubq_s16(t[8], t[10]);
x[11] = vqsubq_s16(t[9], t[11]);
x[12] = vqaddq_s16(s12, s14);
x[13] = vqaddq_s16(s13, s15);
x[14] = vqsubq_s16(s12, s14);
x[15] = vqsubq_s16(s13, s15);
// Stage 8
btf_16_half_neon(x + 2, c);
btf_16_half_neon(x + 6, c);
btf_16_half_neon(x + 10, c);
btf_16_half_neon(x + 14, c);
// Stage 9
out[0] = x[0];
out[1] = vnegq_s16(x[8]);
out[2] = x[12];
out[3] = vnegq_s16(x[4]);
out[4] = x[6];
out[5] = vnegq_s16(x[14]);
out[6] = x[10];
out[7] = vnegq_s16(x[2]);
out[8] = x[3];
out[9] = vnegq_s16(x[11]);
out[10] = x[15];
out[11] = vnegq_s16(x[7]);
out[12] = x[5];
out[13] = vnegq_s16(x[13]);
out[14] = x[9];
out[15] = vnegq_s16(x[1]);
}
static INLINE void idct32_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1[32], step2[32];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 2), (int16_t *)(cospi + 62),
(int16_t *)(cospi + 34), (int16_t *)(cospi + 30));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 18), (int16_t *)(cospi + 46),
(int16_t *)(cospi + 50), (int16_t *)(cospi + 14));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 10), (int16_t *)(cospi + 54),
(int16_t *)(cospi + 42), (int16_t *)(cospi + 22));
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 26), (int16_t *)(cospi + 38),
(int16_t *)(cospi + 58), (int16_t *)(cospi + 6));
const int16x4_t c4 =
create_s16x4_neon((int16_t *)(cospi + 4), (int16_t *)(cospi + 60),
(int16_t *)(cospi + 36), (int16_t *)(cospi + 28));
const int16x4_t c5 =
create_s16x4_neon((int16_t *)(cospi + 20), (int16_t *)(cospi + 44),
(int16_t *)(cospi + 52), (int16_t *)(cospi + 12));
const int16x4_t c6 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c7 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 2
btf_16_lane_0_1_neon(in[1], in[31], c0, &step2[31], &step2[16]);
btf_16_lane_2_3_neon(in[17], in[15], c0, &step2[30], &step2[17]);
btf_16_lane_0_1_neon(in[9], in[23], c1, &step2[29], &step2[18]);
btf_16_lane_2_3_neon(in[25], in[7], c1, &step2[28], &step2[19]);
btf_16_lane_0_1_neon(in[5], in[27], c2, &step2[27], &step2[20]);
btf_16_lane_2_3_neon(in[21], in[11], c2, &step2[26], &step2[21]);
btf_16_lane_0_1_neon(in[13], in[19], c3, &step2[25], &step2[22]);
btf_16_lane_2_3_neon(in[29], in[3], c3, &step2[24], &step2[23]);
step2[0] = in[0];
step2[1] = in[16];
step2[2] = in[8];
step2[3] = in[24];
step2[4] = in[4];
step2[5] = in[20];
step2[6] = in[12];
step2[7] = in[28];
step2[8] = in[2];
step2[9] = in[18];
step2[10] = in[10];
step2[11] = in[26];
step2[12] = in[6];
step2[13] = in[22];
step2[14] = in[14];
step2[15] = in[30];
// stage 3
btf_16_lane_0_1_neon(step2[8], step2[15], c4, &step1[15], &step1[8]);
btf_16_lane_2_3_neon(step2[9], step2[14], c4, &step1[14], &step1[9]);
btf_16_lane_0_1_neon(step2[10], step2[13], c5, &step1[13], &step1[10]);
btf_16_lane_2_3_neon(step2[11], step2[12], c5, &step1[12], &step1[11]);
step1[0] = step2[0];
step1[1] = step2[1];
step1[2] = step2[2];
step1[3] = step2[3];
step1[4] = step2[4];
step1[5] = step2[5];
step1[6] = step2[6];
step1[7] = step2[7];
step1[16] = vqaddq_s16(step2[16], step2[17]);
step1[17] = vqsubq_s16(step2[16], step2[17]);
step1[18] = vqsubq_s16(step2[19], step2[18]);
step1[19] = vqaddq_s16(step2[19], step2[18]);
step1[20] = vqaddq_s16(step2[20], step2[21]);
step1[21] = vqsubq_s16(step2[20], step2[21]);
step1[22] = vqsubq_s16(step2[23], step2[22]);
step1[23] = vqaddq_s16(step2[23], step2[22]);
step1[24] = vqaddq_s16(step2[24], step2[25]);
step1[25] = vqsubq_s16(step2[24], step2[25]);
step1[26] = vqsubq_s16(step2[27], step2[26]);
step1[27] = vqaddq_s16(step2[27], step2[26]);
step1[28] = vqaddq_s16(step2[28], step2[29]);
step1[29] = vqsubq_s16(step2[28], step2[29]);
step1[30] = vqsubq_s16(step2[31], step2[30]);
step1[31] = vqaddq_s16(step2[31], step2[30]);
// stage 4
btf_16_lane_0_1_neon(step1[4], step1[7], c6, &step2[7], &step2[4]);
btf_16_lane_2_3_neon(step1[5], step1[6], c6, &step2[6], &step2[5]);
btf_16_lane_0_1_neon(step1[30], step1[17], c6, &step2[30], &step2[17]);
btf_16_lane_1_0_neon(vnegq_s16(step1[18]), vnegq_s16(step1[29]), c6,
&step2[18], &step2[29]);
btf_16_lane_2_3_neon(step1[26], step1[21], c6, &step2[26], &step2[21]);
btf_16_lane_3_2_neon(vnegq_s16(step1[22]), vnegq_s16(step1[25]), c6,
&step2[22], &step2[25]);
step2[0] = step1[0];
step2[1] = step1[1];
step2[2] = step1[2];
step2[3] = step1[3];
step2[8] = vqaddq_s16(step1[8], step1[9]);
step2[9] = vqsubq_s16(step1[8], step1[9]);
step2[10] = vqsubq_s16(step1[11], step1[10]);
step2[11] = vqaddq_s16(step1[11], step1[10]);
step2[12] = vqaddq_s16(step1[12], step1[13]);
step2[13] = vqsubq_s16(step1[12], step1[13]);
step2[14] = vqsubq_s16(step1[15], step1[14]);
step2[15] = vqaddq_s16(step1[15], step1[14]);
step2[16] = step1[16];
step2[19] = step1[19];
step2[20] = step1[20];
step2[23] = step1[23];
step2[24] = step1[24];
step2[27] = step1[27];
step2[28] = step1[28];
step2[31] = step1[31];
// stage 5
btf_16_lane_0_1_neon(step2[0], step2[1], c7, &step1[0], &step1[1]);
btf_16_lane_2_3_neon(step2[2], step2[3], c7, &step1[3], &step1[2]);
btf_16_lane_2_3_neon(step2[14], step2[9], c7, &step1[14], &step1[9]);
btf_16_lane_3_2_neon(vnegq_s16(step2[10]), vnegq_s16(step2[13]), c7,
&step1[10], &step1[13]);
step1[4] = vqaddq_s16(step2[4], step2[5]);
step1[5] = vqsubq_s16(step2[4], step2[5]);
step1[6] = vqsubq_s16(step2[7], step2[6]);
step1[7] = vqaddq_s16(step2[7], step2[6]);
step1[8] = step2[8];
step1[11] = step2[11];
step1[12] = step2[12];
step1[15] = step2[15];
step1[16] = vqaddq_s16(step2[16], step2[19]);
step1[17] = vqaddq_s16(step2[17], step2[18]);
step1[18] = vqsubq_s16(step2[17], step2[18]);
step1[19] = vqsubq_s16(step2[16], step2[19]);
step1[20] = vqsubq_s16(step2[23], step2[20]);
step1[21] = vqsubq_s16(step2[22], step2[21]);
step1[22] = vqaddq_s16(step2[22], step2[21]);
step1[23] = vqaddq_s16(step2[23], step2[20]);
step1[24] = vqaddq_s16(step2[24], step2[27]);
step1[25] = vqaddq_s16(step2[25], step2[26]);
step1[26] = vqsubq_s16(step2[25], step2[26]);
step1[27] = vqsubq_s16(step2[24], step2[27]);
step1[28] = vqsubq_s16(step2[31], step2[28]);
step1[29] = vqsubq_s16(step2[30], step2[29]);
step1[30] = vqaddq_s16(step2[30], step2[29]);
step1[31] = vqaddq_s16(step2[31], step2[28]);
// stage 6
btf_16_lane_0_1_neon(step1[6], step1[5], c7, &step2[6], &step2[5]);
btf_16_lane_2_3_neon(step1[29], step1[18], c7, &step2[29], &step2[18]);
btf_16_lane_2_3_neon(step1[28], step1[19], c7, &step2[28], &step2[19]);
btf_16_lane_3_2_neon(vnegq_s16(step1[20]), vnegq_s16(step1[27]), c7,
&step2[20], &step2[27]);
btf_16_lane_3_2_neon(vnegq_s16(step1[21]), vnegq_s16(step1[26]), c7,
&step2[21], &step2[26]);
step2[0] = vqaddq_s16(step1[0], step1[3]);
step2[1] = vqaddq_s16(step1[1], step1[2]);
step2[2] = vqsubq_s16(step1[1], step1[2]);
step2[3] = vqsubq_s16(step1[0], step1[3]);
step2[4] = step1[4];
step2[7] = step1[7];
step2[8] = vqaddq_s16(step1[8], step1[11]);
step2[9] = vqaddq_s16(step1[9], step1[10]);
step2[10] = vqsubq_s16(step1[9], step1[10]);
step2[11] = vqsubq_s16(step1[8], step1[11]);
step2[12] = vqsubq_s16(step1[15], step1[12]);
step2[13] = vqsubq_s16(step1[14], step1[13]);
step2[14] = vqaddq_s16(step1[14], step1[13]);
step2[15] = vqaddq_s16(step1[15], step1[12]);
step2[16] = step1[16];
step2[17] = step1[17];
step2[22] = step1[22];
step2[23] = step1[23];
step2[24] = step1[24];
step2[25] = step1[25];
step2[30] = step1[30];
step2[31] = step1[31];
// stage 7
btf_16_lane_0_1_neon(step2[13], step2[10], c7, &step1[13], &step1[10]);
btf_16_lane_0_1_neon(step2[12], step2[11], c7, &step1[12], &step1[11]);
step1[0] = vqaddq_s16(step2[0], step2[7]);
step1[1] = vqaddq_s16(step2[1], step2[6]);
step1[2] = vqaddq_s16(step2[2], step2[5]);
step1[3] = vqaddq_s16(step2[3], step2[4]);
step1[4] = vqsubq_s16(step2[3], step2[4]);
step1[5] = vqsubq_s16(step2[2], step2[5]);
step1[6] = vqsubq_s16(step2[1], step2[6]);
step1[7] = vqsubq_s16(step2[0], step2[7]);
step1[8] = step2[8];
step1[9] = step2[9];
step1[14] = step2[14];
step1[15] = step2[15];
step1[16] = vqaddq_s16(step2[16], step2[23]);
step1[17] = vqaddq_s16(step2[17], step2[22]);
step1[18] = vqaddq_s16(step2[18], step2[21]);
step1[19] = vqaddq_s16(step2[19], step2[20]);
step1[20] = vqsubq_s16(step2[19], step2[20]);
step1[21] = vqsubq_s16(step2[18], step2[21]);
step1[22] = vqsubq_s16(step2[17], step2[22]);
step1[23] = vqsubq_s16(step2[16], step2[23]);
step1[24] = vqsubq_s16(step2[31], step2[24]);
step1[25] = vqsubq_s16(step2[30], step2[25]);
step1[26] = vqsubq_s16(step2[29], step2[26]);
step1[27] = vqsubq_s16(step2[28], step2[27]);
step1[28] = vqaddq_s16(step2[27], step2[28]);
step1[29] = vqaddq_s16(step2[26], step2[29]);
step1[30] = vqaddq_s16(step2[25], step2[30]);
step1[31] = vqaddq_s16(step2[24], step2[31]);
// stage 8
btf_16_lane_0_1_neon(step1[27], step1[20], c7, &step2[27], &step2[20]);
btf_16_lane_0_1_neon(step1[26], step1[21], c7, &step2[26], &step2[21]);
btf_16_lane_0_1_neon(step1[25], step1[22], c7, &step2[25], &step2[22]);
btf_16_lane_0_1_neon(step1[24], step1[23], c7, &step2[24], &step2[23]);
step2[0] = vqaddq_s16(step1[0], step1[15]);
step2[1] = vqaddq_s16(step1[1], step1[14]);
step2[2] = vqaddq_s16(step1[2], step1[13]);
step2[3] = vqaddq_s16(step1[3], step1[12]);
step2[4] = vqaddq_s16(step1[4], step1[11]);
step2[5] = vqaddq_s16(step1[5], step1[10]);
step2[6] = vqaddq_s16(step1[6], step1[9]);
step2[7] = vqaddq_s16(step1[7], step1[8]);
step2[8] = vqsubq_s16(step1[7], step1[8]);
step2[9] = vqsubq_s16(step1[6], step1[9]);
step2[10] = vqsubq_s16(step1[5], step1[10]);
step2[11] = vqsubq_s16(step1[4], step1[11]);
step2[12] = vqsubq_s16(step1[3], step1[12]);
step2[13] = vqsubq_s16(step1[2], step1[13]);
step2[14] = vqsubq_s16(step1[1], step1[14]);
step2[15] = vqsubq_s16(step1[0], step1[15]);
step2[16] = step1[16];
step2[17] = step1[17];
step2[18] = step1[18];
step2[19] = step1[19];
step2[28] = step1[28];
step2[29] = step1[29];
step2[30] = step1[30];
step2[31] = step1[31];
// stage 9
out[0] = vqaddq_s16(step2[0], step2[31]);
out[1] = vqaddq_s16(step2[1], step2[30]);
out[2] = vqaddq_s16(step2[2], step2[29]);
out[3] = vqaddq_s16(step2[3], step2[28]);
out[4] = vqaddq_s16(step2[4], step2[27]);
out[5] = vqaddq_s16(step2[5], step2[26]);
out[6] = vqaddq_s16(step2[6], step2[25]);
out[7] = vqaddq_s16(step2[7], step2[24]);
out[8] = vqaddq_s16(step2[8], step2[23]);
out[9] = vqaddq_s16(step2[9], step2[22]);
out[10] = vqaddq_s16(step2[10], step2[21]);
out[11] = vqaddq_s16(step2[11], step2[20]);
out[12] = vqaddq_s16(step2[12], step2[19]);
out[13] = vqaddq_s16(step2[13], step2[18]);
out[14] = vqaddq_s16(step2[14], step2[17]);
out[15] = vqaddq_s16(step2[15], step2[16]);
out[16] = vqsubq_s16(step2[15], step2[16]);
out[17] = vqsubq_s16(step2[14], step2[17]);
out[18] = vqsubq_s16(step2[13], step2[18]);
out[19] = vqsubq_s16(step2[12], step2[19]);
out[20] = vqsubq_s16(step2[11], step2[20]);
out[21] = vqsubq_s16(step2[10], step2[21]);
out[22] = vqsubq_s16(step2[9], step2[22]);
out[23] = vqsubq_s16(step2[8], step2[23]);
out[24] = vqsubq_s16(step2[7], step2[24]);
out[25] = vqsubq_s16(step2[6], step2[25]);
out[26] = vqsubq_s16(step2[5], step2[26]);
out[27] = vqsubq_s16(step2[4], step2[27]);
out[28] = vqsubq_s16(step2[3], step2[28]);
out[29] = vqsubq_s16(step2[2], step2[29]);
out[30] = vqsubq_s16(step2[1], step2[30]);
out[31] = vqsubq_s16(step2[0], step2[31]);
}
static INLINE void idct32_low1_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1;
int32x4_t t32[2];
// stage 1
// stage 2
// stage 3
// stage 4
// stage 5
t32[0] = vmull_n_s16(vget_low_s16(in[0]), cospi[32]);
t32[1] = vmull_n_s16(vget_high_s16(in[0]), cospi[32]);
step1 = vcombine_s16(vrshrn_n_s32(t32[0], INV_COS_BIT),
vrshrn_n_s32(t32[1], INV_COS_BIT));
// stage 6
// stage 7
// stage 8
// stage 9
out[0] = step1;
out[1] = step1;
out[2] = step1;
out[3] = step1;
out[4] = step1;
out[5] = step1;
out[6] = step1;
out[7] = step1;
out[8] = step1;
out[9] = step1;
out[10] = step1;
out[11] = step1;
out[12] = step1;
out[13] = step1;
out[14] = step1;
out[15] = step1;
out[16] = step1;
out[17] = step1;
out[18] = step1;
out[19] = step1;
out[20] = step1;
out[21] = step1;
out[22] = step1;
out[23] = step1;
out[24] = step1;
out[25] = step1;
out[26] = step1;
out[27] = step1;
out[28] = step1;
out[29] = step1;
out[30] = step1;
out[31] = step1;
}
static INLINE void idct32_low8_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1[32], step2[32];
int32x4_t t32[16];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 1
// stage 2
step2[0] = in[0];
step2[4] = in[4];
step2[8] = in[2];
step2[12] = in[6];
btf_16_neon(in[1], cospi[62], cospi[2], &step2[16], &step2[31]);
btf_16_neon(in[7], -cospi[50], cospi[14], &step2[19], &step2[28]);
btf_16_neon(in[5], cospi[54], cospi[10], &step2[20], &step2[27]);
btf_16_neon(in[3], -cospi[58], cospi[6], &step2[23], &step2[24]);
// stage 3
step1[0] = step2[0];
step1[4] = step2[4];
btf_16_neon(step2[8], cospi[60], cospi[4], &step1[8], &step1[15]);
btf_16_neon(step2[12], -cospi[52], cospi[12], &step1[11], &step1[12]);
step1[16] = step2[16];
step1[17] = step2[16];
step1[18] = step2[19];
step1[19] = step2[19];
step1[20] = step2[20];
step1[21] = step2[20];
step1[22] = step2[23];
step1[23] = step2[23];
step1[24] = step2[24];
step1[25] = step2[24];
step1[26] = step2[27];
step1[27] = step2[27];
step1[28] = step2[28];
step1[29] = step2[28];
step1[30] = step2[31];
step1[31] = step2[31];
// stage 4
btf_16_neon(step1[4], cospi[56], cospi[8], &step2[4], &step2[7]);
btf_16_lane_0_1_neon(step1[30], step1[17], c0, &step2[30], &step2[17]);
btf_16_lane_1_0_neon(vnegq_s16(step1[18]), vnegq_s16(step1[29]), c0,
&step2[18], &step2[29]);
btf_16_lane_2_3_neon(step1[26], step1[21], c0, &step2[26], &step2[21]);
btf_16_lane_3_2_neon(vnegq_s16(step1[22]), vnegq_s16(step1[25]), c0,
&step2[22], &step2[25]);
step2[0] = step1[0];
step2[8] = step1[8];
step2[9] = step1[8];
step2[10] = step1[11];
step2[11] = step1[11];
step2[12] = step1[12];
step2[13] = step1[12];
step2[14] = step1[15];
step2[15] = step1[15];
step2[16] = step1[16];
step2[19] = step1[19];
step2[20] = step1[20];
step2[23] = step1[23];
step2[24] = step1[24];
step2[27] = step1[27];
step2[28] = step1[28];
step2[31] = step1[31];
// stage 5
t32[0] = vmull_n_s16(vget_low_s16(step2[0]), cospi[32]);
t32[1] = vmull_n_s16(vget_high_s16(step2[0]), cospi[32]);
step1[0] = vcombine_s16(vrshrn_n_s32(t32[0], INV_COS_BIT),
vrshrn_n_s32(t32[1], INV_COS_BIT));
btf_16_lane_2_3_neon(step2[14], step2[9], c1, &step1[14], &step1[9]);
btf_16_lane_3_2_neon(vnegq_s16(step2[10]), vnegq_s16(step2[13]), c1,
&step1[10], &step1[13]);
step1[4] = step2[4];
step1[5] = step2[4];
step1[6] = step2[7];
step1[7] = step2[7];
step1[8] = step2[8];
step1[11] = step2[11];
step1[12] = step2[12];
step1[15] = step2[15];
step1[16] = vqaddq_s16(step2[16], step2[19]);
step1[17] = vqaddq_s16(step2[17], step2[18]);
step1[18] = vqsubq_s16(step2[17], step2[18]);
step1[19] = vqsubq_s16(step2[16], step2[19]);
step1[20] = vqsubq_s16(step2[23], step2[20]);
step1[21] = vqsubq_s16(step2[22], step2[21]);
step1[22] = vqaddq_s16(step2[22], step2[21]);
step1[23] = vqaddq_s16(step2[23], step2[20]);
step1[24] = vqaddq_s16(step2[24], step2[27]);
step1[25] = vqaddq_s16(step2[25], step2[26]);
step1[26] = vqsubq_s16(step2[25], step2[26]);
step1[27] = vqsubq_s16(step2[24], step2[27]);
step1[28] = vqsubq_s16(step2[31], step2[28]);
step1[29] = vqsubq_s16(step2[30], step2[29]);
step1[30] = vqaddq_s16(step2[30], step2[29]);
step1[31] = vqaddq_s16(step2[31], step2[28]);
// stage 6
btf_16_lane_0_1_neon(step1[6], step1[5], c1, &step2[6], &step2[5]);
btf_16_lane_2_3_neon(step1[29], step1[18], c1, &step2[29], &step2[18]);
btf_16_lane_2_3_neon(step1[28], step1[19], c1, &step2[28], &step2[19]);
btf_16_lane_3_2_neon(vnegq_s16(step1[20]), vnegq_s16(step1[27]), c1,
&step2[20], &step2[27]);
btf_16_lane_3_2_neon(vnegq_s16(step1[21]), vnegq_s16(step1[26]), c1,
&step2[21], &step2[26]);
step2[0] = step1[0];
step2[1] = step1[0];
step2[2] = step1[0];
step2[3] = step1[0];
step2[4] = step1[4];
step2[7] = step1[7];
step2[8] = vqaddq_s16(step1[8], step1[11]);
step2[9] = vqaddq_s16(step1[9], step1[10]);
step2[10] = vqsubq_s16(step1[9], step1[10]);
step2[11] = vqsubq_s16(step1[8], step1[11]);
step2[12] = vqsubq_s16(step1[15], step1[12]);
step2[13] = vqsubq_s16(step1[14], step1[13]);
step2[14] = vqaddq_s16(step1[14], step1[13]);
step2[15] = vqaddq_s16(step1[15], step1[12]);
step2[16] = step1[16];
step2[17] = step1[17];
step2[22] = step1[22];
step2[23] = step1[23];
step2[24] = step1[24];
step2[25] = step1[25];
step2[30] = step1[30];
step2[31] = step1[31];
// stage 7
btf_16_lane_0_1_neon(step2[13], step2[10], c1, &step1[13], &step1[10]);
btf_16_lane_0_1_neon(step2[12], step2[11], c1, &step1[12], &step1[11]);
step1[0] = vqaddq_s16(step2[0], step2[7]);
step1[1] = vqaddq_s16(step2[1], step2[6]);
step1[2] = vqaddq_s16(step2[2], step2[5]);
step1[3] = vqaddq_s16(step2[3], step2[4]);
step1[4] = vqsubq_s16(step2[3], step2[4]);
step1[5] = vqsubq_s16(step2[2], step2[5]);
step1[6] = vqsubq_s16(step2[1], step2[6]);
step1[7] = vqsubq_s16(step2[0], step2[7]);
step1[8] = step2[8];
step1[9] = step2[9];
step1[14] = step2[14];
step1[15] = step2[15];
step1[16] = vqaddq_s16(step2[16], step2[23]);
step1[17] = vqaddq_s16(step2[17], step2[22]);
step1[18] = vqaddq_s16(step2[18], step2[21]);
step1[19] = vqaddq_s16(step2[19], step2[20]);
step1[20] = vqsubq_s16(step2[19], step2[20]);
step1[21] = vqsubq_s16(step2[18], step2[21]);
step1[22] = vqsubq_s16(step2[17], step2[22]);
step1[23] = vqsubq_s16(step2[16], step2[23]);
step1[24] = vqsubq_s16(step2[31], step2[24]);
step1[25] = vqsubq_s16(step2[30], step2[25]);
step1[26] = vqsubq_s16(step2[29], step2[26]);
step1[27] = vqsubq_s16(step2[28], step2[27]);
step1[28] = vqaddq_s16(step2[27], step2[28]);
step1[29] = vqaddq_s16(step2[26], step2[29]);
step1[30] = vqaddq_s16(step2[25], step2[30]);
step1[31] = vqaddq_s16(step2[24], step2[31]);
// stage 8
btf_16_lane_0_1_neon(step1[27], step1[20], c1, &step2[27], &step2[20]);
btf_16_lane_0_1_neon(step1[26], step1[21], c1, &step2[26], &step2[21]);
btf_16_lane_0_1_neon(step1[25], step1[22], c1, &step2[25], &step2[22]);
btf_16_lane_0_1_neon(step1[24], step1[23], c1, &step2[24], &step2[23]);
step2[0] = vqaddq_s16(step1[0], step1[15]);
step2[1] = vqaddq_s16(step1[1], step1[14]);
step2[2] = vqaddq_s16(step1[2], step1[13]);
step2[3] = vqaddq_s16(step1[3], step1[12]);
step2[4] = vqaddq_s16(step1[4], step1[11]);
step2[5] = vqaddq_s16(step1[5], step1[10]);
step2[6] = vqaddq_s16(step1[6], step1[9]);
step2[7] = vqaddq_s16(step1[7], step1[8]);
step2[8] = vqsubq_s16(step1[7], step1[8]);
step2[9] = vqsubq_s16(step1[6], step1[9]);
step2[10] = vqsubq_s16(step1[5], step1[10]);
step2[11] = vqsubq_s16(step1[4], step1[11]);
step2[12] = vqsubq_s16(step1[3], step1[12]);
step2[13] = vqsubq_s16(step1[2], step1[13]);
step2[14] = vqsubq_s16(step1[1], step1[14]);
step2[15] = vqsubq_s16(step1[0], step1[15]);
step2[16] = step1[16];
step2[17] = step1[17];
step2[18] = step1[18];
step2[19] = step1[19];
step2[28] = step1[28];
step2[29] = step1[29];
step2[30] = step1[30];
step2[31] = step1[31];
// stage 9
out[0] = vqaddq_s16(step2[0], step2[31]);
out[1] = vqaddq_s16(step2[1], step2[30]);
out[2] = vqaddq_s16(step2[2], step2[29]);
out[3] = vqaddq_s16(step2[3], step2[28]);
out[4] = vqaddq_s16(step2[4], step2[27]);
out[5] = vqaddq_s16(step2[5], step2[26]);
out[6] = vqaddq_s16(step2[6], step2[25]);
out[7] = vqaddq_s16(step2[7], step2[24]);
out[8] = vqaddq_s16(step2[8], step2[23]);
out[9] = vqaddq_s16(step2[9], step2[22]);
out[10] = vqaddq_s16(step2[10], step2[21]);
out[11] = vqaddq_s16(step2[11], step2[20]);
out[12] = vqaddq_s16(step2[12], step2[19]);
out[13] = vqaddq_s16(step2[13], step2[18]);
out[14] = vqaddq_s16(step2[14], step2[17]);
out[15] = vqaddq_s16(step2[15], step2[16]);
out[16] = vqsubq_s16(step2[15], step2[16]);
out[17] = vqsubq_s16(step2[14], step2[17]);
out[18] = vqsubq_s16(step2[13], step2[18]);
out[19] = vqsubq_s16(step2[12], step2[19]);
out[20] = vqsubq_s16(step2[11], step2[20]);
out[21] = vqsubq_s16(step2[10], step2[21]);
out[22] = vqsubq_s16(step2[9], step2[22]);
out[23] = vqsubq_s16(step2[8], step2[23]);
out[24] = vqsubq_s16(step2[7], step2[24]);
out[25] = vqsubq_s16(step2[6], step2[25]);
out[26] = vqsubq_s16(step2[5], step2[26]);
out[27] = vqsubq_s16(step2[4], step2[27]);
out[28] = vqsubq_s16(step2[3], step2[28]);
out[29] = vqsubq_s16(step2[2], step2[29]);
out[30] = vqsubq_s16(step2[1], step2[30]);
out[31] = vqsubq_s16(step2[0], step2[31]);
}
static INLINE void idct32_low16_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1[32], step2[32];
int32x4_t t32[16];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 1
// stage 2
btf_16_neon(in[1], cospi[62], cospi[2], &step2[16], &step2[31]);
btf_16_neon(in[15], -cospi[34], cospi[30], &step2[17], &step2[30]);
btf_16_neon(in[9], cospi[46], cospi[18], &step2[18], &step2[29]);
btf_16_neon(in[7], -cospi[50], cospi[14], &step2[19], &step2[28]);
btf_16_neon(in[5], cospi[54], cospi[10], &step2[20], &step2[27]);
btf_16_neon(in[11], -cospi[42], cospi[22], &step2[21], &step2[26]);
btf_16_neon(in[13], cospi[38], cospi[26], &step2[22], &step2[25]);
btf_16_neon(in[3], -cospi[58], cospi[6], &step2[23], &step2[24]);
step2[0] = in[0];
step2[2] = in[8];
step2[4] = in[4];
step2[6] = in[12];
step2[8] = in[2];
step2[10] = in[10];
step2[12] = in[6];
step2[14] = in[14];
// stage 3
btf_16_neon(step2[8], cospi[60], cospi[4], &step1[8], &step1[15]);
btf_16_neon(step2[14], -cospi[36], cospi[28], &step1[9], &step1[14]);
btf_16_neon(step2[10], cospi[44], cospi[20], &step1[10], &step1[13]);
btf_16_neon(step2[12], -cospi[52], cospi[12], &step1[11], &step1[12]);
step1[0] = step2[0];
step1[2] = step2[2];
step1[4] = step2[4];
step1[6] = step2[6];
step1[16] = vqaddq_s16(step2[16], step2[17]);
step1[17] = vqsubq_s16(step2[16], step2[17]);
step1[18] = vqsubq_s16(step2[19], step2[18]);
step1[19] = vqaddq_s16(step2[19], step2[18]);
step1[20] = vqaddq_s16(step2[20], step2[21]);
step1[21] = vqsubq_s16(step2[20], step2[21]);
step1[22] = vqsubq_s16(step2[23], step2[22]);
step1[23] = vqaddq_s16(step2[23], step2[22]);
step1[24] = vqaddq_s16(step2[24], step2[25]);
step1[25] = vqsubq_s16(step2[24], step2[25]);
step1[26] = vqsubq_s16(step2[27], step2[26]);
step1[27] = vqaddq_s16(step2[27], step2[26]);
step1[28] = vqaddq_s16(step2[28], step2[29]);
step1[29] = vqsubq_s16(step2[28], step2[29]);
step1[30] = vqsubq_s16(step2[31], step2[30]);
step1[31] = vqaddq_s16(step2[31], step2[30]);
// stage 4
btf_16_neon(step1[4], cospi[56], cospi[8], &step2[4], &step2[7]);
btf_16_neon(step1[6], -cospi[40], cospi[24], &step2[5], &step2[6]);
btf_16_lane_0_1_neon(step1[30], step1[17], c0, &step2[30], &step2[17]);
btf_16_lane_1_0_neon(vnegq_s16(step1[18]), vnegq_s16(step1[29]), c0,
&step2[18], &step2[29]);
btf_16_lane_2_3_neon(step1[26], step1[21], c0, &step2[26], &step2[21]);
btf_16_lane_3_2_neon(vnegq_s16(step1[22]), vnegq_s16(step1[25]), c0,
&step2[22], &step2[25]);
step2[0] = step1[0];
step2[2] = step1[2];
step2[8] = vqaddq_s16(step1[8], step1[9]);
step2[9] = vqsubq_s16(step1[8], step1[9]);
step2[10] = vqsubq_s16(step1[11], step1[10]);
step2[11] = vqaddq_s16(step1[11], step1[10]);
step2[12] = vqaddq_s16(step1[12], step1[13]);
step2[13] = vqsubq_s16(step1[12], step1[13]);
step2[14] = vqsubq_s16(step1[15], step1[14]);
step2[15] = vqaddq_s16(step1[15], step1[14]);
step2[16] = step1[16];
step2[19] = step1[19];
step2[20] = step1[20];
step2[23] = step1[23];
step2[24] = step1[24];
step2[27] = step1[27];
step2[28] = step1[28];
step2[31] = step1[31];
// stage 5
t32[0] = vmull_n_s16(vget_low_s16(step2[0]), cospi[32]);
t32[1] = vmull_n_s16(vget_high_s16(step2[0]), cospi[32]);
step1[0] = vcombine_s16(vrshrn_n_s32(t32[0], INV_COS_BIT),
vrshrn_n_s32(t32[1], INV_COS_BIT));
btf_16_neon(step2[2], cospi[48], cospi[16], &step1[2], &step1[3]);
btf_16_lane_2_3_neon(step2[14], step2[9], c1, &step1[14], &step1[9]);
btf_16_lane_3_2_neon(vnegq_s16(step2[10]), vnegq_s16(step2[13]), c1,
&step1[10], &step1[13]);
step1[4] = vqaddq_s16(step2[4], step2[5]);
step1[5] = vqsubq_s16(step2[4], step2[5]);
step1[6] = vqsubq_s16(step2[7], step2[6]);
step1[7] = vqaddq_s16(step2[7], step2[6]);
step1[8] = step2[8];
step1[11] = step2[11];
step1[12] = step2[12];
step1[15] = step2[15];
step1[16] = vqaddq_s16(step2[16], step2[19]);
step1[17] = vqaddq_s16(step2[17], step2[18]);
step1[18] = vqsubq_s16(step2[17], step2[18]);
step1[19] = vqsubq_s16(step2[16], step2[19]);
step1[20] = vqsubq_s16(step2[23], step2[20]);
step1[21] = vqsubq_s16(step2[22], step2[21]);
step1[22] = vqaddq_s16(step2[22], step2[21]);
step1[23] = vqaddq_s16(step2[23], step2[20]);
step1[24] = vqaddq_s16(step2[24], step2[27]);
step1[25] = vqaddq_s16(step2[25], step2[26]);
step1[26] = vqsubq_s16(step2[25], step2[26]);
step1[27] = vqsubq_s16(step2[24], step2[27]);
step1[28] = vqsubq_s16(step2[31], step2[28]);
step1[29] = vqsubq_s16(step2[30], step2[29]);
step1[30] = vqaddq_s16(step2[30], step2[29]);
step1[31] = vqaddq_s16(step2[31], step2[28]);
// stage 6
btf_16_lane_0_1_neon(step1[6], step1[5], c1, &step2[6], &step2[5]);
btf_16_lane_2_3_neon(step1[29], step1[18], c1, &step2[29], &step2[18]);
btf_16_lane_2_3_neon(step1[28], step1[19], c1, &step2[28], &step2[19]);
btf_16_lane_3_2_neon(vnegq_s16(step1[20]), vnegq_s16(step1[27]), c1,
&step2[20], &step2[27]);
btf_16_lane_3_2_neon(vnegq_s16(step1[21]), vnegq_s16(step1[26]), c1,
&step2[21], &step2[26]);
step2[0] = vqaddq_s16(step1[0], step1[3]);
step2[1] = vqaddq_s16(step1[0], step1[2]);
step2[2] = vqsubq_s16(step1[0], step1[2]);
step2[3] = vqsubq_s16(step1[0], step1[3]);
step2[4] = step1[4];
step2[7] = step1[7];
step2[8] = vqaddq_s16(step1[8], step1[11]);
step2[9] = vqaddq_s16(step1[9], step1[10]);
step2[10] = vqsubq_s16(step1[9], step1[10]);
step2[11] = vqsubq_s16(step1[8], step1[11]);
step2[12] = vqsubq_s16(step1[15], step1[12]);
step2[13] = vqsubq_s16(step1[14], step1[13]);
step2[14] = vqaddq_s16(step1[14], step1[13]);
step2[15] = vqaddq_s16(step1[15], step1[12]);
step2[16] = step1[16];
step2[17] = step1[17];
step2[22] = step1[22];
step2[23] = step1[23];
step2[24] = step1[24];
step2[25] = step1[25];
step2[30] = step1[30];
step2[31] = step1[31];
// stage 7
btf_16_lane_0_1_neon(step2[13], step2[10], c1, &step1[13], &step1[10]);
btf_16_lane_0_1_neon(step2[12], step2[11], c1, &step1[12], &step1[11]);
step1[0] = vqaddq_s16(step2[0], step2[7]);
step1[1] = vqaddq_s16(step2[1], step2[6]);
step1[2] = vqaddq_s16(step2[2], step2[5]);
step1[3] = vqaddq_s16(step2[3], step2[4]);
step1[4] = vqsubq_s16(step2[3], step2[4]);
step1[5] = vqsubq_s16(step2[2], step2[5]);
step1[6] = vqsubq_s16(step2[1], step2[6]);
step1[7] = vqsubq_s16(step2[0], step2[7]);
step1[8] = step2[8];
step1[9] = step2[9];
step1[14] = step2[14];
step1[15] = step2[15];
step1[16] = vqaddq_s16(step2[16], step2[23]);
step1[17] = vqaddq_s16(step2[17], step2[22]);
step1[18] = vqaddq_s16(step2[18], step2[21]);
step1[19] = vqaddq_s16(step2[19], step2[20]);
step1[20] = vqsubq_s16(step2[19], step2[20]);
step1[21] = vqsubq_s16(step2[18], step2[21]);
step1[22] = vqsubq_s16(step2[17], step2[22]);
step1[23] = vqsubq_s16(step2[16], step2[23]);
step1[24] = vqsubq_s16(step2[31], step2[24]);
step1[25] = vqsubq_s16(step2[30], step2[25]);
step1[26] = vqsubq_s16(step2[29], step2[26]);
step1[27] = vqsubq_s16(step2[28], step2[27]);
step1[28] = vqaddq_s16(step2[27], step2[28]);
step1[29] = vqaddq_s16(step2[26], step2[29]);
step1[30] = vqaddq_s16(step2[25], step2[30]);
step1[31] = vqaddq_s16(step2[24], step2[31]);
// stage 8
btf_16_lane_0_1_neon(step1[27], step1[20], c1, &step2[27], &step2[20]);
btf_16_lane_0_1_neon(step1[26], step1[21], c1, &step2[26], &step2[21]);
btf_16_lane_0_1_neon(step1[25], step1[22], c1, &step2[25], &step2[22]);
btf_16_lane_0_1_neon(step1[24], step1[23], c1, &step2[24], &step2[23]);
step2[0] = vqaddq_s16(step1[0], step1[15]);
step2[1] = vqaddq_s16(step1[1], step1[14]);
step2[2] = vqaddq_s16(step1[2], step1[13]);
step2[3] = vqaddq_s16(step1[3], step1[12]);
step2[4] = vqaddq_s16(step1[4], step1[11]);
step2[5] = vqaddq_s16(step1[5], step1[10]);
step2[6] = vqaddq_s16(step1[6], step1[9]);
step2[7] = vqaddq_s16(step1[7], step1[8]);
step2[8] = vqsubq_s16(step1[7], step1[8]);
step2[9] = vqsubq_s16(step1[6], step1[9]);
step2[10] = vqsubq_s16(step1[5], step1[10]);
step2[11] = vqsubq_s16(step1[4], step1[11]);
step2[12] = vqsubq_s16(step1[3], step1[12]);
step2[13] = vqsubq_s16(step1[2], step1[13]);
step2[14] = vqsubq_s16(step1[1], step1[14]);
step2[15] = vqsubq_s16(step1[0], step1[15]);
step2[16] = step1[16];
step2[17] = step1[17];
step2[18] = step1[18];
step2[19] = step1[19];
step2[28] = step1[28];
step2[29] = step1[29];
step2[30] = step1[30];
step2[31] = step1[31];
// stage 9
out[0] = vqaddq_s16(step2[0], step2[31]);
out[1] = vqaddq_s16(step2[1], step2[30]);
out[2] = vqaddq_s16(step2[2], step2[29]);
out[3] = vqaddq_s16(step2[3], step2[28]);
out[4] = vqaddq_s16(step2[4], step2[27]);
out[5] = vqaddq_s16(step2[5], step2[26]);
out[6] = vqaddq_s16(step2[6], step2[25]);
out[7] = vqaddq_s16(step2[7], step2[24]);
out[8] = vqaddq_s16(step2[8], step2[23]);
out[9] = vqaddq_s16(step2[9], step2[22]);
out[10] = vqaddq_s16(step2[10], step2[21]);
out[11] = vqaddq_s16(step2[11], step2[20]);
out[12] = vqaddq_s16(step2[12], step2[19]);
out[13] = vqaddq_s16(step2[13], step2[18]);
out[14] = vqaddq_s16(step2[14], step2[17]);
out[15] = vqaddq_s16(step2[15], step2[16]);
out[16] = vqsubq_s16(step2[15], step2[16]);
out[17] = vqsubq_s16(step2[14], step2[17]);
out[18] = vqsubq_s16(step2[13], step2[18]);
out[19] = vqsubq_s16(step2[12], step2[19]);
out[20] = vqsubq_s16(step2[11], step2[20]);
out[21] = vqsubq_s16(step2[10], step2[21]);
out[22] = vqsubq_s16(step2[9], step2[22]);
out[23] = vqsubq_s16(step2[8], step2[23]);
out[24] = vqsubq_s16(step2[7], step2[24]);
out[25] = vqsubq_s16(step2[6], step2[25]);
out[26] = vqsubq_s16(step2[5], step2[26]);
out[27] = vqsubq_s16(step2[4], step2[27]);
out[28] = vqsubq_s16(step2[3], step2[28]);
out[29] = vqsubq_s16(step2[2], step2[29]);
out[30] = vqsubq_s16(step2[1], step2[30]);
out[31] = vqsubq_s16(step2[0], step2[31]);
}
static INLINE void idct64_stage9_neon(int16x8_t *step2, int16x8_t *step1,
int8_t cos_bit) {
const int32_t *cospi = cospi_arr(cos_bit);
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
btf_16_lane_0_1_neon(step2[27], step2[20], c3, &step1[27], &step1[20]);
btf_16_lane_0_1_neon(step2[26], step2[21], c3, &step1[26], &step1[21]);
btf_16_lane_0_1_neon(step2[25], step2[22], c3, &step1[25], &step1[22]);
btf_16_lane_0_1_neon(step2[24], step2[23], c3, &step1[24], &step1[23]);
step1[0] = vqaddq_s16(step2[0], step2[15]);
step1[1] = vqaddq_s16(step2[1], step2[14]);
step1[2] = vqaddq_s16(step2[2], step2[13]);
step1[3] = vqaddq_s16(step2[3], step2[12]);
step1[4] = vqaddq_s16(step2[4], step2[11]);
step1[5] = vqaddq_s16(step2[5], step2[10]);
step1[6] = vqaddq_s16(step2[6], step2[9]);
step1[7] = vqaddq_s16(step2[7], step2[8]);
step1[8] = vqsubq_s16(step2[7], step2[8]);
step1[9] = vqsubq_s16(step2[6], step2[9]);
step1[10] = vqsubq_s16(step2[5], step2[10]);
step1[11] = vqsubq_s16(step2[4], step2[11]);
step1[12] = vqsubq_s16(step2[3], step2[12]);
step1[13] = vqsubq_s16(step2[2], step2[13]);
step1[14] = vqsubq_s16(step2[1], step2[14]);
step1[15] = vqsubq_s16(step2[0], step2[15]);
step1[16] = step2[16];
step1[17] = step2[17];
step1[18] = step2[18];
step1[19] = step2[19];
step1[28] = step2[28];
step1[29] = step2[29];
step1[30] = step2[30];
step1[31] = step2[31];
step1[32] = vqaddq_s16(step2[32], step2[47]);
step1[33] = vqaddq_s16(step2[33], step2[46]);
step1[34] = vqaddq_s16(step2[34], step2[45]);
step1[35] = vqaddq_s16(step2[35], step2[44]);
step1[36] = vqaddq_s16(step2[36], step2[43]);
step1[37] = vqaddq_s16(step2[37], step2[42]);
step1[38] = vqaddq_s16(step2[38], step2[41]);
step1[39] = vqaddq_s16(step2[39], step2[40]);
step1[40] = vqsubq_s16(step2[39], step2[40]);
step1[41] = vqsubq_s16(step2[38], step2[41]);
step1[42] = vqsubq_s16(step2[37], step2[42]);
step1[43] = vqsubq_s16(step2[36], step2[43]);
step1[44] = vqsubq_s16(step2[35], step2[44]);
step1[45] = vqsubq_s16(step2[34], step2[45]);
step1[46] = vqsubq_s16(step2[33], step2[46]);
step1[47] = vqsubq_s16(step2[32], step2[47]);
step1[48] = vqsubq_s16(step2[63], step2[48]);
step1[49] = vqsubq_s16(step2[62], step2[49]);
step1[50] = vqsubq_s16(step2[61], step2[50]);
step1[51] = vqsubq_s16(step2[60], step2[51]);
step1[52] = vqsubq_s16(step2[59], step2[52]);
step1[53] = vqsubq_s16(step2[58], step2[53]);
step1[54] = vqsubq_s16(step2[57], step2[54]);
step1[55] = vqsubq_s16(step2[56], step2[55]);
step1[56] = vqaddq_s16(step2[56], step2[55]);
step1[57] = vqaddq_s16(step2[57], step2[54]);
step1[58] = vqaddq_s16(step2[58], step2[53]);
step1[59] = vqaddq_s16(step2[59], step2[52]);
step1[60] = vqaddq_s16(step2[60], step2[51]);
step1[61] = vqaddq_s16(step2[61], step2[50]);
step1[62] = vqaddq_s16(step2[62], step2[49]);
step1[63] = vqaddq_s16(step2[63], step2[48]);
}
static INLINE void idct64_stage10_neon(int16x8_t *step1, int16x8_t *step2,
int8_t cos_bit) {
const int32_t *cospi = cospi_arr(cos_bit);
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
btf_16_lane_0_1_neon(step1[55], step1[40], c3, &step2[55], &step2[40]);
btf_16_lane_0_1_neon(step1[54], step1[41], c3, &step2[54], &step2[41]);
btf_16_lane_0_1_neon(step1[53], step1[42], c3, &step2[53], &step2[42]);
btf_16_lane_0_1_neon(step1[52], step1[43], c3, &step2[52], &step2[43]);
btf_16_lane_0_1_neon(step1[51], step1[44], c3, &step2[51], &step2[44]);
btf_16_lane_0_1_neon(step1[50], step1[45], c3, &step2[50], &step2[45]);
btf_16_lane_0_1_neon(step1[49], step1[46], c3, &step2[49], &step2[46]);
btf_16_lane_0_1_neon(step1[48], step1[47], c3, &step2[48], &step2[47]);
step2[0] = vqaddq_s16(step1[0], step1[31]);
step2[1] = vqaddq_s16(step1[1], step1[30]);
step2[2] = vqaddq_s16(step1[2], step1[29]);
step2[3] = vqaddq_s16(step1[3], step1[28]);
step2[4] = vqaddq_s16(step1[4], step1[27]);
step2[5] = vqaddq_s16(step1[5], step1[26]);
step2[6] = vqaddq_s16(step1[6], step1[25]);
step2[7] = vqaddq_s16(step1[7], step1[24]);
step2[8] = vqaddq_s16(step1[8], step1[23]);
step2[9] = vqaddq_s16(step1[9], step1[22]);
step2[10] = vqaddq_s16(step1[10], step1[21]);
step2[11] = vqaddq_s16(step1[11], step1[20]);
step2[12] = vqaddq_s16(step1[12], step1[19]);
step2[13] = vqaddq_s16(step1[13], step1[18]);
step2[14] = vqaddq_s16(step1[14], step1[17]);
step2[15] = vqaddq_s16(step1[15], step1[16]);
step2[16] = vqsubq_s16(step1[15], step1[16]);
step2[17] = vqsubq_s16(step1[14], step1[17]);
step2[18] = vqsubq_s16(step1[13], step1[18]);
step2[19] = vqsubq_s16(step1[12], step1[19]);
step2[20] = vqsubq_s16(step1[11], step1[20]);
step2[21] = vqsubq_s16(step1[10], step1[21]);
step2[22] = vqsubq_s16(step1[9], step1[22]);
step2[23] = vqsubq_s16(step1[8], step1[23]);
step2[24] = vqsubq_s16(step1[7], step1[24]);
step2[25] = vqsubq_s16(step1[6], step1[25]);
step2[26] = vqsubq_s16(step1[5], step1[26]);
step2[27] = vqsubq_s16(step1[4], step1[27]);
step2[28] = vqsubq_s16(step1[3], step1[28]);
step2[29] = vqsubq_s16(step1[2], step1[29]);
step2[30] = vqsubq_s16(step1[1], step1[30]);
step2[31] = vqsubq_s16(step1[0], step1[31]);
step2[32] = step1[32];
step2[33] = step1[33];
step2[34] = step1[34];
step2[35] = step1[35];
step2[36] = step1[36];
step2[37] = step1[37];
step2[38] = step1[38];
step2[39] = step1[39];
step2[56] = step1[56];
step2[57] = step1[57];
step2[58] = step1[58];
step2[59] = step1[59];
step2[60] = step1[60];
step2[61] = step1[61];
step2[62] = step1[62];
step2[63] = step1[63];
}
static INLINE void idct64_low32_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step2[64], step1[64];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 4), (int16_t *)(cospi + 60),
(int16_t *)(cospi + 36), (int16_t *)(cospi + 28));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 20), (int16_t *)(cospi + 44),
(int16_t *)(cospi + 52), (int16_t *)(cospi + 12));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 1
// stage 2
step2[0] = in[0];
step2[2] = in[16];
step2[4] = in[8];
step2[6] = in[24];
step2[8] = in[4];
step2[10] = in[20];
step2[12] = in[12];
step2[14] = in[28];
step2[16] = in[2];
step2[18] = in[18];
step2[20] = in[10];
step2[22] = in[26];
step2[24] = in[6];
step2[26] = in[22];
step2[28] = in[14];
step2[30] = in[30];
btf_16_neon(in[1], cospi[63], cospi[1], &step2[32], &step2[63]);
btf_16_neon(in[31], -cospi[33], cospi[31], &step2[33], &step2[62]);
btf_16_neon(in[17], cospi[47], cospi[17], &step2[34], &step2[61]);
btf_16_neon(in[15], -cospi[49], cospi[15], &step2[35], &step2[60]);
btf_16_neon(in[9], cospi[55], cospi[9], &step2[36], &step2[59]);
btf_16_neon(in[23], -cospi[41], cospi[23], &step2[37], &step2[58]);
btf_16_neon(in[25], cospi[39], cospi[25], &step2[38], &step2[57]);
btf_16_neon(in[7], -cospi[57], cospi[7], &step2[39], &step2[56]);
btf_16_neon(in[5], cospi[59], cospi[5], &step2[40], &step2[55]);
btf_16_neon(in[27], -cospi[37], cospi[27], &step2[41], &step2[54]);
btf_16_neon(in[21], cospi[43], cospi[21], &step2[42], &step2[53]);
btf_16_neon(in[11], -cospi[53], cospi[11], &step2[43], &step2[52]);
btf_16_neon(in[13], cospi[51], cospi[13], &step2[44], &step2[51]);
btf_16_neon(in[19], -cospi[45], cospi[19], &step2[45], &step2[50]);
btf_16_neon(in[29], cospi[35], cospi[29], &step2[46], &step2[49]);
btf_16_neon(in[3], -cospi[61], cospi[3], &step2[47], &step2[48]);
// stage 3
step1[0] = step2[0];
step1[2] = step2[2];
step1[4] = step2[4];
step1[6] = step2[6];
step1[8] = step2[8];
step1[10] = step2[10];
step1[12] = step2[12];
step1[14] = step2[14];
btf_16_neon(step2[16], cospi[62], cospi[2], &step1[16], &step1[31]);
btf_16_neon(step2[30], -cospi[34], cospi[30], &step1[17], &step1[30]);
btf_16_neon(step2[18], cospi[46], cospi[18], &step1[18], &step1[29]);
btf_16_neon(step2[28], -cospi[50], cospi[14], &step1[19], &step1[28]);
btf_16_neon(step2[20], cospi[54], cospi[10], &step1[20], &step1[27]);
btf_16_neon(step2[26], -cospi[42], cospi[22], &step1[21], &step1[26]);
btf_16_neon(step2[22], cospi[38], cospi[26], &step1[22], &step1[25]);
btf_16_neon(step2[24], -cospi[58], cospi[6], &step1[23], &step1[24]);
step1[32] = vqaddq_s16(step2[32], step2[33]);
step1[33] = vqsubq_s16(step2[32], step2[33]);
step1[34] = vqsubq_s16(step2[35], step2[34]);
step1[35] = vqaddq_s16(step2[35], step2[34]);
step1[36] = vqaddq_s16(step2[36], step2[37]);
step1[37] = vqsubq_s16(step2[36], step2[37]);
step1[38] = vqsubq_s16(step2[39], step2[38]);
step1[39] = vqaddq_s16(step2[39], step2[38]);
step1[40] = vqaddq_s16(step2[40], step2[41]);
step1[41] = vqsubq_s16(step2[40], step2[41]);
step1[42] = vqsubq_s16(step2[43], step2[42]);
step1[43] = vqaddq_s16(step2[43], step2[42]);
step1[44] = vqaddq_s16(step2[44], step2[45]);
step1[45] = vqsubq_s16(step2[44], step2[45]);
step1[46] = vqsubq_s16(step2[47], step2[46]);
step1[47] = vqaddq_s16(step2[47], step2[46]);
step1[48] = vqaddq_s16(step2[48], step2[49]);
step1[49] = vqsubq_s16(step2[48], step2[49]);
step1[50] = vqsubq_s16(step2[51], step2[50]);
step1[51] = vqaddq_s16(step2[51], step2[50]);
step1[52] = vqaddq_s16(step2[52], step2[53]);
step1[53] = vqsubq_s16(step2[52], step2[53]);
step1[54] = vqsubq_s16(step2[55], step2[54]);
step1[55] = vqaddq_s16(step2[55], step2[54]);
step1[56] = vqaddq_s16(step2[56], step2[57]);
step1[57] = vqsubq_s16(step2[56], step2[57]);
step1[58] = vqsubq_s16(step2[59], step2[58]);
step1[59] = vqaddq_s16(step2[59], step2[58]);
step1[60] = vqaddq_s16(step2[60], step2[61]);
step1[61] = vqsubq_s16(step2[60], step2[61]);
step1[62] = vqsubq_s16(step2[63], step2[62]);
step1[63] = vqaddq_s16(step2[63], step2[62]);
// stage 4
step2[0] = step1[0];
step2[2] = step1[2];
step2[4] = step1[4];
step2[6] = step1[6];
btf_16_neon(step1[8], cospi[60], cospi[4], &step2[8], &step2[15]);
btf_16_neon(step1[14], -cospi[36], cospi[28], &step2[9], &step2[14]);
btf_16_neon(step1[10], cospi[44], cospi[20], &step2[10], &step2[13]);
btf_16_neon(step1[12], -cospi[52], cospi[12], &step2[11], &step2[12]);
btf_16_lane_0_1_neon(step1[62], step1[33], c0, &step2[62], &step2[33]);
btf_16_lane_1_0_neon(vnegq_s16(step1[34]), vnegq_s16(step1[61]), c0,
&step2[34], &step2[61]);
btf_16_lane_2_3_neon(step1[58], step1[37], c0, &step2[58], &step2[37]);
btf_16_lane_3_2_neon(vnegq_s16(step1[38]), vnegq_s16(step1[57]), c0,
&step2[38], &step2[57]);
btf_16_lane_0_1_neon(step1[54], step1[41], c1, &step2[54], &step2[41]);
btf_16_lane_1_0_neon(vnegq_s16(step1[42]), vnegq_s16(step1[53]), c1,
&step2[42], &step2[53]);
btf_16_lane_2_3_neon(step1[50], step1[45], c1, &step2[50], &step2[45]);
btf_16_lane_3_2_neon(vnegq_s16(step1[46]), vnegq_s16(step1[49]), c1,
&step2[46], &step2[49]);
step2[16] = vqaddq_s16(step1[16], step1[17]);
step2[17] = vqsubq_s16(step1[16], step1[17]);
step2[18] = vqsubq_s16(step1[19], step1[18]);
step2[19] = vqaddq_s16(step1[19], step1[18]);
step2[20] = vqaddq_s16(step1[20], step1[21]);
step2[21] = vqsubq_s16(step1[20], step1[21]);
step2[22] = vqsubq_s16(step1[23], step1[22]);
step2[23] = vqaddq_s16(step1[23], step1[22]);
step2[24] = vqaddq_s16(step1[24], step1[25]);
step2[25] = vqsubq_s16(step1[24], step1[25]);
step2[26] = vqsubq_s16(step1[27], step1[26]);
step2[27] = vqaddq_s16(step1[27], step1[26]);
step2[28] = vqaddq_s16(step1[28], step1[29]);
step2[29] = vqsubq_s16(step1[28], step1[29]);
step2[30] = vqsubq_s16(step1[31], step1[30]);
step2[31] = vqaddq_s16(step1[31], step1[30]);
step2[32] = step1[32];
step2[35] = step1[35];
step2[36] = step1[36];
step2[39] = step1[39];
step2[40] = step1[40];
step2[43] = step1[43];
step2[44] = step1[44];
step2[47] = step1[47];
step2[48] = step1[48];
step2[51] = step1[51];
step2[52] = step1[52];
step2[55] = step1[55];
step2[56] = step1[56];
step2[59] = step1[59];
step2[60] = step1[60];
step2[63] = step1[63];
// stage 5
step1[0] = step2[0];
step1[2] = step2[2];
btf_16_neon(step2[4], cospi[56], cospi[8], &step1[4], &step1[7]);
btf_16_neon(step2[6], -cospi[40], cospi[24], &step1[5], &step1[6]);
btf_16_lane_0_1_neon(step2[30], step2[17], c2, &step1[30], &step1[17]);
btf_16_lane_1_0_neon(vnegq_s16(step2[18]), vnegq_s16(step2[29]), c2,
&step1[18], &step1[29]);
btf_16_lane_2_3_neon(step2[26], step2[21], c2, &step1[26], &step1[21]);
btf_16_lane_3_2_neon(vnegq_s16(step2[22]), vnegq_s16(step2[25]), c2,
&step1[22], &step1[25]);
step1[8] = vqaddq_s16(step2[8], step2[9]);
step1[9] = vqsubq_s16(step2[8], step2[9]);
step1[10] = vqsubq_s16(step2[11], step2[10]);
step1[11] = vqaddq_s16(step2[11], step2[10]);
step1[12] = vqaddq_s16(step2[12], step2[13]);
step1[13] = vqsubq_s16(step2[12], step2[13]);
step1[14] = vqsubq_s16(step2[15], step2[14]);
step1[15] = vqaddq_s16(step2[15], step2[14]);
step1[16] = step2[16];
step1[19] = step2[19];
step1[20] = step2[20];
step1[23] = step2[23];
step1[24] = step2[24];
step1[27] = step2[27];
step1[28] = step2[28];
step1[31] = step2[31];
step1[32] = vqaddq_s16(step2[32], step2[35]);
step1[33] = vqaddq_s16(step2[33], step2[34]);
step1[34] = vqsubq_s16(step2[33], step2[34]);
step1[35] = vqsubq_s16(step2[32], step2[35]);
step1[36] = vqsubq_s16(step2[39], step2[36]);
step1[37] = vqsubq_s16(step2[38], step2[37]);
step1[38] = vqaddq_s16(step2[38], step2[37]);
step1[39] = vqaddq_s16(step2[39], step2[36]);
step1[40] = vqaddq_s16(step2[40], step2[43]);
step1[41] = vqaddq_s16(step2[41], step2[42]);
step1[42] = vqsubq_s16(step2[41], step2[42]);
step1[43] = vqsubq_s16(step2[40], step2[43]);
step1[44] = vqsubq_s16(step2[47], step2[44]);
step1[45] = vqsubq_s16(step2[46], step2[45]);
step1[46] = vqaddq_s16(step2[46], step2[45]);
step1[47] = vqaddq_s16(step2[47], step2[44]);
step1[48] = vqaddq_s16(step2[48], step2[51]);
step1[49] = vqaddq_s16(step2[49], step2[50]);
step1[50] = vqsubq_s16(step2[49], step2[50]);
step1[51] = vqsubq_s16(step2[48], step2[51]);
step1[52] = vqsubq_s16(step2[55], step2[52]);
step1[53] = vqsubq_s16(step2[54], step2[53]);
step1[54] = vqaddq_s16(step2[54], step2[53]);
step1[55] = vqaddq_s16(step2[55], step2[52]);
step1[56] = vqaddq_s16(step2[56], step2[59]);
step1[57] = vqaddq_s16(step2[57], step2[58]);
step1[58] = vqsubq_s16(step2[57], step2[58]);
step1[59] = vqsubq_s16(step2[56], step2[59]);
step1[60] = vqsubq_s16(step2[63], step2[60]);
step1[61] = vqsubq_s16(step2[62], step2[61]);
step1[62] = vqaddq_s16(step2[62], step2[61]);
step1[63] = vqaddq_s16(step2[63], step2[60]);
// stage 6
btf_16_neon(step1[0], cospi[32], cospi[32], &step2[0], &step2[1]);
btf_16_neon(step1[2], cospi[48], cospi[16], &step2[2], &step2[3]);
btf_16_lane_2_3_neon(step1[14], step1[9], c3, &step2[14], &step2[9]);
btf_16_lane_3_2_neon(vnegq_s16(step1[10]), vnegq_s16(step1[13]), c3,
&step2[10], &step2[13]);
btf_16_lane_0_1_neon(step1[61], step1[34], c2, &step2[61], &step2[34]);
btf_16_lane_0_1_neon(step1[60], step1[35], c2, &step2[60], &step2[35]);
btf_16_lane_1_0_neon(vnegq_s16(step1[36]), vnegq_s16(step1[59]), c2,
&step2[36], &step2[59]);
btf_16_lane_1_0_neon(vnegq_s16(step1[37]), vnegq_s16(step1[58]), c2,
&step2[37], &step2[58]);
btf_16_lane_2_3_neon(step1[53], step1[42], c2, &step2[53], &step2[42]);
btf_16_lane_2_3_neon(step1[52], step1[43], c2, &step2[52], &step2[43]);
btf_16_lane_3_2_neon(vnegq_s16(step1[44]), vnegq_s16(step1[51]), c2,
&step2[44], &step2[51]);
btf_16_lane_3_2_neon(vnegq_s16(step1[45]), vnegq_s16(step1[50]), c2,
&step2[45], &step2[50]);
step2[4] = vqaddq_s16(step1[4], step1[5]);
step2[5] = vqsubq_s16(step1[4], step1[5]);
step2[6] = vqsubq_s16(step1[7], step1[6]);
step2[7] = vqaddq_s16(step1[7], step1[6]);
step2[8] = step1[8];
step2[11] = step1[11];
step2[12] = step1[12];
step2[15] = step1[15];
step2[16] = vqaddq_s16(step1[16], step1[19]);
step2[17] = vqaddq_s16(step1[17], step1[18]);
step2[18] = vqsubq_s16(step1[17], step1[18]);
step2[19] = vqsubq_s16(step1[16], step1[19]);
step2[20] = vqsubq_s16(step1[23], step1[20]);
step2[21] = vqsubq_s16(step1[22], step1[21]);
step2[22] = vqaddq_s16(step1[22], step1[21]);
step2[23] = vqaddq_s16(step1[23], step1[20]);
step2[24] = vqaddq_s16(step1[24], step1[27]);
step2[25] = vqaddq_s16(step1[25], step1[26]);
step2[26] = vqsubq_s16(step1[25], step1[26]);
step2[27] = vqsubq_s16(step1[24], step1[27]);
step2[28] = vqsubq_s16(step1[31], step1[28]);
step2[29] = vqsubq_s16(step1[30], step1[29]);
step2[30] = vqaddq_s16(step1[30], step1[29]);
step2[31] = vqaddq_s16(step1[31], step1[28]);
step2[32] = step1[32];
step2[33] = step1[33];
step2[38] = step1[38];
step2[39] = step1[39];
step2[40] = step1[40];
step2[41] = step1[41];
step2[46] = step1[46];
step2[47] = step1[47];
step2[48] = step1[48];
step2[49] = step1[49];
step2[54] = step1[54];
step2[55] = step1[55];
step2[56] = step1[56];
step2[57] = step1[57];
step2[62] = step1[62];
step2[63] = step1[63];
// stage 7
btf_16_lane_0_1_neon(step2[6], step2[5], c3, &step1[6], &step1[5]);
btf_16_lane_2_3_neon(step2[29], step2[18], c3, &step1[29], &step1[18]);
btf_16_lane_2_3_neon(step2[28], step2[19], c3, &step1[28], &step1[19]);
btf_16_lane_3_2_neon(vnegq_s16(step2[20]), vnegq_s16(step2[27]), c3,
&step1[20], &step1[27]);
btf_16_lane_3_2_neon(vnegq_s16(step2[21]), vnegq_s16(step2[26]), c3,
&step1[21], &step1[26]);
step1[0] = vqaddq_s16(step2[0], step2[3]);
step1[1] = vqaddq_s16(step2[1], step2[2]);
step1[2] = vqsubq_s16(step2[1], step2[2]);
step1[3] = vqsubq_s16(step2[0], step2[3]);
step1[4] = step2[4];
step1[7] = step2[7];
step1[8] = vqaddq_s16(step2[8], step2[11]);
step1[9] = vqaddq_s16(step2[9], step2[10]);
step1[10] = vqsubq_s16(step2[9], step2[10]);
step1[11] = vqsubq_s16(step2[8], step2[11]);
step1[12] = vqsubq_s16(step2[15], step2[12]);
step1[13] = vqsubq_s16(step2[14], step2[13]);
step1[14] = vqaddq_s16(step2[14], step2[13]);
step1[15] = vqaddq_s16(step2[15], step2[12]);
step1[16] = step2[16];
step1[17] = step2[17];
step1[22] = step2[22];
step1[23] = step2[23];
step1[24] = step2[24];
step1[25] = step2[25];
step1[30] = step2[30];
step1[31] = step2[31];
step1[32] = vqaddq_s16(step2[32], step2[39]);
step1[33] = vqaddq_s16(step2[33], step2[38]);
step1[34] = vqaddq_s16(step2[34], step2[37]);
step1[35] = vqaddq_s16(step2[35], step2[36]);
step1[36] = vqsubq_s16(step2[35], step2[36]);
step1[37] = vqsubq_s16(step2[34], step2[37]);
step1[38] = vqsubq_s16(step2[33], step2[38]);
step1[39] = vqsubq_s16(step2[32], step2[39]);
step1[40] = vqsubq_s16(step2[47], step2[40]);
step1[41] = vqsubq_s16(step2[46], step2[41]);
step1[42] = vqsubq_s16(step2[45], step2[42]);
step1[43] = vqsubq_s16(step2[44], step2[43]);
step1[44] = vqaddq_s16(step2[43], step2[44]);
step1[45] = vqaddq_s16(step2[42], step2[45]);
step1[46] = vqaddq_s16(step2[41], step2[46]);
step1[47] = vqaddq_s16(step2[40], step2[47]);
step1[48] = vqaddq_s16(step2[48], step2[55]);
step1[49] = vqaddq_s16(step2[49], step2[54]);
step1[50] = vqaddq_s16(step2[50], step2[53]);
step1[51] = vqaddq_s16(step2[51], step2[52]);
step1[52] = vqsubq_s16(step2[51], step2[52]);
step1[53] = vqsubq_s16(step2[50], step2[53]);
step1[54] = vqsubq_s16(step2[49], step2[54]);
step1[55] = vqsubq_s16(step2[48], step2[55]);
step1[56] = vqsubq_s16(step2[63], step2[56]);
step1[57] = vqsubq_s16(step2[62], step2[57]);
step1[58] = vqsubq_s16(step2[61], step2[58]);
step1[59] = vqsubq_s16(step2[60], step2[59]);
step1[60] = vqaddq_s16(step2[59], step2[60]);
step1[61] = vqaddq_s16(step2[58], step2[61]);
step1[62] = vqaddq_s16(step2[57], step2[62]);
step1[63] = vqaddq_s16(step2[56], step2[63]);
// stage 8
btf_16_lane_0_1_neon(step1[13], step1[10], c3, &step2[13], &step2[10]);
btf_16_lane_0_1_neon(step1[12], step1[11], c3, &step2[12], &step2[11]);
btf_16_lane_2_3_neon(step1[59], step1[36], c3, &step2[59], &step2[36]);
btf_16_lane_2_3_neon(step1[58], step1[37], c3, &step2[58], &step2[37]);
btf_16_lane_2_3_neon(step1[57], step1[38], c3, &step2[57], &step2[38]);
btf_16_lane_2_3_neon(step1[56], step1[39], c3, &step2[56], &step2[39]);
btf_16_lane_3_2_neon(vnegq_s16(step1[40]), vnegq_s16(step1[55]), c3,
&step2[40], &step2[55]);
btf_16_lane_3_2_neon(vnegq_s16(step1[41]), vnegq_s16(step1[54]), c3,
&step2[41], &step2[54]);
btf_16_lane_3_2_neon(vnegq_s16(step1[42]), vnegq_s16(step1[53]), c3,
&step2[42], &step2[53]);
btf_16_lane_3_2_neon(vnegq_s16(step1[43]), vnegq_s16(step1[52]), c3,
&step2[43], &step2[52]);
step2[0] = vqaddq_s16(step1[0], step1[7]);
step2[1] = vqaddq_s16(step1[1], step1[6]);
step2[2] = vqaddq_s16(step1[2], step1[5]);
step2[3] = vqaddq_s16(step1[3], step1[4]);
step2[4] = vqsubq_s16(step1[3], step1[4]);
step2[5] = vqsubq_s16(step1[2], step1[5]);
step2[6] = vqsubq_s16(step1[1], step1[6]);
step2[7] = vqsubq_s16(step1[0], step1[7]);
step2[8] = step1[8];
step2[9] = step1[9];
step2[14] = step1[14];
step2[15] = step1[15];
step2[16] = vqaddq_s16(step1[16], step1[23]);
step2[17] = vqaddq_s16(step1[17], step1[22]);
step2[18] = vqaddq_s16(step1[18], step1[21]);
step2[19] = vqaddq_s16(step1[19], step1[20]);
step2[20] = vqsubq_s16(step1[19], step1[20]);
step2[21] = vqsubq_s16(step1[18], step1[21]);
step2[22] = vqsubq_s16(step1[17], step1[22]);
step2[23] = vqsubq_s16(step1[16], step1[23]);
step2[24] = vqsubq_s16(step1[31], step1[24]);
step2[25] = vqsubq_s16(step1[30], step1[25]);
step2[26] = vqsubq_s16(step1[29], step1[26]);
step2[27] = vqsubq_s16(step1[28], step1[27]);
step2[28] = vqaddq_s16(step1[28], step1[27]);
step2[29] = vqaddq_s16(step1[29], step1[26]);
step2[30] = vqaddq_s16(step1[30], step1[25]);
step2[31] = vqaddq_s16(step1[31], step1[24]);
step2[32] = step1[32];
step2[33] = step1[33];
step2[34] = step1[34];
step2[35] = step1[35];
step2[44] = step1[44];
step2[45] = step1[45];
step2[46] = step1[46];
step2[47] = step1[47];
step2[48] = step1[48];
step2[49] = step1[49];
step2[50] = step1[50];
step2[51] = step1[51];
step2[60] = step1[60];
step2[61] = step1[61];
step2[62] = step1[62];
step2[63] = step1[63];
// stage 9
idct64_stage9_neon(step2, step1, cos_bit);
// stage 10
idct64_stage10_neon(step1, step2, cos_bit);
// stage 11
out[0] = vqaddq_s16(step2[0], step2[63]);
out[1] = vqaddq_s16(step2[1], step2[62]);
out[2] = vqaddq_s16(step2[2], step2[61]);
out[3] = vqaddq_s16(step2[3], step2[60]);
out[4] = vqaddq_s16(step2[4], step2[59]);
out[5] = vqaddq_s16(step2[5], step2[58]);
out[6] = vqaddq_s16(step2[6], step2[57]);
out[7] = vqaddq_s16(step2[7], step2[56]);
out[8] = vqaddq_s16(step2[8], step2[55]);
out[9] = vqaddq_s16(step2[9], step2[54]);
out[10] = vqaddq_s16(step2[10], step2[53]);
out[11] = vqaddq_s16(step2[11], step2[52]);
out[12] = vqaddq_s16(step2[12], step2[51]);
out[13] = vqaddq_s16(step2[13], step2[50]);
out[14] = vqaddq_s16(step2[14], step2[49]);
out[15] = vqaddq_s16(step2[15], step2[48]);
out[16] = vqaddq_s16(step2[16], step2[47]);
out[17] = vqaddq_s16(step2[17], step2[46]);
out[18] = vqaddq_s16(step2[18], step2[45]);
out[19] = vqaddq_s16(step2[19], step2[44]);
out[20] = vqaddq_s16(step2[20], step2[43]);
out[21] = vqaddq_s16(step2[21], step2[42]);
out[22] = vqaddq_s16(step2[22], step2[41]);
out[23] = vqaddq_s16(step2[23], step2[40]);
out[24] = vqaddq_s16(step2[24], step2[39]);
out[25] = vqaddq_s16(step2[25], step2[38]);
out[26] = vqaddq_s16(step2[26], step2[37]);
out[27] = vqaddq_s16(step2[27], step2[36]);
out[28] = vqaddq_s16(step2[28], step2[35]);
out[29] = vqaddq_s16(step2[29], step2[34]);
out[30] = vqaddq_s16(step2[30], step2[33]);
out[31] = vqaddq_s16(step2[31], step2[32]);
out[32] = vqsubq_s16(step2[31], step2[32]);
out[33] = vqsubq_s16(step2[30], step2[33]);
out[34] = vqsubq_s16(step2[29], step2[34]);
out[35] = vqsubq_s16(step2[28], step2[35]);
out[36] = vqsubq_s16(step2[27], step2[36]);
out[37] = vqsubq_s16(step2[26], step2[37]);
out[38] = vqsubq_s16(step2[25], step2[38]);
out[39] = vqsubq_s16(step2[24], step2[39]);
out[40] = vqsubq_s16(step2[23], step2[40]);
out[41] = vqsubq_s16(step2[22], step2[41]);
out[42] = vqsubq_s16(step2[21], step2[42]);
out[43] = vqsubq_s16(step2[20], step2[43]);
out[44] = vqsubq_s16(step2[19], step2[44]);
out[45] = vqsubq_s16(step2[18], step2[45]);
out[46] = vqsubq_s16(step2[17], step2[46]);
out[47] = vqsubq_s16(step2[16], step2[47]);
out[48] = vqsubq_s16(step2[15], step2[48]);
out[49] = vqsubq_s16(step2[14], step2[49]);
out[50] = vqsubq_s16(step2[13], step2[50]);
out[51] = vqsubq_s16(step2[12], step2[51]);
out[52] = vqsubq_s16(step2[11], step2[52]);
out[53] = vqsubq_s16(step2[10], step2[53]);
out[54] = vqsubq_s16(step2[9], step2[54]);
out[55] = vqsubq_s16(step2[8], step2[55]);
out[56] = vqsubq_s16(step2[7], step2[56]);
out[57] = vqsubq_s16(step2[6], step2[57]);
out[58] = vqsubq_s16(step2[5], step2[58]);
out[59] = vqsubq_s16(step2[4], step2[59]);
out[60] = vqsubq_s16(step2[3], step2[60]);
out[61] = vqsubq_s16(step2[2], step2[61]);
out[62] = vqsubq_s16(step2[1], step2[62]);
out[63] = vqsubq_s16(step2[0], step2[63]);
}
static INLINE void idct64_low1_new_neon(int16x8_t *input, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step1;
int32x4_t t32[2];
// stage 1
// stage 2
// stage 3
// stage 4
// stage 5
// stage 6
t32[0] = vmull_n_s16(vget_low_s16(input[0]), cospi[32]);
t32[1] = vmull_n_s16(vget_high_s16(input[0]), cospi[32]);
step1 = vcombine_s16(vrshrn_n_s32(t32[0], INV_COS_BIT),
vrshrn_n_s32(t32[1], INV_COS_BIT));
// stage 7
// stage 8
// stage 9
// stage 10
// stage 11
out[0] = step1;
out[1] = step1;
out[2] = step1;
out[3] = step1;
out[4] = step1;
out[5] = step1;
out[6] = step1;
out[7] = step1;
out[8] = step1;
out[9] = step1;
out[10] = step1;
out[11] = step1;
out[12] = step1;
out[13] = step1;
out[14] = step1;
out[15] = step1;
out[16] = step1;
out[17] = step1;
out[18] = step1;
out[19] = step1;
out[20] = step1;
out[21] = step1;
out[22] = step1;
out[23] = step1;
out[24] = step1;
out[25] = step1;
out[26] = step1;
out[27] = step1;
out[28] = step1;
out[29] = step1;
out[30] = step1;
out[31] = step1;
out[32] = step1;
out[33] = step1;
out[34] = step1;
out[35] = step1;
out[36] = step1;
out[37] = step1;
out[38] = step1;
out[39] = step1;
out[40] = step1;
out[41] = step1;
out[42] = step1;
out[43] = step1;
out[44] = step1;
out[45] = step1;
out[46] = step1;
out[47] = step1;
out[48] = step1;
out[49] = step1;
out[50] = step1;
out[51] = step1;
out[52] = step1;
out[53] = step1;
out[54] = step1;
out[55] = step1;
out[56] = step1;
out[57] = step1;
out[58] = step1;
out[59] = step1;
out[60] = step1;
out[61] = step1;
out[62] = step1;
out[63] = step1;
}
static INLINE void idct64_low8_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step2[64], step1[64];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 4), (int16_t *)(cospi + 60),
(int16_t *)(cospi + 36), (int16_t *)(cospi + 28));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 20), (int16_t *)(cospi + 44),
(int16_t *)(cospi + 52), (int16_t *)(cospi + 12));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 1
// stage 2
step2[0] = in[0];
step2[8] = in[4];
step2[16] = in[2];
step2[24] = in[6];
btf_16_neon(in[1], cospi[63], cospi[1], &step2[32], &step2[63]);
btf_16_neon(in[7], -cospi[57], cospi[7], &step2[39], &step2[56]);
btf_16_neon(in[5], cospi[59], cospi[5], &step2[40], &step2[55]);
btf_16_neon(in[3], -cospi[61], cospi[3], &step2[47], &step2[48]);
// stage 3
step1[0] = step2[0];
step1[8] = step2[8];
btf_16_neon(step2[16], cospi[62], cospi[2], &step1[16], &step1[31]);
btf_16_neon(step2[24], -cospi[58], cospi[6], &step1[23], &step1[24]);
step1[32] = step2[32];
step1[33] = step2[32];
step1[38] = step2[39];
step1[39] = step2[39];
step1[40] = step2[40];
step1[41] = step2[40];
step1[46] = step2[47];
step1[47] = step2[47];
step1[48] = step2[48];
step1[49] = step2[48];
step1[54] = step2[55];
step1[55] = step2[55];
step1[56] = step2[56];
step1[57] = step2[56];
step1[62] = step2[63];
step1[63] = step2[63];
// stage 4
step2[0] = step1[0];
btf_16_neon(step1[8], cospi[60], cospi[4], &step2[8], &step2[15]);
btf_16_lane_0_1_neon(step1[62], step1[33], c0, &step2[62], &step2[33]);
btf_16_lane_3_2_neon(vnegq_s16(step1[38]), vnegq_s16(step1[57]), c0,
&step2[38], &step2[57]);
btf_16_lane_0_1_neon(step1[54], step1[41], c1, &step2[54], &step2[41]);
btf_16_lane_3_2_neon(vnegq_s16(step1[46]), vnegq_s16(step1[49]), c1,
&step2[46], &step2[49]);
step2[16] = step1[16];
step2[17] = step1[16];
step2[22] = step1[23];
step2[23] = step1[23];
step2[24] = step1[24];
step2[25] = step1[24];
step2[30] = step1[31];
step2[31] = step1[31];
step2[32] = step1[32];
step2[39] = step1[39];
step2[40] = step1[40];
step2[47] = step1[47];
step2[48] = step1[48];
step2[55] = step1[55];
step2[56] = step1[56];
step2[63] = step1[63];
// stage 5
step1[0] = step2[0];
btf_16_lane_0_1_neon(step2[30], step2[17], c2, &step1[30], &step1[17]);
btf_16_lane_3_2_neon(vnegq_s16(step2[22]), vnegq_s16(step2[25]), c2,
&step1[22], &step1[25]);
step1[8] = step2[8];
step1[9] = step2[8];
step1[14] = step2[15];
step1[15] = step2[15];
step1[16] = step2[16];
step1[23] = step2[23];
step1[24] = step2[24];
step1[31] = step2[31];
step1[32] = step2[32];
step1[33] = step2[33];
step1[34] = step2[33];
step1[35] = step2[32];
step1[36] = step2[39];
step1[37] = step2[38];
step1[38] = step2[38];
step1[39] = step2[39];
step1[40] = step2[40];
step1[41] = step2[41];
step1[42] = step2[41];
step1[43] = step2[40];
step1[44] = step2[47];
step1[45] = step2[46];
step1[46] = step2[46];
step1[47] = step2[47];
step1[48] = step2[48];
step1[49] = step2[49];
step1[50] = step2[49];
step1[51] = step2[48];
step1[52] = step2[55];
step1[53] = step2[54];
step1[54] = step2[54];
step1[55] = step2[55];
step1[56] = step2[56];
step1[57] = step2[57];
step1[58] = step2[57];
step1[59] = step2[56];
step1[60] = step2[63];
step1[61] = step2[62];
step1[62] = step2[62];
step1[63] = step2[63];
// stage 6
btf_16_neon(step1[0], cospi[32], cospi[32], &step2[0], &step2[1]);
btf_16_lane_2_3_neon(step1[14], step1[9], c3, &step2[14], &step2[9]);
btf_16_lane_0_1_neon(step1[61], step1[34], c2, &step2[61], &step2[34]);
btf_16_lane_0_1_neon(step1[60], step1[35], c2, &step2[60], &step2[35]);
btf_16_lane_1_0_neon(vnegq_s16(step1[36]), vnegq_s16(step1[59]), c2,
&step2[36], &step2[59]);
btf_16_lane_1_0_neon(vnegq_s16(step1[37]), vnegq_s16(step1[58]), c2,
&step2[37], &step2[58]);
btf_16_lane_2_3_neon(step1[53], step1[42], c2, &step2[53], &step2[42]);
btf_16_lane_2_3_neon(step1[52], step1[43], c2, &step2[52], &step2[43]);
btf_16_lane_3_2_neon(vnegq_s16(step1[44]), vnegq_s16(step1[51]), c2,
&step2[44], &step2[51]);
btf_16_lane_3_2_neon(vnegq_s16(step1[45]), vnegq_s16(step1[50]), c2,
&step2[45], &step2[50]);
step2[8] = step1[8];
step2[15] = step1[15];
step2[16] = step1[16];
step2[17] = step1[17];
step2[18] = step1[17];
step2[19] = step1[16];
step2[20] = step1[23];
step2[21] = step1[22];
step2[22] = step1[22];
step2[23] = step1[23];
step2[24] = step1[24];
step2[25] = step1[25];
step2[26] = step1[25];
step2[27] = step1[24];
step2[28] = step1[31];
step2[29] = step1[30];
step2[30] = step1[30];
step2[31] = step1[31];
step2[32] = step1[32];
step2[33] = step1[33];
step2[38] = step1[38];
step2[39] = step1[39];
step2[40] = step1[40];
step2[41] = step1[41];
step2[46] = step1[46];
step2[47] = step1[47];
step2[48] = step1[48];
step2[49] = step1[49];
step2[54] = step1[54];
step2[55] = step1[55];
step2[56] = step1[56];
step2[57] = step1[57];
step2[62] = step1[62];
step2[63] = step1[63];
// stage 7
btf_16_lane_2_3_neon(step2[29], step2[18], c3, &step1[29], &step1[18]);
btf_16_lane_2_3_neon(step2[28], step2[19], c3, &step1[28], &step1[19]);
btf_16_lane_3_2_neon(vnegq_s16(step2[20]), vnegq_s16(step2[27]), c3,
&step1[20], &step1[27]);
btf_16_lane_3_2_neon(vnegq_s16(step2[21]), vnegq_s16(step2[26]), c3,
&step1[21], &step1[26]);
step1[0] = step2[0];
step1[1] = step2[1];
step1[2] = step2[1];
step1[3] = step2[0];
step1[8] = step2[8];
step1[9] = step2[9];
step1[10] = step2[9];
step1[11] = step2[8];
step1[12] = step2[15];
step1[13] = step2[14];
step1[14] = step2[14];
step1[15] = step2[15];
step1[16] = step2[16];
step1[17] = step2[17];
step1[22] = step2[22];
step1[23] = step2[23];
step1[24] = step2[24];
step1[25] = step2[25];
step1[30] = step2[30];
step1[31] = step2[31];
step1[32] = vqaddq_s16(step2[32], step2[39]);
step1[33] = vqaddq_s16(step2[33], step2[38]);
step1[34] = vqaddq_s16(step2[34], step2[37]);
step1[35] = vqaddq_s16(step2[35], step2[36]);
step1[36] = vqsubq_s16(step2[35], step2[36]);
step1[37] = vqsubq_s16(step2[34], step2[37]);
step1[38] = vqsubq_s16(step2[33], step2[38]);
step1[39] = vqsubq_s16(step2[32], step2[39]);
step1[40] = vqsubq_s16(step2[47], step2[40]);
step1[41] = vqsubq_s16(step2[46], step2[41]);
step1[42] = vqsubq_s16(step2[45], step2[42]);
step1[43] = vqsubq_s16(step2[44], step2[43]);
step1[44] = vqaddq_s16(step2[43], step2[44]);
step1[45] = vqaddq_s16(step2[42], step2[45]);
step1[46] = vqaddq_s16(step2[41], step2[46]);
step1[47] = vqaddq_s16(step2[40], step2[47]);
step1[48] = vqaddq_s16(step2[48], step2[55]);
step1[49] = vqaddq_s16(step2[49], step2[54]);
step1[50] = vqaddq_s16(step2[50], step2[53]);
step1[51] = vqaddq_s16(step2[51], step2[52]);
step1[52] = vqsubq_s16(step2[51], step2[52]);
step1[53] = vqsubq_s16(step2[50], step2[53]);
step1[54] = vqsubq_s16(step2[49], step2[54]);
step1[55] = vqsubq_s16(step2[48], step2[55]);
step1[56] = vqsubq_s16(step2[63], step2[56]);
step1[57] = vqsubq_s16(step2[62], step2[57]);
step1[58] = vqsubq_s16(step2[61], step2[58]);
step1[59] = vqsubq_s16(step2[60], step2[59]);
step1[60] = vqaddq_s16(step2[59], step2[60]);
step1[61] = vqaddq_s16(step2[58], step2[61]);
step1[62] = vqaddq_s16(step2[57], step2[62]);
step1[63] = vqaddq_s16(step2[56], step2[63]);
// stage 8
btf_16_lane_0_1_neon(step1[13], step1[10], c3, &step2[13], &step2[10]);
btf_16_lane_0_1_neon(step1[12], step1[11], c3, &step2[12], &step2[11]);
btf_16_lane_2_3_neon(step1[59], step1[36], c3, &step2[59], &step2[36]);
btf_16_lane_2_3_neon(step1[58], step1[37], c3, &step2[58], &step2[37]);
btf_16_lane_2_3_neon(step1[57], step1[38], c3, &step2[57], &step2[38]);
btf_16_lane_2_3_neon(step1[56], step1[39], c3, &step2[56], &step2[39]);
btf_16_lane_3_2_neon(vnegq_s16(step1[40]), vnegq_s16(step1[55]), c3,
&step2[40], &step2[55]);
btf_16_lane_3_2_neon(vnegq_s16(step1[41]), vnegq_s16(step1[54]), c3,
&step2[41], &step2[54]);
btf_16_lane_3_2_neon(vnegq_s16(step1[42]), vnegq_s16(step1[53]), c3,
&step2[42], &step2[53]);
btf_16_lane_3_2_neon(vnegq_s16(step1[43]), vnegq_s16(step1[52]), c3,
&step2[43], &step2[52]);
step2[0] = step1[0];
step2[1] = step1[1];
step2[2] = step1[2];
step2[3] = step1[3];
step2[4] = step1[3];
step2[5] = step1[2];
step2[6] = step1[1];
step2[7] = step1[0];
step2[8] = step1[8];
step2[9] = step1[9];
step2[14] = step1[14];
step2[15] = step1[15];
step2[16] = vqaddq_s16(step1[16], step1[23]);
step2[17] = vqaddq_s16(step1[17], step1[22]);
step2[18] = vqaddq_s16(step1[18], step1[21]);
step2[19] = vqaddq_s16(step1[19], step1[20]);
step2[20] = vqsubq_s16(step1[19], step1[20]);
step2[21] = vqsubq_s16(step1[18], step1[21]);
step2[22] = vqsubq_s16(step1[17], step1[22]);
step2[23] = vqsubq_s16(step1[16], step1[23]);
step2[24] = vqsubq_s16(step1[31], step1[24]);
step2[25] = vqsubq_s16(step1[30], step1[25]);
step2[26] = vqsubq_s16(step1[29], step1[26]);
step2[27] = vqsubq_s16(step1[28], step1[27]);
step2[28] = vqaddq_s16(step1[28], step1[27]);
step2[29] = vqaddq_s16(step1[29], step1[26]);
step2[30] = vqaddq_s16(step1[30], step1[25]);
step2[31] = vqaddq_s16(step1[31], step1[24]);
step2[32] = step1[32];
step2[33] = step1[33];
step2[34] = step1[34];
step2[35] = step1[35];
step2[44] = step1[44];
step2[45] = step1[45];
step2[46] = step1[46];
step2[47] = step1[47];
step2[48] = step1[48];
step2[49] = step1[49];
step2[50] = step1[50];
step2[51] = step1[51];
step2[60] = step1[60];
step2[61] = step1[61];
step2[62] = step1[62];
step2[63] = step1[63];
// stage 9
idct64_stage9_neon(step2, step1, cos_bit);
// stage 10
idct64_stage10_neon(step1, step2, cos_bit);
// stage 11
out[0] = vqaddq_s16(step2[0], step2[63]);
out[1] = vqaddq_s16(step2[1], step2[62]);
out[2] = vqaddq_s16(step2[2], step2[61]);
out[3] = vqaddq_s16(step2[3], step2[60]);
out[4] = vqaddq_s16(step2[4], step2[59]);
out[5] = vqaddq_s16(step2[5], step2[58]);
out[6] = vqaddq_s16(step2[6], step2[57]);
out[7] = vqaddq_s16(step2[7], step2[56]);
out[8] = vqaddq_s16(step2[8], step2[55]);
out[9] = vqaddq_s16(step2[9], step2[54]);
out[10] = vqaddq_s16(step2[10], step2[53]);
out[11] = vqaddq_s16(step2[11], step2[52]);
out[12] = vqaddq_s16(step2[12], step2[51]);
out[13] = vqaddq_s16(step2[13], step2[50]);
out[14] = vqaddq_s16(step2[14], step2[49]);
out[15] = vqaddq_s16(step2[15], step2[48]);
out[16] = vqaddq_s16(step2[16], step2[47]);
out[17] = vqaddq_s16(step2[17], step2[46]);
out[18] = vqaddq_s16(step2[18], step2[45]);
out[19] = vqaddq_s16(step2[19], step2[44]);
out[20] = vqaddq_s16(step2[20], step2[43]);
out[21] = vqaddq_s16(step2[21], step2[42]);
out[22] = vqaddq_s16(step2[22], step2[41]);
out[23] = vqaddq_s16(step2[23], step2[40]);
out[24] = vqaddq_s16(step2[24], step2[39]);
out[25] = vqaddq_s16(step2[25], step2[38]);
out[26] = vqaddq_s16(step2[26], step2[37]);
out[27] = vqaddq_s16(step2[27], step2[36]);
out[28] = vqaddq_s16(step2[28], step2[35]);
out[29] = vqaddq_s16(step2[29], step2[34]);
out[30] = vqaddq_s16(step2[30], step2[33]);
out[31] = vqaddq_s16(step2[31], step2[32]);
out[32] = vqsubq_s16(step2[31], step2[32]);
out[33] = vqsubq_s16(step2[30], step2[33]);
out[34] = vqsubq_s16(step2[29], step2[34]);
out[35] = vqsubq_s16(step2[28], step2[35]);
out[36] = vqsubq_s16(step2[27], step2[36]);
out[37] = vqsubq_s16(step2[26], step2[37]);
out[38] = vqsubq_s16(step2[25], step2[38]);
out[39] = vqsubq_s16(step2[24], step2[39]);
out[40] = vqsubq_s16(step2[23], step2[40]);
out[41] = vqsubq_s16(step2[22], step2[41]);
out[42] = vqsubq_s16(step2[21], step2[42]);
out[43] = vqsubq_s16(step2[20], step2[43]);
out[44] = vqsubq_s16(step2[19], step2[44]);
out[45] = vqsubq_s16(step2[18], step2[45]);
out[46] = vqsubq_s16(step2[17], step2[46]);
out[47] = vqsubq_s16(step2[16], step2[47]);
out[48] = vqsubq_s16(step2[15], step2[48]);
out[49] = vqsubq_s16(step2[14], step2[49]);
out[50] = vqsubq_s16(step2[13], step2[50]);
out[51] = vqsubq_s16(step2[12], step2[51]);
out[52] = vqsubq_s16(step2[11], step2[52]);
out[53] = vqsubq_s16(step2[10], step2[53]);
out[54] = vqsubq_s16(step2[9], step2[54]);
out[55] = vqsubq_s16(step2[8], step2[55]);
out[56] = vqsubq_s16(step2[7], step2[56]);
out[57] = vqsubq_s16(step2[6], step2[57]);
out[58] = vqsubq_s16(step2[5], step2[58]);
out[59] = vqsubq_s16(step2[4], step2[59]);
out[60] = vqsubq_s16(step2[3], step2[60]);
out[61] = vqsubq_s16(step2[2], step2[61]);
out[62] = vqsubq_s16(step2[1], step2[62]);
out[63] = vqsubq_s16(step2[0], step2[63]);
}
static INLINE void idct64_low16_new_neon(int16x8_t *in, int16x8_t *out,
int8_t cos_bit, int bit) {
(void)bit;
const int32_t *cospi = cospi_arr(cos_bit);
int16x8_t step2[64], step1[64];
const int16x4_t c0 =
create_s16x4_neon((int16_t *)(cospi + 4), (int16_t *)(cospi + 60),
(int16_t *)(cospi + 36), (int16_t *)(cospi + 28));
const int16x4_t c1 =
create_s16x4_neon((int16_t *)(cospi + 20), (int16_t *)(cospi + 44),
(int16_t *)(cospi + 52), (int16_t *)(cospi + 12));
const int16x4_t c2 =
create_s16x4_neon((int16_t *)(cospi + 8), (int16_t *)(cospi + 56),
(int16_t *)(cospi + 40), (int16_t *)(cospi + 24));
const int16x4_t c3 =
create_s16x4_neon((int16_t *)(cospi + 32), (int16_t *)(cospi + 32),
(int16_t *)(cospi + 16), (int16_t *)(cospi + 48));
// stage 1
// stage 2
step2[0] = in[0];
step2[4] = in[8];
step2[8] = in[4];
step2[12] = in[12];
step2[16] = in[2];
step2[20] = in[10];
step2[24] = in[6];
step2[28] = in[14];
btf_16_neon(in[1], cospi[63], cospi[1], &step2[32], &step2[63]);
btf_16_neon(in[15], -cospi[49], cospi[15], &step2[35], &step2[60]);
btf_16_neon(in[9], cospi[55], cospi[9], &step2[36], &step2[59]);
btf_16_neon(in[7], -cospi[57], cospi[7], &step2[39], &step2[56]);
btf_16_neon(in[5], cospi[59], cospi[5], &step2[40], &step2[55]);
btf_16_neon(in[11], -cospi[53], cospi[11], &step2[43], &step2[52]);
btf_16_neon(in[13], cospi[51], cospi[13], &step2[44], &step2[51]);
btf_16_neon(in[3], -cospi[61], cospi[3], &step2[47], &step2[48]);
// stage 3
step1[0] = step2[0];
step1[4] = step2[4];
step1[8] = step2[8];
step1[12] = step2[12];
btf_16_neon(step2[16], cospi[62], cospi[2], &step1[16], &step1[31]);
btf_16_neon(step2[20], cospi[54], cospi[10], &step1[20], &step1[27]);
btf_16_neon(step2[24], -cospi[58], cospi[6], &step1[23], &step1[24]);
btf_16_neon(step2[28], -cospi[50], cospi[14], &step1[19], &step1[28]);
step1[32] = step2[32];
step1[33] = step2[32];
step1[34] = step2[35];
step1[35] = step2[35];
step1[36] = step2[36];
step1[37] = step2[36];
step1[38] = step2[39];
step1[39] = step2[39];
step1[40] = step2[40];
step1[41] = step2[40];
step1[42] = step2[43];
step1[43] = step2[43];
step1[44] = step2[44];
step1[45] = step2[44];
step1[46] = step2[47];
step1[47] = step2[47];
step1[48] = step2[48];
step1[49] = step2[48];
step1[50] = step2[51];
step1[51] = step2[51];
step1[52] = step2[52];
step1[53] = step2[52];
step1[54] = step2[55];
step1[55] = step2[55];
step1[56] = step2[56];
step1[57] = step2[56];
step1[58] = step2[59];
step1[59] = step2[59];
step1[60] = step2[60];
step1[61] = step2[60];
step1[62] = step2[63];
step1[63] = step2[63];
// stage 4
step2[0] = step1[0];
step2[4] = step1[4];
btf_16_neon(step1[8], cospi[60], cospi[4], &step2[8], &step2[15]);
btf_16_neon(step1[12], -cospi[52], cospi[12], &step2[11], &step2[12]);
btf_16_lane_0_1_neon(step1[62], step1[33], c0, &step2[62], &step2[33]);
btf_16_lane_1_0_neon(vnegq_s16(step1[34]), vnegq_s16(step1[61]), c0,
&step2[34], &step2[61]);
btf_16_lane_2_3_neon(step1[58], step1[37], c0, &step2[58], &step2[37]);
btf_16_lane_3_2_neon(vnegq_s16(step1[38]), vnegq_s16(step1[57]), c0,
&step2[38], &step2[57]);
btf_16_lane_0_1_neon(step1[54], step1[41], c1, &step2[54], &step2[41]);
btf_16_lane_1_0_neon(vnegq_s16(step1[42]), vnegq_s16(step1[53]), c1,
&step2[42], &step2[53]);
btf_16_lane_2_3_neon(step1[50], step1[45], c1, &step2[50], &step2[45]);
btf_16_lane_3_2_neon(vnegq_s16(step1[46]), vnegq_s16(step1[49]), c1,
&step2[46], &step2[49]);
step2[16] = step1[16];
step2[17] = step1[16];
step2[18] = step1[19];
step2[19] = step1[19];
step2[20] = step1[20];
step2[21] = step1[20];
step2[22] = step1[23];
step2[23] = step1[23];
step2[24] = step1[24];
step2[25] = step1[24];
step2[26] = step1[27];
step2[27] = step1[27];
step2[28] = step1[28];
step2[29] = step1[28];
step2[30] = step1[31];
step2[31] = step1[31];
step2[32] = step1[32];
step2[35] = step1[35];
step2[36] = step1[36];
step2[39] = step1[39];
step2[40] = step1[40];
step2[43] = step1[43];
step2[44] = step1[44];
step2[47] = step1[47];
step2[48] = step1[48];
step2[51] = step1[51];
step2[52] = step1[52];
step2[55] = step1[55];
step2[56] = step1[56];
step2[59] = step1[59];
step2[60] = step1[60];
step2[63] = step1[63];
// stage 5
step1[0] = step2[0];
btf_16_neon(step2[4], cospi[56], cospi[8], &step1[4], &step1[7]);
btf_16_lane_0_1_neon(step2[30], step2[17], c2, &step1[30], &step1[17]);
btf_16_lane_1_0_neon(vnegq_s16(step2[18]), vnegq_s16(step2[29]), c2,
&step1[18], &step1[29]);
btf_16_lane_2_3_neon(step2[26], step2[21], c2, &step1[26], &step1[21]);
btf_16_lane_3_2_neon(vnegq_s16(step2[22]), vnegq_s16(step2[25]), c2,
&step1[22], &step1[25]);
step1[8] = step2[8];
step1[9] = step2[8];
step1[10] = step2[11];
step1[11] = step2[11];
step1[12] = step2[12];
step1[13] = step2[12];
step1[14] = step2[15];
step1[15] = step2[15];
step1[16] = step2[16];
step1[19] = step2[19];
step1[20] = step2[20];
step1[23] = step2[23];
step1[24] = step2[24];
step1[27] = step2[27];
step1[28] = step2[28];
step1[31] = step2[31];
step1[32] = vqaddq_s16(step2[32], step2[35]);
step1[33] = vqaddq_s16(step2[33], step2[34]);
step1[34] = vqsubq_s16(step2[33], step2[34]);
step1[35] = vqsubq_s16(step2[32], step2[35]);
step1[36] = vqsubq_s16(step2[39], step2[36]);
step1[37] = vqsubq_s16(step2[38], step2[37]);
step1[38] = vqaddq_s16(step2[38], step2[37]);
step1[39] = vqaddq_s16(step2[39], step2[36]);
step1[40] = vqaddq_s16(step2[40], step2[43]);
step1[41] = vqaddq_s16(step2[41], step2[42]);
step1[42] = vqsubq_s16(step2[41], step2[42]);
step1[43] = vqsubq_s16(step2[40], step2[43]);
step1[44] = vqsubq_s16(step2[47], step2[44]);
step1[45] = vqsubq_s16(step2[46], step2[45]);
step1[46] = vqaddq_s16(step2[46], step2[45]);
step1[47] = vqaddq_s16(step2[47], step2[44]);
step1[48] = vqaddq_s16(step2[48], step2[51]);
step1[49] = vqaddq_s16(step2[49], step2[50]);
step1[50] = vqsubq_s16(step2[49], step2[50]);
step1[51] = vqsubq_s16(step2[48], step2[51]);
step1[52] = vqsubq_s16(step2[55], step2[52]);
step1[53] = vqsubq_s16(step2[54], step2[53]);
step1[54] = vqaddq_s16(step2[54], step2[53]);
step1[55] = vqaddq_s16(step2[55], step2[52]);
step1[56] = vqaddq_s16(step2[56], step2[59]);
step1[57] = vqaddq_s16(step2[57], step2[58]);
step1[58] = vqsubq_s16(step2[57], step2[58]);
step1[59] = vqsubq_s16(step2[56], step2[59]);
step1[60] = vqsubq_s16(step2[63], step2[60]);
step1[61] = vqsubq_s16(step2[62], step2[61]);
step1[62] = vqaddq_s16(step2[62], step2[61]);
step1[63] = vqaddq_s16(step2[63], step2[60]);
// stage 6
btf_16_neon(step1[0], cospi[32], cospi[32], &step2[0], &step2[1]);
btf_16_lane_2_3_neon(step1[14], step1[9], c3, &step2[14], &step2[9]);
btf_16_lane_3_2_neon(vnegq_s16(step1[10]), vnegq_s16(step1[13]), c3,
&step2[10], &step2[13]);
btf_16_lane_0_1_neon(step1[61], step1[34], c2, &step2[61], &step2[34]);
btf_16_lane_0_1_neon(step1[60], step1[35], c2, &step2[60], &step2[35]);
btf_16_lane_1_0_neon(vnegq_s16(step1[36]), vnegq_s16(step1[59]), c2,
&step2[36], &step2[59]);
btf_16_lane_1_0_neon(vnegq_s16(step1[37]), vnegq_s16(step1[58]), c2,
&step2[37], &step2[58]);
btf_16_lane_2_3_neon(step1[53], step1[42], c2, &step2[53], &step2[42]);
btf_16_lane_2_3_neon(step1[52], step1[43], c2, &step2[52], &step2[43]);
btf_16_lane_3_2_neon(vnegq_s16(step1[44]), vnegq_s16(step1[51]), c2,
&step2[44], &step2[51]);
btf_16_lane_3_2_neon(vnegq_s16(step1[45]), vnegq_s16(step1[50]), c2,
&step2[45], &step2[50]);
step2[4] = step1[4];
step2[5] = step1[4];
step2[6] = step1[7];
step2[7] = step1[7];
step2[8] = step1[8];
step2[11] = step1[11];
step2[12] = step1[12];
step2[15] = step1[15];
step2[16] = vqaddq_s16(step1[16], step1[19]);
step2[17] = vqaddq_s16(step1[17], step1[18]);
step2[18] = vqsubq_s16(step1[17], step1[18]);
step2[19] = vqsubq_s16(step1[16], step1[19]);
step2[20] = vqsubq_s16(step1[23], step1[20]);
step2[21] = vqsubq_s16(step1[22], step1[21]);
step2[22] = vqaddq_s16(step1[22], step1[21]);
step2[23] = vqaddq_s16(step1[23], step1[20]);
step2[24] = vqaddq_s16(step1[24], step1[27]);
step2[25] = vqaddq_s16(step1[25], step1[26]);
step2[26] = vqsubq_s16(step1[25], step1[26]);
step2[27] = vqsubq_s16(step1[24], step1[27]);
step2[28] = vqsubq_s16(step1[31], step1[28]);
step2[29] = vqsubq_s16(step1[30], step1[29]);
step2[30] = vqaddq_s16(step1[30], step1[29]);
step2[31] = vqaddq_s16(step1[31], step1[28]);
step2[32] = step1[32];
step2[33] = step1[33];
step2[38] = step1[38];
step2[39] = step1[39];
step2[40] = step1[40];
step2[41] = step1[41];
step2[46] = step1[46];
step2[47] = step1[47];
step2[48] = step1[48];
step2[49] = step1[49];
step2[54] = step1[54];
step2[55] = step1[55];
step2[56] = step1[56];
step2[57] = step1[57];
step2[62] = step1[62];
step2[63] = step1[63];
// stage 7
btf_16_lane_0_1_neon(step2[6], step2[5], c3, &step1[6], &step1[5]);
btf_16_lane_2_3_neon(step2[29], step2[18], c3, &step1[29], &step1[18]);
btf_16_lane_2_3_neon(step2[28], step2[19], c3, &step1[28], &step1[19]);
btf_16_lane_3_2_neon(vnegq_s16(step2[20]), vnegq_s16(step2[27]), c3,
&step1[20], &step1[27]);
btf_16_lane_3_2_neon(vnegq_s16(step2[21]), vnegq_s16(step2[26]), c3,
&step1[21], &step1[26]);
step1[0] = step2[0];
step1[1] = step2[1];
step1[2] = step2[1];
step1[3] = step2[0];
step1[4] = step2[4];
step1[7] = step2[7];
step1[8] = vqaddq_s16(step2[8], step2[11]);
step1[9] = vqaddq_s16(step2[9], step2[10]);
step1[10] = vqsubq_s16(step2[9], step2[10]);
step1[11] = vqsubq_s16(step2[8], step2[11]);
step1[12] = vqsubq_s16(step2[15], step2[12]);
step1[13] = vqsubq_s16(step2[14], step2[13]);
step1[14] = vqaddq_s16(step2[14], step2[13]);
step1[15] = vqaddq_s16(step2[15], step2[12]);
step1[16] = step2[16];
step1[17] = step2[17];
step1[22] = step2[22];
step1[23] = step2[23];
step1[24] = step2[24];
step1[25] = step2[25];
step1[30] = step2[30];
step1[31] = step2[31];
step1[32] = vqaddq_s16(step2[32], step2[39]);
step1[33] = vqaddq_s16(step2[33], step2[38]);
step1[34] = vqaddq_s16(step2[34], step2[37]);
step1[35] = vqaddq_s16(step2[35], step2[36]);
step1[36] = vqsubq_s16(step2[35], step2[36]);
step1[37] = vqsubq_s16(step2[34], step2[37]);
step1[38] = vqsubq_s16(step2[33], step2[38]);
step1[39] = vqsubq_s16(step2[32], step2[39]);
step1[40] = vqsubq_s16(step2[47], step2[40]);
step1[41] = vqsubq_s16(step2[46], step2[41]);
step1[42] = vqsubq_s16(step2[45], step2[42]);
step1[43] = vqsubq_s16(step2[44], step2[43]);
step1[44] = vqaddq_s16(step2[43], step2[44]);
step1[45] = vqaddq_s16(step2[42], step2[45]);
step1[46] = vqaddq_s16(step2[41], step2[46]);
step1[47] = vqaddq_s16(step2[40], step2[47]);
step1[48] = vqaddq_s16(step2[48], step2[55]);
step1[49] = vqaddq_s16(step2[49], step2[54]);
step1[50] = vqaddq_s16(step2[50], step2[53]);
step1[51] = vqaddq_s16(step2[51], step2[52]);
step1[52] = vqsubq_s16(step2[51], step2[52]);
step1[53] = vqsubq_s16(step2[50], step2[53]);
step1[54] = vqsubq_s16(step2[49], step2[54]);
step1[55] = vqsubq_s16(step2[48], step2[55]);
step1[56] = vqsubq_s16(step2[63], step2[56]);
step1[57] = vqsubq_s16(step2[62], step2[57]);
step1[58] = vqsubq_s16(step2[61], step2[58]);
step1[59] = vqsubq_s16(step2[60], step2[59]);
step1[60] = vqaddq_s16(step2[59], step2[60]);
step1[61] = vqaddq_s16(step2[58], step2[61]);
step1[62] = vqaddq_s16(step2[57], step2[62]);
step1[63] = vqaddq_s16(step2[56], step2[63]);
// stage 8
btf_16_lane_0_1_neon(step1[13], step1[10], c3, &step2[13], &step2[10]);
btf_16_lane_0_1_neon(step1[12], step1[11], c3, &step2[12], &step2[11]);
btf_16_lane_2_3_neon(step1[59], step1[36], c3, &step2[59], &step2[36]);
btf_16_lane_2_3_neon(step1[58], step1[37], c3, &step2[58], &step2[37]);
btf_16_lane_2_3_neon(step1[57], step1[38], c3, &step2[57], &step2[38]);
btf_16_lane_2_3_neon(step1[56], step1[39], c3, &step2[56], &step2[39]);
btf_16_lane_3_2_neon(vnegq_s16(step1[40]), vnegq_s16(step1[55]), c3,
&step2[40], &step2[55]);
btf_16_lane_3_2_neon(vnegq_s16(step1[41]), vnegq_s16(step1[54]), c3,
&step2[41], &step2[54]);
btf_16_lane_3_2_neon(vnegq_s16(step1[42]), vnegq_s16(step1[53]), c3,
&step2[42], &step2[53]);
btf_16_lane_3_2_neon(vnegq_s16(step1[43]), vnegq_s16(step1[52]), c3,
&step2[43], &step2[52]);
step2[0] = vqaddq_s16(step1[0], step1[7]);
step2[1] = vqaddq_s16(step1[1], step1[6]);
step2[2] = vqaddq_s16(step1[2], step1[5]);
step2[3] = vqaddq_s16(step1[3], step1[4]);
step2[4] = vqsubq_s16(step1[3], step1[4]);
step2[5] = vqsubq_s16(step1[2], step1[5]);
step2[6] = vqsubq_s16(step1[1], step1[6]);
step2[7] = vqsubq_s16(step1[0], step1[7]);
step2[8] = step1[8];
step2[9] = step1[9];
step2[14] = step1[14];
step2[15] = step1[15];
step2[16] = vqaddq_s16(step1[16], step1[23]);
step2[17] = vqaddq_s16(step1[17], step1[22]);
step2[18] = vqaddq_s16(step1[18], step1[21]);
step2[19] = vqaddq_s16(step1[19], step1[20]);
step2[20] = vqsubq_s16(step1[19], step1[20]);
step2[21] = vqsubq_s16(step1[18], step1[21]);
step2[22] = vqsubq_s16(step1[17], step1[22]);
step2[23] = vqsubq_s16(step1[16], step1[23]);
step2[24] = vqsubq_s16(step1[31], step1[24]);
step2[25] = vqsubq_s16(step1[30], step1[25]);
step2[26] = vqsubq_s16(step1[29], step1[26]);
step2[27] = vqsubq_s16(step1[28], step1[27]);
step2[28] = vqaddq_s16(step1[28], step1[27]);
step2[29] = vqaddq_s16(step1[29], step1[26]);
step2[30] = vqaddq_s16(step1[30], step1[25]);
step2[31] = vqaddq_s16(step1[31], step1[24]);
step2[32] = step1[32];
step2[33] = step1[33];
step2[34] = step1[34];
step2[35] = step1[35];
step2[44] = step1[44];
step2[45] = step1[45];
step2[46] = step1[46];
step2[47] = step1[47];
step2[48] = step1[48];
step2[49] = step1[49];
step2[50] = step1[50];
step2[51] = step1[51];
step2[60] = step1[60];
step2[61] = step1[61];
step2[62] = step1[62];
step2[63] = step1[63];
// stage 9
idct64_stage9_neon(step2, step1, cos_bit);
// stage 10
idct64_stage10_neon(step1, step2, cos_bit);
// stage 11
out[0] = vqaddq_s16(step2[0], step2[63]);
out[1] = vqaddq_s16(step2[1], step2[62]);
out[2] = vqaddq_s16(step2[2], step2[61]);
out[3] = vqaddq_s16(step2[3], step2[60]);
out[4] = vqaddq_s16(step2[4], step2[59]);
out[5] = vqaddq_s16(step2[5], step2[58]);
out[6] = vqaddq_s16(step2[6], step2[57]);
out[7] = vqaddq_s16(step2[7], step2[56]);
out[8] = vqaddq_s16(step2[8], step2[55]);
out[9] = vqaddq_s16(step2[9], step2[54]);
out[10] = vqaddq_s16(step2[10], step2[53]);
out[11] = vqaddq_s16(step2[11], step2[52]);
out[12] = vqaddq_s16(step2[12], step2[51]);
out[13] = vqaddq_s16(step2[13], step2[50]);
out[14] = vqaddq_s16(step2[14], step2[49]);
out[15] = vqaddq_s16(step2[15], step2[48]);
out[16] = vqaddq_s16(step2[16], step2[47]);
out[17] = vqaddq_s16(step2[17], step2[46]);
out[18] = vqaddq_s16(step2[18], step2[45]);
out[19] = vqaddq_s16(step2[19], step2[44]);
out[20] = vqaddq_s16(step2[20], step2[43]);
out[21] = vqaddq_s16(step2[21], step2[42]);
out[22] = vqaddq_s16(step2[22], step2[41]);
out[23] = vqaddq_s16(step2[23], step2[40]);
out[24] = vqaddq_s16(step2[24], step2[39]);
out[25] = vqaddq_s16(step2[25], step2[38]);
out[26] = vqaddq_s16(step2[26], step2[37]);
out[27] = vqaddq_s16(step2[27], step2[36]);
out[28] = vqaddq_s16(step2[28], step2[35]);
out[29] = vqaddq_s16(step2[29], step2[34]);
out[30] = vqaddq_s16(step2[30], step2[33]);
out[31] = vqaddq_s16(step2[31], step2[32]);
out[32] = vqsubq_s16(step2[31], step2[32]);
out[33] = vqsubq_s16(step2[30], step2[33]);
out[34] = vqsubq_s16(step2[29], step2[34]);
out[35] = vqsubq_s16(step2[28], step2[35]);
out[36] = vqsubq_s16(step2[27], step2[36]);
out[37] = vqsubq_s16(step2[26], step2[37]);
out[38] = vqsubq_s16(step2[25], step2[38]);
out[39] = vqsubq_s16(step2[24], step2[39]);
out[40] = vqsubq_s16(step2[23], step2[40]);
out[41] = vqsubq_s16(step2[22], step2[41]);
out[42] = vqsubq_s16(step2[21], step2[42]);
out[43] = vqsubq_s16(step2[20], step2[43]);
out[44] = vqsubq_s16(step2[19], step2[44]);
out[45] = vqsubq_s16(step2[18], step2[45]);
out[46] = vqsubq_s16(step2[17], step2[46]);
out[47] = vqsubq_s16(step2[16], step2[47]);
out[48] = vqsubq_s16(step2[15], step2[48]);
out[49] = vqsubq_s16(step2[14], step2[49]);
out[50] = vqsubq_s16(step2[13], step2[50]);
out[51] = vqsubq_s16(step2[12], step2[51]);
out[52] = vqsubq_s16(step2[11], step2[52]);
out[53] = vqsubq_s16(step2[10], step2[53]);
out[54] = vqsubq_s16(step2[9], step2[54]);
out[55] = vqsubq_s16(step2[8], step2[55]);
out[56] = vqsubq_s16(step2[7], step2[56]);
out[57] = vqsubq_s16(step2[6], step2[57]);
out[58] = vqsubq_s16(step2[5], step2[58]);
out[59] = vqsubq_s16(step2[4], step2[59]);
out[60] = vqsubq_s16(step2[3], step2[60]);
out[61] = vqsubq_s16(step2[2], step2[61]);
out[62] = vqsubq_s16(step2[1], step2[62]);
out[63] = vqsubq_s16(step2[0], step2[63]);
}
// Functions for blocks with eob at DC and within
// topleft 8x8, 16x16, 32x32 corner
static const transform_neon
lowbd_txfm_all_1d_zeros_w_arr[TX_SIZES][ITX_TYPES_1D][4] = {
{
{ NULL, NULL, NULL, NULL },
{ NULL, NULL, NULL, NULL },
{ NULL, NULL, NULL, NULL },
},
{ { idct8_low1_new_neon, idct8_new_neon, NULL, NULL },
{ iadst8_low1_new_neon, iadst8_new_neon, NULL, NULL },
{ identity8_new_neon, identity8_new_neon, NULL, NULL } },
{
{ idct16_low1_new_neon, idct16_low8_new_neon, idct16_new_neon, NULL },
{ iadst16_low1_new_neon, iadst16_low8_new_neon, iadst16_new_neon,
NULL },
{ identity16_new_neon, identity16_new_neon, identity16_new_neon,
NULL },
},
{ { idct32_low1_new_neon, idct32_low8_new_neon, idct32_low16_new_neon,
idct32_new_neon },
{ NULL, NULL, NULL, NULL },
{ identity32_new_neon, identity32_new_neon, identity32_new_neon,
identity32_new_neon } },
{ { idct64_low1_new_neon, idct64_low8_new_neon, idct64_low16_new_neon,
idct64_low32_new_neon },
{ NULL, NULL, NULL, NULL },
{ NULL, NULL, NULL, NULL } }
};
static INLINE void lowbd_inv_txfm2d_add_idtx_neon(const int32_t *input,
uint8_t *output, int stride,
TX_TYPE tx_type,
TX_SIZE tx_size, int eob) {
int16x8_t a[32 * 4];
int16x8_t b[32 * 4];
int eobx, eoby;
get_eobx_eoby_scan_default(&eobx, &eoby, tx_size, eob);
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
lowbd_inv_txfm2d_memset_neon(&a[0], (txfm_size_col * (txfm_size_row) >> 3),
0);
lowbd_inv_txfm2d_memset_neon(&b[0], (txfm_size_col * (txfm_size_row) >> 3),
0);
const int buf_size_w_div8 = txfm_size_col >> 3;
const int rect_type = get_rect_tx_log_ratio(txfm_size_col, txfm_size_row);
const int buf_size_nonzero_h_div8 = (eoby + 8) >> 3;
const int buf_size_nonzero_w_div8 = (eobx + 8) >> 3;
const int fun_idx_x = lowbd_txfm_all_1d_zeros_idx[eobx];
const int fun_idx_y = lowbd_txfm_all_1d_zeros_idx[eoby];
const int32_t *input_1;
int temp_b = 0;
const transform_neon row_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txw_idx][hitx_1d_tab[tx_type]][fun_idx_x];
const transform_neon col_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txh_idx][vitx_1d_tab[tx_type]][fun_idx_y];
assert(col_txfm != NULL);
assert(row_txfm != NULL);
for (int i = 0; i < buf_size_nonzero_h_div8; i++) {
input_1 = input;
for (int j = 0; j < buf_size_nonzero_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
load_buffer_32bit_to_16bit_neon(input_1, &a[k], txfm_size_col);
transpose_s16_8x8q(&a[k], &a[k]);
input_1 += 8;
}
input += (txfm_size_col * 8);
if (abs(rect_type) == 1) {
int y = i * txfm_size_col;
round_shift_for_rect(&a[y], &a[y], txfm_size_col);
}
row_txfm(&a[i * txfm_size_col], &a[i * txfm_size_col], cos_bit_row, 0);
av1_round_shift_array_16_neon(&a[i * txfm_size_col], txfm_size_col,
-shift[0]);
for (int j = 0; j < buf_size_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
transpose_s16_8x8q(&a[k], &b[temp_b + txfm_size_row * j]);
}
temp_b += 8;
}
for (int j = 0; j < buf_size_w_div8; ++j) {
col_txfm(&b[j * txfm_size_row], &b[j * txfm_size_row], cos_bit_col, 0);
av1_round_shift_array_16_neon(&b[j * txfm_size_row], txfm_size_row,
-shift[1]);
}
if (txfm_size_col >= 16) {
for (int i = 0; i < (txfm_size_col >> 4); i++) {
lowbd_add_flip_buffer_16xn_neon(
&b[i * txfm_size_row * 2], output + 16 * i, stride, 0, txfm_size_row);
}
} else if (txfm_size_col == 8) {
lowbd_add_flip_buffer_8xn_neon(b, output, stride, 0, txfm_size_row);
}
}
static INLINE void lowbd_inv_txfm2d_add_v_identity_neon(
const int32_t *input, uint8_t *output, int stride, TX_TYPE tx_type,
TX_SIZE tx_size, int eob) {
int16x8_t a[16 * 2];
int16x8_t b[16 * 2];
int eobx, eoby, ud_flip, lr_flip;
get_eobx_eoby_scan_v_identity(&eobx, &eoby, tx_size, eob);
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
lowbd_inv_txfm2d_memset_neon(&b[0], (txfm_size_col * (txfm_size_row) >> 3),
0);
const int rect_type = get_rect_tx_log_ratio(txfm_size_col, txfm_size_row);
const int buf_size_w_div8 = txfm_size_col >> 3;
const int buf_size_nonzero_h_div8 = (eoby + 8) >> 3;
const int buf_size_nonzero_w_div8 = (eobx + 8) >> 3;
const int fun_idx_x = lowbd_txfm_all_1d_zeros_idx[eobx];
const int fun_idx_y = lowbd_txfm_all_1d_zeros_idx[eoby];
const int32_t *input_1;
int temp_b = 0;
const transform_neon row_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txw_idx][hitx_1d_tab[tx_type]][fun_idx_x];
const transform_neon col_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txh_idx][vitx_1d_tab[tx_type]][fun_idx_y];
assert(col_txfm != NULL);
assert(row_txfm != NULL);
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < buf_size_nonzero_h_div8; i++) {
input_1 = input;
for (int j = 0; j < buf_size_nonzero_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
load_buffer_32bit_to_16bit_neon(input_1, &a[k], txfm_size_col);
transpose_s16_8x8q(&a[k], &a[k]);
input_1 += 8;
}
input += (txfm_size_col * 8);
if (abs(rect_type) == 1) {
int y = i * txfm_size_col;
round_shift_for_rect(&a[y], &a[y], txfm_size_col);
}
row_txfm(&a[i * txfm_size_col], &a[i * txfm_size_col], cos_bit_row, 0);
av1_round_shift_array_16_neon(&a[i * txfm_size_col], txfm_size_col,
-shift[0]);
if (lr_flip == 1) {
for (int j = 0; j < buf_size_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
flip_buf_ud_neon(&a[k], 8);
transpose_s16_8x8q(
&a[k], &b[temp_b + txfm_size_row * (buf_size_w_div8 - 1 - j)]);
}
temp_b += 8;
} else {
for (int j = 0; j < buf_size_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
transpose_s16_8x8q(&a[k], &b[temp_b + txfm_size_row * j]);
}
temp_b += 8;
}
}
for (int j = 0; j < buf_size_w_div8; ++j) {
col_txfm(&b[j * txfm_size_row], &b[j * txfm_size_row], cos_bit_col, 0);
av1_round_shift_array_16_neon(&b[j * txfm_size_row], txfm_size_row,
-shift[1]);
}
if (txfm_size_col >= 16) {
for (int i = 0; i < (txfm_size_col >> 4); i++) {
lowbd_add_flip_buffer_16xn_neon(
&b[i * txfm_size_row * 2], output + 16 * i, stride, 0, txfm_size_row);
}
} else if (txfm_size_col == 8) {
lowbd_add_flip_buffer_8xn_neon(b, output, stride, 0, txfm_size_row);
}
}
static INLINE void lowbd_inv_txfm2d_add_h_identity_neon(
const int32_t *input, uint8_t *output, int stride, TX_TYPE tx_type,
TX_SIZE tx_size, int eob) {
int16x8_t a[16 * 2];
int16x8_t b[16 * 2];
int eobx, eoby, ud_flip, lr_flip;
get_eobx_eoby_scan_h_identity(&eobx, &eoby, tx_size, eob);
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
lowbd_inv_txfm2d_memset_neon(&a[0], (txfm_size_col * (txfm_size_row) >> 3),
0);
const int buf_size_w_div8 = txfm_size_col >> 3;
const int rect_type = get_rect_tx_log_ratio(txfm_size_col, txfm_size_row);
const int buf_size_nonzero_h_div8 = (eoby + 8) >> 3;
const int buf_size_nonzero_w_div8 = (eobx + 8) >> 3;
const int fun_idx_x = lowbd_txfm_all_1d_zeros_idx[eobx];
const int fun_idx_y = lowbd_txfm_all_1d_zeros_idx[eoby];
const int32_t *input_1;
int temp_b = 0;
const transform_neon row_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txw_idx][hitx_1d_tab[tx_type]][fun_idx_x];
const transform_neon col_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txh_idx][vitx_1d_tab[tx_type]][fun_idx_y];
assert(col_txfm != NULL);
assert(row_txfm != NULL);
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < buf_size_nonzero_h_div8; i++) {
input_1 = input;
for (int j = 0; j < buf_size_nonzero_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
load_buffer_32bit_to_16bit_neon(input_1, &a[k], txfm_size_col);
transpose_s16_8x8q(&a[k], &a[k]);
input_1 += 8;
}
input += (txfm_size_col * 8);
if (abs(rect_type) == 1) {
int y = i * txfm_size_col;
round_shift_for_rect(&a[y], &a[y], txfm_size_col);
}
row_txfm(&a[i * txfm_size_col], &a[i * txfm_size_col], cos_bit_row, 0);
av1_round_shift_array_16_neon(&a[i * txfm_size_col], txfm_size_col,
-shift[0]);
for (int j = 0; j < buf_size_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
transpose_s16_8x8q(&a[k], &b[temp_b + txfm_size_row * j]);
}
temp_b += 8;
}
for (int j = 0; j < buf_size_w_div8; ++j) {
col_txfm(&b[j * txfm_size_row], &b[j * txfm_size_row], cos_bit_col, 0);
av1_round_shift_array_16_neon(&b[j * txfm_size_row], txfm_size_row,
-shift[1]);
}
if (txfm_size_col >= 16) {
for (int i = 0; i < (txfm_size_col >> 4); i++) {
lowbd_add_flip_buffer_16xn_neon(&b[i * txfm_size_row * 2],
output + 16 * i, stride, ud_flip,
txfm_size_row);
}
} else if (txfm_size_col == 8) {
lowbd_add_flip_buffer_8xn_neon(b, output, stride, ud_flip, txfm_size_row);
}
}
static INLINE void lowbd_inv_txfm2d_add_4x4_neon(const int32_t *input,
uint8_t *output, int stride,
TX_TYPE tx_type, int eob) {
(void)eob;
TX_SIZE tx_size = TX_4X4;
DECLARE_ALIGNED(32, int, txfm_buf[4 * 4 + 8 + 8]);
int32_t *temp_in = txfm_buf;
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
const int buf_offset = AOMMAX(txfm_size_row, txfm_size_col);
int32_t *temp_out = temp_in + buf_offset;
int32_t *buf = temp_out + buf_offset;
int32_t *buf_ptr = buf;
const int8_t stage_range[MAX_TXFM_STAGE_NUM] = { 16 };
int r, bd = 8;
const transform_1d_neon row_txfm =
lowbd_txfm_all_1d_arr[txw_idx][hitx_1d_tab[tx_type]];
const transform_1d_neon col_txfm =
lowbd_txfm_all_1d_arr[txh_idx][vitx_1d_tab[tx_type]];
int ud_flip, lr_flip;
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < txfm_size_row; i++) {
row_txfm(input, buf_ptr, cos_bit_row, stage_range);
input += txfm_size_col;
buf_ptr += txfm_size_col;
}
for (int c = 0; c < txfm_size_col; ++c) {
if (lr_flip == 0) {
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + c];
} else {
// flip left right
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + (txfm_size_col - c - 1)];
}
col_txfm(temp_in, temp_out, cos_bit_col, stage_range);
av1_round_shift_array(temp_out, txfm_size_row, -shift[1]);
if (ud_flip == 0) {
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] =
highbd_clip_pixel_add(output[r * stride + c], temp_out[r], bd);
}
} else {
// flip upside down
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] = highbd_clip_pixel_add(
output[r * stride + c], temp_out[txfm_size_row - r - 1], bd);
}
}
}
}
void lowbd_inv_txfm2d_add_4x8_neon(const int32_t *input, uint8_t *output,
int stride, TX_TYPE tx_type, int eob) {
(void)eob;
TX_SIZE tx_size = TX_4X8;
DECLARE_ALIGNED(32, int, txfm_buf[4 * 8 + 8 + 8]);
int32_t *temp_in = txfm_buf;
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
const int buf_offset = AOMMAX(txfm_size_row, txfm_size_col);
int32_t *temp_out = temp_in + buf_offset;
int32_t *buf = temp_out + buf_offset;
int32_t *buf_ptr = buf;
const int8_t stage_range[MAX_TXFM_STAGE_NUM] = { 16 };
int r, bd = 8;
const transform_1d_neon row_txfm =
lowbd_txfm_all_1d_arr[txw_idx][hitx_1d_tab[tx_type]];
const transform_1d_neon col_txfm =
lowbd_txfm_all_1d_arr[txh_idx][vitx_1d_tab[tx_type]];
int ud_flip, lr_flip;
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < txfm_size_row; i++) {
for (int j = 0; j < txfm_size_col; j++)
temp_in[j] = round_shift((int64_t)input[j] * NewInvSqrt2, NewSqrt2Bits);
row_txfm(temp_in, buf_ptr, cos_bit_row, stage_range);
input += txfm_size_col;
buf_ptr += txfm_size_col;
}
for (int c = 0; c < txfm_size_col; ++c) {
if (lr_flip == 0) {
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + c];
} else {
// flip left right
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + (txfm_size_col - c - 1)];
}
col_txfm(temp_in, temp_out, cos_bit_col, stage_range);
av1_round_shift_array(temp_out, txfm_size_row, -shift[1]);
if (ud_flip == 0) {
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] =
highbd_clip_pixel_add(output[r * stride + c], temp_out[r], bd);
}
} else {
// flip upside down
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] = highbd_clip_pixel_add(
output[r * stride + c], temp_out[txfm_size_row - r - 1], bd);
}
}
}
}
void lowbd_inv_txfm2d_add_8x4_neon(const int32_t *input, uint8_t *output,
int stride, TX_TYPE tx_type, int eob) {
(void)eob;
TX_SIZE tx_size = TX_8X4;
DECLARE_ALIGNED(32, int, txfm_buf[8 * 4 + 8 + 8]);
int32_t *temp_in = txfm_buf;
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
const int buf_offset = AOMMAX(txfm_size_row, txfm_size_col);
int32_t *temp_out = temp_in + buf_offset;
int32_t *buf = temp_out + buf_offset;
int32_t *buf_ptr = buf;
const int8_t stage_range[MAX_TXFM_STAGE_NUM] = { 16 };
int r, bd = 8;
const transform_1d_neon row_txfm =
lowbd_txfm_all_1d_arr[txw_idx][hitx_1d_tab[tx_type]];
const transform_1d_neon col_txfm =
lowbd_txfm_all_1d_arr[txh_idx][vitx_1d_tab[tx_type]];
int ud_flip, lr_flip;
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < txfm_size_row; i++) {
for (int j = 0; j < txfm_size_col; j++)
temp_in[j] = round_shift((int64_t)input[j] * NewInvSqrt2, NewSqrt2Bits);
row_txfm(temp_in, buf_ptr, cos_bit_row, stage_range);
input += txfm_size_col;
buf_ptr += txfm_size_col;
}
for (int c = 0; c < txfm_size_col; ++c) {
if (lr_flip == 0) {
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + c];
} else {
// flip left right
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + (txfm_size_col - c - 1)];
}
col_txfm(temp_in, temp_out, cos_bit_col, stage_range);
av1_round_shift_array(temp_out, txfm_size_row, -shift[1]);
if (ud_flip == 0) {
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] =
highbd_clip_pixel_add(output[r * stride + c], temp_out[r], bd);
}
} else {
// flip upside down
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] = highbd_clip_pixel_add(
output[r * stride + c], temp_out[txfm_size_row - r - 1], bd);
}
}
}
}
void lowbd_inv_txfm2d_add_4x16_neon(const int32_t *input, uint8_t *output,
int stride, TX_TYPE tx_type, int eob) {
(void)eob;
TX_SIZE tx_size = TX_4X16;
DECLARE_ALIGNED(32, int, txfm_buf[4 * 16 + 16 + 16]);
int32_t *temp_in = txfm_buf;
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
const int buf_offset = AOMMAX(txfm_size_row, txfm_size_col);
int32_t *temp_out = temp_in + buf_offset;
int32_t *buf = temp_out + buf_offset;
int32_t *buf_ptr = buf;
const int8_t stage_range[MAX_TXFM_STAGE_NUM] = { 16 };
int r, bd = 8;
const transform_1d_neon row_txfm =
lowbd_txfm_all_1d_arr[txw_idx][hitx_1d_tab[tx_type]];
const transform_1d_neon col_txfm =
lowbd_txfm_all_1d_arr[txh_idx][vitx_1d_tab[tx_type]];
int ud_flip, lr_flip;
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < txfm_size_row; i++) {
row_txfm(input, buf_ptr, cos_bit_row, stage_range);
av1_round_shift_array(buf_ptr, txfm_size_col, -shift[0]);
input += txfm_size_col;
buf_ptr += txfm_size_col;
}
for (int c = 0; c < txfm_size_col; ++c) {
if (lr_flip == 0) {
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + c];
} else {
// flip left right
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + (txfm_size_col - c - 1)];
}
col_txfm(temp_in, temp_out, cos_bit_col, stage_range);
av1_round_shift_array(temp_out, txfm_size_row, -shift[1]);
if (ud_flip == 0) {
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] =
highbd_clip_pixel_add(output[r * stride + c], temp_out[r], bd);
}
} else {
// flip upside down
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] = highbd_clip_pixel_add(
output[r * stride + c], temp_out[txfm_size_row - r - 1], bd);
}
}
}
}
void lowbd_inv_txfm2d_add_16x4_neon(const int32_t *input, uint8_t *output,
int stride, TX_TYPE tx_type, int eob) {
(void)eob;
TX_SIZE tx_size = TX_16X4;
DECLARE_ALIGNED(32, int, txfm_buf[16 * 4 + 16 + 16]);
int32_t *temp_in = txfm_buf;
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
const int buf_offset = AOMMAX(txfm_size_row, txfm_size_col);
int32_t *temp_out = temp_in + buf_offset;
int32_t *buf = temp_out + buf_offset;
int32_t *buf_ptr = buf;
const int8_t stage_range[MAX_TXFM_STAGE_NUM] = { 16 };
int r, bd = 8;
const transform_1d_neon row_txfm =
lowbd_txfm_all_1d_arr[txw_idx][hitx_1d_tab[tx_type]];
const transform_1d_neon col_txfm =
lowbd_txfm_all_1d_arr[txh_idx][vitx_1d_tab[tx_type]];
int ud_flip, lr_flip;
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < txfm_size_row; i++) {
row_txfm(input, buf_ptr, cos_bit_row, stage_range);
av1_round_shift_array(buf_ptr, txfm_size_col, -shift[0]);
input += txfm_size_col;
buf_ptr += txfm_size_col;
}
for (int c = 0; c < txfm_size_col; ++c) {
if (lr_flip == 0) {
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + c];
} else {
// flip left right
for (r = 0; r < txfm_size_row; ++r)
temp_in[r] = buf[r * txfm_size_col + (txfm_size_col - c - 1)];
}
col_txfm(temp_in, temp_out, cos_bit_col, stage_range);
av1_round_shift_array(temp_out, txfm_size_row, -shift[1]);
if (ud_flip == 0) {
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] =
highbd_clip_pixel_add(output[r * stride + c], temp_out[r], bd);
}
} else {
// flip upside down
for (r = 0; r < txfm_size_row; ++r) {
output[r * stride + c] = highbd_clip_pixel_add(
output[r * stride + c], temp_out[txfm_size_row - r - 1], bd);
}
}
}
}
static INLINE void lowbd_inv_txfm2d_add_no_identity_neon(
const int32_t *input, uint8_t *output, int stride, TX_TYPE tx_type,
TX_SIZE tx_size, int eob) {
int16x8_t a[64 * 8];
int16x8_t b[64 * 8];
int eobx, eoby, ud_flip, lr_flip;
get_eobx_eoby_scan_default(&eobx, &eoby, tx_size, eob);
const int8_t *shift = inv_txfm_shift_ls[tx_size];
const int txw_idx = get_txw_idx(tx_size);
const int txh_idx = get_txh_idx(tx_size);
const int cos_bit_col = inv_cos_bit_col[txw_idx][txh_idx];
const int cos_bit_row = inv_cos_bit_row[txw_idx][txh_idx];
const int txfm_size_col = tx_size_wide[tx_size];
const int txfm_size_row = tx_size_high[tx_size];
const int rect_type = get_rect_tx_log_ratio(txfm_size_col, txfm_size_row);
const int buf_size_w_div8 = txfm_size_col >> 3;
const int buf_size_nonzero_h_div8 = (eoby + 8) >> 3;
const int buf_size_nonzero_w_div8 = (eobx + 8) >> 3;
const int input_stride = AOMMIN(32, txfm_size_col);
const int fun_idx_x = lowbd_txfm_all_1d_zeros_idx[eobx];
const int fun_idx_y = lowbd_txfm_all_1d_zeros_idx[eoby];
const int32_t *input_1;
int temp_b = 0;
const transform_neon row_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txw_idx][hitx_1d_tab[tx_type]][fun_idx_x];
const transform_neon col_txfm =
lowbd_txfm_all_1d_zeros_w_arr[txh_idx][vitx_1d_tab[tx_type]][fun_idx_y];
assert(col_txfm != NULL);
assert(row_txfm != NULL);
get_flip_cfg(tx_type, &ud_flip, &lr_flip);
for (int i = 0; i < buf_size_nonzero_h_div8; i++) {
input_1 = input;
for (int j = 0; j < buf_size_nonzero_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
load_buffer_32bit_to_16bit_neon(input_1, &a[k], input_stride);
transpose_s16_8x8q(&a[k], &a[k]);
input_1 += 8;
}
input += (input_stride * 8);
if (abs(rect_type) == 1) {
int y = i * txfm_size_col;
round_shift_for_rect(&a[y], &a[y], input_stride);
}
row_txfm(&a[i * txfm_size_col], &a[i * txfm_size_col], cos_bit_row, 0);
av1_round_shift_array_16_neon(&a[i * txfm_size_col], txfm_size_col,
-shift[0]);
if (lr_flip == 1) {
for (int j = 0; j < buf_size_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
flip_buf_ud_neon(&a[k], 8);
transpose_s16_8x8q(
&a[k], &b[temp_b + txfm_size_row * (buf_size_w_div8 - 1 - j)]);
}
temp_b += 8;
} else {
for (int j = 0; j < buf_size_w_div8; ++j) {
int k = j * 8 + i * txfm_size_col;
transpose_s16_8x8q(&a[k], &b[temp_b + txfm_size_row * j]);
}
temp_b += 8;
}
}
for (int j = 0; j < buf_size_w_div8; ++j) {
col_txfm(&b[j * txfm_size_row], &b[j * txfm_size_row], cos_bit_col, 0);
av1_round_shift_array_16_neon(&b[j * txfm_size_row], txfm_size_row,
-shift[1]);
}
if (txfm_size_col >= 16) {
for (int i = 0; i < (txfm_size_col >> 4); i++) {
lowbd_add_flip_buffer_16xn_neon(&b[i * txfm_size_row * 2],
output + 16 * i, stride, ud_flip,
txfm_size_row);
}
} else if (txfm_size_col == 8) {
lowbd_add_flip_buffer_8xn_neon(b, output, stride, ud_flip, txfm_size_row);
}
}
static INLINE void lowbd_inv_txfm2d_add_universe_neon(
const int32_t *input, uint8_t *output, int stride, TX_TYPE tx_type,
TX_SIZE tx_size, int eob) {
switch (tx_type) {
case IDTX:
lowbd_inv_txfm2d_add_idtx_neon(input, output, stride, tx_type, tx_size,
eob);
break;
case H_DCT:
case H_ADST:
case H_FLIPADST:
lowbd_inv_txfm2d_add_v_identity_neon(input, output, stride, tx_type,
tx_size, eob);
break;
case V_DCT:
case V_ADST:
case V_FLIPADST:
lowbd_inv_txfm2d_add_h_identity_neon(input, output, stride, tx_type,
tx_size, eob);
break;
default:
lowbd_inv_txfm2d_add_no_identity_neon(input, output, stride, tx_type,
tx_size, eob);
break;
}
}
void av1_lowbd_inv_txfm2d_add_neon(const int32_t *input, uint8_t *output,
int stride, TX_TYPE tx_type, TX_SIZE tx_size,
int eob) {
switch (tx_size) {
case TX_4X4:
lowbd_inv_txfm2d_add_4x4_neon(input, output, stride, tx_type, eob);
break;
case TX_4X8:
lowbd_inv_txfm2d_add_4x8_neon(input, output, stride, tx_type, eob);
break;
case TX_8X4:
lowbd_inv_txfm2d_add_8x4_neon(input, output, stride, tx_type, eob);
break;
case TX_4X16:
lowbd_inv_txfm2d_add_4x16_neon(input, output, stride, tx_type, eob);
break;
case TX_16X4:
lowbd_inv_txfm2d_add_16x4_neon(input, output, stride, tx_type, eob);
break;
default:
lowbd_inv_txfm2d_add_universe_neon(input, output, stride, tx_type,
tx_size, eob);
break;
}
}
void av1_inv_txfm_add_neon(const tran_low_t *dqcoeff, uint8_t *dst, int stride,
const TxfmParam *txfm_param) {
const TX_TYPE tx_type = txfm_param->tx_type;
if (!txfm_param->lossless) {
av1_lowbd_inv_txfm2d_add_neon(dqcoeff, dst, stride, tx_type,
txfm_param->tx_size, txfm_param->eob);
} else {
av1_inv_txfm_add_c(dqcoeff, dst, stride, txfm_param);
}
}