diff options
Diffstat (limited to 'third_party/aom/test/av1_fht16x16_test.cc')
-rw-r--r-- | third_party/aom/test/av1_fht16x16_test.cc | 164 |
1 files changed, 82 insertions, 82 deletions
diff --git a/third_party/aom/test/av1_fht16x16_test.cc b/third_party/aom/test/av1_fht16x16_test.cc index c0f6974c6..21235a837 100644 --- a/third_party/aom/test/av1_fht16x16_test.cc +++ b/third_party/aom/test/av1_fht16x16_test.cc @@ -28,7 +28,7 @@ typedef void (*IhtFunc)(const tran_low_t *in, uint8_t *out, int stride, const TxfmParam *txfm_param); using std::tr1::tuple; using libaom_test::FhtFunc; -typedef tuple<FhtFunc, IhtFunc, int, aom_bit_depth_t, int> Ht16x16Param; +typedef tuple<FhtFunc, IhtFunc, TX_TYPE, aom_bit_depth_t, int> Ht16x16Param; void fht16x16_ref(const int16_t *in, tran_low_t *out, int stride, TxfmParam *txfm_param) { @@ -42,15 +42,15 @@ void iht16x16_ref(const tran_low_t *in, uint8_t *dest, int stride, #if CONFIG_HIGHBITDEPTH typedef void (*IHbdHtFunc)(const tran_low_t *in, uint8_t *out, int stride, - int tx_type, int bd); + TX_TYPE tx_type, int bd); typedef void (*HbdHtFunc)(const int16_t *input, int32_t *output, int stride, - int tx_type, int bd); + TX_TYPE tx_type, int bd); // Target optimized function, tx_type, bit depth -typedef tuple<HbdHtFunc, int, int> HighbdHt16x16Param; +typedef tuple<HbdHtFunc, TX_TYPE, int> HighbdHt16x16Param; void highbd_fht16x16_ref(const int16_t *in, int32_t *out, int stride, - int tx_type, int bd) { + TX_TYPE tx_type, int bd) { av1_fwd_txfm2d_16x16_c(in, out, stride, tx_type, bd); } #endif // CONFIG_HIGHBITDEPTH @@ -128,7 +128,7 @@ class AV1HighbdTrans16x16HT private: HbdHtFunc fwd_txfm_; HbdHtFunc fwd_txfm_ref_; - int tx_type_; + TX_TYPE tx_type_; int bit_depth_; int mask_; int num_coeffs_; @@ -164,113 +164,113 @@ TEST_P(AV1HighbdTrans16x16HT, HighbdCoeffCheck) { RunBitexactCheck(); } using std::tr1::make_tuple; -#if HAVE_SSE2 +#if HAVE_SSE2 && !CONFIG_DAALA_DCT16 const Ht16x16Param kArrayHt16x16Param_sse2[] = { - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 0, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 1, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 2, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 3, AOM_BITS_8, - 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, DCT_DCT, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, ADST_DCT, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, DCT_ADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, ADST_ADST, + AOM_BITS_8, 256), #if CONFIG_EXT_TX - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 4, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 5, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 6, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 7, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 8, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 9, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 10, AOM_BITS_8, + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, FLIPADST_DCT, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, DCT_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, FLIPADST_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, ADST_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, FLIPADST_ADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, IDTX, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 11, AOM_BITS_8, + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, V_DCT, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 12, AOM_BITS_8, + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, H_DCT, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 13, AOM_BITS_8, + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, V_ADST, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 14, AOM_BITS_8, + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, H_ADST, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, 15, AOM_BITS_8, - 256) + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, V_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_sse2, &av1_iht16x16_256_add_sse2, H_FLIPADST, + AOM_BITS_8, 256) #endif // CONFIG_EXT_TX }; INSTANTIATE_TEST_CASE_P(SSE2, AV1Trans16x16HT, ::testing::ValuesIn(kArrayHt16x16Param_sse2)); #endif // HAVE_SSE2 -#if HAVE_AVX2 +#if HAVE_AVX2 && !CONFIG_DAALA_DCT16 const Ht16x16Param kArrayHt16x16Param_avx2[] = { - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 0, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 1, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 2, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 3, AOM_BITS_8, - 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, DCT_DCT, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, ADST_DCT, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, DCT_ADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, ADST_ADST, + AOM_BITS_8, 256), #if CONFIG_EXT_TX - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 4, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 5, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 6, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 7, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 8, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 9, AOM_BITS_8, - 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 10, AOM_BITS_8, + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, FLIPADST_DCT, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, DCT_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, FLIPADST_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, ADST_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, FLIPADST_ADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, IDTX, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 11, AOM_BITS_8, + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, V_DCT, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 12, AOM_BITS_8, + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, H_DCT, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 13, AOM_BITS_8, + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, V_ADST, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 14, AOM_BITS_8, + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, H_ADST, AOM_BITS_8, 256), - make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, 15, AOM_BITS_8, - 256) + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, V_FLIPADST, + AOM_BITS_8, 256), + make_tuple(&av1_fht16x16_avx2, &av1_iht16x16_256_add_avx2, H_FLIPADST, + AOM_BITS_8, 256) #endif // CONFIG_EXT_TX }; INSTANTIATE_TEST_CASE_P(AVX2, AV1Trans16x16HT, ::testing::ValuesIn(kArrayHt16x16Param_avx2)); #endif // HAVE_AVX2 -#if HAVE_SSE4_1 && CONFIG_HIGHBITDEPTH +#if HAVE_SSE4_1 && CONFIG_HIGHBITDEPTH && !CONFIG_DAALA_DCT16 const HighbdHt16x16Param kArrayHBDHt16x16Param_sse4_1[] = { - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 0, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 0, 12), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 1, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 1, 12), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 2, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 2, 12), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 3, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 3, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, DCT_DCT, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, DCT_DCT, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, ADST_DCT, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, ADST_DCT, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, DCT_ADST, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, DCT_ADST, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, ADST_ADST, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, ADST_ADST, 12), #if CONFIG_EXT_TX - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 4, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 4, 12), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 5, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 5, 12), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 6, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 6, 12), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 7, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 7, 12), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 8, 10), - make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, 8, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, FLIPADST_DCT, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, FLIPADST_DCT, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, DCT_FLIPADST, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, DCT_FLIPADST, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, FLIPADST_FLIPADST, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, FLIPADST_FLIPADST, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, ADST_FLIPADST, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, ADST_FLIPADST, 12), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, FLIPADST_ADST, 10), + make_tuple(&av1_fwd_txfm2d_16x16_sse4_1, FLIPADST_ADST, 12), #endif // CONFIG_EXT_TX }; INSTANTIATE_TEST_CASE_P(SSE4_1, AV1HighbdTrans16x16HT, ::testing::ValuesIn(kArrayHBDHt16x16Param_sse4_1)); -#endif // HAVE_SSE4_1 && CONFIG_HIGHBITDEPTH +#endif // HAVE_SSE4_1 && CONFIG_HIGHBITDEPTH && !CONFIG_DAALA_DCT16 } // namespace |