summaryrefslogtreecommitdiffstats
path: root/third_party/aom/test/masked_variance_test.cc
diff options
context:
space:
mode:
Diffstat (limited to 'third_party/aom/test/masked_variance_test.cc')
-rw-r--r--third_party/aom/test/masked_variance_test.cc594
1 files changed, 155 insertions, 439 deletions
diff --git a/third_party/aom/test/masked_variance_test.cc b/third_party/aom/test/masked_variance_test.cc
index 65e852aea..bf113c69e 100644
--- a/third_party/aom/test/masked_variance_test.cc
+++ b/third_party/aom/test/masked_variance_test.cc
@@ -29,107 +29,12 @@
using libaom_test::ACMRandom;
namespace {
-const int number_of_iterations = 500;
-
-typedef unsigned int (*MaskedVarianceFunc)(const uint8_t *a, int a_stride,
- const uint8_t *b, int b_stride,
- const uint8_t *m, int m_stride,
- unsigned int *sse);
-
-typedef std::tr1::tuple<MaskedVarianceFunc, MaskedVarianceFunc>
- MaskedVarianceParam;
-
-class MaskedVarianceTest
- : public ::testing::TestWithParam<MaskedVarianceParam> {
- public:
- virtual ~MaskedVarianceTest() {}
- virtual void SetUp() {
- opt_func_ = GET_PARAM(0);
- ref_func_ = GET_PARAM(1);
- }
-
- virtual void TearDown() { libaom_test::ClearSystemState(); }
-
- protected:
- MaskedVarianceFunc opt_func_;
- MaskedVarianceFunc ref_func_;
-};
-
-TEST_P(MaskedVarianceTest, OperationCheck) {
- unsigned int ref_ret, opt_ret;
- unsigned int ref_sse, opt_sse;
- ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint8_t, src_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint8_t, ref_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- int err_count = 0;
- int first_failure = -1;
- int src_stride = MAX_SB_SIZE;
- int ref_stride = MAX_SB_SIZE;
- int msk_stride = MAX_SB_SIZE;
-
- for (int i = 0; i < number_of_iterations; ++i) {
- for (int j = 0; j < MAX_SB_SIZE * MAX_SB_SIZE; j++) {
- src_ptr[j] = rnd.Rand8();
- ref_ptr[j] = rnd.Rand8();
- msk_ptr[j] = rnd(65);
- }
-
- ref_ret = ref_func_(src_ptr, src_stride, ref_ptr, ref_stride, msk_ptr,
- msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(opt_ret = opt_func_(src_ptr, src_stride, ref_ptr,
- ref_stride, msk_ptr,
- msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) first_failure = i;
- }
- }
-
- EXPECT_EQ(0, err_count) << "Error: Masked Variance Test OperationCheck,"
- << "C output doesn't match SSSE3 output. "
- << "First failed at test case " << first_failure;
-}
-
-TEST_P(MaskedVarianceTest, ExtremeValues) {
- unsigned int ref_ret, opt_ret;
- unsigned int ref_sse, opt_sse;
- ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint8_t, src_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint8_t, ref_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- int err_count = 0;
- int first_failure = -1;
- int src_stride = MAX_SB_SIZE;
- int ref_stride = MAX_SB_SIZE;
- int msk_stride = MAX_SB_SIZE;
-
- for (int i = 0; i < 8; ++i) {
- memset(src_ptr, (i & 0x1) ? 255 : 0, MAX_SB_SIZE * MAX_SB_SIZE);
- memset(ref_ptr, (i & 0x2) ? 255 : 0, MAX_SB_SIZE * MAX_SB_SIZE);
- memset(msk_ptr, (i & 0x4) ? 64 : 0, MAX_SB_SIZE * MAX_SB_SIZE);
-
- ref_ret = ref_func_(src_ptr, src_stride, ref_ptr, ref_stride, msk_ptr,
- msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(opt_ret = opt_func_(src_ptr, src_stride, ref_ptr,
- ref_stride, msk_ptr,
- msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) first_failure = i;
- }
- }
-
- EXPECT_EQ(0, err_count) << "Error: Masked Variance Test ExtremeValues,"
- << "C output doesn't match SSSE3 output. "
- << "First failed at test case " << first_failure;
-}
+const int number_of_iterations = 200;
typedef unsigned int (*MaskedSubPixelVarianceFunc)(
- const uint8_t *a, int a_stride, int xoffset, int yoffset, const uint8_t *b,
- int b_stride, const uint8_t *m, int m_stride, unsigned int *sse);
+ const uint8_t *src, int src_stride, int xoffset, int yoffset,
+ const uint8_t *ref, int ref_stride, const uint8_t *second_pred,
+ const uint8_t *msk, int msk_stride, int invert_mask, unsigned int *sse);
typedef std::tr1::tuple<MaskedSubPixelVarianceFunc, MaskedSubPixelVarianceFunc>
MaskedSubPixelVarianceParam;
@@ -154,9 +59,18 @@ TEST_P(MaskedSubPixelVarianceTest, OperationCheck) {
unsigned int ref_ret, opt_ret;
unsigned int ref_sse, opt_sse;
ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint8_t, src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint8_t, ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
+ // Note: We pad the input arrays out with 15 extra elements, since the SSE
+ // implementations can read up to 15 elements off the end of the main data.
+ // The extra data is never actually used, but it simplifies the code
+ // if we can do this.
+ DECLARE_ALIGNED(16, uint8_t,
+ src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
+ DECLARE_ALIGNED(16, uint8_t,
+ ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
+ DECLARE_ALIGNED(16, uint8_t,
+ second_pred_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
+ DECLARE_ALIGNED(16, uint8_t,
+ msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
int err_count = 0;
int first_failure = -1;
int src_stride = (MAX_SB_SIZE + 1);
@@ -171,23 +85,26 @@ TEST_P(MaskedSubPixelVarianceTest, OperationCheck) {
for (int j = 0; j < (MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1); j++) {
src_ptr[j] = rnd.Rand8();
ref_ptr[j] = rnd.Rand8();
+ second_pred_ptr[j] = rnd.Rand8();
msk_ptr[j] = rnd(65);
}
for (int k = 0; k < 3; k++) {
- xoffset = xoffsets[k];
for (int l = 0; l < 3; l++) {
xoffset = xoffsets[k];
yoffset = yoffsets[l];
-
- ref_ret = ref_func_(src_ptr, src_stride, xoffset, yoffset, ref_ptr,
- ref_stride, msk_ptr, msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(
- opt_ret = opt_func_(src_ptr, src_stride, xoffset, yoffset, ref_ptr,
- ref_stride, msk_ptr, msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) first_failure = i;
+ for (int invert_mask = 0; invert_mask < 2; ++invert_mask) {
+ ref_ret = ref_func_(src_ptr, src_stride, xoffset, yoffset, ref_ptr,
+ ref_stride, second_pred_ptr, msk_ptr, msk_stride,
+ invert_mask, &ref_sse);
+ ASM_REGISTER_STATE_CHECK(
+ opt_ret = opt_func_(src_ptr, src_stride, xoffset, yoffset,
+ ref_ptr, ref_stride, second_pred_ptr, msk_ptr,
+ msk_stride, invert_mask, &opt_sse));
+
+ if (opt_ret != ref_ret || opt_sse != ref_sse) {
+ err_count++;
+ if (first_failure == -1) first_failure = i;
+ }
}
}
}
@@ -203,9 +120,14 @@ TEST_P(MaskedSubPixelVarianceTest, ExtremeValues) {
unsigned int ref_ret, opt_ret;
unsigned int ref_sse, opt_sse;
ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint8_t, src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint8_t, ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
+ DECLARE_ALIGNED(16, uint8_t,
+ src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
+ DECLARE_ALIGNED(16, uint8_t,
+ ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
+ DECLARE_ALIGNED(16, uint8_t,
+ second_pred_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
+ DECLARE_ALIGNED(16, uint8_t,
+ msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 15]);
int first_failure_x = -1;
int first_failure_y = -1;
int err_count = 0;
@@ -216,26 +138,32 @@ TEST_P(MaskedSubPixelVarianceTest, ExtremeValues) {
for (int xoffset = 0; xoffset < BIL_SUBPEL_SHIFTS; xoffset++) {
for (int yoffset = 0; yoffset < BIL_SUBPEL_SHIFTS; yoffset++) {
- for (int i = 0; i < 8; ++i) {
+ for (int i = 0; i < 16; ++i) {
memset(src_ptr, (i & 0x1) ? 255 : 0,
(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
memset(ref_ptr, (i & 0x2) ? 255 : 0,
(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
- memset(msk_ptr, (i & 0x4) ? 64 : 0,
+ memset(second_pred_ptr, (i & 0x4) ? 255 : 0,
+ (MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
+ memset(msk_ptr, (i & 0x8) ? 64 : 0,
(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
- ref_ret = ref_func_(src_ptr, src_stride, xoffset, yoffset, ref_ptr,
- ref_stride, msk_ptr, msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(
- opt_ret = opt_func_(src_ptr, src_stride, xoffset, yoffset, ref_ptr,
- ref_stride, msk_ptr, msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) {
- first_failure = i;
- first_failure_x = xoffset;
- first_failure_y = yoffset;
+ for (int invert_mask = 0; invert_mask < 2; ++invert_mask) {
+ ref_ret = ref_func_(src_ptr, src_stride, xoffset, yoffset, ref_ptr,
+ ref_stride, second_pred_ptr, msk_ptr, msk_stride,
+ invert_mask, &ref_sse);
+ ASM_REGISTER_STATE_CHECK(
+ opt_ret = opt_func_(src_ptr, src_stride, xoffset, yoffset,
+ ref_ptr, ref_stride, second_pred_ptr, msk_ptr,
+ msk_stride, invert_mask, &opt_sse));
+
+ if (opt_ret != ref_ret || opt_sse != ref_sse) {
+ err_count++;
+ if (first_failure == -1) {
+ first_failure = i;
+ first_failure_x = xoffset;
+ first_failure_y = yoffset;
+ }
}
}
}
@@ -250,105 +178,6 @@ TEST_P(MaskedSubPixelVarianceTest, ExtremeValues) {
}
#if CONFIG_HIGHBITDEPTH
-typedef std::tr1::tuple<MaskedVarianceFunc, MaskedVarianceFunc, aom_bit_depth_t>
- HighbdMaskedVarianceParam;
-
-class HighbdMaskedVarianceTest
- : public ::testing::TestWithParam<HighbdMaskedVarianceParam> {
- public:
- virtual ~HighbdMaskedVarianceTest() {}
- virtual void SetUp() {
- opt_func_ = GET_PARAM(0);
- ref_func_ = GET_PARAM(1);
- bit_depth_ = GET_PARAM(2);
- }
-
- virtual void TearDown() { libaom_test::ClearSystemState(); }
-
- protected:
- MaskedVarianceFunc opt_func_;
- MaskedVarianceFunc ref_func_;
- aom_bit_depth_t bit_depth_;
-};
-
-TEST_P(HighbdMaskedVarianceTest, OperationCheck) {
- unsigned int ref_ret, opt_ret;
- unsigned int ref_sse, opt_sse;
- ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint16_t, src_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint16_t, ref_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- uint8_t *src8_ptr = CONVERT_TO_BYTEPTR(src_ptr);
- uint8_t *ref8_ptr = CONVERT_TO_BYTEPTR(ref_ptr);
- int err_count = 0;
- int first_failure = -1;
- int src_stride = MAX_SB_SIZE;
- int ref_stride = MAX_SB_SIZE;
- int msk_stride = MAX_SB_SIZE;
-
- for (int i = 0; i < number_of_iterations; ++i) {
- for (int j = 0; j < MAX_SB_SIZE * MAX_SB_SIZE; j++) {
- src_ptr[j] = rnd.Rand16() & ((1 << bit_depth_) - 1);
- ref_ptr[j] = rnd.Rand16() & ((1 << bit_depth_) - 1);
- msk_ptr[j] = rnd(65);
- }
-
- ref_ret = ref_func_(src8_ptr, src_stride, ref8_ptr, ref_stride, msk_ptr,
- msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(opt_ret = opt_func_(src8_ptr, src_stride, ref8_ptr,
- ref_stride, msk_ptr,
- msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) first_failure = i;
- }
- }
-
- EXPECT_EQ(0, err_count) << "Error: Masked Variance Test OperationCheck,"
- << "C output doesn't match SSSE3 output. "
- << "First failed at test case " << first_failure;
-}
-
-TEST_P(HighbdMaskedVarianceTest, ExtremeValues) {
- unsigned int ref_ret, opt_ret;
- unsigned int ref_sse, opt_sse;
- ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint16_t, src_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint16_t, ref_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[MAX_SB_SIZE * MAX_SB_SIZE]);
- uint8_t *src8_ptr = CONVERT_TO_BYTEPTR(src_ptr);
- uint8_t *ref8_ptr = CONVERT_TO_BYTEPTR(ref_ptr);
- int err_count = 0;
- int first_failure = -1;
- int src_stride = MAX_SB_SIZE;
- int ref_stride = MAX_SB_SIZE;
- int msk_stride = MAX_SB_SIZE;
-
- for (int i = 0; i < 8; ++i) {
- aom_memset16(src_ptr, (i & 0x1) ? ((1 << bit_depth_) - 1) : 0,
- MAX_SB_SIZE * MAX_SB_SIZE);
- aom_memset16(ref_ptr, (i & 0x2) ? ((1 << bit_depth_) - 1) : 0,
- MAX_SB_SIZE * MAX_SB_SIZE);
- memset(msk_ptr, (i & 0x4) ? 64 : 0, MAX_SB_SIZE * MAX_SB_SIZE);
-
- ref_ret = ref_func_(src8_ptr, src_stride, ref8_ptr, ref_stride, msk_ptr,
- msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(opt_ret = opt_func_(src8_ptr, src_stride, ref8_ptr,
- ref_stride, msk_ptr,
- msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) first_failure = i;
- }
- }
-
- EXPECT_EQ(0, err_count) << "Error: Masked Variance Test ExtremeValues,"
- << "C output doesn't match SSSE3 output. "
- << "First failed at test case " << first_failure;
-}
-
typedef std::tr1::tuple<MaskedSubPixelVarianceFunc, MaskedSubPixelVarianceFunc,
aom_bit_depth_t>
HighbdMaskedSubPixelVarianceParam;
@@ -375,11 +204,21 @@ TEST_P(HighbdMaskedSubPixelVarianceTest, OperationCheck) {
unsigned int ref_ret, opt_ret;
unsigned int ref_sse, opt_sse;
ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint16_t, src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint16_t, ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
+ // Note: We pad the input arrays out with 7 extra elements, since the SSE
+ // implementations can read up to 7 elements off the end of the main data.
+ // The extra data is never actually used, but it simplifies the code
+ // if we can do this.
+ DECLARE_ALIGNED(16, uint16_t,
+ src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
+ DECLARE_ALIGNED(16, uint16_t,
+ ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
+ DECLARE_ALIGNED(16, uint16_t,
+ second_pred_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
+ DECLARE_ALIGNED(16, uint8_t,
+ msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
uint8_t *src8_ptr = CONVERT_TO_BYTEPTR(src_ptr);
uint8_t *ref8_ptr = CONVERT_TO_BYTEPTR(ref_ptr);
+ uint8_t *second_pred8_ptr = CONVERT_TO_BYTEPTR(second_pred_ptr);
int err_count = 0;
int first_failure = -1;
int first_failure_x = -1;
@@ -390,27 +229,30 @@ TEST_P(HighbdMaskedSubPixelVarianceTest, OperationCheck) {
int xoffset, yoffset;
for (int i = 0; i < number_of_iterations; ++i) {
+ for (int j = 0; j < (MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1); j++) {
+ src_ptr[j] = rnd.Rand16() & ((1 << bit_depth_) - 1);
+ ref_ptr[j] = rnd.Rand16() & ((1 << bit_depth_) - 1);
+ second_pred_ptr[j] = rnd.Rand16() & ((1 << bit_depth_) - 1);
+ msk_ptr[j] = rnd(65);
+ }
for (xoffset = 0; xoffset < BIL_SUBPEL_SHIFTS; xoffset++) {
for (yoffset = 0; yoffset < BIL_SUBPEL_SHIFTS; yoffset++) {
- for (int j = 0; j < (MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1); j++) {
- src_ptr[j] = rnd.Rand16() & ((1 << bit_depth_) - 1);
- ref_ptr[j] = rnd.Rand16() & ((1 << bit_depth_) - 1);
- msk_ptr[j] = rnd(65);
- }
-
- ref_ret = ref_func_(src8_ptr, src_stride, xoffset, yoffset, ref8_ptr,
- ref_stride, msk_ptr, msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(opt_ret =
- opt_func_(src8_ptr, src_stride, xoffset,
- yoffset, ref8_ptr, ref_stride,
- msk_ptr, msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) {
- first_failure = i;
- first_failure_x = xoffset;
- first_failure_y = yoffset;
+ for (int invert_mask = 0; invert_mask < 2; ++invert_mask) {
+ ref_ret = ref_func_(src8_ptr, src_stride, xoffset, yoffset, ref8_ptr,
+ ref_stride, second_pred8_ptr, msk_ptr, msk_stride,
+ invert_mask, &ref_sse);
+ ASM_REGISTER_STATE_CHECK(
+ opt_ret = opt_func_(src8_ptr, src_stride, xoffset, yoffset,
+ ref8_ptr, ref_stride, second_pred8_ptr,
+ msk_ptr, msk_stride, invert_mask, &opt_sse));
+
+ if (opt_ret != ref_ret || opt_sse != ref_sse) {
+ err_count++;
+ if (first_failure == -1) {
+ first_failure = i;
+ first_failure_x = xoffset;
+ first_failure_y = yoffset;
+ }
}
}
}
@@ -428,11 +270,17 @@ TEST_P(HighbdMaskedSubPixelVarianceTest, ExtremeValues) {
unsigned int ref_ret, opt_ret;
unsigned int ref_sse, opt_sse;
ACMRandom rnd(ACMRandom::DeterministicSeed());
- DECLARE_ALIGNED(16, uint16_t, src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint16_t, ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
- DECLARE_ALIGNED(16, uint8_t, msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1)]);
+ DECLARE_ALIGNED(16, uint16_t,
+ src_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
+ DECLARE_ALIGNED(16, uint16_t,
+ ref_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
+ DECLARE_ALIGNED(16, uint8_t,
+ msk_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
+ DECLARE_ALIGNED(16, uint16_t,
+ second_pred_ptr[(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1) + 7]);
uint8_t *src8_ptr = CONVERT_TO_BYTEPTR(src_ptr);
uint8_t *ref8_ptr = CONVERT_TO_BYTEPTR(ref_ptr);
+ uint8_t *second_pred8_ptr = CONVERT_TO_BYTEPTR(second_pred_ptr);
int first_failure_x = -1;
int first_failure_y = -1;
int err_count = 0;
@@ -443,27 +291,32 @@ TEST_P(HighbdMaskedSubPixelVarianceTest, ExtremeValues) {
for (int xoffset = 0; xoffset < BIL_SUBPEL_SHIFTS; xoffset++) {
for (int yoffset = 0; yoffset < BIL_SUBPEL_SHIFTS; yoffset++) {
- for (int i = 0; i < 8; ++i) {
+ for (int i = 0; i < 16; ++i) {
aom_memset16(src_ptr, (i & 0x1) ? ((1 << bit_depth_) - 1) : 0,
(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
aom_memset16(ref_ptr, (i & 0x2) ? ((1 << bit_depth_) - 1) : 0,
(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
- memset(msk_ptr, (i & 0x4) ? 64 : 0,
+ aom_memset16(second_pred_ptr, (i & 0x4) ? ((1 << bit_depth_) - 1) : 0,
+ (MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
+ memset(msk_ptr, (i & 0x8) ? 64 : 0,
(MAX_SB_SIZE + 1) * (MAX_SB_SIZE + 1));
- ref_ret = ref_func_(src8_ptr, src_stride, xoffset, yoffset, ref8_ptr,
- ref_stride, msk_ptr, msk_stride, &ref_sse);
- ASM_REGISTER_STATE_CHECK(opt_ret =
- opt_func_(src8_ptr, src_stride, xoffset,
- yoffset, ref8_ptr, ref_stride,
- msk_ptr, msk_stride, &opt_sse));
-
- if (opt_ret != ref_ret || opt_sse != ref_sse) {
- err_count++;
- if (first_failure == -1) {
- first_failure = i;
- first_failure_x = xoffset;
- first_failure_y = yoffset;
+ for (int invert_mask = 0; invert_mask < 2; ++invert_mask) {
+ ref_ret = ref_func_(src8_ptr, src_stride, xoffset, yoffset, ref8_ptr,
+ ref_stride, second_pred8_ptr, msk_ptr, msk_stride,
+ invert_mask, &ref_sse);
+ ASM_REGISTER_STATE_CHECK(
+ opt_ret = opt_func_(src8_ptr, src_stride, xoffset, yoffset,
+ ref8_ptr, ref_stride, second_pred8_ptr,
+ msk_ptr, msk_stride, invert_mask, &opt_sse));
+
+ if (opt_ret != ref_ret || opt_sse != ref_sse) {
+ err_count++;
+ if (first_failure == -1) {
+ first_failure = i;
+ first_failure_x = xoffset;
+ first_failure_y = yoffset;
+ }
}
}
}
@@ -482,38 +335,6 @@ using std::tr1::make_tuple;
#if HAVE_SSSE3
INSTANTIATE_TEST_CASE_P(
- SSSE3_C_COMPARE, MaskedVarianceTest,
- ::testing::Values(
-#if CONFIG_EXT_PARTITION
- make_tuple(&aom_masked_variance128x128_ssse3,
- &aom_masked_variance128x128_c),
- make_tuple(&aom_masked_variance128x64_ssse3,
- &aom_masked_variance128x64_c),
- make_tuple(&aom_masked_variance64x128_ssse3,
- &aom_masked_variance64x128_c),
-#endif // CONFIG_EXT_PARTITION
- make_tuple(&aom_masked_variance64x64_ssse3,
- &aom_masked_variance64x64_c),
- make_tuple(&aom_masked_variance64x32_ssse3,
- &aom_masked_variance64x32_c),
- make_tuple(&aom_masked_variance32x64_ssse3,
- &aom_masked_variance32x64_c),
- make_tuple(&aom_masked_variance32x32_ssse3,
- &aom_masked_variance32x32_c),
- make_tuple(&aom_masked_variance32x16_ssse3,
- &aom_masked_variance32x16_c),
- make_tuple(&aom_masked_variance16x32_ssse3,
- &aom_masked_variance16x32_c),
- make_tuple(&aom_masked_variance16x16_ssse3,
- &aom_masked_variance16x16_c),
- make_tuple(&aom_masked_variance16x8_ssse3, &aom_masked_variance16x8_c),
- make_tuple(&aom_masked_variance8x16_ssse3, &aom_masked_variance8x16_c),
- make_tuple(&aom_masked_variance8x8_ssse3, &aom_masked_variance8x8_c),
- make_tuple(&aom_masked_variance8x4_ssse3, &aom_masked_variance8x4_c),
- make_tuple(&aom_masked_variance4x8_ssse3, &aom_masked_variance4x8_c),
- make_tuple(&aom_masked_variance4x4_ssse3, &aom_masked_variance4x4_c)));
-
-INSTANTIATE_TEST_CASE_P(
SSSE3_C_COMPARE, MaskedSubPixelVarianceTest,
::testing::Values(
#if CONFIG_EXT_PARTITION
@@ -553,148 +374,43 @@ INSTANTIATE_TEST_CASE_P(
#if CONFIG_HIGHBITDEPTH
INSTANTIATE_TEST_CASE_P(
- SSSE3_C_COMPARE, HighbdMaskedVarianceTest,
- ::testing::Values(
-#if CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_masked_variance128x128_ssse3,
- &aom_highbd_masked_variance128x128_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance128x64_ssse3,
- &aom_highbd_masked_variance128x64_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance64x128_ssse3,
- &aom_highbd_masked_variance64x128_c, AOM_BITS_8),
-#endif // CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_masked_variance64x64_ssse3,
- &aom_highbd_masked_variance64x64_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance64x32_ssse3,
- &aom_highbd_masked_variance64x32_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance32x64_ssse3,
- &aom_highbd_masked_variance32x64_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance32x32_ssse3,
- &aom_highbd_masked_variance32x32_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance32x16_ssse3,
- &aom_highbd_masked_variance32x16_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance16x32_ssse3,
- &aom_highbd_masked_variance16x32_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance16x16_ssse3,
- &aom_highbd_masked_variance16x16_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance16x8_ssse3,
- &aom_highbd_masked_variance16x8_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance8x16_ssse3,
- &aom_highbd_masked_variance8x16_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance8x8_ssse3,
- &aom_highbd_masked_variance8x8_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance8x4_ssse3,
- &aom_highbd_masked_variance8x4_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance4x8_ssse3,
- &aom_highbd_masked_variance4x8_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_variance4x4_ssse3,
- &aom_highbd_masked_variance4x4_c, AOM_BITS_8),
-#if CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_10_masked_variance128x128_ssse3,
- &aom_highbd_10_masked_variance128x128_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance128x64_ssse3,
- &aom_highbd_10_masked_variance128x64_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance64x128_ssse3,
- &aom_highbd_10_masked_variance64x128_c, AOM_BITS_10),
-#endif // CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_10_masked_variance64x64_ssse3,
- &aom_highbd_10_masked_variance64x64_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance64x32_ssse3,
- &aom_highbd_10_masked_variance64x32_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance32x64_ssse3,
- &aom_highbd_10_masked_variance32x64_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance32x32_ssse3,
- &aom_highbd_10_masked_variance32x32_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance32x16_ssse3,
- &aom_highbd_10_masked_variance32x16_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance16x32_ssse3,
- &aom_highbd_10_masked_variance16x32_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance16x16_ssse3,
- &aom_highbd_10_masked_variance16x16_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance16x8_ssse3,
- &aom_highbd_10_masked_variance16x8_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance8x16_ssse3,
- &aom_highbd_10_masked_variance8x16_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance8x8_ssse3,
- &aom_highbd_10_masked_variance8x8_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance8x4_ssse3,
- &aom_highbd_10_masked_variance8x4_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance4x8_ssse3,
- &aom_highbd_10_masked_variance4x8_c, AOM_BITS_10),
- make_tuple(&aom_highbd_10_masked_variance4x4_ssse3,
- &aom_highbd_10_masked_variance4x4_c, AOM_BITS_10),
-#if CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_12_masked_variance128x128_ssse3,
- &aom_highbd_12_masked_variance128x128_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance128x64_ssse3,
- &aom_highbd_12_masked_variance128x64_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance64x128_ssse3,
- &aom_highbd_12_masked_variance64x128_c, AOM_BITS_12),
-#endif // CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_12_masked_variance64x64_ssse3,
- &aom_highbd_12_masked_variance64x64_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance64x32_ssse3,
- &aom_highbd_12_masked_variance64x32_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance32x64_ssse3,
- &aom_highbd_12_masked_variance32x64_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance32x32_ssse3,
- &aom_highbd_12_masked_variance32x32_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance32x16_ssse3,
- &aom_highbd_12_masked_variance32x16_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance16x32_ssse3,
- &aom_highbd_12_masked_variance16x32_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance16x16_ssse3,
- &aom_highbd_12_masked_variance16x16_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance16x8_ssse3,
- &aom_highbd_12_masked_variance16x8_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance8x16_ssse3,
- &aom_highbd_12_masked_variance8x16_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance8x8_ssse3,
- &aom_highbd_12_masked_variance8x8_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance8x4_ssse3,
- &aom_highbd_12_masked_variance8x4_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance4x8_ssse3,
- &aom_highbd_12_masked_variance4x8_c, AOM_BITS_12),
- make_tuple(&aom_highbd_12_masked_variance4x4_ssse3,
- &aom_highbd_12_masked_variance4x4_c, AOM_BITS_12)));
-
-INSTANTIATE_TEST_CASE_P(
SSSE3_C_COMPARE, HighbdMaskedSubPixelVarianceTest,
::testing::Values(
#if CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_masked_sub_pixel_variance128x128_ssse3,
- &aom_highbd_masked_sub_pixel_variance128x128_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance128x64_ssse3,
- &aom_highbd_masked_sub_pixel_variance128x64_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance64x128_ssse3,
- &aom_highbd_masked_sub_pixel_variance64x128_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance128x128_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance128x128_c,
+ AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance128x64_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance128x64_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance64x128_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance64x128_c, AOM_BITS_8),
#endif // CONFIG_EXT_PARTITION
- make_tuple(&aom_highbd_masked_sub_pixel_variance64x64_ssse3,
- &aom_highbd_masked_sub_pixel_variance64x64_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance64x32_ssse3,
- &aom_highbd_masked_sub_pixel_variance64x32_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance32x64_ssse3,
- &aom_highbd_masked_sub_pixel_variance32x64_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance32x32_ssse3,
- &aom_highbd_masked_sub_pixel_variance32x32_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance32x16_ssse3,
- &aom_highbd_masked_sub_pixel_variance32x16_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance16x32_ssse3,
- &aom_highbd_masked_sub_pixel_variance16x32_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance16x16_ssse3,
- &aom_highbd_masked_sub_pixel_variance16x16_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance16x8_ssse3,
- &aom_highbd_masked_sub_pixel_variance16x8_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance8x16_ssse3,
- &aom_highbd_masked_sub_pixel_variance8x16_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance8x8_ssse3,
- &aom_highbd_masked_sub_pixel_variance8x8_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance8x4_ssse3,
- &aom_highbd_masked_sub_pixel_variance8x4_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance4x8_ssse3,
- &aom_highbd_masked_sub_pixel_variance4x8_c, AOM_BITS_8),
- make_tuple(&aom_highbd_masked_sub_pixel_variance4x4_ssse3,
- &aom_highbd_masked_sub_pixel_variance4x4_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance64x64_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance64x64_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance64x32_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance64x32_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance32x64_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance32x64_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance32x32_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance32x32_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance32x16_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance32x16_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance16x32_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance16x32_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance16x16_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance16x16_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance16x8_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance16x8_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance8x16_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance8x16_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance8x8_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance8x8_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance8x4_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance8x4_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance4x8_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance4x8_c, AOM_BITS_8),
+ make_tuple(&aom_highbd_8_masked_sub_pixel_variance4x4_ssse3,
+ &aom_highbd_8_masked_sub_pixel_variance4x4_c, AOM_BITS_8),
#if CONFIG_EXT_PARTITION
make_tuple(&aom_highbd_10_masked_sub_pixel_variance128x128_ssse3,
&aom_highbd_10_masked_sub_pixel_variance128x128_c,