/* * Copyright (c) 2016, Alliance for Open Media. All rights reserved * * This source code is subject to the terms of the BSD 2 Clause License and * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License * was not distributed with this source code in the LICENSE file, you can * obtain it at www.aomedia.org/license/software. If the Alliance for Open * Media Patent License 1.0 was not distributed with this source code in the * PATENTS file, you can obtain it at www.aomedia.org/license/patent. */ #include #include "./aom_config.h" #include "./aom_dsp_rtcd.h" #include "aom/aom_integer.h" #include "aom_ports/mem.h" #include "aom_dsp/blend.h" /* Sum the difference between every corresponding element of the buffers. */ static INLINE unsigned int sad(const uint8_t *a, int a_stride, const uint8_t *b, int b_stride, int width, int height) { int y, x; unsigned int sad = 0; for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += abs(a[x] - b[x]); a += a_stride; b += b_stride; } return sad; } #define sadMxN(m, n) \ unsigned int aom_sad##m##x##n##_c(const uint8_t *src, int src_stride, \ const uint8_t *ref, int ref_stride) { \ return sad(src, src_stride, ref, ref_stride, m, n); \ } \ unsigned int aom_sad##m##x##n##_avg_c(const uint8_t *src, int src_stride, \ const uint8_t *ref, int ref_stride, \ const uint8_t *second_pred) { \ uint8_t comp_pred[m * n]; \ aom_comp_avg_pred_c(comp_pred, second_pred, m, n, ref, ref_stride); \ return sad(src, src_stride, comp_pred, m, m, n); \ } // depending on call sites, pass **ref_array to avoid & in subsequent call and // de-dup with 4D below. #define sadMxNxK(m, n, k) \ void aom_sad##m##x##n##x##k##_c(const uint8_t *src, int src_stride, \ const uint8_t *ref_array, int ref_stride, \ uint32_t *sad_array) { \ int i; \ for (i = 0; i < k; ++i) \ sad_array[i] = \ aom_sad##m##x##n##_c(src, src_stride, &ref_array[i], ref_stride); \ } // This appears to be equivalent to the above when k == 4 and refs is const #define sadMxNx4D(m, n) \ void aom_sad##m##x##n##x4d_c(const uint8_t *src, int src_stride, \ const uint8_t *const ref_array[], \ int ref_stride, uint32_t *sad_array) { \ int i; \ for (i = 0; i < 4; ++i) \ sad_array[i] = \ aom_sad##m##x##n##_c(src, src_stride, ref_array[i], ref_stride); \ } /* clang-format off */ #if CONFIG_AV1 && CONFIG_EXT_PARTITION // 128x128 sadMxN(128, 128) sadMxNxK(128, 128, 3) sadMxNxK(128, 128, 8) sadMxNx4D(128, 128) // 128x64 sadMxN(128, 64) sadMxNx4D(128, 64) // 64x128 sadMxN(64, 128) sadMxNx4D(64, 128) #endif // CONFIG_AV1 && CONFIG_EXT_PARTITION // 64x64 sadMxN(64, 64) sadMxNxK(64, 64, 3) sadMxNxK(64, 64, 8) sadMxNx4D(64, 64) // 64x32 sadMxN(64, 32) sadMxNx4D(64, 32) // 32x64 sadMxN(32, 64) sadMxNx4D(32, 64) // 32x32 sadMxN(32, 32) sadMxNxK(32, 32, 3) sadMxNxK(32, 32, 8) sadMxNx4D(32, 32) // 32x16 sadMxN(32, 16) sadMxNx4D(32, 16) // 16x32 sadMxN(16, 32) sadMxNx4D(16, 32) // 16x16 sadMxN(16, 16) sadMxNxK(16, 16, 3) sadMxNxK(16, 16, 8) sadMxNx4D(16, 16) // 16x8 sadMxN(16, 8) sadMxNxK(16, 8, 3) sadMxNxK(16, 8, 8) sadMxNx4D(16, 8) // 8x16 sadMxN(8, 16) sadMxNxK(8, 16, 3) sadMxNxK(8, 16, 8) sadMxNx4D(8, 16) // 8x8 sadMxN(8, 8) sadMxNxK(8, 8, 3) sadMxNxK(8, 8, 8) sadMxNx4D(8, 8) // 8x4 sadMxN(8, 4) sadMxNxK(8, 4, 8) sadMxNx4D(8, 4) // 4x8 sadMxN(4, 8) sadMxNxK(4, 8, 8) sadMxNx4D(4, 8) // 4x4 sadMxN(4, 4) sadMxNxK(4, 4, 3) sadMxNxK(4, 4, 8) sadMxNx4D(4, 4) /* clang-format on */ #if CONFIG_HIGHBITDEPTH static INLINE unsigned int highbd_sad(const uint8_t *a8, int a_stride, const uint8_t *b8, int b_stride, int width, int height) { int y, x; unsigned int sad = 0; const uint16_t *a = CONVERT_TO_SHORTPTR(a8); const uint16_t *b = CONVERT_TO_SHORTPTR(b8); for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += abs(a[x] - b[x]); a += a_stride; b += b_stride; } return sad; } static INLINE unsigned int highbd_sadb(const uint8_t *a8, int a_stride, const uint16_t *b, int b_stride, int width, int height) { int y, x; unsigned int sad = 0; const uint16_t *a = CONVERT_TO_SHORTPTR(a8); for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += abs(a[x] - b[x]); a += a_stride; b += b_stride; } return sad; } #define highbd_sadMxN(m, n) \ unsigned int aom_highbd_sad##m##x##n##_c(const uint8_t *src, int src_stride, \ const uint8_t *ref, \ int ref_stride) { \ return highbd_sad(src, src_stride, ref, ref_stride, m, n); \ } \ unsigned int aom_highbd_sad##m##x##n##_avg_c( \ const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, \ const uint8_t *second_pred) { \ uint16_t comp_pred[m * n]; \ aom_highbd_comp_avg_pred_c(comp_pred, second_pred, m, n, ref, ref_stride); \ return highbd_sadb(src, src_stride, comp_pred, m, m, n); \ } #define highbd_sadMxNxK(m, n, k) \ void aom_highbd_sad##m##x##n##x##k##_c( \ const uint8_t *src, int src_stride, const uint8_t *ref_array, \ int ref_stride, uint32_t *sad_array) { \ int i; \ for (i = 0; i < k; ++i) { \ sad_array[i] = aom_highbd_sad##m##x##n##_c(src, src_stride, \ &ref_array[i], ref_stride); \ } \ } #define highbd_sadMxNx4D(m, n) \ void aom_highbd_sad##m##x##n##x4d_c(const uint8_t *src, int src_stride, \ const uint8_t *const ref_array[], \ int ref_stride, uint32_t *sad_array) { \ int i; \ for (i = 0; i < 4; ++i) { \ sad_array[i] = aom_highbd_sad##m##x##n##_c(src, src_stride, \ ref_array[i], ref_stride); \ } \ } /* clang-format off */ #if CONFIG_AV1 && CONFIG_EXT_PARTITION // 128x128 highbd_sadMxN(128, 128) highbd_sadMxNxK(128, 128, 3) highbd_sadMxNxK(128, 128, 8) highbd_sadMxNx4D(128, 128) // 128x64 highbd_sadMxN(128, 64) highbd_sadMxNx4D(128, 64) // 64x128 highbd_sadMxN(64, 128) highbd_sadMxNx4D(64, 128) #endif // CONFIG_AV1 && CONFIG_EXT_PARTITION // 64x64 highbd_sadMxN(64, 64) highbd_sadMxNxK(64, 64, 3) highbd_sadMxNxK(64, 64, 8) highbd_sadMxNx4D(64, 64) // 64x32 highbd_sadMxN(64, 32) highbd_sadMxNx4D(64, 32) // 32x64 highbd_sadMxN(32, 64) highbd_sadMxNx4D(32, 64) // 32x32 highbd_sadMxN(32, 32) highbd_sadMxNxK(32, 32, 3) highbd_sadMxNxK(32, 32, 8) highbd_sadMxNx4D(32, 32) // 32x16 highbd_sadMxN(32, 16) highbd_sadMxNx4D(32, 16) // 16x32 highbd_sadMxN(16, 32) highbd_sadMxNx4D(16, 32) // 16x16 highbd_sadMxN(16, 16) highbd_sadMxNxK(16, 16, 3) highbd_sadMxNxK(16, 16, 8) highbd_sadMxNx4D(16, 16) // 16x8 highbd_sadMxN(16, 8) highbd_sadMxNxK(16, 8, 3) highbd_sadMxNxK(16, 8, 8) highbd_sadMxNx4D(16, 8) // 8x16 highbd_sadMxN(8, 16) highbd_sadMxNxK(8, 16, 3) highbd_sadMxNxK(8, 16, 8) highbd_sadMxNx4D(8, 16) // 8x8 highbd_sadMxN(8, 8) highbd_sadMxNxK(8, 8, 3) highbd_sadMxNxK(8, 8, 8) highbd_sadMxNx4D(8, 8) // 8x4 highbd_sadMxN(8, 4) highbd_sadMxNxK(8, 4, 8) highbd_sadMxNx4D(8, 4) // 4x8 highbd_sadMxN(4, 8) highbd_sadMxNxK(4, 8, 8) highbd_sadMxNx4D(4, 8) // 4x4 highbd_sadMxN(4, 4) highbd_sadMxNxK(4, 4, 3) highbd_sadMxNxK(4, 4, 8) highbd_sadMxNx4D(4, 4) /* clang-format on */ #endif // CONFIG_HIGHBITDEPTH #if CONFIG_AV1 && CONFIG_EXT_INTER static INLINE unsigned int masked_sad(const uint8_t *src, int src_stride, const uint8_t *a, int a_stride, const uint8_t *b, int b_stride, const uint8_t *m, int m_stride, int width, int height) { int y, x; unsigned int sad = 0; for (y = 0; y < height; y++) { for (x = 0; x < width; x++) { const uint8_t pred = AOM_BLEND_A64(m[x], a[x], b[x]); sad += abs(pred - src[x]); } src += src_stride; a += a_stride; b += b_stride; m += m_stride; } sad = (sad + 31) >> 6; return sad; } #define MASKSADMxN(m, n) \ unsigned int aom_masked_sad##m##x##n##_c( \ const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, \ const uint8_t *second_pred, const uint8_t *msk, int msk_stride, \ int invert_mask) { \ if (!invert_mask) \ return masked_sad(src, src_stride, ref, ref_stride, second_pred, m, msk, \ msk_stride, m, n); \ else \ return masked_sad(src, src_stride, second_pred, m, ref, ref_stride, msk, \ msk_stride, m, n); \ } /* clang-format off */ #if CONFIG_EXT_PARTITION MASKSADMxN(128, 128) MASKSADMxN(128, 64) MASKSADMxN(64, 128) #endif // CONFIG_EXT_PARTITION MASKSADMxN(64, 64) MASKSADMxN(64, 32) MASKSADMxN(32, 64) MASKSADMxN(32, 32) MASKSADMxN(32, 16) MASKSADMxN(16, 32) MASKSADMxN(16, 16) MASKSADMxN(16, 8) MASKSADMxN(8, 16) MASKSADMxN(8, 8) MASKSADMxN(8, 4) MASKSADMxN(4, 8) MASKSADMxN(4, 4) /* clang-format on */ #if CONFIG_HIGHBITDEPTH static INLINE unsigned int highbd_masked_sad(const uint8_t *src8, int src_stride, const uint8_t *a8, int a_stride, const uint8_t *b8, int b_stride, const uint8_t *m, int m_stride, int width, int height) { int y, x; unsigned int sad = 0; const uint16_t *src = CONVERT_TO_SHORTPTR(src8); const uint16_t *a = CONVERT_TO_SHORTPTR(a8); const uint16_t *b = CONVERT_TO_SHORTPTR(b8); for (y = 0; y < height; y++) { for (x = 0; x < width; x++) { const uint16_t pred = AOM_BLEND_A64(m[x], a[x], b[x]); sad += abs(pred - src[x]); } src += src_stride; a += a_stride; b += b_stride; m += m_stride; } sad = (sad + 31) >> 6; return sad; } #define HIGHBD_MASKSADMXN(m, n) \ unsigned int aom_highbd_masked_sad##m##x##n##_c( \ const uint8_t *src8, int src_stride, const uint8_t *ref8, \ int ref_stride, const uint8_t *second_pred8, const uint8_t *msk, \ int msk_stride, int invert_mask) { \ if (!invert_mask) \ return highbd_masked_sad(src8, src_stride, ref8, ref_stride, \ second_pred8, m, msk, msk_stride, m, n); \ else \ return highbd_masked_sad(src8, src_stride, second_pred8, m, ref8, \ ref_stride, msk, msk_stride, m, n); \ } #if CONFIG_EXT_PARTITION HIGHBD_MASKSADMXN(128, 128) HIGHBD_MASKSADMXN(128, 64) HIGHBD_MASKSADMXN(64, 128) #endif // CONFIG_EXT_PARTITION HIGHBD_MASKSADMXN(64, 64) HIGHBD_MASKSADMXN(64, 32) HIGHBD_MASKSADMXN(32, 64) HIGHBD_MASKSADMXN(32, 32) HIGHBD_MASKSADMXN(32, 16) HIGHBD_MASKSADMXN(16, 32) HIGHBD_MASKSADMXN(16, 16) HIGHBD_MASKSADMXN(16, 8) HIGHBD_MASKSADMXN(8, 16) HIGHBD_MASKSADMXN(8, 8) HIGHBD_MASKSADMXN(8, 4) HIGHBD_MASKSADMXN(4, 8) HIGHBD_MASKSADMXN(4, 4) #endif // CONFIG_HIGHBITDEPTH #endif // CONFIG_AV1 && CONFIG_EXT_INTER #if CONFIG_AV1 && CONFIG_MOTION_VAR // pre: predictor being evaluated // wsrc: target weighted prediction (has been *4096 to keep precision) // mask: 2d weights (scaled by 4096) static INLINE unsigned int obmc_sad(const uint8_t *pre, int pre_stride, const int32_t *wsrc, const int32_t *mask, int width, int height) { int y, x; unsigned int sad = 0; for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += ROUND_POWER_OF_TWO(abs(wsrc[x] - pre[x] * mask[x]), 12); pre += pre_stride; wsrc += width; mask += width; } return sad; } #define OBMCSADMxN(m, n) \ unsigned int aom_obmc_sad##m##x##n##_c(const uint8_t *ref, int ref_stride, \ const int32_t *wsrc, \ const int32_t *mask) { \ return obmc_sad(ref, ref_stride, wsrc, mask, m, n); \ } /* clang-format off */ #if CONFIG_EXT_PARTITION OBMCSADMxN(128, 128) OBMCSADMxN(128, 64) OBMCSADMxN(64, 128) #endif // CONFIG_EXT_PARTITION OBMCSADMxN(64, 64) OBMCSADMxN(64, 32) OBMCSADMxN(32, 64) OBMCSADMxN(32, 32) OBMCSADMxN(32, 16) OBMCSADMxN(16, 32) OBMCSADMxN(16, 16) OBMCSADMxN(16, 8) OBMCSADMxN(8, 16) OBMCSADMxN(8, 8) OBMCSADMxN(8, 4) OBMCSADMxN(4, 8) OBMCSADMxN(4, 4) /* clang-format on */ #if CONFIG_HIGHBITDEPTH static INLINE unsigned int highbd_obmc_sad(const uint8_t *pre8, int pre_stride, const int32_t *wsrc, const int32_t *mask, int width, int height) { int y, x; unsigned int sad = 0; const uint16_t *pre = CONVERT_TO_SHORTPTR(pre8); for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += ROUND_POWER_OF_TWO(abs(wsrc[x] - pre[x] * mask[x]), 12); pre += pre_stride; wsrc += width; mask += width; } return sad; } #define HIGHBD_OBMCSADMXN(m, n) \ unsigned int aom_highbd_obmc_sad##m##x##n##_c( \ const uint8_t *ref, int ref_stride, const int32_t *wsrc, \ const int32_t *mask) { \ return highbd_obmc_sad(ref, ref_stride, wsrc, mask, m, n); \ } /* clang-format off */ #if CONFIG_EXT_PARTITION HIGHBD_OBMCSADMXN(128, 128) HIGHBD_OBMCSADMXN(128, 64) HIGHBD_OBMCSADMXN(64, 128) #endif // CONFIG_EXT_PARTITION HIGHBD_OBMCSADMXN(64, 64) HIGHBD_OBMCSADMXN(64, 32) HIGHBD_OBMCSADMXN(32, 64) HIGHBD_OBMCSADMXN(32, 32) HIGHBD_OBMCSADMXN(32, 16) HIGHBD_OBMCSADMXN(16, 32) HIGHBD_OBMCSADMXN(16, 16) HIGHBD_OBMCSADMXN(16, 8) HIGHBD_OBMCSADMXN(8, 16) HIGHBD_OBMCSADMXN(8, 8) HIGHBD_OBMCSADMXN(8, 4) HIGHBD_OBMCSADMXN(4, 8) HIGHBD_OBMCSADMXN(4, 4) /* clang-format on */ #endif // CONFIG_HIGHBITDEPTH #endif // CONFIG_AV1 && CONFIG_MOTION_VAR