/* * Copyright (c) 2016, Alliance for Open Media. All rights reserved * * This source code is subject to the terms of the BSD 2 Clause License and * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License * was not distributed with this source code in the LICENSE file, you can * obtain it at www.aomedia.org/license/software. If the Alliance for Open * Media Patent License 1.0 was not distributed with this source code in the * PATENTS file, you can obtain it at www.aomedia.org/license/patent. */ #include #include "./aom_dsp_rtcd.h" typedef void (*get_var_avx2)(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse, int *sum); void aom_get32x32var_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse, int *sum); static void variance_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, int w, int h, unsigned int *sse, int *sum, get_var_avx2 var_fn, int block_size) { int i, j; *sse = 0; *sum = 0; for (i = 0; i < h; i += 16) { for (j = 0; j < w; j += block_size) { unsigned int sse0; int sum0; var_fn(&src[src_stride * i + j], src_stride, &ref[ref_stride * i + j], ref_stride, &sse0, &sum0); *sse += sse0; *sum += sum0; } } } unsigned int aom_variance16x16_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse) { int sum; unsigned int variance; variance_avx2(src, src_stride, ref, ref_stride, 16, 16, sse, &sum, aom_get16x16var_avx2, 16); variance = *sse - (((uint32_t)((int64_t)sum * sum)) >> 8); _mm256_zeroupper(); return variance; } unsigned int aom_mse16x16_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse) { int sum; aom_get16x16var_avx2(src, src_stride, ref, ref_stride, sse, &sum); _mm256_zeroupper(); return *sse; } unsigned int aom_variance32x16_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse) { int sum; unsigned int variance; variance_avx2(src, src_stride, ref, ref_stride, 32, 16, sse, &sum, aom_get32x32var_avx2, 32); variance = *sse - (uint32_t)(((int64_t)sum * sum) >> 9); _mm256_zeroupper(); return variance; } unsigned int aom_variance32x32_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse) { int sum; unsigned int variance; variance_avx2(src, src_stride, ref, ref_stride, 32, 32, sse, &sum, aom_get32x32var_avx2, 32); variance = *sse - (uint32_t)(((int64_t)sum * sum) >> 10); _mm256_zeroupper(); return variance; } unsigned int aom_variance64x64_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse) { int sum; unsigned int variance; variance_avx2(src, src_stride, ref, ref_stride, 64, 64, sse, &sum, aom_get32x32var_avx2, 32); variance = *sse - (uint32_t)(((int64_t)sum * sum) >> 12); _mm256_zeroupper(); return variance; } unsigned int aom_variance64x32_avx2(const uint8_t *src, int src_stride, const uint8_t *ref, int ref_stride, unsigned int *sse) { int sum; unsigned int variance; variance_avx2(src, src_stride, ref, ref_stride, 64, 32, sse, &sum, aom_get32x32var_avx2, 32); variance = *sse - (uint32_t)(((int64_t)sum * sum) >> 11); _mm256_zeroupper(); return variance; } unsigned int aom_sub_pixel_variance32xh_avx2(const uint8_t *src, int src_stride, int x_offset, int y_offset, const uint8_t *dst, int dst_stride, int height, unsigned int *sse); unsigned int aom_sub_pixel_avg_variance32xh_avx2( const uint8_t *src, int src_stride, int x_offset, int y_offset, const uint8_t *dst, int dst_stride, const uint8_t *sec, int sec_stride, int height, unsigned int *sseptr); unsigned int aom_sub_pixel_variance64x64_avx2(const uint8_t *src, int src_stride, int x_offset, int y_offset, const uint8_t *dst, int dst_stride, unsigned int *sse) { unsigned int sse1; const int se1 = aom_sub_pixel_variance32xh_avx2( src, src_stride, x_offset, y_offset, dst, dst_stride, 64, &sse1); unsigned int sse2; const int se2 = aom_sub_pixel_variance32xh_avx2(src + 32, src_stride, x_offset, y_offset, dst + 32, dst_stride, 64, &sse2); const int se = se1 + se2; unsigned int variance; *sse = sse1 + sse2; variance = *sse - (uint32_t)(((int64_t)se * se) >> 12); _mm256_zeroupper(); return variance; } unsigned int aom_sub_pixel_variance32x32_avx2(const uint8_t *src, int src_stride, int x_offset, int y_offset, const uint8_t *dst, int dst_stride, unsigned int *sse) { const int se = aom_sub_pixel_variance32xh_avx2( src, src_stride, x_offset, y_offset, dst, dst_stride, 32, sse); const unsigned int variance = *sse - (uint32_t)(((int64_t)se * se) >> 10); _mm256_zeroupper(); return variance; } unsigned int aom_sub_pixel_avg_variance64x64_avx2( const uint8_t *src, int src_stride, int x_offset, int y_offset, const uint8_t *dst, int dst_stride, unsigned int *sse, const uint8_t *sec) { unsigned int sse1; const int se1 = aom_sub_pixel_avg_variance32xh_avx2( src, src_stride, x_offset, y_offset, dst, dst_stride, sec, 64, 64, &sse1); unsigned int sse2; const int se2 = aom_sub_pixel_avg_variance32xh_avx2( src + 32, src_stride, x_offset, y_offset, dst + 32, dst_stride, sec + 32, 64, 64, &sse2); const int se = se1 + se2; unsigned int variance; *sse = sse1 + sse2; variance = *sse - (uint32_t)(((int64_t)se * se) >> 12); _mm256_zeroupper(); return variance; } unsigned int aom_sub_pixel_avg_variance32x32_avx2( const uint8_t *src, int src_stride, int x_offset, int y_offset, const uint8_t *dst, int dst_stride, unsigned int *sse, const uint8_t *sec) { // Process 32 elements in parallel. const int se = aom_sub_pixel_avg_variance32xh_avx2( src, src_stride, x_offset, y_offset, dst, dst_stride, sec, 32, 32, sse); const unsigned int variance = *sse - (uint32_t)(((int64_t)se * se) >> 10); _mm256_zeroupper(); return variance; }