From 5f8de423f190bbb79a62f804151bc24824fa32d8 Mon Sep 17 00:00:00 2001 From: "Matt A. Tobin" Date: Fri, 2 Feb 2018 04:16:08 -0500 Subject: Add m-esr52 at 52.6.0 --- gfx/2d/convolverLS3.cpp | 927 ++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 927 insertions(+) create mode 100644 gfx/2d/convolverLS3.cpp (limited to 'gfx/2d/convolverLS3.cpp') diff --git a/gfx/2d/convolverLS3.cpp b/gfx/2d/convolverLS3.cpp new file mode 100644 index 000000000..a1a41c98b --- /dev/null +++ b/gfx/2d/convolverLS3.cpp @@ -0,0 +1,927 @@ +// Copyright (c) 2014-2015 The Chromium Authors. All rights reserved. +// +// Redistribution and use in source and binary forms, with or without +// modification, are permitted provided that the following conditions +// are met: +// * Redistributions of source code must retain the above copyright +// notice, this list of conditions and the following disclaimer. +// * Redistributions in binary form must reproduce the above copyright +// notice, this list of conditions and the following disclaimer in +// the documentation and/or other materials provided with the +// distribution. +// * Neither the name of Google, Inc. nor the names of its contributors +// may be used to endorse or promote products derived from this +// software without specific prior written permission. +// +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS +// FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE +// COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, +// INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, +// BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS +// OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED +// AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, +// OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT +// OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF +// SUCH DAMAGE. + +#include "convolver.h" +#include +#include "skia/include/core/SkTypes.h" + +#if defined(_MIPS_ARCH_LOONGSON3A) + +#include "MMIHelpers.h" + +namespace skia { + +// Convolves horizontally along a single row. The row data is given in +// |src_data| and continues for the num_values() of the filter. +void ConvolveHorizontally_LS3(const unsigned char* src_data, + const ConvolutionFilter1D& filter, + unsigned char* out_row) { + int num_values = filter.num_values(); + int tmp, filter_offset, filter_length; + double zero, mask[4], shuf_50, shuf_fa; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "xor %[zero], %[zero], %[zero] \n\t" + // |mask| will be used to decimate all extra filter coefficients that are + // loaded by SIMD when |filter_length| is not divisible by 4. + // mask[0] is not used in following algorithm. + "li %[tmp], 1 \n\t" + "dsll32 %[tmp], 0x10 \n\t" + "daddiu %[tmp], -1 \n\t" + "dmtc1 %[tmp], %[mask3] \n\t" + "dsrl %[tmp], 0x10 \n\t" + "mtc1 %[tmp], %[mask2] \n\t" + "dsrl %[tmp], 0x10 \n\t" + "mtc1 %[tmp], %[mask1] \n\t" + "ori %[tmp], $0, 0x50 \n\t" + "mtc1 %[tmp], %[shuf_50] \n\t" + "ori %[tmp], $0, 0xfa \n\t" + "mtc1 %[tmp], %[shuf_fa] \n\t" + ".set pop \n\t" + :[zero]"=f"(zero), [mask1]"=f"(mask[1]), + [mask2]"=f"(mask[2]), [mask3]"=f"(mask[3]), + [shuf_50]"=f"(shuf_50), [shuf_fa]"=f"(shuf_fa), + [tmp]"=&r"(tmp) + ); + + // Output one pixel each iteration, calculating all channels (RGBA) together. + for (int out_x = 0; out_x < num_values; out_x++) { + const ConvolutionFilter1D::Fixed* filter_values = + filter.FilterForValue(out_x, &filter_offset, &filter_length); + double accumh, accuml; + // Compute the first pixel in this row that the filter affects. It will + // touch |filter_length| pixels (4 bytes each) after this. + const void *row_to_filter = + reinterpret_cast(&src_data[filter_offset << 2]); + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + _mm_xor(accum, accum, accum) + ".set pop \n\t" + :[accumh]"=f"(accumh), [accuml]"=f"(accuml) + ); + + // We will load and accumulate with four coefficients per iteration. + for (int filter_x = 0; filter_x < filter_length >> 2; filter_x++) { + double src16h, src16l, mul_hih, mul_hil, mul_loh, mul_lol; + double coeffh, coeffl, src8h, src8l, th, tl, coeff16h, coeff16l; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + // Load 4 coefficients => duplicate 1st and 2nd of them for all channels. + // [16] xx xx xx xx c3 c2 c1 c0 + "gsldlc1 %[coeffl], 7(%[fval]) \n\t" + "gsldrc1 %[coeffl], (%[fval]) \n\t" + "xor %[coeffh], %[coeffh], %[coeffh] \n\t" + // [16] xx xx xx xx c1 c1 c0 c0 + _mm_pshuflh(coeff16, coeff, shuf_50) + // [16] c1 c1 c1 c1 c0 c0 c0 c0 + _mm_punpcklhw(coeff16, coeff16, coeff16) + // Load four pixels => unpack the first two pixels to 16 bits => + // multiply with coefficients => accumulate the convolution result. + // [8] a3 b3 g3 r3 a2 b2 g2 r2 a1 b1 g1 r1 a0 b0 g0 r0 + "gsldlc1 %[src8h], 0xf(%[rtf]) \n\t" + "gsldrc1 %[src8h], 0x8(%[rtf]) \n\t" + "gsldlc1 %[src8l], 0x7(%[rtf]) \n\t" + "gsldrc1 %[src8l], 0x0(%[rtf]) \n\t" + // [16] a1 b1 g1 r1 a0 b0 g0 r0 + _mm_punpcklbh(src16, src8, zero) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + // [32] a0*c0 b0*c0 g0*c0 r0*c0 + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum, accum, t) + // [32] a1*c1 b1*c1 g1*c1 r1*c1 + _mm_punpckhhw(t, mul_lo, mul_hi) + _mm_paddw(accum, accum, t) + // Duplicate 3rd and 4th coefficients for all channels => + // unpack the 3rd and 4th pixels to 16 bits => multiply with coefficients + // => accumulate the convolution results. + // [16] xx xx xx xx c3 c3 c2 c2 + _mm_pshuflh(coeff16, coeff, shuf_fa) + // [16] c3 c3 c3 c3 c2 c2 c2 c2 + _mm_punpcklhw(coeff16, coeff16, coeff16) + // [16] a3 g3 b3 r3 a2 g2 b2 r2 + _mm_punpckhbh(src16, src8, zero) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + // [32] a2*c2 b2*c2 g2*c2 r2*c2 + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum, accum, t) + // [32] a3*c3 b3*c3 g3*c3 r3*c3 + _mm_punpckhhw(t, mul_lo, mul_hi) + _mm_paddw(accum, accum, t) + ".set pop \n\t" + :[th]"=&f"(th), [tl]"=&f"(tl), + [src8h]"=&f"(src8h), [src8l]"=&f"(src8l), + [accumh]"+f"(accumh), [accuml]"+f"(accuml), + [src16h]"=&f"(src16h), [src16l]"=&f"(src16l), + [coeffh]"=&f"(coeffh), [coeffl]"=&f"(coeffl), + [coeff16h]"=&f"(coeff16h), [coeff16l]"=&f"(coeff16l), + [mul_hih]"=&f"(mul_hih), [mul_hil]"=&f"(mul_hil), + [mul_loh]"=&f"(mul_loh), [mul_lol]"=&f"(mul_lol) + :[zeroh]"f"(zero), [zerol]"f"(zero), + [shuf_50]"f"(shuf_50), [shuf_fa]"f"(shuf_fa), + [fval]"r"(filter_values), [rtf]"r"(row_to_filter) + ); + + // Advance the pixel and coefficients pointers. + row_to_filter += 16; + filter_values += 4; + } + + // When |filter_length| is not divisible by 4, we need to decimate some of + // the filter coefficient that was loaded incorrectly to zero; Other than + // that the algorithm is same with above, except that the 4th pixel will be + // always absent. + int r = filter_length & 3; + if (r) { + double coeffh, coeffl, th, tl, coeff16h, coeff16l; + double src8h, src8l, src16h, src16l, mul_hih, mul_hil, mul_loh, mul_lol; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "gsldlc1 %[coeffl], 7(%[fval]) \n\t" + "gsldrc1 %[coeffl], (%[fval]) \n\t" + "xor %[coeffh], %[coeffh], %[coeffh] \n\t" + // Mask out extra filter taps. + "and %[coeffl], %[coeffl], %[mask] \n\t" + _mm_pshuflh(coeff16, coeff, shuf_50) + _mm_punpcklhw(coeff16, coeff16, coeff16) + "gsldlc1 %[src8h], 0xf(%[rtf]) \n\t" + "gsldrc1 %[src8h], 0x8(%[rtf]) \n\t" + "gsldlc1 %[src8l], 0x7(%[rtf]) \n\t" + "gsldrc1 %[src8l], 0x0(%[rtf]) \n\t" + _mm_punpcklbh(src16, src8, zero) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum, accum, t) + _mm_punpckhhw(t, mul_lo, mul_hi) + _mm_paddw(accum, accum, t) + _mm_punpckhbh(src16, src8, zero) + _mm_pshuflh(coeff16, coeff, shuf_fa) + _mm_punpcklhw(coeff16, coeff16, coeff16) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum, accum, t) + ".set pop \n\t" + :[th]"=&f"(th), [tl]"=&f"(tl), + [src8h]"=&f"(src8h), [src8l]"=&f"(src8l), + [accumh]"+f"(accumh), [accuml]"+f"(accuml), + [src16h]"=&f"(src16h), [src16l]"=&f"(src16l), + [coeffh]"=&f"(coeffh), [coeffl]"=&f"(coeffl), + [coeff16h]"=&f"(coeff16h), [coeff16l]"=&f"(coeff16l), + [mul_hih]"=&f"(mul_hih), [mul_hil]"=&f"(mul_hil), + [mul_loh]"=&f"(mul_loh), [mul_lol]"=&f"(mul_lol) + :[fval]"r"(filter_values), [rtf]"r"(row_to_filter), + [zeroh]"f"(zero), [zerol]"f"(zero), [mask]"f"(mask[r]), + [shuf_50]"f"(shuf_50), [shuf_fa]"f"(shuf_fa) + ); + } + + double t, sra; + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "ori %[tmp], $0, %[sk_sra] \n\t" + "mtc1 %[tmp], %[sra] \n\t" + // Shift right for fixed point implementation. + _mm_psraw(accum, accum, sra) + // Packing 32 bits |accum| to 16 bits per channel (signed saturation). + _mm_packsswh(accum, accum, zero, t) + // Packing 16 bits |accum| to 8 bits per channel (unsigned saturation). + _mm_packushb(accum, accum, zero, t) + // Store the pixel value of 32 bits. + "swc1 %[accuml], (%[out_row]) \n\t" + ".set pop \n\t" + :[sra]"=&f"(sra), [t]"=&f"(t), [tmp]"=&r"(tmp), + [accumh]"+f"(accumh), [accuml]"+f"(accuml) + :[sk_sra]"i"(ConvolutionFilter1D::kShiftBits), + [out_row]"r"(out_row), [zeroh]"f"(zero), [zerol]"f"(zero) + :"memory" + ); + + out_row += 4; + } +} + +// Convolves horizontally along a single row. The row data is given in +// |src_data| and continues for the [begin, end) of the filter. +// Process one pixel at a time. +void ConvolveHorizontally1_LS3(const unsigned char* src_data, + const ConvolutionFilter1D& filter, + unsigned char* out_row) { + int num_values = filter.num_values(); + double zero; + double sra; + + asm volatile ( + ".set push \n" + ".set arch=loongson3a \n" + "xor %[zero], %[zero], %[zero] \n" + "mtc1 %[sk_sra], %[sra] \n" + ".set pop \n" + :[zero]"=&f"(zero), [sra]"=&f"(sra) + :[sk_sra]"r"(ConvolutionFilter1D::kShiftBits) + ); + // Loop over each pixel on this row in the output image. + for (int out_x = 0; out_x < num_values; out_x++) { + // Get the filter that determines the current output pixel. + int filter_offset; + int filter_length; + const ConvolutionFilter1D::Fixed* filter_values = + filter.FilterForValue(out_x, &filter_offset, &filter_length); + + // Compute the first pixel in this row that the filter affects. It will + // touch |filter_length| pixels (4 bytes each) after this. + const unsigned char* row_to_filter = &src_data[filter_offset * 4]; + + // Apply the filter to the row to get the destination pixel in |accum|. + double accuml; + double accumh; + asm volatile ( + ".set push \n" + ".set arch=loongson3a \n" + "xor %[accuml], %[accuml], %[accuml] \n" + "xor %[accumh], %[accumh], %[accumh] \n" + ".set pop \n" + :[accuml]"=&f"(accuml), [accumh]"=&f"(accumh) + ); + for (int filter_x = 0; filter_x < filter_length; filter_x++) { + double src8; + double src16; + double coeff; + double coeff16; + asm volatile ( + ".set push \n" + ".set arch=loongson3a \n" + "lwc1 %[src8], %[rtf] \n" + "mtc1 %[fv], %[coeff] \n" + "pshufh %[coeff16], %[coeff], %[zero] \n" + "punpcklbh %[src16], %[src8], %[zero] \n" + "pmullh %[src8], %[src16], %[coeff16] \n" + "pmulhh %[coeff], %[src16], %[coeff16] \n" + "punpcklhw %[src16], %[src8], %[coeff] \n" + "punpckhhw %[coeff16], %[src8], %[coeff] \n" + "paddw %[accuml], %[accuml], %[src16] \n" + "paddw %[accumh], %[accumh], %[coeff16] \n" + ".set pop \n" + :[accuml]"+f"(accuml), [accumh]"+f"(accumh), + [src8]"=&f"(src8), [src16]"=&f"(src16), + [coeff]"=&f"(coeff), [coeff16]"=&f"(coeff16) + :[rtf]"m"(row_to_filter[filter_x * 4]), + [fv]"r"(filter_values[filter_x]), [zero]"f"(zero) + ); + } + + asm volatile ( + ".set push \n" + ".set arch=loongson3a \n" + // Bring this value back in range. All of the filter scaling factors + // are in fixed point with kShiftBits bits of fractional part. + "psraw %[accuml], %[accuml], %[sra] \n" + "psraw %[accumh], %[accumh], %[sra] \n" + // Store the new pixel. + "packsswh %[accuml], %[accuml], %[accumh] \n" + "packushb %[accuml], %[accuml], %[zero] \n" + "swc1 %[accuml], %[out_row] \n" + ".set pop \n" + :[accuml]"+f"(accuml), [accumh]"+f"(accumh) + :[sra]"f"(sra), [zero]"f"(zero), [out_row]"m"(out_row[out_x * 4]) + :"memory" + ); + } +} + +// Convolves horizontally along four rows. The row data is given in +// |src_data| and continues for the num_values() of the filter. +// The algorithm is almost same as |ConvolveHorizontally_LS3|. Please +// refer to that function for detailed comments. +void ConvolveHorizontally4_LS3(const unsigned char* src_data[4], + const ConvolutionFilter1D& filter, + unsigned char* out_row[4]) { + int num_values = filter.num_values(); + int tmp, filter_offset, filter_length; + double zero, mask[4], shuf_50, shuf_fa; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "xor %[zero], %[zero], %[zero] \n\t" + // |mask| will be used to decimate all extra filter coefficients that are + // loaded by SIMD when |filter_length| is not divisible by 4. + // mask[0] is not used in following algorithm. + "li %[tmp], 1 \n\t" + "dsll32 %[tmp], 0x10 \n\t" + "daddiu %[tmp], -1 \n\t" + "dmtc1 %[tmp], %[mask3] \n\t" + "dsrl %[tmp], 0x10 \n\t" + "mtc1 %[tmp], %[mask2] \n\t" + "dsrl %[tmp], 0x10 \n\t" + "mtc1 %[tmp], %[mask1] \n\t" + "ori %[tmp], $0, 0x50 \n\t" + "mtc1 %[tmp], %[shuf_50] \n\t" + "ori %[tmp], $0, 0xfa \n\t" + "mtc1 %[tmp], %[shuf_fa] \n\t" + ".set pop \n\t" + :[zero]"=f"(zero), [mask1]"=f"(mask[1]), + [mask2]"=f"(mask[2]), [mask3]"=f"(mask[3]), + [shuf_50]"=f"(shuf_50), [shuf_fa]"=f"(shuf_fa), + [tmp]"=&r"(tmp) + ); + + // Output one pixel each iteration, calculating all channels (RGBA) together. + for (int out_x = 0; out_x < num_values; out_x++) { + const ConvolutionFilter1D::Fixed* filter_values = + filter.FilterForValue(out_x, &filter_offset, &filter_length); + double accum0h, accum0l, accum1h, accum1l; + double accum2h, accum2l, accum3h, accum3l; + + // four pixels in a column per iteration. + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + _mm_xor(accum0, accum0, accum0) + _mm_xor(accum1, accum1, accum1) + _mm_xor(accum2, accum2, accum2) + _mm_xor(accum3, accum3, accum3) + ".set pop \n\t" + :[accum0h]"=f"(accum0h), [accum0l]"=f"(accum0l), + [accum1h]"=f"(accum1h), [accum1l]"=f"(accum1l), + [accum2h]"=f"(accum2h), [accum2l]"=f"(accum2l), + [accum3h]"=f"(accum3h), [accum3l]"=f"(accum3l) + ); + + int start = (filter_offset<<2); + // We will load and accumulate with four coefficients per iteration. + for (int filter_x = 0; filter_x < (filter_length >> 2); filter_x++) { + double src8h, src8l, src16h, src16l; + double mul_hih, mul_hil, mul_loh, mul_lol, th, tl; + double coeffh, coeffl, coeff16loh, coeff16lol, coeff16hih, coeff16hil; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + // [16] xx xx xx xx c3 c2 c1 c0 + "gsldlc1 %[coeffl], 7(%[fval]) \n\t" + "gsldrc1 %[coeffl], (%[fval]) \n\t" + "xor %[coeffh], %[coeffh], %[coeffh] \n\t" + // [16] xx xx xx xx c1 c1 c0 c0 + _mm_pshuflh(coeff16lo, coeff, shuf_50) + // [16] c1 c1 c1 c1 c0 c0 c0 c0 + _mm_punpcklhw(coeff16lo, coeff16lo, coeff16lo) + // [16] xx xx xx xx c3 c3 c2 c2 + _mm_pshuflh(coeff16hi, coeff, shuf_fa) + // [16] c3 c3 c3 c3 c2 c2 c2 c2 + _mm_punpcklhw(coeff16hi, coeff16hi, coeff16hi) + ".set pop \n\t" + :[coeffh]"=&f"(coeffh), [coeffl]"=&f"(coeffl), + [coeff16loh]"=&f"(coeff16loh), [coeff16lol]"=&f"(coeff16lol), + [coeff16hih]"=&f"(coeff16hih), [coeff16hil]"=&f"(coeff16hil) + :[fval]"r"(filter_values), [shuf_50]"f"(shuf_50), [shuf_fa]"f"(shuf_fa) + ); + +#define ITERATION(_src, _accumh, _accuml) \ + asm volatile ( \ + ".set push \n\t" \ + ".set arch=loongson3a \n\t" \ + "gsldlc1 %[src8h], 0xf(%[src]) \n\t" \ + "gsldrc1 %[src8h], 0x8(%[src]) \n\t" \ + "gsldlc1 %[src8l], 0x7(%[src]) \n\t" \ + "gsldrc1 %[src8l], 0x0(%[src]) \n\t" \ + _mm_punpcklbh(src16, src8, zero) \ + _mm_pmulhh(mul_hi, src16, coeff16lo) \ + _mm_pmullh(mul_lo, src16, coeff16lo) \ + _mm_punpcklhw(t, mul_lo, mul_hi) \ + _mm_paddw(accum, accum, t) \ + _mm_punpckhhw(t, mul_lo, mul_hi) \ + _mm_paddw(accum, accum, t) \ + _mm_punpckhbh(src16, src8, zero) \ + _mm_pmulhh(mul_hi, src16, coeff16hi) \ + _mm_pmullh(mul_lo, src16, coeff16hi) \ + _mm_punpcklhw(t, mul_lo, mul_hi) \ + _mm_paddw(accum, accum, t) \ + _mm_punpckhhw(t, mul_lo, mul_hi) \ + _mm_paddw(accum, accum, t) \ + ".set pop \n\t" \ + :[th]"=&f"(th), [tl]"=&f"(tl), \ + [src8h]"=&f"(src8h), [src8l]"=&f"(src8l), \ + [src16h]"=&f"(src16h), [src16l]"=&f"(src16l), \ + [mul_hih]"=&f"(mul_hih), [mul_hil]"=&f"(mul_hil), \ + [mul_loh]"=&f"(mul_loh), [mul_lol]"=&f"(mul_lol), \ + [accumh]"+f"(_accumh), [accuml]"+f"(_accuml) \ + :[zeroh]"f"(zero), [zerol]"f"(zero), [src]"r"(_src), \ + [coeff16loh]"f"(coeff16loh), [coeff16lol]"f"(coeff16lol), \ + [coeff16hih]"f"(coeff16hih), [coeff16hil]"f"(coeff16hil) \ + ); + + ITERATION(src_data[0] + start, accum0h, accum0l); + ITERATION(src_data[1] + start, accum1h, accum1l); + ITERATION(src_data[2] + start, accum2h, accum2l); + ITERATION(src_data[3] + start, accum3h, accum3l); + + start += 16; + filter_values += 4; + } + + int r = filter_length & 3; + if (r) { + double src8h, src8l, src16h, src16l; + double mul_hih, mul_hil, mul_loh, mul_lol, th, tl; + double coeffh, coeffl, coeff16loh, coeff16lol, coeff16hih, coeff16hil; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "gsldlc1 %[coeffl], 7(%[fval]) \n\t" + "gsldrc1 %[coeffl], (%[fval]) \n\t" + "xor %[coeffh], %[coeffh], %[coeffh] \n\t" + // Mask out extra filter taps. + "and %[coeffl], %[coeffl], %[mask] \n\t" + _mm_pshuflh(coeff16lo, coeff, shuf_50) + /* c1 c1 c1 c1 c0 c0 c0 c0 */ + _mm_punpcklhw(coeff16lo, coeff16lo, coeff16lo) + _mm_pshuflh(coeff16hi, coeff, shuf_fa) + _mm_punpcklhw(coeff16hi, coeff16hi, coeff16hi) + ".set pop \n\t" + :[coeffh]"=&f"(coeffh), [coeffl]"=&f"(coeffl), + [coeff16loh]"=&f"(coeff16loh), [coeff16lol]"=&f"(coeff16lol), + [coeff16hih]"=&f"(coeff16hih), [coeff16hil]"=&f"(coeff16hil) + :[fval]"r"(filter_values), [mask]"f"(mask[r]), + [shuf_50]"f"(shuf_50), [shuf_fa]"f"(shuf_fa) + ); + + ITERATION(src_data[0] + start, accum0h, accum0l); + ITERATION(src_data[1] + start, accum1h, accum1l); + ITERATION(src_data[2] + start, accum2h, accum2l); + ITERATION(src_data[3] + start, accum3h, accum3l); + } + + double t, sra; + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "ori %[tmp], $0, %[sk_sra] \n\t" + "mtc1 %[tmp], %[sra] \n\t" + _mm_psraw(accum0, accum0, sra) + _mm_packsswh(accum0, accum0, zero, t) + _mm_packushb(accum0, accum0, zero, t) + _mm_psraw(accum1, accum1, sra) + _mm_packsswh(accum1, accum1, zero, t) + _mm_packushb(accum1, accum1, zero, t) + _mm_psraw(accum2, accum2, sra) + _mm_packsswh(accum2, accum2, zero, t) + _mm_packushb(accum2, accum2, zero, t) + _mm_psraw(accum3, accum3, sra) + _mm_packsswh(accum3, accum3, zero, t) + _mm_packushb(accum3, accum3, zero, t) + "swc1 %[accum0l], (%[out_row0]) \n\t" + "swc1 %[accum1l], (%[out_row1]) \n\t" + "swc1 %[accum2l], (%[out_row2]) \n\t" + "swc1 %[accum3l], (%[out_row3]) \n\t" + ".set pop \n\t" + :[accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [accum1h]"+f"(accum1h), [accum1l]"+f"(accum1l), + [accum2h]"+f"(accum2h), [accum2l]"+f"(accum2l), + [accum3h]"+f"(accum3h), [accum3l]"+f"(accum3l), + [sra]"=&f"(sra), [t]"=&f"(t), [tmp]"=&r"(tmp) + :[zeroh]"f"(zero), [zerol]"f"(zero), + [out_row0]"r"(out_row[0]), [out_row1]"r"(out_row[1]), + [out_row2]"r"(out_row[2]), [out_row3]"r"(out_row[3]), + [sk_sra]"i"(ConvolutionFilter1D::kShiftBits) + :"memory" + ); + + out_row[0] += 4; + out_row[1] += 4; + out_row[2] += 4; + out_row[3] += 4; + } +} + +// Does vertical convolution to produce one output row. The filter values and +// length are given in the first two parameters. These are applied to each +// of the rows pointed to in the |source_data_rows| array, with each row +// being |pixel_width| wide. +// +// The output must have room for |pixel_width * 4| bytes. +template +void ConvolveVertically_LS3_impl(const ConvolutionFilter1D::Fixed* filter_values, + int filter_length, + unsigned char* const* source_data_rows, + int pixel_width, + unsigned char* out_row) { + uint64_t tmp; + int width = pixel_width & ~3; + double zero, sra, coeff16h, coeff16l; + double accum0h, accum0l, accum1h, accum1l; + double accum2h, accum2l, accum3h, accum3l; + const void *src; + int out_x; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "xor %[zero], %[zero], %[zero] \n\t" + "ori %[tmp], $0, %[sk_sra] \n\t" + "mtc1 %[tmp], %[sra] \n\t" + ".set pop \n\t" + :[zero]"=f"(zero), [sra]"=f"(sra), [tmp]"=&r"(tmp) + :[sk_sra]"i"(ConvolutionFilter1D::kShiftBits) + ); + + // Output four pixels per iteration (16 bytes). + for (out_x = 0; out_x < width; out_x += 4) { + // Accumulated result for each pixel. 32 bits per RGBA channel. + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + _mm_xor(accum0, accum0, accum0) + _mm_xor(accum1, accum1, accum1) + _mm_xor(accum2, accum2, accum2) + _mm_xor(accum3, accum3, accum3) + ".set pop \n\t" + :[accum0h]"=f"(accum0h), [accum0l]"=f"(accum0l), + [accum1h]"=f"(accum1h), [accum1l]"=f"(accum1l), + [accum2h]"=f"(accum2h), [accum2l]"=f"(accum2l), + [accum3h]"=f"(accum3h), [accum3l]"=f"(accum3l) + ); + + // Convolve with one filter coefficient per iteration. + for (int filter_y = 0; filter_y < filter_length; filter_y++) { + double src8h, src8l, src16h, src16l; + double mul_hih, mul_hil, mul_loh, mul_lol, th, tl; + + src = reinterpret_cast( + &source_data_rows[filter_y][out_x << 2]); + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + // Duplicate the filter coefficient 8 times. + // [16] cj cj cj cj cj cj cj cj + "gsldlc1 %[coeff16l], 7+%[fval] \n\t" + "gsldrc1 %[coeff16l], %[fval] \n\t" + "pshufh %[coeff16l], %[coeff16l], %[zerol] \n\t" + "mov.d %[coeff16h], %[coeff16l] \n\t" + // Load four pixels (16 bytes) together. + // [8] a3 b3 g3 r3 a2 b2 g2 r2 a1 b1 g1 r1 a0 b0 g0 r0 + "gsldlc1 %[src8h], 0xf(%[src]) \n\t" + "gsldrc1 %[src8h], 0x8(%[src]) \n\t" + "gsldlc1 %[src8l], 0x7(%[src]) \n\t" + "gsldrc1 %[src8l], 0x0(%[src]) \n\t" + // Unpack 1st and 2nd pixels from 8 bits to 16 bits for each channels => + // multiply with current coefficient => accumulate the result. + // [16] a1 b1 g1 r1 a0 b0 g0 r0 + _mm_punpcklbh(src16, src8, zero) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + // [32] a0 b0 g0 r0 + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum0, accum0, t) + // [32] a1 b1 g1 r1 + _mm_punpckhhw(t, mul_lo, mul_hi) + _mm_paddw(accum1, accum1, t) + // Unpack 3rd and 4th pixels from 8 bits to 16 bits for each channels => + // multiply with current coefficient => accumulate the result. + // [16] a3 b3 g3 r3 a2 b2 g2 r2 + _mm_punpckhbh(src16, src8, zero) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + ".set pop \n\t" + :[th]"=&f"(th), [tl]"=&f"(tl), + [src8h]"=&f"(src8h), [src8l]"=&f"(src8l), + [src16h]"=&f"(src16h), [src16l]"=&f"(src16l), + [mul_hih]"=&f"(mul_hih), [mul_hil]"=&f"(mul_hil), + [mul_loh]"=&f"(mul_loh), [mul_lol]"=&f"(mul_lol), + [accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [accum1h]"+f"(accum1h), [accum1l]"+f"(accum1l), + [coeff16h]"=&f"(coeff16h), [coeff16l]"=&f"(coeff16l) + :[zeroh]"f"(zero), [zerol]"f"(zero), + [fval]"m"(filter_values[filter_y]), + [src]"r"(src) + ); + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + // [32] a2 b2 g2 r2 + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum2, accum2, t) + // [32] a3 b3 g3 r3 + _mm_punpckhhw(t, mul_lo, mul_hi) + _mm_paddw(accum3, accum3, t) + ".set pop \n\t" + :[th]"=&f"(th), [tl]"=&f"(tl), + [mul_hih]"+f"(mul_hih), [mul_hil]"+f"(mul_hil), + [mul_loh]"+f"(mul_loh), [mul_lol]"+f"(mul_lol), + [accum2h]"+f"(accum2h), [accum2l]"+f"(accum2l), + [accum3h]"+f"(accum3h), [accum3l]"+f"(accum3l) + ); + } + + double t; + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + // Shift right for fixed point implementation. + _mm_psraw(accum0, accum0, sra) + _mm_psraw(accum1, accum1, sra) + _mm_psraw(accum2, accum2, sra) + _mm_psraw(accum3, accum3, sra) + // Packing 32 bits |accum| to 16 bits per channel (signed saturation). + // [16] a1 b1 g1 r1 a0 b0 g0 r0 + _mm_packsswh(accum0, accum0, accum1, t) + // [16] a3 b3 g3 r3 a2 b2 g2 r2 + _mm_packsswh(accum2, accum2, accum3, t) + // Packing 16 bits |accum| to 8 bits per channel (unsigned saturation). + // [8] a3 b3 g3 r3 a2 b2 g2 r2 a1 b1 g1 r1 a0 b0 g0 r0 + _mm_packushb(accum0, accum0, accum2, t) + ".set pop \n\t" + :[accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [accum1h]"+f"(accum1h), [accum1l]"+f"(accum1l), + [accum2h]"+f"(accum2h), [accum2l]"+f"(accum2l), + [accum3h]"+f"(accum3h), [accum3l]"+f"(accum3l), + [t]"=&f"(t) + :[sra]"f"(sra) + ); + + if (has_alpha) { + double ah, al, bh, bl, srl8, srl16, sll24; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "li %[tmp], 8 \n\t" + "mtc1 %[tmp], %[srl8] \n\t" + "li %[tmp], 16 \n\t" + "mtc1 %[tmp], %[srl16] \n\t" + "li %[tmp], 24 \n\t" + "mtc1 %[tmp], %[sll24] \n\t" + // Compute the max(ri, gi, bi) for each pixel. + // [8] xx a3 b3 g3 xx a2 b2 g2 xx a1 b1 g1 xx a0 b0 g0 + _mm_psraw(a, accum0, srl8) + // [8] xx xx xx max3 xx xx xx max2 xx xx xx max1 xx xx xx max0 + _mm_pmaxub(b, a, accum0) // Max of r and g. + // [8] xx xx a3 b3 xx xx a2 b2 xx xx a1 b1 xx xx a0 b0 + _mm_psrlw(a, accum0, srl16) + // [8] xx xx xx max3 xx xx xx max2 xx xx xx max1 xx xx xx max0 + _mm_pmaxub(b, a, b) // Max of r and g and b. + // [8] max3 00 00 00 max2 00 00 00 max1 00 00 00 max0 00 00 00 + _mm_psllw(b, b, sll24) + // Make sure the value of alpha channel is always larger than maximum + // value of color channels. + _mm_pmaxub(accum0, b, accum0) + ".set pop \n\t" + :[accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [tmp]"=&r"(tmp), [ah]"=&f"(ah), [al]"=&f"(al), + [bh]"=&f"(bh), [bl]"=&f"(bl), [srl8]"=&f"(srl8), + [srl16]"=&f"(srl16), [sll24]"=&f"(sll24) + ); + } else { + double maskh, maskl; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + // Set value of alpha channels to 0xFF. + "li %[tmp], 0xff000000 \n\t" + "mtc1 %[tmp], %[maskl] \n\t" + "punpcklwd %[maskl], %[maskl], %[maskl] \n\t" + "mov.d %[maskh], %[maskl] \n\t" + _mm_or(accum0, accum0, mask) + ".set pop \n\t" + :[maskh]"=&f"(maskh), [maskl]"=&f"(maskl), + [accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [tmp]"=&r"(tmp) + ); + } + + // Store the convolution result (16 bytes) and advance the pixel pointers. + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "gssdlc1 %[accum0h], 0xf(%[out_row]) \n\t" + "gssdrc1 %[accum0h], 0x8(%[out_row]) \n\t" + "gssdlc1 %[accum0l], 0x7(%[out_row]) \n\t" + "gssdrc1 %[accum0l], 0x0(%[out_row]) \n\t" + ".set pop \n\t" + ::[accum0h]"f"(accum0h), [accum0l]"f"(accum0l), + [out_row]"r"(out_row) + :"memory" + ); + out_row += 16; + } + + // When the width of the output is not divisible by 4, We need to save one + // pixel (4 bytes) each time. And also the fourth pixel is always absent. + if (pixel_width & 3) { + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + _mm_xor(accum0, accum0, accum0) + _mm_xor(accum1, accum1, accum1) + _mm_xor(accum2, accum2, accum2) + ".set pop \n\t" + :[accum0h]"=&f"(accum0h), [accum0l]"=&f"(accum0l), + [accum1h]"=&f"(accum1h), [accum1l]"=&f"(accum1l), + [accum2h]"=&f"(accum2h), [accum2l]"=&f"(accum2l) + ); + for (int filter_y = 0; filter_y < filter_length; ++filter_y) { + double src8h, src8l, src16h, src16l; + double th, tl, mul_hih, mul_hil, mul_loh, mul_lol; + src = reinterpret_cast( + &source_data_rows[filter_y][out_x<<2]); + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "gsldlc1 %[coeff16l], 7+%[fval] \n\t" + "gsldrc1 %[coeff16l], %[fval] \n\t" + "pshufh %[coeff16l], %[coeff16l], %[zerol] \n\t" + "mov.d %[coeff16h], %[coeff16l] \n\t" + // [8] a3 b3 g3 r3 a2 b2 g2 r2 a1 b1 g1 r1 a0 b0 g0 r0 + "gsldlc1 %[src8h], 0xf(%[src]) \n\t" + "gsldrc1 %[src8h], 0x8(%[src]) \n\t" + "gsldlc1 %[src8l], 0x7(%[src]) \n\t" + "gsldrc1 %[src8l], 0x0(%[src]) \n\t" + // [16] a1 b1 g1 r1 a0 b0 g0 r0 + _mm_punpcklbh(src16, src8, zero) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + // [32] a0 b0 g0 r0 + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum0, accum0, t) + // [32] a1 b1 g1 r1 + _mm_punpckhhw(t, mul_lo, mul_hi) + _mm_paddw(accum1, accum1, t) + // [16] a3 b3 g3 r3 a2 b2 g2 r2 + _mm_punpckhbh(src16, src8, zero) + _mm_pmulhh(mul_hi, src16, coeff16) + _mm_pmullh(mul_lo, src16, coeff16) + // [32] a2 b2 g2 r2 + _mm_punpcklhw(t, mul_lo, mul_hi) + _mm_paddw(accum2, accum2, t) + ".set pop \n\t" + :[th]"=&f"(th), [tl]"=&f"(tl), + [src8h]"=&f"(src8h), [src8l]"=&f"(src8l), + [src16h]"=&f"(src16h), [src16l]"=&f"(src16l), + [mul_hih]"=&f"(mul_hih), [mul_hil]"=&f"(mul_hil), + [mul_loh]"=&f"(mul_loh), [mul_lol]"=&f"(mul_lol), + [accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [accum1h]"+f"(accum1h), [accum1l]"+f"(accum1l), + [accum2h]"+f"(accum2h), [accum2l]"+f"(accum2l), + [coeff16h]"=&f"(coeff16h), [coeff16l]"=&f"(coeff16l) + :[zeroh]"f"(zero), [zerol]"f"(zero), + [fval]"m"(filter_values[filter_y]), + [src]"r"(src) + ); + } + + double t; + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + _mm_psraw(accum0, accum0, sra) + _mm_psraw(accum1, accum1, sra) + _mm_psraw(accum2, accum2, sra) + // [16] a1 b1 g1 r1 a0 b0 g0 r0 + _mm_packsswh(accum0, accum0, accum1, t) + // [16] a3 b3 g3 r3 a2 b2 g2 r2 + _mm_packsswh(accum2, accum2, zero, t) + // [8] a3 b3 g3 r3 a2 b2 g2 r2 a1 b1 g1 r1 a0 b0 g0 r0 + _mm_packushb(accum0, accum0, accum2, t) + ".set pop \n\t" + :[accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [accum1h]"+f"(accum1h), [accum1l]"+f"(accum1l), + [accum2h]"+f"(accum2h), [accum2l]"+f"(accum2l), + [t]"=&f"(t) + :[zeroh]"f"(zero), [zerol]"f"(zero), [sra]"f"(sra) + ); + if (has_alpha) { + double ah, al, bh, bl, srl8, srl16, sll24; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "li %[tmp], 8 \n\t" + "mtc1 %[tmp], %[srl8] \n\t" + "li %[tmp], 16 \n\t" + "mtc1 %[tmp], %[srl16] \n\t" + "li %[tmp], 24 \n\t" + "mtc1 %[tmp], %[sll24] \n\t" + // [8] xx a3 b3 g3 xx a2 b2 g2 xx a1 b1 g1 xx a0 b0 g0 + _mm_psrlw(a, accum0, srl8) + // [8] xx xx xx max3 xx xx xx max2 xx xx xx max1 xx xx xx max0 + _mm_pmaxub(b, a, accum0) // Max of r and g. + // [8] xx xx a3 b3 xx xx a2 b2 xx xx a1 b1 xx xx a0 b0 + _mm_psrlw(a, accum0, srl16) + // [8] xx xx xx max3 xx xx xx max2 xx xx xx max1 xx xx xx max0 + _mm_pmaxub(b, a, b) // Max of r and g and b. + // [8] max3 00 00 00 max2 00 00 00 max1 00 00 00 max0 00 00 00 + _mm_psllw(b, b, sll24) + _mm_pmaxub(accum0, b, accum0) + ".set pop \n\t" + :[ah]"=&f"(ah), [al]"=&f"(al), [bh]"=&f"(bh), [bl]"=&f"(bl), + [accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), [tmp]"=&r"(tmp), + [srl8]"=&f"(srl8), [srl16]"=&f"(srl16), [sll24]"=&f"(sll24) + ); + } else { + double maskh, maskl; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + // Set value of alpha channels to 0xFF. + "li %[tmp], 0xff000000 \n\t" + "mtc1 %[tmp], %[maskl] \n\t" + "punpcklwd %[maskl], %[maskl], %[maskl] \n\t" + "mov.d %[maskh], %[maskl] \n\t" + _mm_or(accum0, accum0, mask) + ".set pop \n\t" + :[maskh]"=&f"(maskh), [maskl]"=&f"(maskl), + [accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l), + [tmp]"=&r"(tmp) + ); + } + + double s4, s64; + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "li %[tmp], 4 \n\t" + "mtc1 %[tmp], %[s4] \n\t" + "li %[tmp], 64 \n\t" + "mtc1 %[tmp], %[s64] \n\t" + ".set pop \n\t" + :[s4]"=f"(s4), [s64]"=f"(s64), + [tmp]"=&r"(tmp) + ); + for (int out_x = width; out_x < pixel_width; out_x++) { + double t; + + asm volatile ( + ".set push \n\t" + ".set arch=loongson3a \n\t" + "swc1 %[accum0l], (%[out_row]) \n\t" + _mm_psrlq(accum0, accum0, s4, s64, t) + ".set pop \n\t" + :[t]"=&f"(t), + [accum0h]"+f"(accum0h), [accum0l]"+f"(accum0l) + :[out_row]"r"(out_row), [s4]"f"(s4), [s64]"f"(s64) + :"memory" + ); + out_row += 4; + } + } +} + +void ConvolveVertically_LS3(const ConvolutionFilter1D::Fixed* filter_values, + int filter_length, + unsigned char* const* source_data_rows, + int pixel_width, + unsigned char* out_row, bool has_alpha) { + if (has_alpha) { + ConvolveVertically_LS3_impl(filter_values, filter_length, + source_data_rows, pixel_width, out_row); + } else { + ConvolveVertically_LS3_impl(filter_values, filter_length, + source_data_rows, pixel_width, out_row); + } +} + +} // namespace skia + +#endif /* _MIPS_ARCH_LOONGSON3A */ -- cgit v1.2.3