diff options
Diffstat (limited to 'gfx/skia/skia/src/utils/SkTextureCompressor_LATC.cpp')
-rw-r--r-- | gfx/skia/skia/src/utils/SkTextureCompressor_LATC.cpp | 519 |
1 files changed, 519 insertions, 0 deletions
diff --git a/gfx/skia/skia/src/utils/SkTextureCompressor_LATC.cpp b/gfx/skia/skia/src/utils/SkTextureCompressor_LATC.cpp new file mode 100644 index 000000000..50aaf0b27 --- /dev/null +++ b/gfx/skia/skia/src/utils/SkTextureCompressor_LATC.cpp @@ -0,0 +1,519 @@ +/* + * Copyright 2014 Google Inc. + * + * Use of this source code is governed by a BSD-style license that can be + * found in the LICENSE file. + */ + +#include "SkTextureCompressor_LATC.h" +#include "SkTextureCompressor_Blitter.h" +#include "SkTextureCompressor_Utils.h" + +#include "SkBlitter.h" +#include "SkEndian.h" + +// Compression options. In general, the slow version is much more accurate, but +// much slower. The fast option is much faster, but much less accurate. YMMV. +#define COMPRESS_LATC_SLOW 0 +#define COMPRESS_LATC_FAST 1 + +//////////////////////////////////////////////////////////////////////////////// + +// Generates an LATC palette. LATC constructs +// a palette of eight colors from LUM0 and LUM1 using the algorithm: +// +// LUM0, if lum0 > lum1 and code(x,y) == 0 +// LUM1, if lum0 > lum1 and code(x,y) == 1 +// (6*LUM0+ LUM1)/7, if lum0 > lum1 and code(x,y) == 2 +// (5*LUM0+2*LUM1)/7, if lum0 > lum1 and code(x,y) == 3 +// (4*LUM0+3*LUM1)/7, if lum0 > lum1 and code(x,y) == 4 +// (3*LUM0+4*LUM1)/7, if lum0 > lum1 and code(x,y) == 5 +// (2*LUM0+5*LUM1)/7, if lum0 > lum1 and code(x,y) == 6 +// ( LUM0+6*LUM1)/7, if lum0 > lum1 and code(x,y) == 7 +// +// LUM0, if lum0 <= lum1 and code(x,y) == 0 +// LUM1, if lum0 <= lum1 and code(x,y) == 1 +// (4*LUM0+ LUM1)/5, if lum0 <= lum1 and code(x,y) == 2 +// (3*LUM0+2*LUM1)/5, if lum0 <= lum1 and code(x,y) == 3 +// (2*LUM0+3*LUM1)/5, if lum0 <= lum1 and code(x,y) == 4 +// ( LUM0+4*LUM1)/5, if lum0 <= lum1 and code(x,y) == 5 +// 0, if lum0 <= lum1 and code(x,y) == 6 +// 255, if lum0 <= lum1 and code(x,y) == 7 + +static const int kLATCPaletteSize = 8; +static void generate_latc_palette(uint8_t palette[], uint8_t lum0, uint8_t lum1) { + palette[0] = lum0; + palette[1] = lum1; + if (lum0 > lum1) { + for (int i = 1; i < 7; i++) { + palette[i+1] = ((7-i)*lum0 + i*lum1) / 7; + } + } else { + for (int i = 1; i < 5; i++) { + palette[i+1] = ((5-i)*lum0 + i*lum1) / 5; + } + palette[6] = 0; + palette[7] = 255; + } +} + +//////////////////////////////////////////////////////////////////////////////// + +#if COMPRESS_LATC_SLOW + +//////////////////////////////////////////////////////////////////////////////// +// +// Utility Functions +// +//////////////////////////////////////////////////////////////////////////////// + +// Absolute difference between two values. More correct than SkTAbs(a - b) +// because it works on unsigned values. +template <typename T> inline T abs_diff(const T &a, const T &b) { + return (a > b) ? (a - b) : (b - a); +} + +static bool is_extremal(uint8_t pixel) { + return 0 == pixel || 255 == pixel; +} + +typedef uint64_t (*A84x4To64BitProc)(const uint8_t block[]); + +// This function is used by both R11 EAC and LATC to compress 4x4 blocks +// of 8-bit alpha into 64-bit values that comprise the compressed data. +// For both formats, we need to make sure that the dimensions of the +// src pixels are divisible by 4, and copy 4x4 blocks one at a time +// for compression. +static bool compress_4x4_a8_to_64bit(uint8_t* dst, const uint8_t* src, + int width, int height, size_t rowBytes, + A84x4To64BitProc proc) { + // Make sure that our data is well-formed enough to be considered for compression + if (0 == width || 0 == height || (width % 4) != 0 || (height % 4) != 0) { + return false; + } + + int blocksX = width >> 2; + int blocksY = height >> 2; + + uint8_t block[16]; + uint64_t* encPtr = reinterpret_cast<uint64_t*>(dst); + for (int y = 0; y < blocksY; ++y) { + for (int x = 0; x < blocksX; ++x) { + // Load block + for (int k = 0; k < 4; ++k) { + memcpy(block + k*4, src + k*rowBytes + 4*x, 4); + } + + // Compress it + *encPtr = proc(block); + ++encPtr; + } + src += 4 * rowBytes; + } + + return true; +} + +//////////////////////////////////////////////////////////////////////////////// +// +// LATC compressor +// +//////////////////////////////////////////////////////////////////////////////// + +// LATC compressed texels down into square 4x4 blocks +static const int kLATCBlockSize = 4; +static const int kLATCPixelsPerBlock = kLATCBlockSize * kLATCBlockSize; + +// Compress a block by using the bounding box of the pixels. It is assumed that +// there are no extremal pixels in this block otherwise we would have used +// compressBlockBBIgnoreExtremal. +static uint64_t compress_latc_block_bb(const uint8_t pixels[]) { + uint8_t minVal = 255; + uint8_t maxVal = 0; + for (int i = 0; i < kLATCPixelsPerBlock; ++i) { + minVal = SkTMin(pixels[i], minVal); + maxVal = SkTMax(pixels[i], maxVal); + } + + SkASSERT(!is_extremal(minVal)); + SkASSERT(!is_extremal(maxVal)); + + uint8_t palette[kLATCPaletteSize]; + generate_latc_palette(palette, maxVal, minVal); + + uint64_t indices = 0; + for (int i = kLATCPixelsPerBlock - 1; i >= 0; --i) { + + // Find the best palette index + uint8_t bestError = abs_diff(pixels[i], palette[0]); + uint8_t idx = 0; + for (int j = 1; j < kLATCPaletteSize; ++j) { + uint8_t error = abs_diff(pixels[i], palette[j]); + if (error < bestError) { + bestError = error; + idx = j; + } + } + + indices <<= 3; + indices |= idx; + } + + return + SkEndian_SwapLE64( + static_cast<uint64_t>(maxVal) | + (static_cast<uint64_t>(minVal) << 8) | + (indices << 16)); +} + +// Compress a block by using the bounding box of the pixels without taking into +// account the extremal values. The generated palette will contain extremal values +// and fewer points along the line segment to interpolate. +static uint64_t compress_latc_block_bb_ignore_extremal(const uint8_t pixels[]) { + uint8_t minVal = 255; + uint8_t maxVal = 0; + for (int i = 0; i < kLATCPixelsPerBlock; ++i) { + if (is_extremal(pixels[i])) { + continue; + } + + minVal = SkTMin(pixels[i], minVal); + maxVal = SkTMax(pixels[i], maxVal); + } + + SkASSERT(!is_extremal(minVal)); + SkASSERT(!is_extremal(maxVal)); + + uint8_t palette[kLATCPaletteSize]; + generate_latc_palette(palette, minVal, maxVal); + + uint64_t indices = 0; + for (int i = kLATCPixelsPerBlock - 1; i >= 0; --i) { + + // Find the best palette index + uint8_t idx = 0; + if (is_extremal(pixels[i])) { + if (0xFF == pixels[i]) { + idx = 7; + } else if (0 == pixels[i]) { + idx = 6; + } else { + SkFAIL("Pixel is extremal but not really?!"); + } + } else { + uint8_t bestError = abs_diff(pixels[i], palette[0]); + for (int j = 1; j < kLATCPaletteSize - 2; ++j) { + uint8_t error = abs_diff(pixels[i], palette[j]); + if (error < bestError) { + bestError = error; + idx = j; + } + } + } + + indices <<= 3; + indices |= idx; + } + + return + SkEndian_SwapLE64( + static_cast<uint64_t>(minVal) | + (static_cast<uint64_t>(maxVal) << 8) | + (indices << 16)); +} + + +// Compress LATC block. Each 4x4 block of pixels is decompressed by LATC from two +// values LUM0 and LUM1, and an index into the generated palette. Details of how +// the palette is generated can be found in the comments of generatePalette above. +// +// We choose which palette type to use based on whether or not 'pixels' contains +// any extremal values (0 or 255). If there are extremal values, then we use the +// palette that has the extremal values built in. Otherwise, we use the full bounding +// box. + +static uint64_t compress_latc_block(const uint8_t pixels[]) { + // Collect unique pixels + int nUniquePixels = 0; + uint8_t uniquePixels[kLATCPixelsPerBlock]; + for (int i = 0; i < kLATCPixelsPerBlock; ++i) { + bool foundPixel = false; + for (int j = 0; j < nUniquePixels; ++j) { + foundPixel = foundPixel || uniquePixels[j] == pixels[i]; + } + + if (!foundPixel) { + uniquePixels[nUniquePixels] = pixels[i]; + ++nUniquePixels; + } + } + + // If there's only one unique pixel, then our compression is easy. + if (1 == nUniquePixels) { + return SkEndian_SwapLE64(pixels[0] | (pixels[0] << 8)); + + // Similarly, if there are only two unique pixels, then our compression is + // easy again: place the pixels in the block header, and assign the indices + // with one or zero depending on which pixel they belong to. + } else if (2 == nUniquePixels) { + uint64_t outBlock = 0; + for (int i = kLATCPixelsPerBlock - 1; i >= 0; --i) { + int idx = 0; + if (pixels[i] == uniquePixels[1]) { + idx = 1; + } + + outBlock <<= 3; + outBlock |= idx; + } + outBlock <<= 16; + outBlock |= (uniquePixels[0] | (uniquePixels[1] << 8)); + return SkEndian_SwapLE64(outBlock); + } + + // Count non-maximal pixel values + int nonExtremalPixels = 0; + for (int i = 0; i < nUniquePixels; ++i) { + if (!is_extremal(uniquePixels[i])) { + ++nonExtremalPixels; + } + } + + // If all the pixels are nonmaximal then compute the palette using + // the bounding box of all the pixels. + if (nonExtremalPixels == nUniquePixels) { + // This is really just for correctness, in all of my tests we + // never take this step. We don't lose too much perf here because + // most of the processing in this function is worth it for the + // 1 == nUniquePixels optimization. + return compress_latc_block_bb(pixels); + } else { + return compress_latc_block_bb_ignore_extremal(pixels); + } +} + +#endif // COMPRESS_LATC_SLOW + +//////////////////////////////////////////////////////////////////////////////// + +#if COMPRESS_LATC_FAST + +// Take the top three bits of each index and pack them into the low 12 +// bits of the integer. +static inline uint32_t pack_index(uint32_t x) { + // Pack it in... +#if defined (SK_CPU_BENDIAN) + return + (x >> 24) | + ((x >> 13) & 0x38) | + ((x >> 2) & 0x1C0) | + ((x << 9) & 0xE00); +#else + return + (x & 0x7) | + ((x >> 5) & 0x38) | + ((x >> 10) & 0x1C0) | + ((x >> 15) & 0xE00); +#endif +} + +// Converts each 8-bit byte in the integer into an LATC index, and then packs +// the indices into the low 12 bits of the integer. +static inline uint32_t convert_index(uint32_t x) { + // Since the palette is + // 255, 0, 219, 182, 146, 109, 73, 36 + // we need to map the high three bits of each byte in the integer + // from + // 0 1 2 3 4 5 6 7 + // to + // 1 7 6 5 4 3 2 0 + // + // This first operation takes the mapping from + // 0 1 2 3 4 5 6 7 --> 7 6 5 4 3 2 1 0 + x = 0x07070707 - SkTextureCompressor::ConvertToThreeBitIndex(x); + + // mask is 1 if index is non-zero + const uint32_t mask = (x | (x >> 1) | (x >> 2)) & 0x01010101; + + // add mask: + // 7 6 5 4 3 2 1 0 --> 8 7 6 5 4 3 2 0 + x = (x + mask); + + // Handle overflow: + // 8 7 6 5 4 3 2 0 --> 9 7 6 5 4 3 2 0 + x |= (x >> 3) & 0x01010101; + + // Mask out high bits: + // 9 7 6 5 4 3 2 0 --> 1 7 6 5 4 3 2 0 + x &= 0x07070707; + + return pack_index(x); +} + +typedef uint64_t (*PackIndicesProc)(const uint8_t* alpha, size_t rowBytes); +template<PackIndicesProc packIndicesProc> +static void compress_a8_latc_block(uint8_t** dstPtr, const uint8_t* src, size_t rowBytes) { + *(reinterpret_cast<uint64_t*>(*dstPtr)) = + SkEndian_SwapLE64(0xFF | (packIndicesProc(src, rowBytes) << 16)); + *dstPtr += 8; +} + +inline uint64_t PackRowMajor(const uint8_t *indices, size_t rowBytes) { + uint64_t result = 0; + for (int i = 0; i < 4; ++i) { + const uint32_t idx = *(reinterpret_cast<const uint32_t*>(indices + i*rowBytes)); + result |= static_cast<uint64_t>(convert_index(idx)) << 12*i; + } + return result; +} + +inline uint64_t PackColumnMajor(const uint8_t *indices, size_t rowBytes) { + // !SPEED! Blarg, this is kind of annoying. SSE4 can make this + // a LOT faster. + uint8_t transposed[16]; + for (int i = 0; i < 4; ++i) { + for (int j = 0; j < 4; ++j) { + transposed[j*4+i] = indices[i*rowBytes + j]; + } + } + + return PackRowMajor(transposed, 4); +} + +static bool compress_4x4_a8_latc(uint8_t* dst, const uint8_t* src, + int width, int height, size_t rowBytes) { + + if (width < 0 || ((width % 4) != 0) || height < 0 || ((height % 4) != 0)) { + return false; + } + + uint8_t** dstPtr = &dst; + for (int y = 0; y < height; y += 4) { + for (int x = 0; x < width; x += 4) { + compress_a8_latc_block<PackRowMajor>(dstPtr, src + y*rowBytes + x, rowBytes); + } + } + + return true; +} + +void CompressA8LATCBlockVertical(uint8_t* dst, const uint8_t block[]) { + compress_a8_latc_block<PackColumnMajor>(&dst, block, 4); +} + +#endif // COMPRESS_LATC_FAST + +void decompress_latc_block(uint8_t* dst, int dstRowBytes, const uint8_t* src) { + uint64_t block = SkEndian_SwapLE64(*(reinterpret_cast<const uint64_t *>(src))); + uint8_t lum0 = block & 0xFF; + uint8_t lum1 = (block >> 8) & 0xFF; + + uint8_t palette[kLATCPaletteSize]; + generate_latc_palette(palette, lum0, lum1); + + block >>= 16; + for (int j = 0; j < 4; ++j) { + for (int i = 0; i < 4; ++i) { + dst[i] = palette[block & 0x7]; + block >>= 3; + } + dst += dstRowBytes; + } +} + +// This is the type passed as the CompressorType argument of the compressed +// blitter for the LATC format. The static functions required to be in this +// struct are documented in SkTextureCompressor_Blitter.h +struct CompressorLATC { + static inline void CompressA8Vertical(uint8_t* dst, const uint8_t block[]) { + compress_a8_latc_block<PackColumnMajor>(&dst, block, 4); + } + + static inline void CompressA8Horizontal(uint8_t* dst, const uint8_t* src, + int srcRowBytes) { + compress_a8_latc_block<PackRowMajor>(&dst, src, srcRowBytes); + } + +#if PEDANTIC_BLIT_RECT + static inline void UpdateBlock(uint8_t* dst, const uint8_t* src, int srcRowBytes, + const uint8_t* mask) { + // Pack the mask + uint64_t cmpMask = 0; + for (int i = 0; i < 4; ++i) { + const uint32_t idx = *(reinterpret_cast<const uint32_t*>(src + i*srcRowBytes)); + cmpMask |= static_cast<uint64_t>(pack_index(idx)) << 12*i; + } + cmpMask = SkEndian_SwapLE64(cmpMask << 16); // avoid header + + uint64_t cmpSrc; + uint8_t *cmpSrcPtr = reinterpret_cast<uint8_t*>(&cmpSrc); + compress_a8_latc_block<PackRowMajor>(&cmpSrcPtr, src, srcRowBytes); + + // Mask out header + cmpSrc = cmpSrc & cmpMask; + + // Read destination encoding + uint64_t *cmpDst = reinterpret_cast<uint64_t*>(dst); + + // If the destination is the encoding for a blank block, then we need + // to properly set the header + if (0 == cmpDst) { + *cmpDst = SkTEndian_SwapLE64(0x24924924924900FFULL); + } + + // Set the new indices + *cmpDst &= ~cmpMask; + *cmpDst |= cmpSrc; + } +#endif // PEDANTIC_BLIT_RECT +}; + +//////////////////////////////////////////////////////////////////////////////// + +namespace SkTextureCompressor { + +bool CompressA8ToLATC(uint8_t* dst, const uint8_t* src, int width, int height, size_t rowBytes) { +#if COMPRESS_LATC_FAST + return compress_4x4_a8_latc(dst, src, width, height, rowBytes); +#elif COMPRESS_LATC_SLOW + return compress_4x4_a8_to_64bit(dst, src, width, height, rowBytes, compress_latc_block); +#else +#error "Must choose either fast or slow LATC compression" +#endif +} + +SkBlitter* CreateLATCBlitter(int width, int height, void* outputBuffer, + SkTBlitterAllocator* allocator) { + if ((width % 4) != 0 || (height % 4) != 0) { + return nullptr; + } + +#if COMPRESS_LATC_FAST + // Memset the output buffer to an encoding that decodes to zero. We must do this + // in order to avoid having uninitialized values in the buffer if the blitter + // decides not to write certain scanlines (and skip entire rows of blocks). + // In the case of LATC, if everything is zero, then LUM0 and LUM1 are also zero, + // and they will only be non-zero (0xFF) if the index is 7. So bzero will do just fine. + // (8 bytes per block) * (w * h / 16 blocks) = w * h / 2 + sk_bzero(outputBuffer, width * height / 2); + + return allocator->createT< + SkTCompressedAlphaBlitter<4, 8, CompressorLATC>, int, int, void* > + (width, height, outputBuffer); +#elif COMPRESS_LATC_SLOW + // TODO (krajcevski) + return nullptr; +#endif +} + +void DecompressLATC(uint8_t* dst, int dstRowBytes, const uint8_t* src, int width, int height) { + for (int j = 0; j < height; j += 4) { + for (int i = 0; i < width; i += 4) { + decompress_latc_block(dst + i, dstRowBytes, src); + src += 8; + } + dst += 4 * dstRowBytes; + } +} + +} // SkTextureCompressor |