| // |
| // Copyright 2014 The ANGLE Project Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| // |
| |
| // copyvertex.inc.h: Implementation of vertex buffer copying and conversion functions |
| |
| namespace rx |
| { |
| |
| // Returns an aligned buffer to read the input from |
| template <typename T, size_t inputComponentCount> |
| inline const T *GetAlignedOffsetInput(const T *offsetInput, T *alignedElement) |
| { |
| if (reinterpret_cast<uintptr_t>(offsetInput) % sizeof(T) != 0) |
| { |
| // Applications may pass in arbitrarily aligned buffers as input. |
| // Certain architectures have restrictions regarding unaligned reads. Specifically, we crash |
| // on armeabi-v7a devices with a SIGBUS error when performing such operations. arm64 and |
| // x86-64 devices do not appear to have such issues. |
| // |
| // The workaround is to detect if the input buffer is unaligned and if so, perform a |
| // byte-wise copy of the unaligned portion and a memcpy of the rest of the buffer. |
| uint8_t *alignedBuffer = reinterpret_cast<uint8_t *>(&alignedElement[0]); |
| uintptr_t unalignedInputStartAddress = reinterpret_cast<uintptr_t>(offsetInput); |
| constexpr size_t kAlignmentMinusOne = sizeof(T) - 1; |
| uintptr_t alignedInputStartAddress = |
| (reinterpret_cast<uintptr_t>(offsetInput) + kAlignmentMinusOne) & ~(kAlignmentMinusOne); |
| ASSERT(alignedInputStartAddress >= unalignedInputStartAddress); |
| |
| const size_t totalBytesToCopy = sizeof(T) * inputComponentCount; |
| const size_t unalignedBytesToCopy = alignedInputStartAddress - unalignedInputStartAddress; |
| ASSERT(totalBytesToCopy >= unalignedBytesToCopy); |
| |
| // byte-wise copy of unaligned portion |
| for (size_t i = 0; i < unalignedBytesToCopy; i++) |
| { |
| alignedBuffer[i] = reinterpret_cast<const uint8_t *>(&offsetInput[0])[i]; |
| } |
| |
| // memcpy remaining buffer |
| memcpy(&alignedBuffer[unalignedBytesToCopy], |
| &reinterpret_cast<const uint8_t *>(&offsetInput[0])[unalignedBytesToCopy], |
| totalBytesToCopy - unalignedBytesToCopy); |
| |
| return alignedElement; |
| } |
| else |
| { |
| return offsetInput; |
| } |
| } |
| |
| template <typename T, |
| size_t inputComponentCount, |
| size_t outputComponentCount, |
| uint32_t alphaDefaultValueBits> |
| inline void CopyNativeVertexData(const uint8_t *input, size_t stride, size_t count, uint8_t *output) |
| { |
| const size_t attribSize = sizeof(T) * inputComponentCount; |
| |
| if (attribSize == stride && inputComponentCount == outputComponentCount) |
| { |
| memcpy(output, input, count * attribSize); |
| return; |
| } |
| |
| if (inputComponentCount == outputComponentCount) |
| { |
| for (size_t i = 0; i < count; i++) |
| { |
| const T *offsetInput = reinterpret_cast<const T *>(input + (i * stride)); |
| T offsetInputAligned[inputComponentCount]; |
| offsetInput = |
| GetAlignedOffsetInput<T, inputComponentCount>(offsetInput, &offsetInputAligned[0]); |
| |
| T *offsetOutput = reinterpret_cast<T *>(output) + i * outputComponentCount; |
| |
| memcpy(offsetOutput, offsetInput, attribSize); |
| } |
| return; |
| } |
| |
| const T defaultAlphaValue = gl::bitCast<T>(alphaDefaultValueBits); |
| const size_t lastNonAlphaOutputComponent = std::min<size_t>(outputComponentCount, 3); |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| const T *offsetInput = reinterpret_cast<const T *>(input + (i * stride)); |
| T offsetInputAligned[inputComponentCount]; |
| ASSERT(sizeof(offsetInputAligned) == attribSize); |
| offsetInput = |
| GetAlignedOffsetInput<T, inputComponentCount>(offsetInput, &offsetInputAligned[0]); |
| |
| T *offsetOutput = reinterpret_cast<T *>(output) + i * outputComponentCount; |
| |
| memcpy(offsetOutput, offsetInput, attribSize); |
| |
| if (inputComponentCount < lastNonAlphaOutputComponent) |
| { |
| // Set the remaining G/B channels to 0. |
| size_t numComponents = (lastNonAlphaOutputComponent - inputComponentCount); |
| memset(&offsetOutput[inputComponentCount], 0, numComponents * sizeof(T)); |
| } |
| |
| if (inputComponentCount < outputComponentCount && outputComponentCount == 4) |
| { |
| // Set the remaining alpha channel to the defaultAlphaValue. |
| offsetOutput[3] = defaultAlphaValue; |
| } |
| } |
| } |
| |
| template <size_t inputComponentCount, size_t outputComponentCount> |
| inline void Copy8SintTo16SintVertexData(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| const size_t lastNonAlphaOutputComponent = std::min<size_t>(outputComponentCount, 3); |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| const GLbyte *offsetInput = reinterpret_cast<const GLbyte *>(input + i * stride); |
| GLshort *offsetOutput = reinterpret_cast<GLshort *>(output) + i * outputComponentCount; |
| |
| for (size_t j = 0; j < inputComponentCount; j++) |
| { |
| offsetOutput[j] = static_cast<GLshort>(offsetInput[j]); |
| } |
| |
| for (size_t j = inputComponentCount; j < lastNonAlphaOutputComponent; j++) |
| { |
| // Set remaining G/B channels to 0. |
| offsetOutput[j] = 0; |
| } |
| |
| if (inputComponentCount < outputComponentCount && outputComponentCount == 4) |
| { |
| // On integer formats, we must set the Alpha channel to 1 if it's unused. |
| offsetOutput[3] = 1; |
| } |
| } |
| } |
| |
| template <size_t inputComponentCount, size_t outputComponentCount> |
| inline void Copy8SnormTo16SnormVertexData(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| for (size_t i = 0; i < count; i++) |
| { |
| const GLbyte *offsetInput = reinterpret_cast<const GLbyte *>(input + i * stride); |
| GLshort *offsetOutput = reinterpret_cast<GLshort *>(output) + i * outputComponentCount; |
| |
| for (size_t j = 0; j < inputComponentCount; j++) |
| { |
| // The original GLbyte value ranges from -128 to +127 (INT8_MAX). |
| // When converted to GLshort, the value must be scaled to between -32768 and +32767 |
| // (INT16_MAX). |
| if (offsetInput[j] > 0) |
| { |
| offsetOutput[j] = |
| offsetInput[j] << 8 | offsetInput[j] << 1 | ((offsetInput[j] & 0x40) >> 6); |
| } |
| else |
| { |
| offsetOutput[j] = offsetInput[j] << 8; |
| } |
| } |
| |
| for (size_t j = inputComponentCount; j < std::min<size_t>(outputComponentCount, 3); j++) |
| { |
| // Set remaining G/B channels to 0. |
| offsetOutput[j] = 0; |
| } |
| |
| if (inputComponentCount < outputComponentCount && outputComponentCount == 4) |
| { |
| // On normalized formats, we must set the Alpha channel to the max value if it's unused. |
| offsetOutput[3] = INT16_MAX; |
| } |
| } |
| } |
| |
| template <size_t inputComponentCount, size_t outputComponentCount> |
| inline void Copy32FixedTo32FVertexData(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| static const float divisor = 1.0f / (1 << 16); |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| const uint8_t *offsetInput = input + i * stride; |
| float *offsetOutput = reinterpret_cast<float *>(output) + i * outputComponentCount; |
| |
| // GLfixed access must be 4-byte aligned on arm32, input and stride sometimes are not |
| if (reinterpret_cast<uintptr_t>(offsetInput) % sizeof(GLfixed) == 0) |
| { |
| for (size_t j = 0; j < inputComponentCount; j++) |
| { |
| offsetOutput[j] = |
| static_cast<float>(reinterpret_cast<const GLfixed *>(offsetInput)[j]) * divisor; |
| } |
| } |
| else |
| { |
| for (size_t j = 0; j < inputComponentCount; j++) |
| { |
| GLfixed alignedInput; |
| memcpy(&alignedInput, offsetInput + j * sizeof(GLfixed), sizeof(GLfixed)); |
| offsetOutput[j] = static_cast<float>(alignedInput) * divisor; |
| } |
| } |
| |
| // 4-component output formats would need special padding in the alpha channel. |
| static_assert(!(inputComponentCount < 4 && outputComponentCount == 4), |
| "An inputComponentCount less than 4 and an outputComponentCount equal to 4 " |
| "is not supported."); |
| |
| for (size_t j = inputComponentCount; j < outputComponentCount; j++) |
| { |
| offsetOutput[j] = 0.0f; |
| } |
| } |
| } |
| |
| template <typename T, |
| size_t inputComponentCount, |
| size_t outputComponentCount, |
| bool normalized, |
| bool toHalf> |
| inline void CopyToFloatVertexData(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| typedef std::numeric_limits<T> NL; |
| typedef typename std::conditional<toHalf, GLhalf, float>::type outputType; |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| const T *offsetInput = reinterpret_cast<const T *>(input + (stride * i)); |
| outputType *offsetOutput = |
| reinterpret_cast<outputType *>(output) + i * outputComponentCount; |
| |
| T offsetInputAligned[inputComponentCount]; |
| offsetInput = |
| GetAlignedOffsetInput<T, inputComponentCount>(offsetInput, &offsetInputAligned[0]); |
| |
| for (size_t j = 0; j < inputComponentCount; j++) |
| { |
| float result = 0; |
| |
| if (normalized) |
| { |
| if (NL::is_signed) |
| { |
| result = static_cast<float>(offsetInput[j]) / static_cast<float>(NL::max()); |
| result = result >= -1.0f ? result : -1.0f; |
| } |
| else |
| { |
| result = static_cast<float>(offsetInput[j]) / static_cast<float>(NL::max()); |
| } |
| } |
| else |
| { |
| result = static_cast<float>(offsetInput[j]); |
| } |
| |
| if (toHalf) |
| { |
| offsetOutput[j] = gl::float32ToFloat16(result); |
| } |
| else |
| { |
| offsetOutput[j] = static_cast<outputType>(result); |
| } |
| } |
| |
| for (size_t j = inputComponentCount; j < outputComponentCount; j++) |
| { |
| offsetOutput[j] = 0; |
| } |
| |
| if (inputComponentCount < 4 && outputComponentCount == 4) |
| { |
| if (toHalf) |
| { |
| offsetOutput[3] = gl::Float16One; |
| } |
| else |
| { |
| offsetOutput[3] = static_cast<outputType>(gl::Float32One); |
| } |
| } |
| } |
| } |
| |
| template <size_t inputComponentCount, size_t outputComponentCount> |
| void Copy32FTo16FVertexData(const uint8_t *input, size_t stride, size_t count, uint8_t *output) |
| { |
| const unsigned short kZero = gl::float32ToFloat16(0.0f); |
| const unsigned short kOne = gl::float32ToFloat16(1.0f); |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| const float *offsetInput = reinterpret_cast<const float *>(input + (stride * i)); |
| unsigned short *offsetOutput = |
| reinterpret_cast<unsigned short *>(output) + i * outputComponentCount; |
| |
| for (size_t j = 0; j < inputComponentCount; j++) |
| { |
| offsetOutput[j] = gl::float32ToFloat16(offsetInput[j]); |
| } |
| |
| for (size_t j = inputComponentCount; j < outputComponentCount; j++) |
| { |
| offsetOutput[j] = (j == 3) ? kOne : kZero; |
| } |
| } |
| } |
| |
| inline void CopyXYZ32FToXYZ9E5(const uint8_t *input, size_t stride, size_t count, uint8_t *output) |
| { |
| for (size_t i = 0; i < count; i++) |
| { |
| const float *offsetInput = reinterpret_cast<const float *>(input + (stride * i)); |
| unsigned int *offsetOutput = reinterpret_cast<unsigned int *>(output) + i; |
| |
| *offsetOutput = gl::convertRGBFloatsTo999E5(offsetInput[0], offsetInput[1], offsetInput[2]); |
| } |
| } |
| |
| inline void CopyXYZ32FToX11Y11B10F(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| for (size_t i = 0; i < count; i++) |
| { |
| const float *offsetInput = reinterpret_cast<const float *>(input + (stride * i)); |
| unsigned int *offsetOutput = reinterpret_cast<unsigned int *>(output) + i; |
| |
| *offsetOutput = gl::float32ToFloat11(offsetInput[0]) << 0 | |
| gl::float32ToFloat11(offsetInput[1]) << 11 | |
| gl::float32ToFloat10(offsetInput[2]) << 22; |
| } |
| } |
| |
| namespace priv |
| { |
| |
| template <bool isSigned, bool normalized, bool toFloat, bool toHalf> |
| static inline void CopyPackedRGB(uint32_t data, uint8_t *output) |
| { |
| const uint32_t rgbSignMask = 0x200; // 1 set at the 9 bit |
| const uint32_t negativeMask = 0xFFFFFC00; // All bits from 10 to 31 set to 1 |
| |
| if (toFloat || toHalf) |
| { |
| GLfloat finalValue = static_cast<GLfloat>(data); |
| if (isSigned) |
| { |
| if (data & rgbSignMask) |
| { |
| int negativeNumber = data | negativeMask; |
| finalValue = static_cast<GLfloat>(negativeNumber); |
| } |
| |
| if (normalized) |
| { |
| const int32_t maxValue = 0x1FF; // 1 set in bits 0 through 8 |
| const int32_t minValue = 0xFFFFFE01; // Inverse of maxValue |
| |
| // A 10-bit two's complement number has the possibility of being minValue - 1 but |
| // OpenGL's normalization rules dictate that it should be clamped to minValue in |
| // this case. |
| if (finalValue < minValue) |
| { |
| finalValue = minValue; |
| } |
| |
| const int32_t halfRange = (maxValue - minValue) >> 1; |
| finalValue = ((finalValue - minValue) / halfRange) - 1.0f; |
| } |
| } |
| else |
| { |
| if (normalized) |
| { |
| const uint32_t maxValue = 0x3FF; // 1 set in bits 0 through 9 |
| finalValue /= static_cast<GLfloat>(maxValue); |
| } |
| } |
| |
| if (toHalf) |
| { |
| *reinterpret_cast<GLhalf *>(output) = gl::float32ToFloat16(finalValue); |
| } |
| else |
| { |
| *reinterpret_cast<GLfloat *>(output) = finalValue; |
| } |
| } |
| else |
| { |
| if (isSigned) |
| { |
| GLshort *intOutput = reinterpret_cast<GLshort *>(output); |
| |
| if (data & rgbSignMask) |
| { |
| *intOutput = static_cast<GLshort>(data | negativeMask); |
| } |
| else |
| { |
| *intOutput = static_cast<GLshort>(data); |
| } |
| } |
| else |
| { |
| GLushort *uintOutput = reinterpret_cast<GLushort *>(output); |
| *uintOutput = static_cast<GLushort>(data); |
| } |
| } |
| } |
| |
| template <bool isSigned, bool normalized, bool toFloat, bool toHalf> |
| inline void CopyPackedAlpha(uint32_t data, uint8_t *output) |
| { |
| ASSERT(data >= 0 && data <= 3); |
| |
| if (toFloat || toHalf) |
| { |
| GLfloat finalValue = 0; |
| if (isSigned) |
| { |
| if (normalized) |
| { |
| switch (data) |
| { |
| case 0x0: |
| finalValue = 0.0f; |
| break; |
| case 0x1: |
| finalValue = 1.0f; |
| break; |
| case 0x2: |
| finalValue = -1.0f; |
| break; |
| case 0x3: |
| finalValue = -1.0f; |
| break; |
| default: |
| UNREACHABLE(); |
| } |
| } |
| else |
| { |
| switch (data) |
| { |
| case 0x0: |
| finalValue = 0.0f; |
| break; |
| case 0x1: |
| finalValue = 1.0f; |
| break; |
| case 0x2: |
| finalValue = -2.0f; |
| break; |
| case 0x3: |
| finalValue = -1.0f; |
| break; |
| default: |
| UNREACHABLE(); |
| } |
| } |
| } |
| else |
| { |
| if (normalized) |
| { |
| finalValue = data / 3.0f; |
| } |
| else |
| { |
| finalValue = static_cast<float>(data); |
| } |
| } |
| |
| if (toHalf) |
| { |
| *reinterpret_cast<GLhalf *>(output) = gl::float32ToFloat16(finalValue); |
| } |
| else |
| { |
| *reinterpret_cast<GLfloat *>(output) = finalValue; |
| } |
| } |
| else |
| { |
| if (isSigned) |
| { |
| GLshort *intOutput = reinterpret_cast<GLshort *>(output); |
| switch (data) |
| { |
| case 0x0: |
| *intOutput = 0; |
| break; |
| case 0x1: |
| *intOutput = 1; |
| break; |
| case 0x2: |
| *intOutput = -2; |
| break; |
| case 0x3: |
| *intOutput = -1; |
| break; |
| default: |
| UNREACHABLE(); |
| } |
| } |
| else |
| { |
| *reinterpret_cast<GLushort *>(output) = static_cast<GLushort>(data); |
| } |
| } |
| } |
| |
| } // namespace priv |
| |
| template <bool isSigned, bool normalized, bool toFloat, bool toHalf> |
| inline void CopyXYZ10W2ToXYZWFloatVertexData(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| const size_t outputComponentSize = toFloat && !toHalf ? 4 : 2; |
| const size_t componentCount = 4; |
| |
| const uint32_t rgbMask = 0x3FF; // 1 set in bits 0 through 9 |
| const size_t redShift = 0; // red is bits 0 through 9 |
| const size_t greenShift = 10; // green is bits 10 through 19 |
| const size_t blueShift = 20; // blue is bits 20 through 29 |
| |
| const uint32_t alphaMask = 0x3; // 1 set in bits 0 and 1 |
| const size_t alphaShift = 30; // Alpha is the 30 and 31 bits |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| GLuint packedValue = *reinterpret_cast<const GLuint *>(input + (i * stride)); |
| uint8_t *offsetOutput = output + (i * outputComponentSize * componentCount); |
| |
| priv::CopyPackedRGB<isSigned, normalized, toFloat, toHalf>( |
| (packedValue >> redShift) & rgbMask, offsetOutput + (0 * outputComponentSize)); |
| priv::CopyPackedRGB<isSigned, normalized, toFloat, toHalf>( |
| (packedValue >> greenShift) & rgbMask, offsetOutput + (1 * outputComponentSize)); |
| priv::CopyPackedRGB<isSigned, normalized, toFloat, toHalf>( |
| (packedValue >> blueShift) & rgbMask, offsetOutput + (2 * outputComponentSize)); |
| priv::CopyPackedAlpha<isSigned, normalized, toFloat, toHalf>( |
| (packedValue >> alphaShift) & alphaMask, offsetOutput + (3 * outputComponentSize)); |
| } |
| } |
| |
| template <bool isSigned, bool normalized, bool toHalf> |
| inline void CopyXYZ10ToXYZWFloatVertexData(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| const size_t outputComponentSize = toHalf ? 2 : 4; |
| const size_t componentCount = 4; |
| |
| const uint32_t rgbMask = 0x3FF; // 1 set in bits 0 through 9 |
| const size_t redShift = 22; // red is bits 22 through 31 |
| const size_t greenShift = 12; // green is bits 12 through 21 |
| const size_t blueShift = 2; // blue is bits 2 through 11 |
| |
| const uint32_t alphaDefaultValueBits = normalized ? (isSigned ? 0x1 : 0x3) : 0x1; |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| GLuint packedValue = *reinterpret_cast<const GLuint *>(input + (i * stride)); |
| uint8_t *offsetOutput = output + (i * outputComponentSize * componentCount); |
| |
| priv::CopyPackedRGB<isSigned, normalized, true, toHalf>( |
| (packedValue >> redShift) & rgbMask, offsetOutput + (0 * outputComponentSize)); |
| priv::CopyPackedRGB<isSigned, normalized, true, toHalf>( |
| (packedValue >> greenShift) & rgbMask, offsetOutput + (1 * outputComponentSize)); |
| priv::CopyPackedRGB<isSigned, normalized, true, toHalf>( |
| (packedValue >> blueShift) & rgbMask, offsetOutput + (2 * outputComponentSize)); |
| priv::CopyPackedAlpha<isSigned, normalized, true, toHalf>( |
| alphaDefaultValueBits, offsetOutput + (3 * outputComponentSize)); |
| } |
| } |
| |
| template <bool isSigned, bool normalized, bool toHalf> |
| inline void CopyW2XYZ10ToXYZWFloatVertexData(const uint8_t *input, |
| size_t stride, |
| size_t count, |
| uint8_t *output) |
| { |
| const size_t outputComponentSize = toHalf ? 2 : 4; |
| const size_t componentCount = 4; |
| |
| const uint32_t rgbMask = 0x3FF; // 1 set in bits 0 through 9 |
| const size_t redShift = 22; // red is bits 22 through 31 |
| const size_t greenShift = 12; // green is bits 12 through 21 |
| const size_t blueShift = 2; // blue is bits 2 through 11 |
| |
| const uint32_t alphaMask = 0x3; // 1 set in bits 0 and 1 |
| const size_t alphaShift = 0; // Alpha is the 30 and 31 bits |
| |
| for (size_t i = 0; i < count; i++) |
| { |
| GLuint packedValue = *reinterpret_cast<const GLuint *>(input + (i * stride)); |
| uint8_t *offsetOutput = output + (i * outputComponentSize * componentCount); |
| |
| priv::CopyPackedRGB<isSigned, normalized, true, toHalf>( |
| (packedValue >> redShift) & rgbMask, offsetOutput + (0 * outputComponentSize)); |
| priv::CopyPackedRGB<isSigned, normalized, true, toHalf>( |
| (packedValue >> greenShift) & rgbMask, offsetOutput + (1 * outputComponentSize)); |
| priv::CopyPackedRGB<isSigned, normalized, true, toHalf>( |
| (packedValue >> blueShift) & rgbMask, offsetOutput + (2 * outputComponentSize)); |
| priv::CopyPackedAlpha<isSigned, normalized, true, toHalf>( |
| (packedValue >> alphaShift) & alphaMask, offsetOutput + (3 * outputComponentSize)); |
| } |
| } |
| } // namespace rx |