blob: 3d1ca5f2474b23f7ea4171d526cb5d2316d2c565 [file] [log] [blame]
/*
* Copyright (C) 2021 Apple Inc. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS''
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
* THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
* THE POSSIBILITY OF SUCH DAMAGE.
*/
#import "config.h"
#import "CVUtilities.h"
#import "ColorSpaceCG.h"
#import "IOSurface.h"
#import "Logging.h"
#import "RealtimeVideoUtilities.h"
#import <wtf/StdLibExtras.h>
#import <wtf/cf/TypeCastsCF.h>
#import "CoreVideoSoftLink.h"
namespace WebCore {
static Expected<RetainPtr<CVPixelBufferPoolRef>, CVReturn> createBufferPool(unsigned minimumBufferCount, NSDictionary *pixelAttributes)
{
NSDictionary *poolOptions = nullptr;
if (minimumBufferCount) {
poolOptions = @{
(__bridge NSString *)kCVPixelBufferPoolMinimumBufferCountKey : @(minimumBufferCount)
};
}
CVPixelBufferPoolRef pool = nullptr;
auto status = CVPixelBufferPoolCreate(kCFAllocatorDefault, (__bridge CFDictionaryRef)poolOptions, (__bridge CFDictionaryRef)pixelAttributes, &pool);
if (status != kCVReturnSuccess || !pool)
return makeUnexpected(status);
return adoptCF(pool);
}
Expected<RetainPtr<CVPixelBufferPoolRef>, CVReturn> createIOSurfaceCVPixelBufferPool(size_t width, size_t height, OSType format, unsigned minimumBufferCount, bool isCGImageCompatible)
{
return createBufferPool(minimumBufferCount, @{
(__bridge NSString *)kCVPixelBufferWidthKey : @(width),
(__bridge NSString *)kCVPixelBufferHeightKey : @(height),
(__bridge NSString *)kCVPixelBufferPixelFormatTypeKey : @(format),
(__bridge NSString *)kCVPixelBufferCGImageCompatibilityKey : isCGImageCompatible ? @YES : @NO,
#if PLATFORM(MAC)
(__bridge NSString *)kCVPixelBufferOpenGLCompatibilityKey : @YES,
#endif
(__bridge NSString *)kCVPixelBufferIOSurfacePropertiesKey : @{ }
});
}
Expected<RetainPtr<CVPixelBufferPoolRef>, CVReturn> createInMemoryCVPixelBufferPool(size_t width, size_t height, OSType format, unsigned minimumBufferCount, bool isCGImageCompatible)
{
return createBufferPool(minimumBufferCount, @{
(__bridge NSString *)kCVPixelBufferWidthKey : @(width),
(__bridge NSString *)kCVPixelBufferHeightKey : @(height),
(__bridge NSString *)kCVPixelBufferPixelFormatTypeKey : @(format),
(__bridge NSString *)kCVPixelBufferCGImageCompatibilityKey : isCGImageCompatible ? @YES : @NO,
#if PLATFORM(MAC)
(__bridge NSString *)kCVPixelBufferOpenGLCompatibilityKey : @YES,
#endif
});
}
Expected<RetainPtr<CVPixelBufferPoolRef>, CVReturn> createCVPixelBufferPool(size_t width, size_t height, OSType format, unsigned minimumBufferCount, bool isCGImageCompatible, bool shouldUseIOSurfacePool)
{
return shouldUseIOSurfacePool ? createIOSurfaceCVPixelBufferPool(width, height, format, minimumBufferCount, isCGImageCompatible) : createInMemoryCVPixelBufferPool(width, height, format, minimumBufferCount, isCGImageCompatible);
}
Expected<RetainPtr<CVPixelBufferRef>, CVReturn> createCVPixelBufferFromPool(CVPixelBufferPoolRef pixelBufferPool)
{
CVPixelBufferRef pixelBuffer = nullptr;
auto status = CVPixelBufferPoolCreatePixelBuffer(kCFAllocatorDefault, pixelBufferPool, &pixelBuffer);
if (status != kCVReturnSuccess || !pixelBuffer)
return makeUnexpected(status);
return adoptCF(pixelBuffer);
}
static CFDictionaryRef pixelBufferCreationOptions(IOSurfaceRef surface)
{
#if PLATFORM(MAC)
auto format = IOSurfaceGetPixelFormat(surface);
if (format == kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange || format == kCVPixelFormatType_420YpCbCr8BiPlanarFullRange) {
// YUV formats might contain extra pixels due to block size. Define the picture position
// in the buffer as the default top-left position. The caller might use the pixel buffer
// with APIs that expect the properties.
constexpr size_t macroblockSize = 16;
auto width = IOSurfaceGetWidth(surface);
auto height = IOSurfaceGetHeight(surface);
auto extendedRight = roundUpToMultipleOf(macroblockSize, width) - width;
auto extendedBottom = roundUpToMultipleOf(macroblockSize, height) - height;
if ((IOSurfaceGetBytesPerRowOfPlane(surface, 0) >= width + extendedRight)
&& (IOSurfaceGetBytesPerRowOfPlane(surface, 1) >= width + extendedRight)
&& (IOSurfaceGetAllocSize(surface) >= (height + extendedBottom) * IOSurfaceGetBytesPerRowOfPlane(surface, 0) * 3 / 2)) {
return (__bridge CFDictionaryRef) @{
(__bridge NSString *)kCVPixelBufferOpenGLCompatibilityKey : @YES,
(__bridge NSString *)kCVPixelBufferExtendedPixelsRightKey : @(extendedRight),
(__bridge NSString *)kCVPixelBufferExtendedPixelsBottomKey : @(extendedBottom)
};
}
}
#else
UNUSED_PARAM(surface);
#endif
return (__bridge CFDictionaryRef) @{
#if PLATFORM(MAC)
(__bridge NSString *)kCVPixelBufferOpenGLCompatibilityKey : @YES
#endif
};
}
Expected<RetainPtr<CVPixelBufferRef>, CVReturn> createCVPixelBuffer(IOSurfaceRef surface)
{
CVPixelBufferRef pixelBuffer = nullptr;
auto status = CVPixelBufferCreateWithIOSurface(kCFAllocatorDefault, surface, pixelBufferCreationOptions(surface), &pixelBuffer);
if (status != kCVReturnSuccess || !pixelBuffer) {
RELEASE_LOG_ERROR(WebRTC, "createCVPixelBuffer failed with IOSurface status=%d, pixelBuffer=%p", (int)status, pixelBuffer);
return makeUnexpected(status);
}
return adoptCF(pixelBuffer);
}
RetainPtr<CGColorSpaceRef> createCGColorSpaceForCVPixelBuffer(CVPixelBufferRef buffer)
{
if (CGColorSpaceRef colorSpace = dynamic_cf_cast<CGColorSpaceRef>(CVBufferGetAttachment(buffer, kCVImageBufferCGColorSpaceKey, nullptr)))
return colorSpace;
RetainPtr<CFDictionaryRef> attachments;
#if HAVE(CVBUFFERCOPYATTACHMENTS)
attachments = adoptCF(CVBufferCopyAttachments(buffer, kCVAttachmentMode_ShouldPropagate));
#else
attachments = CVBufferGetAttachments(buffer, kCVAttachmentMode_ShouldPropagate);
#endif
if (auto colorSpace = adoptCF(CVImageBufferCreateColorSpaceFromAttachments(attachments.get())))
return colorSpace;
// We should only get here with content that has a broken embedded ICC
// profile; in all other cases VideoToolbox should have put either known
// accurate or guessed color space attachments on the pixel buffer. Content
// that requires an embedded ICC profile is unlikely to be presented
// correctly with any particular fallback color space we choose, so we
// choose sRGB for ease.
return sRGBColorSpaceRef();
}
void setOwnershipIdentityForCVPixelBuffer(CVPixelBufferRef pixelBuffer, const ProcessIdentity& owner)
{
auto surface = CVPixelBufferGetIOSurface(pixelBuffer);
ASSERT(surface);
IOSurface::setOwnershipIdentity(surface, owner);
}
RetainPtr<CVPixelBufferRef> createBlackPixelBuffer(size_t width, size_t height, bool shouldUseIOSurface)
{
OSType format = preferedPixelBufferFormat();
ASSERT(format == kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange || format == kCVPixelFormatType_420YpCbCr8BiPlanarFullRange);
NSDictionary *pixelAttributes = @{ (__bridge NSString *)kCVPixelBufferIOSurfacePropertiesKey : @{ } };
CVPixelBufferRef pixelBuffer = nullptr;
auto status = CVPixelBufferCreate(kCFAllocatorDefault, width, height, format, shouldUseIOSurface ? (__bridge CFDictionaryRef)pixelAttributes : nullptr, &pixelBuffer);
ASSERT_UNUSED(status, status == noErr);
status = CVPixelBufferLockBaseAddress(pixelBuffer, 0);
ASSERT(status == noErr);
auto* yPlane = static_cast<uint8_t*>(CVPixelBufferGetBaseAddressOfPlane(pixelBuffer, 0));
size_t yStride = CVPixelBufferGetBytesPerRowOfPlane(pixelBuffer, 0);
for (unsigned i = 0; i < height; ++i)
memset(&yPlane[i * yStride], 0, width);
auto* uvPlane = static_cast<uint8_t*>(CVPixelBufferGetBaseAddressOfPlane(pixelBuffer, 1));
size_t uvStride = CVPixelBufferGetBytesPerRowOfPlane(pixelBuffer, 1);
for (unsigned i = 0; i < height / 2; ++i)
memset(&uvPlane[i * uvStride], 128, width);
status = CVPixelBufferUnlockBaseAddress(pixelBuffer, 0);
ASSERT(!status);
return adoptCF(pixelBuffer);
}
}