blob: b81dbb23eff4533a07255788a26eea7c33e05ae3 [file] [log] [blame]
/*
* Copyright (C) 2020 Apple Inc. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS''
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
* THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
* THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "config.h"
#include "SpeechRecognitionRealtimeMediaSourceManager.h"
#if ENABLE(MEDIA_STREAM)
#include "Logging.h"
#include "SpeechRecognitionRealtimeMediaSourceManagerMessages.h"
#include "SpeechRecognitionRemoteRealtimeMediaSourceManagerMessages.h"
#include "WebProcess.h"
#include <WebCore/RealtimeMediaSource.h>
#include <WebCore/SpeechRecognitionCaptureSource.h>
#if PLATFORM(COCOA)
#include "SharedRingBufferStorage.h"
#include <WebCore/CAAudioStreamDescription.h>
#include <WebCore/WebAudioBufferList.h>
#else
#include <WebCore/AudioStreamDescription.h>
#include <WebCore/PlatformAudioData.h>
#endif
#if USE(AUDIO_SESSION)
#include <WebCore/AudioSession.h>
#endif
namespace WebKit {
using namespace WebCore;
class SpeechRecognitionRealtimeMediaSourceManager::Source
: private RealtimeMediaSource::Observer
, private RealtimeMediaSource::AudioSampleObserver
{
WTF_MAKE_FAST_ALLOCATED;
public:
Source(RealtimeMediaSourceIdentifier identifier, Ref<RealtimeMediaSource>&& source, Ref<IPC::Connection>&& connection)
: m_identifier(identifier)
, m_source(WTFMove(source))
, m_connection(WTFMove(connection))
#if PLATFORM(COCOA)
, m_ringBuffer(makeUniqueRef<SharedRingBufferStorage>(std::bind(&Source::storageChanged, this, std::placeholders::_1, std::placeholders::_2, std::placeholders::_3)))
#endif
{
m_source->addObserver(*this);
m_source->addAudioSampleObserver(*this);
}
~Source()
{
#if PLATFORM(COCOA)
storage().invalidate();
#endif
m_source->removeAudioSampleObserver(*this);
m_source->removeObserver(*this);
}
void start()
{
m_source->start();
}
void stop()
{
m_source->stop();
}
#if PLATFORM(COCOA)
SharedRingBufferStorage& storage()
{
return static_cast<SharedRingBufferStorage&>(m_ringBuffer.storage());
}
#endif
private:
void sourceStopped() final
{
if (m_source->captureDidFail()) {
m_connection->send(Messages::SpeechRecognitionRemoteRealtimeMediaSourceManager::RemoteCaptureFailed(m_identifier), 0);
return;
}
m_connection->send(Messages::SpeechRecognitionRemoteRealtimeMediaSourceManager::RemoteSourceStopped(m_identifier), 0);
}
void audioSamplesAvailable(const MediaTime& time, const PlatformAudioData& audioData, const AudioStreamDescription& description, size_t numberOfFrames) final
{
#if PLATFORM(COCOA)
DisableMallocRestrictionsForCurrentThreadScope scope;
if (m_description != description) {
ASSERT(description.platformDescription().type == PlatformDescription::CAAudioStreamBasicType);
m_description = *std::get<const AudioStreamBasicDescription*>(description.platformDescription().description);
m_ringBuffer.allocate(m_description.streamDescription(), m_description.sampleRate() * 2);
}
ASSERT(is<WebAudioBufferList>(audioData));
m_ringBuffer.store(downcast<WebAudioBufferList>(audioData).list(), numberOfFrames, time.timeValue());
m_connection->send(Messages::SpeechRecognitionRemoteRealtimeMediaSourceManager::RemoteAudioSamplesAvailable(m_identifier, time, numberOfFrames), 0);
#else
UNUSED_PARAM(time);
UNUSED_PARAM(audioData);
UNUSED_PARAM(description);
UNUSED_PARAM(numberOfFrames);
#endif
}
#if PLATFORM(COCOA)
void storageChanged(SharedMemory* storage, const WebCore::CAAudioStreamDescription& description, size_t numberOfFrames)
{
DisableMallocRestrictionsForCurrentThreadScope scope;
SharedMemory::Handle handle;
if (storage)
storage->createHandle(handle, SharedMemory::Protection::ReadOnly);
#if OS(DARWIN) || OS(WINDOWS)
uint64_t dataSize = handle.size();
#else
uint64_t dataSize = 0;
#endif
m_connection->send(Messages::SpeechRecognitionRemoteRealtimeMediaSourceManager::SetStorage(m_identifier, SharedMemory::IPCHandle { WTFMove(handle), dataSize }, description, numberOfFrames), 0);
}
#endif
void audioUnitWillStart() final
{
#if USE(AUDIO_SESSION)
auto bufferSize = AudioSession::sharedSession().sampleRate() / 50;
if (AudioSession::sharedSession().preferredBufferSize() > bufferSize)
AudioSession::sharedSession().setPreferredBufferSize(bufferSize);
AudioSession::sharedSession().setCategory(AudioSession::CategoryType::PlayAndRecord, RouteSharingPolicy::Default);
#endif
}
RealtimeMediaSourceIdentifier m_identifier;
Ref<RealtimeMediaSource> m_source;
Ref<IPC::Connection> m_connection;
#if PLATFORM(COCOA)
CARingBuffer m_ringBuffer;
CAAudioStreamDescription m_description { };
#endif
};
SpeechRecognitionRealtimeMediaSourceManager::SpeechRecognitionRealtimeMediaSourceManager(Ref<IPC::Connection>&& connection)
: m_connection(WTFMove(connection))
{
WebProcess::singleton().addMessageReceiver(Messages::SpeechRecognitionRealtimeMediaSourceManager::messageReceiverName(), *this);
}
SpeechRecognitionRealtimeMediaSourceManager::~SpeechRecognitionRealtimeMediaSourceManager()
{
WebProcess::singleton().removeMessageReceiver(*this);
}
#if ENABLE(SANDBOX_EXTENSIONS)
void SpeechRecognitionRealtimeMediaSourceManager::grantSandboxExtensions(SandboxExtension::Handle&& sandboxHandleForTCCD, SandboxExtension::Handle&& sandboxHandleForMicrophone)
{
m_sandboxExtensionForTCCD = SandboxExtension::create(WTFMove(sandboxHandleForTCCD));
if (!m_sandboxExtensionForTCCD)
RELEASE_LOG_ERROR(Media, "Failed to create sandbox extension for tccd");
else
m_sandboxExtensionForTCCD->consume();
m_sandboxExtensionForMicrophone = SandboxExtension::create(WTFMove(sandboxHandleForMicrophone));
if (!m_sandboxExtensionForMicrophone)
RELEASE_LOG_ERROR(Media, "Failed to create sandbox extension for microphone");
else
m_sandboxExtensionForMicrophone->consume();
}
void SpeechRecognitionRealtimeMediaSourceManager::revokeSandboxExtensions()
{
if (m_sandboxExtensionForTCCD) {
m_sandboxExtensionForTCCD->revoke();
m_sandboxExtensionForTCCD = nullptr;
}
if (m_sandboxExtensionForMicrophone) {
m_sandboxExtensionForMicrophone->revoke();
m_sandboxExtensionForMicrophone = nullptr;
}
}
#endif
void SpeechRecognitionRealtimeMediaSourceManager::createSource(RealtimeMediaSourceIdentifier identifier, const CaptureDevice& device)
{
auto result = SpeechRecognitionCaptureSource::createRealtimeMediaSource(device);
if (!result) {
RELEASE_LOG_ERROR(Media, "Failed to create realtime source");
send(Messages::SpeechRecognitionRemoteRealtimeMediaSourceManager::RemoteCaptureFailed(identifier), 0);
return;
}
ASSERT(!m_sources.contains(identifier));
m_sources.add(identifier, makeUnique<Source>(identifier, result.source(), *messageSenderConnection()));
}
void SpeechRecognitionRealtimeMediaSourceManager::deleteSource(RealtimeMediaSourceIdentifier identifier)
{
m_sources.remove(identifier);
}
void SpeechRecognitionRealtimeMediaSourceManager::start(RealtimeMediaSourceIdentifier identifier)
{
if (auto source = m_sources.get(identifier))
source->start();
}
void SpeechRecognitionRealtimeMediaSourceManager::stop(RealtimeMediaSourceIdentifier identifier)
{
if (auto source = m_sources.get(identifier))
source->stop();
}
IPC::Connection* SpeechRecognitionRealtimeMediaSourceManager::messageSenderConnection() const
{
return m_connection.ptr();
}
uint64_t SpeechRecognitionRealtimeMediaSourceManager::messageSenderDestinationID() const
{
return 0;
}
} // namespace WebKit
#endif