blob: dd5397c2ad1ebacb6e2dd04982397116bc68eeb7 [file] [log] [blame]
/*
* Copyright (C) 2013 Google Inc. All rights reserved.
* Copyright (C) 2013-2019 Apple Inc. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following disclaimer
* in the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Google Inc. nor the names of its
* contributors may be used to endorse or promote products derived from
* this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "config.h"
#include "SourceBuffer.h"
#if ENABLE(MEDIA_SOURCE)
#include "AudioTrackList.h"
#include "BufferSource.h"
#include "Event.h"
#include "EventNames.h"
#include "GenericEventQueue.h"
#include "HTMLMediaElement.h"
#include "InbandTextTrack.h"
#include "InbandTextTrackPrivate.h"
#include "Logging.h"
#include "MediaDescription.h"
#include "MediaSample.h"
#include "MediaSource.h"
#include "SampleMap.h"
#include "SourceBufferList.h"
#include "SourceBufferPrivate.h"
#include "TextTrackList.h"
#include "TimeRanges.h"
#include "VideoTrackList.h"
#include <JavaScriptCore/JSCInlines.h>
#include <JavaScriptCore/JSLock.h>
#include <JavaScriptCore/VM.h>
#include <limits>
#include <wtf/CheckedArithmetic.h>
#include <wtf/IsoMallocInlines.h>
#include <wtf/WeakPtr.h>
namespace WebCore {
WTF_MAKE_ISO_ALLOCATED_IMPL(SourceBuffer);
static const double ExponentialMovingAverageCoefficient = 0.1;
struct SourceBuffer::TrackBuffer {
MediaTime lastDecodeTimestamp;
MediaTime greatestDecodeDuration;
MediaTime lastFrameDuration;
MediaTime highestPresentationTimestamp;
MediaTime lastEnqueuedPresentationTime;
MediaTime minimumEnqueuedPresentationTime;
DecodeOrderSampleMap::KeyType lastEnqueuedDecodeKey;
MediaTime lastEnqueuedDecodeDuration;
MediaTime roundedTimestampOffset;
uint32_t lastFrameTimescale { 0 };
bool needRandomAccessFlag { true };
bool enabled { false };
bool needsReenqueueing { false };
bool needsMinimumUpcomingPresentationTimeUpdating { false };
SampleMap samples;
DecodeOrderSampleMap::MapType decodeQueue;
RefPtr<MediaDescription> description;
PlatformTimeRanges buffered;
TrackBuffer()
: lastDecodeTimestamp(MediaTime::invalidTime())
, greatestDecodeDuration(MediaTime::invalidTime())
, lastFrameDuration(MediaTime::invalidTime())
, highestPresentationTimestamp(MediaTime::invalidTime())
, lastEnqueuedPresentationTime(MediaTime::invalidTime())
, lastEnqueuedDecodeKey({MediaTime::invalidTime(), MediaTime::invalidTime()})
, lastEnqueuedDecodeDuration(MediaTime::invalidTime())
{
}
};
Ref<SourceBuffer> SourceBuffer::create(Ref<SourceBufferPrivate>&& sourceBufferPrivate, MediaSource* source)
{
auto sourceBuffer = adoptRef(*new SourceBuffer(WTFMove(sourceBufferPrivate), source));
sourceBuffer->suspendIfNeeded();
return sourceBuffer;
}
SourceBuffer::SourceBuffer(Ref<SourceBufferPrivate>&& sourceBufferPrivate, MediaSource* source)
: ActiveDOMObject(source->scriptExecutionContext())
, m_private(WTFMove(sourceBufferPrivate))
, m_source(source)
, m_asyncEventQueue(MainThreadGenericEventQueue::create(*this))
, m_appendBufferTimer(*this, &SourceBuffer::appendBufferTimerFired)
, m_appendWindowStart(MediaTime::zeroTime())
, m_appendWindowEnd(MediaTime::positiveInfiniteTime())
, m_groupStartTimestamp(MediaTime::invalidTime())
, m_groupEndTimestamp(MediaTime::zeroTime())
, m_buffered(TimeRanges::create())
, m_appendState(WaitingForSegment)
, m_timeOfBufferingMonitor(MonotonicTime::now())
, m_pendingRemoveStart(MediaTime::invalidTime())
, m_pendingRemoveEnd(MediaTime::invalidTime())
, m_removeTimer(*this, &SourceBuffer::removeTimerFired)
#if !RELEASE_LOG_DISABLED
, m_logger(m_private->sourceBufferLogger())
, m_logIdentifier(m_private->sourceBufferLogIdentifier())
#endif
{
ASSERT(m_source);
ALWAYS_LOG(LOGIDENTIFIER);
m_private->setClient(this);
}
SourceBuffer::~SourceBuffer()
{
ASSERT(isRemoved());
ALWAYS_LOG(LOGIDENTIFIER);
m_private->setClient(nullptr);
}
ExceptionOr<Ref<TimeRanges>> SourceBuffer::buffered() const
{
// Section 3.1 buffered attribute steps.
// https://dvcs.w3.org/hg/html-media/raw-file/default/media-source/media-source.html#attributes-1
// 1. If this object has been removed from the sourceBuffers attribute of the parent media source then throw an
// InvalidStateError exception and abort these steps.
if (isRemoved())
return Exception { InvalidStateError };
// 2. Return a new static normalized TimeRanges object for the media segments buffered.
return m_buffered->copy();
}
double SourceBuffer::timestampOffset() const
{
return m_timestampOffset.toDouble();
}
ExceptionOr<void> SourceBuffer::setTimestampOffset(double offset)
{
// Section 3.1 timestampOffset attribute setter steps.
// https://dvcs.w3.org/hg/html-media/raw-file/default/media-source/media-source.html#attributes-1
// 1. Let new timestamp offset equal the new value being assigned to this attribute.
// 2. If this object has been removed from the sourceBuffers attribute of the parent media source, then throw an
// InvalidStateError exception and abort these steps.
// 3. If the updating attribute equals true, then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || m_updating)
return Exception { InvalidStateError };
// 4. If the readyState attribute of the parent media source is in the "ended" state then run the following steps:
// 4.1 Set the readyState attribute of the parent media source to "open"
// 4.2 Queue a task to fire a simple event named sourceopen at the parent media source.
m_source->openIfInEndedState();
// 5. If the append state equals PARSING_MEDIA_SEGMENT, then throw an InvalidStateError and abort these steps.
if (m_appendState == ParsingMediaSegment)
return Exception { InvalidStateError };
MediaTime newTimestampOffset = MediaTime::createWithDouble(offset);
// 6. If the mode attribute equals "sequence", then set the group start timestamp to new timestamp offset.
if (m_mode == AppendMode::Sequence)
m_groupStartTimestamp = newTimestampOffset;
// 7. Update the attribute to the new value.
m_timestampOffset = newTimestampOffset;
for (auto& trackBuffer : m_trackBufferMap.values()) {
trackBuffer.lastFrameTimescale = 0;
trackBuffer.roundedTimestampOffset = MediaTime::invalidTime();
}
return { };
}
double SourceBuffer::appendWindowStart() const
{
return m_appendWindowStart.toDouble();
}
ExceptionOr<void> SourceBuffer::setAppendWindowStart(double newValue)
{
// Section 3.1 appendWindowStart attribute setter steps.
// W3C Editor's Draft 16 September 2016
// https://rawgit.com/w3c/media-source/45627646344eea0170dd1cbc5a3d508ca751abb8/media-source-respec.html#dom-sourcebuffer-appendwindowstart
// 1. If this object has been removed from the sourceBuffers attribute of the parent media source,
// then throw an InvalidStateError exception and abort these steps.
// 2. If the updating attribute equals true, then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || m_updating)
return Exception { InvalidStateError };
// 3. If the new value is less than 0 or greater than or equal to appendWindowEnd then
// throw an TypeError exception and abort these steps.
if (newValue < 0 || newValue >= m_appendWindowEnd.toDouble())
return Exception { TypeError };
// 4. Update the attribute to the new value.
m_appendWindowStart = MediaTime::createWithDouble(newValue);
return { };
}
double SourceBuffer::appendWindowEnd() const
{
return m_appendWindowEnd.toDouble();
}
ExceptionOr<void> SourceBuffer::setAppendWindowEnd(double newValue)
{
// Section 3.1 appendWindowEnd attribute setter steps.
// W3C Editor's Draft 16 September 2016
// https://rawgit.com/w3c/media-source/45627646344eea0170dd1cbc5a3d508ca751abb8/media-source-respec.html#dom-sourcebuffer-appendwindowend
// 1. If this object has been removed from the sourceBuffers attribute of the parent media source,
// then throw an InvalidStateError exception and abort these steps.
// 2. If the updating attribute equals true, then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || m_updating)
return Exception { InvalidStateError };
// 3. If the new value equals NaN, then throw an TypeError and abort these steps.
// 4. If the new value is less than or equal to appendWindowStart then throw an TypeError exception
// and abort these steps.
if (std::isnan(newValue) || newValue <= m_appendWindowStart.toDouble())
return Exception { TypeError };
// 5.. Update the attribute to the new value.
m_appendWindowEnd = MediaTime::createWithDouble(newValue);
return { };
}
ExceptionOr<void> SourceBuffer::appendBuffer(const BufferSource& data)
{
return appendBufferInternal(static_cast<const unsigned char*>(data.data()), data.length());
}
void SourceBuffer::resetParserState()
{
// Section 3.5.2 Reset Parser State algorithm steps.
// http://www.w3.org/TR/2014/CR-media-source-20140717/#sourcebuffer-reset-parser-state
// 1. If the append state equals PARSING_MEDIA_SEGMENT and the input buffer contains some complete coded frames,
// then run the coded frame processing algorithm until all of these complete coded frames have been processed.
// FIXME: If any implementation will work in pulling mode (instead of async push to SourceBufferPrivate, and forget)
// this should be handled somehow either here, or in m_private->abort();
// 2. Unset the last decode timestamp on all track buffers.
// 3. Unset the last frame duration on all track buffers.
// 4. Unset the highest presentation timestamp on all track buffers.
// 5. Set the need random access point flag on all track buffers to true.
for (auto& trackBufferPair : m_trackBufferMap.values()) {
trackBufferPair.lastDecodeTimestamp = MediaTime::invalidTime();
trackBufferPair.greatestDecodeDuration = MediaTime::invalidTime();
trackBufferPair.lastFrameDuration = MediaTime::invalidTime();
trackBufferPair.highestPresentationTimestamp = MediaTime::invalidTime();
trackBufferPair.needRandomAccessFlag = true;
}
// 6. Remove all bytes from the input buffer.
// Note: this is handled by abortIfUpdating()
// 7. Set append state to WAITING_FOR_SEGMENT.
m_appendState = WaitingForSegment;
m_private->resetParserState();
}
ExceptionOr<void> SourceBuffer::abort()
{
// Section 3.2 abort() method steps.
// https://rawgit.com/w3c/media-source/45627646344eea0170dd1cbc5a3d508ca751abb8/media-source-respec.html#dom-sourcebuffer-abort
// 1. If this object has been removed from the sourceBuffers attribute of the parent media source
// then throw an InvalidStateError exception and abort these steps.
// 2. If the readyState attribute of the parent media source is not in the "open" state
// then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || !m_source->isOpen())
return Exception { InvalidStateError };
// 3. If the range removal algorithm is running, then throw an InvalidStateError exception and abort these steps.
if (m_removeTimer.isActive())
return Exception { InvalidStateError };
// 4. If the sourceBuffer.updating attribute equals true, then run the following steps: ...
abortIfUpdating();
// 5. Run the reset parser state algorithm.
resetParserState();
// 6. Set appendWindowStart to the presentation start time.
m_appendWindowStart = MediaTime::zeroTime();
// 7. Set appendWindowEnd to positive Infinity.
m_appendWindowEnd = MediaTime::positiveInfiniteTime();
return { };
}
ExceptionOr<void> SourceBuffer::remove(double start, double end)
{
return remove(MediaTime::createWithDouble(start), MediaTime::createWithDouble(end));
}
ExceptionOr<void> SourceBuffer::remove(const MediaTime& start, const MediaTime& end)
{
DEBUG_LOG(LOGIDENTIFIER, "start = ", start, ", end = ", end);
// https://rawgit.com/w3c/media-source/45627646344eea0170dd1cbc5a3d508ca751abb8/media-source-respec.html#dom-sourcebuffer-remove
// Section 3.2 remove() method steps.
// 1. If this object has been removed from the sourceBuffers attribute of the parent media source then throw
// an InvalidStateError exception and abort these steps.
// 2. If the updating attribute equals true, then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || m_updating)
return Exception { InvalidStateError };
// 3. If duration equals NaN, then throw a TypeError exception and abort these steps.
// 4. If start is negative or greater than duration, then throw a TypeError exception and abort these steps.
// 5. If end is less than or equal to start or end equals NaN, then throw a TypeError exception and abort these steps.
if (m_source->duration().isInvalid()
|| end.isInvalid()
|| start.isInvalid()
|| start < MediaTime::zeroTime()
|| start > m_source->duration()
|| end <= start) {
return Exception { TypeError };
}
// 6. If the readyState attribute of the parent media source is in the "ended" state then run the following steps:
// 6.1. Set the readyState attribute of the parent media source to "open"
// 6.2. Queue a task to fire a simple event named sourceopen at the parent media source .
m_source->openIfInEndedState();
// 7. Run the range removal algorithm with start and end as the start and end of the removal range.
rangeRemoval(start, end);
return { };
}
void SourceBuffer::rangeRemoval(const MediaTime& start, const MediaTime& end)
{
// 3.5.7 Range Removal
// https://rawgit.com/w3c/media-source/7bbe4aa33c61ec025bc7acbd80354110f6a000f9/media-source.html#sourcebuffer-range-removal
// 1. Let start equal the starting presentation timestamp for the removal range.
// 2. Let end equal the end presentation timestamp for the removal range.
// 3. Set the updating attribute to true.
m_updating = true;
// 4. Queue a task to fire a simple event named updatestart at this SourceBuffer object.
scheduleEvent(eventNames().updatestartEvent);
// 5. Return control to the caller and run the rest of the steps asynchronously.
m_pendingRemoveStart = start;
m_pendingRemoveEnd = end;
m_removeTimer.startOneShot(0_s);
}
ExceptionOr<void> SourceBuffer::changeType(const String& type)
{
// changeType() proposed API. See issue #155: <https://github.com/w3c/media-source/issues/155>
// https://rawgit.com/wicg/media-source/codec-switching/index.html#dom-sourcebuffer-changetype
// 1. If type is an empty string then throw a TypeError exception and abort these steps.
if (type.isEmpty())
return Exception { TypeError };
// 2. If this object has been removed from the sourceBuffers attribute of the parent media source,
// then throw an InvalidStateError exception and abort these steps.
// 3. If the updating attribute equals true, then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || m_updating)
return Exception { InvalidStateError };
// 4. If type contains a MIME type that is not supported or contains a MIME type that is not supported with
// the types specified (currently or previously) of SourceBuffer objects in the sourceBuffers attribute of
// the parent media source, then throw a NotSupportedError exception and abort these steps.
ContentType contentType(type);
if (!m_private->canSwitchToType(contentType))
return Exception { NotSupportedError };
// 5. If the readyState attribute of the parent media source is in the "ended" state then run the following
// steps:
// 5.1. Set the readyState attribute of the parent media source to "open"
// 5.2. Queue a task to fire a simple event named sourceopen at the parent media source.
m_source->openIfInEndedState();
// 6. Run the reset parser state algorithm.
resetParserState();
// 7. Update the generate timestamps flag on this SourceBuffer object to the value in the "Generate Timestamps
// Flag" column of the byte stream format registry [MSE-REGISTRY] entry that is associated with type.
setShouldGenerateTimestamps(MediaSource::contentTypeShouldGenerateTimestamps(contentType));
// ↳ If the generate timestamps flag equals true:
// Set the mode attribute on this SourceBuffer object to "sequence", including running the associated steps
// for that attribute being set.
if (m_shouldGenerateTimestamps)
setMode(AppendMode::Sequence);
// ↳ Otherwise:
// Keep the previous value of the mode attribute on this SourceBuffer object, without running any associated
// steps for that attribute being set.
// NOTE: No-op.
// 9. Set pending initialization segment for changeType flag to true.
m_pendingInitializationSegmentForChangeType = true;
return { };
}
void SourceBuffer::abortIfUpdating()
{
// Section 3.2 abort() method step 4 substeps.
// https://rawgit.com/w3c/media-source/45627646344eea0170dd1cbc5a3d508ca751abb8/media-source-respec.html#dom-sourcebuffer-abort
if (!m_updating)
return;
// 4.1. Abort the buffer append algorithm if it is running.
m_appendBufferTimer.stop();
m_pendingAppendData.clear();
m_private->abort();
// 4.2. Set the updating attribute to false.
m_updating = false;
// 4.3. Queue a task to fire a simple event named abort at this SourceBuffer object.
scheduleEvent(eventNames().abortEvent);
// 4.4. Queue a task to fire a simple event named updateend at this SourceBuffer object.
scheduleEvent(eventNames().updateendEvent);
}
MediaTime SourceBuffer::highestPresentationTimestamp() const
{
MediaTime highestTime;
for (auto& trackBuffer : m_trackBufferMap.values()) {
auto lastSampleIter = trackBuffer.samples.presentationOrder().rbegin();
if (lastSampleIter == trackBuffer.samples.presentationOrder().rend())
continue;
highestTime = std::max(highestTime, lastSampleIter->first);
}
return highestTime;
}
void SourceBuffer::readyStateChanged()
{
updateBufferedFromTrackBuffers();
}
void SourceBuffer::removedFromMediaSource()
{
if (isRemoved())
return;
abortIfUpdating();
for (auto& trackBufferPair : m_trackBufferMap.values()) {
trackBufferPair.samples.clear();
trackBufferPair.decodeQueue.clear();
}
m_private->removedFromMediaSource();
m_source = nullptr;
}
void SourceBuffer::seekToTime(const MediaTime& time)
{
ALWAYS_LOG(LOGIDENTIFIER, time);
for (auto& trackBufferPair : m_trackBufferMap) {
TrackBuffer& trackBuffer = trackBufferPair.value;
const AtomString& trackID = trackBufferPair.key;
trackBuffer.needsReenqueueing = true;
reenqueueMediaForTime(trackBuffer, trackID, time);
}
}
MediaTime SourceBuffer::sourceBufferPrivateFastSeekTimeForMediaTime(const MediaTime& targetTime, const MediaTime& negativeThreshold, const MediaTime& positiveThreshold)
{
MediaTime seekTime = targetTime;
for (auto& trackBuffer : m_trackBufferMap.values()) {
// Find the sample which contains the target time time.
auto futureSyncSampleIterator = trackBuffer.samples.decodeOrder().findSyncSampleAfterPresentationTime(targetTime, positiveThreshold);
auto pastSyncSampleIterator = trackBuffer.samples.decodeOrder().findSyncSamplePriorToPresentationTime(targetTime, negativeThreshold);
auto upperBound = trackBuffer.samples.decodeOrder().end();
auto lowerBound = trackBuffer.samples.decodeOrder().rend();
if (futureSyncSampleIterator == upperBound && pastSyncSampleIterator == lowerBound)
continue;
MediaTime futureSeekTime = MediaTime::positiveInfiniteTime();
if (futureSyncSampleIterator != upperBound) {
RefPtr<MediaSample>& sample = futureSyncSampleIterator->second;
futureSeekTime = sample->presentationTime();
}
MediaTime pastSeekTime = MediaTime::negativeInfiniteTime();
if (pastSyncSampleIterator != lowerBound) {
RefPtr<MediaSample>& sample = pastSyncSampleIterator->second;
pastSeekTime = sample->presentationTime();
}
MediaTime trackSeekTime = abs(targetTime - futureSeekTime) < abs(targetTime - pastSeekTime) ? futureSeekTime : pastSeekTime;
if (abs(targetTime - trackSeekTime) > abs(targetTime - seekTime))
seekTime = trackSeekTime;
}
return seekTime;
}
bool SourceBuffer::hasPendingActivity() const
{
return m_source || m_asyncEventQueue->hasPendingEvents();
}
void SourceBuffer::stop()
{
m_appendBufferTimer.stop();
m_removeTimer.stop();
}
const char* SourceBuffer::activeDOMObjectName() const
{
return "SourceBuffer";
}
bool SourceBuffer::isRemoved() const
{
return !m_source;
}
void SourceBuffer::scheduleEvent(const AtomString& eventName)
{
auto event = Event::create(eventName, Event::CanBubble::No, Event::IsCancelable::No);
event->setTarget(this);
m_asyncEventQueue->enqueueEvent(WTFMove(event));
}
ExceptionOr<void> SourceBuffer::appendBufferInternal(const unsigned char* data, unsigned size)
{
// Section 3.2 appendBuffer()
// https://dvcs.w3.org/hg/html-media/raw-file/default/media-source/media-source.html#widl-SourceBuffer-appendBuffer-void-ArrayBufferView-data
// Step 1 is enforced by the caller.
// 2. Run the prepare append algorithm.
// Section 3.5.4 Prepare AppendAlgorithm
// 1. If the SourceBuffer has been removed from the sourceBuffers attribute of the parent media source
// then throw an InvalidStateError exception and abort these steps.
// 2. If the updating attribute equals true, then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || m_updating)
return Exception { InvalidStateError };
// 3. If the readyState attribute of the parent media source is in the "ended" state then run the following steps:
// 3.1. Set the readyState attribute of the parent media source to "open"
// 3.2. Queue a task to fire a simple event named sourceopen at the parent media source .
m_source->openIfInEndedState();
// 4. Run the coded frame eviction algorithm.
evictCodedFrames(size);
// FIXME: enable this code when MSE libraries have been updated to support it.
#if USE(GSTREAMER)
// 5. If the buffer full flag equals true, then throw a QuotaExceededError exception and abort these step.
if (m_bufferFull) {
ERROR_LOG(LOGIDENTIFIER, "buffer full, failing with QuotaExceededError error");
return Exception { QuotaExceededError };
}
#endif
// NOTE: Return to 3.2 appendBuffer()
// 3. Add data to the end of the input buffer.
m_pendingAppendData.append(data, size);
// 4. Set the updating attribute to true.
m_updating = true;
// 5. Queue a task to fire a simple event named updatestart at this SourceBuffer object.
scheduleEvent(eventNames().updatestartEvent);
// 6. Asynchronously run the buffer append algorithm.
m_appendBufferTimer.startOneShot(0_s);
reportExtraMemoryAllocated();
return { };
}
void SourceBuffer::appendBufferTimerFired()
{
if (isRemoved())
return;
ASSERT(m_updating);
// Section 3.5.5 Buffer Append Algorithm
// https://dvcs.w3.org/hg/html-media/raw-file/default/media-source/media-source.html#sourcebuffer-buffer-append
// 1. Run the segment parser loop algorithm.
// Section 3.5.1 Segment Parser Loop
// https://dvcs.w3.org/hg/html-media/raw-file/tip/media-source/media-source.html#sourcebuffer-segment-parser-loop
// When the segment parser loop algorithm is invoked, run the following steps:
// 1. Loop Top: If the input buffer is empty, then jump to the need more data step below.
if (!m_pendingAppendData.size()) {
sourceBufferPrivateAppendComplete(AppendSucceeded);
return;
}
// Manually clear out the m_pendingAppendData Vector, in case the platform implementation
// rejects appending the buffer for whatever reason.
// FIXME: The implementation should guarantee the move from this Vector, and we should
// assert here to confirm that. See https://bugs.webkit.org/show_bug.cgi?id=178003.
m_private->append(WTFMove(m_pendingAppendData));
m_pendingAppendData.clear();
}
void SourceBuffer::sourceBufferPrivateAppendComplete(AppendResult result)
{
if (isRemoved())
return;
// Resolve the changes it TrackBuffers' buffered ranges
// into the SourceBuffer's buffered ranges
updateBufferedFromTrackBuffers();
// Section 3.5.5 Buffer Append Algorithm, ctd.
// https://dvcs.w3.org/hg/html-media/raw-file/default/media-source/media-source.html#sourcebuffer-buffer-append
// 2. If the input buffer contains bytes that violate the SourceBuffer byte stream format specification,
// then run the append error algorithm with the decode error parameter set to true and abort this algorithm.
if (result == ParsingFailed) {
ERROR_LOG(LOGIDENTIFIER, "ParsingFailed");
appendError(true);
return;
}
// NOTE: Steps 3 - 6 enforced by sourceBufferPrivateDidReceiveInitializationSegment() and
// sourceBufferPrivateDidReceiveSample below.
// 7. Need more data: Return control to the calling algorithm.
// NOTE: return to Section 3.5.5
// 2.If the segment parser loop algorithm in the previous step was aborted, then abort this algorithm.
if (result != AppendSucceeded)
return;
// 3. Set the updating attribute to false.
m_updating = false;
// 4. Queue a task to fire a simple event named update at this SourceBuffer object.
scheduleEvent(eventNames().updateEvent);
// 5. Queue a task to fire a simple event named updateend at this SourceBuffer object.
scheduleEvent(eventNames().updateendEvent);
if (m_source)
m_source->monitorSourceBuffers();
MediaTime currentMediaTime = m_source->currentTime();
for (auto& trackBufferPair : m_trackBufferMap) {
TrackBuffer& trackBuffer = trackBufferPair.value;
const AtomString& trackID = trackBufferPair.key;
if (trackBuffer.needsReenqueueing) {
DEBUG_LOG(LOGIDENTIFIER, "reenqueuing at time ", currentMediaTime);
reenqueueMediaForTime(trackBuffer, trackID, currentMediaTime);
} else
provideMediaData(trackBuffer, trackID);
}
reportExtraMemoryAllocated();
if (extraMemoryCost() > this->maximumBufferSize())
m_bufferFull = true;
DEBUG_LOG(LOGIDENTIFIER);
}
void SourceBuffer::sourceBufferPrivateDidReceiveRenderingError(int error)
{
#if RELEASE_LOG_DISABLED
UNUSED_PARAM(error);
#endif
ERROR_LOG(LOGIDENTIFIER, error);
if (!isRemoved())
m_source->streamEndedWithError(MediaSource::EndOfStreamError::Decode);
}
static bool decodeTimeComparator(const PresentationOrderSampleMap::MapType::value_type& a, const PresentationOrderSampleMap::MapType::value_type& b)
{
return a.second->decodeTime() < b.second->decodeTime();
}
static PlatformTimeRanges removeSamplesFromTrackBuffer(const DecodeOrderSampleMap::MapType& samples, SourceBuffer::TrackBuffer& trackBuffer, const SourceBuffer* buffer, const char* logPrefix)
{
#if !RELEASE_LOG_DISABLED
MediaTime earliestSample = MediaTime::positiveInfiniteTime();
MediaTime latestSample = MediaTime::zeroTime();
size_t bytesRemoved = 0;
auto logIdentifier = WTF::Logger::LogSiteIdentifier(buffer->logClassName(), logPrefix, buffer->logIdentifier());
auto& logger = buffer->logger();
auto willLog = logger.willLog(buffer->logChannel(), WTFLogLevel::Debug);
#else
UNUSED_PARAM(logPrefix);
UNUSED_PARAM(buffer);
#endif
PlatformTimeRanges erasedRanges;
for (const auto& sampleIt : samples) {
const DecodeOrderSampleMap::KeyType& decodeKey = sampleIt.first;
#if !RELEASE_LOG_DISABLED
size_t startBufferSize = trackBuffer.samples.sizeInBytes();
#endif
const RefPtr<MediaSample>& sample = sampleIt.second;
#if !RELEASE_LOG_DISABLED
if (willLog)
logger.debug(buffer->logChannel(), logIdentifier, "removing sample ", *sampleIt.second);
#endif
// Remove the erased samples from the TrackBuffer sample map.
trackBuffer.samples.removeSample(sample.get());
// Also remove the erased samples from the TrackBuffer decodeQueue.
trackBuffer.decodeQueue.erase(decodeKey);
auto startTime = sample->presentationTime();
auto endTime = startTime + sample->duration();
erasedRanges.add(startTime, endTime);
#if !RELEASE_LOG_DISABLED
bytesRemoved += startBufferSize - trackBuffer.samples.sizeInBytes();
if (startTime < earliestSample)
earliestSample = startTime;
if (endTime > latestSample)
latestSample = endTime;
#endif
}
// Because we may have added artificial padding in the buffered ranges when adding samples, we may
// need to remove that padding when removing those same samples. Walk over the erased ranges looking
// for unbuffered areas and expand erasedRanges to encompass those areas.
PlatformTimeRanges additionalErasedRanges;
for (unsigned i = 0; i < erasedRanges.length(); ++i) {
auto erasedStart = erasedRanges.start(i);
auto erasedEnd = erasedRanges.end(i);
auto startIterator = trackBuffer.samples.presentationOrder().reverseFindSampleBeforePresentationTime(erasedStart);
if (startIterator == trackBuffer.samples.presentationOrder().rend())
additionalErasedRanges.add(MediaTime::zeroTime(), erasedStart);
else {
auto& previousSample = *startIterator->second;
if (previousSample.presentationTime() + previousSample.duration() < erasedStart)
additionalErasedRanges.add(previousSample.presentationTime() + previousSample.duration(), erasedStart);
}
auto endIterator = trackBuffer.samples.presentationOrder().findSampleStartingOnOrAfterPresentationTime(erasedEnd);
if (endIterator == trackBuffer.samples.presentationOrder().end())
additionalErasedRanges.add(erasedEnd, MediaTime::positiveInfiniteTime());
else {
auto& nextSample = *endIterator->second;
if (nextSample.presentationTime() > erasedEnd)
additionalErasedRanges.add(erasedEnd, nextSample.presentationTime());
}
}
if (additionalErasedRanges.length())
erasedRanges.unionWith(additionalErasedRanges);
#if !RELEASE_LOG_DISABLED
if (bytesRemoved && willLog)
logger.debug(buffer->logChannel(), logIdentifier, "removed ", bytesRemoved, ", start = ", earliestSample, ", end = ", latestSample);
#endif
return erasedRanges;
}
void SourceBuffer::removeCodedFrames(const MediaTime& start, const MediaTime& end)
{
DEBUG_LOG(LOGIDENTIFIER, "start = ", start, ", end = ", end);
ASSERT(start < end);
if (start >= end)
return;
// 3.5.9 Coded Frame Removal Algorithm
// https://dvcs.w3.org/hg/html-media/raw-file/tip/media-source/media-source.html#sourcebuffer-coded-frame-removal
// 1. Let start be the starting presentation timestamp for the removal range.
MediaTime currentMediaTime = m_source->currentTime();
// 2. Let end be the end presentation timestamp for the removal range.
// 3. For each track buffer in this source buffer, run the following steps:
for (auto& trackBufferKeyValue : m_trackBufferMap) {
TrackBuffer& trackBuffer = trackBufferKeyValue.value;
AtomString trackID = trackBufferKeyValue.key;
// 3.1. Let remove end timestamp be the current value of duration
// 3.2 If this track buffer has a random access point timestamp that is greater than or equal to end, then update
// remove end timestamp to that random access point timestamp.
// NOTE: Step 3.2 will be incorrect for any random access point timestamp whose decode time is later than the sample at end,
// but whose presentation time is less than the sample at end. Skip this step until step 3.3 below.
// NOTE: To handle MediaSamples which may be an amalgamation of multiple shorter samples, find samples whose presentation
// interval straddles the start and end times, and divide them if possible:
auto divideSampleIfPossibleAtPresentationTime = [&] (const MediaTime& time) {
auto sampleIterator = trackBuffer.samples.presentationOrder().findSampleContainingPresentationTime(time);
if (sampleIterator == trackBuffer.samples.presentationOrder().end())
return;
RefPtr<MediaSample> sample = sampleIterator->second;
if (!sample->isDivisable())
return;
std::pair<RefPtr<MediaSample>, RefPtr<MediaSample>> replacementSamples = sample->divide(time);
if (!replacementSamples.first || !replacementSamples.second)
return;
DEBUG_LOG(LOGIDENTIFIER, "splitting sample ", *sample, " into ", *replacementSamples.first, " and ", *replacementSamples.second);
trackBuffer.samples.removeSample(sample.get());
trackBuffer.samples.addSample(*replacementSamples.first);
trackBuffer.samples.addSample(*replacementSamples.second);
};
divideSampleIfPossibleAtPresentationTime(start);
divideSampleIfPossibleAtPresentationTime(end);
auto removePresentationStart = trackBuffer.samples.presentationOrder().findSampleContainingOrAfterPresentationTime(start);
auto removePresentationEnd = trackBuffer.samples.presentationOrder().findSampleStartingOnOrAfterPresentationTime(end);
if (removePresentationStart == removePresentationEnd)
continue;
// 3.3 Remove all media data, from this track buffer, that contain starting timestamps greater than or equal to
// start and less than the remove end timestamp.
// NOTE: frames must be removed in decode order, so that all dependant frames between the frame to be removed
// and the next sync sample frame are removed. But we must start from the first sample in decode order, not
// presentation order.
auto minmaxDecodeTimeIterPair = std::minmax_element(removePresentationStart, removePresentationEnd, decodeTimeComparator);
auto& firstSample = *minmaxDecodeTimeIterPair.first->second;
auto& lastSample = *minmaxDecodeTimeIterPair.second->second;
auto removeDecodeStart = trackBuffer.samples.decodeOrder().findSampleWithDecodeKey({firstSample.decodeTime(), firstSample.presentationTime()});
auto removeDecodeLast = trackBuffer.samples.decodeOrder().findSampleWithDecodeKey({lastSample.decodeTime(), lastSample.presentationTime()});
auto removeDecodeEnd = trackBuffer.samples.decodeOrder().findSyncSampleAfterDecodeIterator(removeDecodeLast);
DecodeOrderSampleMap::MapType erasedSamples(removeDecodeStart, removeDecodeEnd);
PlatformTimeRanges erasedRanges = removeSamplesFromTrackBuffer(erasedSamples, trackBuffer, this, "removeCodedFrames");
// Only force the TrackBuffer to re-enqueue if the removed ranges overlap with enqueued and possibly
// not yet displayed samples.
if (trackBuffer.lastEnqueuedPresentationTime.isValid() && currentMediaTime < trackBuffer.lastEnqueuedPresentationTime) {
PlatformTimeRanges possiblyEnqueuedRanges(currentMediaTime, trackBuffer.lastEnqueuedPresentationTime);
possiblyEnqueuedRanges.intersectWith(erasedRanges);
if (possiblyEnqueuedRanges.length()) {
trackBuffer.needsReenqueueing = true;
DEBUG_LOG(LOGIDENTIFIER, "the range in removeCodedFrames() includes already enqueued samples, reenqueueing from ", currentMediaTime);
reenqueueMediaForTime(trackBuffer, trackID, currentMediaTime);
}
}
erasedRanges.invert();
trackBuffer.buffered.intersectWith(erasedRanges);
setBufferedDirty(true);
// 3.4 If this object is in activeSourceBuffers, the current playback position is greater than or equal to start
// and less than the remove end timestamp, and HTMLMediaElement.readyState is greater than HAVE_METADATA, then set
// the HTMLMediaElement.readyState attribute to HAVE_METADATA and stall playback.
if (m_active && currentMediaTime >= start && currentMediaTime < end && m_private->readyState() > MediaPlayer::ReadyState::HaveMetadata)
m_private->setReadyState(MediaPlayer::ReadyState::HaveMetadata);
}
updateBufferedFromTrackBuffers();
// 4. If buffer full flag equals true and this object is ready to accept more bytes, then set the buffer full flag to false.
// No-op
LOG(Media, "SourceBuffer::removeCodedFrames(%p) - buffered = %s", this, toString(m_buffered->ranges()).utf8().data());
}
void SourceBuffer::removeTimerFired()
{
if (isRemoved())
return;
ASSERT(m_updating);
ASSERT(m_pendingRemoveStart.isValid());
ASSERT(m_pendingRemoveStart < m_pendingRemoveEnd);
// Section 3.5.7 Range Removal
// http://w3c.github.io/media-source/#sourcebuffer-range-removal
// 6. Run the coded frame removal algorithm with start and end as the start and end of the removal range.
removeCodedFrames(m_pendingRemoveStart, m_pendingRemoveEnd);
// 7. Set the updating attribute to false.
m_updating = false;
m_pendingRemoveStart = MediaTime::invalidTime();
m_pendingRemoveEnd = MediaTime::invalidTime();
// 8. Queue a task to fire a simple event named update at this SourceBuffer object.
scheduleEvent(eventNames().updateEvent);
// 9. Queue a task to fire a simple event named updateend at this SourceBuffer object.
scheduleEvent(eventNames().updateendEvent);
}
void SourceBuffer::evictCodedFrames(size_t newDataSize)
{
// 3.5.13 Coded Frame Eviction Algorithm
// http://www.w3.org/TR/media-source/#sourcebuffer-coded-frame-eviction
if (isRemoved())
return;
// This algorithm is run to free up space in this source buffer when new data is appended.
// 1. Let new data equal the data that is about to be appended to this SourceBuffer.
// 2. If the buffer full flag equals false, then abort these steps.
if (!m_bufferFull)
return;
size_t maximumBufferSize = this->maximumBufferSize();
// 3. Let removal ranges equal a list of presentation time ranges that can be evicted from
// the presentation to make room for the new data.
// NOTE: begin by removing data from the beginning of the buffered ranges, 30 seconds at
// a time, up to 30 seconds before currentTime.
MediaTime thirtySeconds = MediaTime(30, 1);
MediaTime currentTime = m_source->currentTime();
MediaTime maximumRangeEnd = currentTime - thirtySeconds;
#if !RELEASE_LOG_DISABLED
DEBUG_LOG(LOGIDENTIFIER, "currentTime = ", m_source->currentTime(), ", require ", extraMemoryCost() + newDataSize, " bytes, maximum buffer size is ", maximumBufferSize);
size_t initialBufferedSize = extraMemoryCost();
#endif
MediaTime rangeStart = MediaTime::zeroTime();
MediaTime rangeEnd = rangeStart + thirtySeconds;
while (rangeStart < maximumRangeEnd) {
// 4. For each range in removal ranges, run the coded frame removal algorithm with start and
// end equal to the removal range start and end timestamp respectively.
removeCodedFrames(rangeStart, std::min(rangeEnd, maximumRangeEnd));
if (extraMemoryCost() + newDataSize < maximumBufferSize) {
m_bufferFull = false;
break;
}
rangeStart += thirtySeconds;
rangeEnd += thirtySeconds;
}
#if !RELEASE_LOG_DISABLED
if (!m_bufferFull) {
DEBUG_LOG(LOGIDENTIFIER, "evicted ", initialBufferedSize - extraMemoryCost());
return;
}
#endif
// If there still isn't enough free space and there buffers in time ranges after the current range (ie. there is a gap after
// the current buffered range), delete 30 seconds at a time from duration back to the current time range or 30 seconds after
// currenTime whichever we hit first.
auto buffered = m_buffered->ranges();
size_t currentTimeRange = buffered.find(currentTime);
if (currentTimeRange == buffered.length() - 1) {
#if !RELEASE_LOG_DISABLED
ERROR_LOG(LOGIDENTIFIER, "FAILED to free enough after evicting ", initialBufferedSize - extraMemoryCost());
#endif
return;
}
MediaTime minimumRangeStart = currentTime + thirtySeconds;
rangeEnd = m_source->duration();
rangeStart = rangeEnd - thirtySeconds;
while (rangeStart > minimumRangeStart) {
// Do not evict data from the time range that contains currentTime.
size_t startTimeRange = buffered.find(rangeStart);
if (currentTimeRange != notFound && startTimeRange == currentTimeRange) {
size_t endTimeRange = buffered.find(rangeEnd);
if (currentTimeRange != notFound && endTimeRange == currentTimeRange)
break;
rangeEnd = buffered.start(endTimeRange);
}
// 4. For each range in removal ranges, run the coded frame removal algorithm with start and
// end equal to the removal range start and end timestamp respectively.
removeCodedFrames(std::max(minimumRangeStart, rangeStart), rangeEnd);
if (extraMemoryCost() + newDataSize < maximumBufferSize) {
m_bufferFull = false;
break;
}
rangeStart -= thirtySeconds;
rangeEnd -= thirtySeconds;
}
#if !RELEASE_LOG_DISABLED
if (m_bufferFull)
ERROR_LOG(LOGIDENTIFIER, "FAILED to free enough after evicting ", initialBufferedSize - extraMemoryCost());
else
DEBUG_LOG(LOGIDENTIFIER, "evicted ", initialBufferedSize - extraMemoryCost());
#endif
}
size_t SourceBuffer::maximumBufferSize() const
{
if (isRemoved())
return 0;
auto* element = m_source->mediaElement();
if (!element)
return 0;
return element->maximumSourceBufferSize(*this);
}
VideoTrackList& SourceBuffer::videoTracks()
{
if (!m_videoTracks)
m_videoTracks = VideoTrackList::create(makeWeakPtr(m_source->mediaElement()), scriptExecutionContext());
return *m_videoTracks;
}
AudioTrackList& SourceBuffer::audioTracks()
{
if (!m_audioTracks)
m_audioTracks = AudioTrackList::create(makeWeakPtr(m_source->mediaElement()), scriptExecutionContext());
return *m_audioTracks;
}
TextTrackList& SourceBuffer::textTracks()
{
if (!m_textTracks)
m_textTracks = TextTrackList::create(makeWeakPtr(m_source->mediaElement()), scriptExecutionContext());
return *m_textTracks;
}
void SourceBuffer::setActive(bool active)
{
if (m_active == active)
return;
m_active = active;
m_private->setActive(active);
if (!isRemoved())
m_source->sourceBufferDidChangeActiveState(*this, active);
}
void SourceBuffer::sourceBufferPrivateDidReceiveInitializationSegment(const InitializationSegment& segment)
{
if (isRemoved())
return;
ALWAYS_LOG(LOGIDENTIFIER);
// 3.5.8 Initialization Segment Received (ctd)
// https://rawgit.com/w3c/media-source/c3ad59c7a370d04430969ba73d18dc9bcde57a33/index.html#sourcebuffer-init-segment-received [Editor's Draft 09 January 2015]
// 1. Update the duration attribute if it currently equals NaN:
if (m_source->duration().isInvalid()) {
// ↳ If the initialization segment contains a duration:
// Run the duration change algorithm with new duration set to the duration in the initialization segment.
// ↳ Otherwise:
// Run the duration change algorithm with new duration set to positive Infinity.
if (segment.duration.isValid() && !segment.duration.isIndefinite())
m_source->setDurationInternal(segment.duration);
else
m_source->setDurationInternal(MediaTime::positiveInfiniteTime());
}
// 2. If the initialization segment has no audio, video, or text tracks, then run the append error algorithm
// with the decode error parameter set to true and abort these steps.
if (segment.audioTracks.isEmpty() && segment.videoTracks.isEmpty() && segment.textTracks.isEmpty()) {
appendError(true);
return;
}
// 3. If the first initialization segment flag is true, then run the following steps:
if (m_receivedFirstInitializationSegment) {
// 3.1. Verify the following properties. If any of the checks fail then run the append error algorithm
// with the decode error parameter set to true and abort these steps.
if (!validateInitializationSegment(segment)) {
appendError(true);
return;
}
// 3.2 Add the appropriate track descriptions from this initialization segment to each of the track buffers.
ASSERT(segment.audioTracks.size() == audioTracks().length());
for (auto& audioTrackInfo : segment.audioTracks) {
if (audioTracks().length() == 1) {
audioTracks().item(0)->setPrivate(*audioTrackInfo.track);
break;
}
auto audioTrack = audioTracks().getTrackById(audioTrackInfo.track->id());
ASSERT(audioTrack);
audioTrack->setPrivate(*audioTrackInfo.track);
}
ASSERT(segment.videoTracks.size() == videoTracks().length());
for (auto& videoTrackInfo : segment.videoTracks) {
if (videoTracks().length() == 1) {
videoTracks().item(0)->setPrivate(*videoTrackInfo.track);
break;
}
auto videoTrack = videoTracks().getTrackById(videoTrackInfo.track->id());
ASSERT(videoTrack);
videoTrack->setPrivate(*videoTrackInfo.track);
}
ASSERT(segment.textTracks.size() == textTracks().length());
for (auto& textTrackInfo : segment.textTracks) {
if (textTracks().length() == 1) {
downcast<InbandTextTrack>(*textTracks().item(0)).setPrivate(*textTrackInfo.track);
break;
}
auto textTrack = textTracks().getTrackById(textTrackInfo.track->id());
ASSERT(textTrack);
downcast<InbandTextTrack>(*textTrack).setPrivate(*textTrackInfo.track);
}
// 3.3 Set the need random access point flag on all track buffers to true.
for (auto& trackBuffer : m_trackBufferMap.values())
trackBuffer.needRandomAccessFlag = true;
}
// 4. Let active track flag equal false.
bool activeTrackFlag = false;
// 5. If the first initialization segment flag is false, then run the following steps:
if (!m_receivedFirstInitializationSegment) {
// 5.1 If the initialization segment contains tracks with codecs the user agent does not support,
// then run the append error algorithm with the decode error parameter set to true and abort these steps.
// NOTE: This check is the responsibility of the SourceBufferPrivate.
// 5.2 For each audio track in the initialization segment, run following steps:
for (auto& audioTrackInfo : segment.audioTracks) {
// FIXME: Implement steps 5.2.1-5.2.8.1 as per Editor's Draft 09 January 2015, and reorder this
// 5.2.1 Let new audio track be a new AudioTrack object.
// 5.2.2 Generate a unique ID and assign it to the id property on new video track.
auto newAudioTrack = AudioTrack::create(*this, *audioTrackInfo.track);
newAudioTrack->setSourceBuffer(this);
// 5.2.3 If audioTracks.length equals 0, then run the following steps:
if (!audioTracks().length()) {
// 5.2.3.1 Set the enabled property on new audio track to true.
newAudioTrack->setEnabled(true);
// 5.2.3.2 Set active track flag to true.
activeTrackFlag = true;
}
// 5.2.4 Add new audio track to the audioTracks attribute on this SourceBuffer object.
// 5.2.5 Queue a task to fire a trusted event named addtrack, that does not bubble and is
// not cancelable, and that uses the TrackEvent interface, at the AudioTrackList object
// referenced by the audioTracks attribute on this SourceBuffer object.
audioTracks().append(newAudioTrack.copyRef());
// 5.2.6 Add new audio track to the audioTracks attribute on the HTMLMediaElement.
// 5.2.7 Queue a task to fire a trusted event named addtrack, that does not bubble and is
// not cancelable, and that uses the TrackEvent interface, at the AudioTrackList object
// referenced by the audioTracks attribute on the HTMLMediaElement.
m_source->mediaElement()->ensureAudioTracks().append(newAudioTrack.copyRef());
// 5.2.8 Create a new track buffer to store coded frames for this track.
ASSERT(!m_trackBufferMap.contains(newAudioTrack->id()));
auto& trackBuffer = m_trackBufferMap.add(newAudioTrack->id(), TrackBuffer()).iterator->value;
// 5.2.9 Add the track description for this track to the track buffer.
trackBuffer.description = audioTrackInfo.description;
m_audioCodecs.append(trackBuffer.description->codec());
}
// 5.3 For each video track in the initialization segment, run following steps:
for (auto& videoTrackInfo : segment.videoTracks) {
// FIXME: Implement steps 5.3.1-5.3.8.1 as per Editor's Draft 09 January 2015, and reorder this
// 5.3.1 Let new video track be a new VideoTrack object.
// 5.3.2 Generate a unique ID and assign it to the id property on new video track.
auto newVideoTrack = VideoTrack::create(*this, *videoTrackInfo.track);
newVideoTrack->setSourceBuffer(this);
// 5.3.3 If videoTracks.length equals 0, then run the following steps:
if (!videoTracks().length()) {
// 5.3.3.1 Set the selected property on new video track to true.
newVideoTrack->setSelected(true);
// 5.3.3.2 Set active track flag to true.
activeTrackFlag = true;
}
// 5.3.4 Add new video track to the videoTracks attribute on this SourceBuffer object.
// 5.3.5 Queue a task to fire a trusted event named addtrack, that does not bubble and is
// not cancelable, and that uses the TrackEvent interface, at the VideoTrackList object
// referenced by the videoTracks attribute on this SourceBuffer object.
videoTracks().append(newVideoTrack.copyRef());
// 5.3.6 Add new video track to the videoTracks attribute on the HTMLMediaElement.
// 5.3.7 Queue a task to fire a trusted event named addtrack, that does not bubble and is
// not cancelable, and that uses the TrackEvent interface, at the VideoTrackList object
// referenced by the videoTracks attribute on the HTMLMediaElement.
m_source->mediaElement()->ensureVideoTracks().append(newVideoTrack.copyRef());
// 5.3.8 Create a new track buffer to store coded frames for this track.
ASSERT(!m_trackBufferMap.contains(newVideoTrack->id()));
auto& trackBuffer = m_trackBufferMap.add(newVideoTrack->id(), TrackBuffer()).iterator->value;
// 5.3.9 Add the track description for this track to the track buffer.
trackBuffer.description = videoTrackInfo.description;
m_videoCodecs.append(trackBuffer.description->codec());
}
// 5.4 For each text track in the initialization segment, run following steps:
for (auto& textTrackInfo : segment.textTracks) {
auto& textTrackPrivate = *textTrackInfo.track;
// FIXME: Implement steps 5.4.1-5.4.8.1 as per Editor's Draft 09 January 2015, and reorder this
// 5.4.1 Let new text track be a new TextTrack object with its properties populated with the
// appropriate information from the initialization segment.
auto newTextTrack = InbandTextTrack::create(*scriptExecutionContext(), *this, textTrackPrivate);
// 5.4.2 If the mode property on new text track equals "showing" or "hidden", then set active
// track flag to true.
if (textTrackPrivate.mode() != InbandTextTrackPrivate::Disabled)
activeTrackFlag = true;
// 5.4.3 Add new text track to the textTracks attribute on this SourceBuffer object.
// 5.4.4 Queue a task to fire a trusted event named addtrack, that does not bubble and is
// not cancelable, and that uses the TrackEvent interface, at textTracks attribute on this
// SourceBuffer object.
textTracks().append(newTextTrack.get());
// 5.4.5 Add new text track to the textTracks attribute on the HTMLMediaElement.
// 5.4.6 Queue a task to fire a trusted event named addtrack, that does not bubble and is
// not cancelable, and that uses the TrackEvent interface, at the TextTrackList object
// referenced by the textTracks attribute on the HTMLMediaElement.
m_source->mediaElement()->ensureTextTracks().append(WTFMove(newTextTrack));
// 5.4.7 Create a new track buffer to store coded frames for this track.
ASSERT(!m_trackBufferMap.contains(textTrackPrivate.id()));
auto& trackBuffer = m_trackBufferMap.add(textTrackPrivate.id(), TrackBuffer()).iterator->value;
// 5.4.8 Add the track description for this track to the track buffer.
trackBuffer.description = textTrackInfo.description;
m_textCodecs.append(trackBuffer.description->codec());
}
// 5.5 If active track flag equals true, then run the following steps:
if (activeTrackFlag) {
// 5.5.1 Add this SourceBuffer to activeSourceBuffers.
// 5.5.2 Queue a task to fire a simple event named addsourcebuffer at activeSourceBuffers
setActive(true);
}
// 5.6 Set first initialization segment flag to true.
m_receivedFirstInitializationSegment = true;
}
// (Note: Issue #155 adds this step after step 5:)
// 6. Set pending initialization segment for changeType flag to false.
m_pendingInitializationSegmentForChangeType = false;
// 6. If the HTMLMediaElement.readyState attribute is HAVE_NOTHING, then run the following steps:
if (m_private->readyState() == MediaPlayer::ReadyState::HaveNothing) {
// 6.1 If one or more objects in sourceBuffers have first initialization segment flag set to false, then abort these steps.
for (auto& sourceBuffer : *m_source->sourceBuffers()) {
if (!sourceBuffer->m_receivedFirstInitializationSegment)
return;
}
// 6.2 Set the HTMLMediaElement.readyState attribute to HAVE_METADATA.
// 6.3 Queue a task to fire a simple event named loadedmetadata at the media element.
m_private->setReadyState(MediaPlayer::ReadyState::HaveMetadata);
}
// 7. If the active track flag equals true and the HTMLMediaElement.readyState
// attribute is greater than HAVE_CURRENT_DATA, then set the HTMLMediaElement.readyState
// attribute to HAVE_METADATA.
if (activeTrackFlag && m_private->readyState() > MediaPlayer::ReadyState::HaveCurrentData)
m_private->setReadyState(MediaPlayer::ReadyState::HaveMetadata);
}
bool SourceBuffer::validateInitializationSegment(const InitializationSegment& segment)
{
// FIXME: ordering of all 3.5.X (X>=7) functions needs to be updated to post-[24 July 2014 Editor's Draft] version
// 3.5.8 Initialization Segment Received (ctd)
// https://rawgit.com/w3c/media-source/c3ad59c7a370d04430969ba73d18dc9bcde57a33/index.html#sourcebuffer-init-segment-received [Editor's Draft 09 January 2015]
// Note: those are checks from step 3.1
// * The number of audio, video, and text tracks match what was in the first initialization segment.
if (segment.audioTracks.size() != audioTracks().length()
|| segment.videoTracks.size() != videoTracks().length()
|| segment.textTracks.size() != textTracks().length())
return false;
// * The codecs for each track, match what was specified in the first initialization segment.
// (Note: Issue #155 strikes out this check. For broad compatibility when this experimental feature
// is not enabled, only perform this check if the "pending initialization segment for changeType flag"
// is not set.)
for (auto& audioTrackInfo : segment.audioTracks) {
if (m_audioCodecs.contains(audioTrackInfo.description->codec()))
continue;
if (!m_pendingInitializationSegmentForChangeType)
return false;
m_audioCodecs.append(audioTrackInfo.description->codec());
}
for (auto& videoTrackInfo : segment.videoTracks) {
if (m_videoCodecs.contains(videoTrackInfo.description->codec()))
continue;
if (!m_pendingInitializationSegmentForChangeType)
return false;
m_videoCodecs.append(videoTrackInfo.description->codec());
}
for (auto& textTrackInfo : segment.textTracks) {
if (m_textCodecs.contains(textTrackInfo.description->codec()))
continue;
if (!m_pendingInitializationSegmentForChangeType)
return false;
m_textCodecs.append(textTrackInfo.description->codec());
}
// * If more than one track for a single type are present (ie 2 audio tracks), then the Track
// IDs match the ones in the first initialization segment.
if (segment.audioTracks.size() >= 2) {
for (auto& audioTrackInfo : segment.audioTracks) {
if (!m_trackBufferMap.contains(audioTrackInfo.track->id()))
return false;
}
}
if (segment.videoTracks.size() >= 2) {
for (auto& videoTrackInfo : segment.videoTracks) {
if (!m_trackBufferMap.contains(videoTrackInfo.track->id()))
return false;
}
}
if (segment.textTracks.size() >= 2) {
for (auto& textTrackInfo : segment.videoTracks) {
if (!m_trackBufferMap.contains(textTrackInfo.track->id()))
return false;
}
}
return true;
}
class SampleLessThanComparator {
public:
bool operator()(std::pair<MediaTime, RefPtr<MediaSample>> value1, std::pair<MediaTime, RefPtr<MediaSample>> value2)
{
return value1.first < value2.first;
}
bool operator()(MediaTime value1, std::pair<MediaTime, RefPtr<MediaSample>> value2)
{
return value1 < value2.first;
}
bool operator()(std::pair<MediaTime, RefPtr<MediaSample>> value1, MediaTime value2)
{
return value1.first < value2;
}
};
void SourceBuffer::appendError(bool decodeErrorParam)
{
// 3.5.3 Append Error Algorithm
// https://rawgit.com/w3c/media-source/c3ad59c7a370d04430969ba73d18dc9bcde57a33/index.html#sourcebuffer-append-error [Editor's Draft 09 January 2015]
ASSERT(m_updating);
// 1. Run the reset parser state algorithm.
resetParserState();
// 2. Set the updating attribute to false.
m_updating = false;
// 3. Queue a task to fire a simple event named error at this SourceBuffer object.
scheduleEvent(eventNames().errorEvent);
// 4. Queue a task to fire a simple event named updateend at this SourceBuffer object.
scheduleEvent(eventNames().updateendEvent);
// 5. If decode error is true, then run the end of stream algorithm with the error parameter set to "decode".
if (decodeErrorParam)
m_source->streamEndedWithError(MediaSource::EndOfStreamError::Decode);
}
void SourceBuffer::sourceBufferPrivateDidReceiveSample(MediaSample& sample)
{
if (isRemoved())
return;
// 3.5.1 Segment Parser Loop
// 6.1 If the first initialization segment received flag is false, (Note: Issue # 155 & changeType()
// algorithm) or the pending initialization segment for changeType flag is true, (End note)
// then run the append error algorithm
// with the decode error parameter set to true and abort this algorithm.
// Note: current design makes SourceBuffer somehow ignorant of append state - it's more a thing
// of SourceBufferPrivate. That's why this check can't really be done in appendInternal.
// unless we force some kind of design with state machine switching.
if (!m_receivedFirstInitializationSegment || m_pendingInitializationSegmentForChangeType) {
appendError(true);
return;
}
// 3.5.8 Coded Frame Processing
// http://www.w3.org/TR/media-source/#sourcebuffer-coded-frame-processing
// When complete coded frames have been parsed by the segment parser loop then the following steps
// are run:
// 1. For each coded frame in the media segment run the following steps:
// 1.1. Loop Top
do {
MediaTime presentationTimestamp;
MediaTime decodeTimestamp;
// NOTE: this is out-of-order, but we need the timescale from the
// sample's duration for timestamp generation.
// 1.2 Let frame duration be a double precision floating point representation of the coded frame's
// duration in seconds.
MediaTime frameDuration = sample.duration();
if (m_shouldGenerateTimestamps) {
// ↳ If generate timestamps flag equals true:
// 1. Let presentation timestamp equal 0.
// NOTE: Use the duration timscale for the presentation timestamp, as this will eliminate
// timescale rounding when generating timestamps.
presentationTimestamp = { 0, frameDuration.timeScale() };
// 2. Let decode timestamp equal 0.
decodeTimestamp = { 0, frameDuration.timeScale() };
} else {
// ↳ Otherwise:
// 1. Let presentation timestamp be a double precision floating point representation of
// the coded frame's presentation timestamp in seconds.
presentationTimestamp = sample.presentationTime();
// 2. Let decode timestamp be a double precision floating point representation of the coded frame's
// decode timestamp in seconds.
decodeTimestamp = sample.decodeTime();
}
// 1.3 If mode equals "sequence" and group start timestamp is set, then run the following steps:
if (m_mode == AppendMode::Sequence && m_groupStartTimestamp.isValid()) {
// 1.3.1 Set timestampOffset equal to group start timestamp - presentation timestamp.
m_timestampOffset = m_groupStartTimestamp;
for (auto& trackBuffer : m_trackBufferMap.values()) {
trackBuffer.lastFrameTimescale = 0;
trackBuffer.roundedTimestampOffset = MediaTime::invalidTime();
}
// 1.3.2 Set group end timestamp equal to group start timestamp.
m_groupEndTimestamp = m_groupStartTimestamp;
// 1.3.3 Set the need random access point flag on all track buffers to true.
for (auto& trackBuffer : m_trackBufferMap.values())
trackBuffer.needRandomAccessFlag = true;
// 1.3.4 Unset group start timestamp.
m_groupStartTimestamp = MediaTime::invalidTime();
}
// NOTE: this is out-of-order, but we need TrackBuffer to be able to cache the results of timestamp offset rounding
// 1.5 Let track buffer equal the track buffer that the coded frame will be added to.
AtomString trackID = sample.trackID();
auto it = m_trackBufferMap.find(trackID);
if (it == m_trackBufferMap.end()) {
// The client managed to append a sample with a trackID not present in the initialization
// segment. This would be a good place to post an message to the developer console.
didDropSample();
return;
}
TrackBuffer& trackBuffer = it->value;
MediaTime microsecond(1, 1000000);
auto roundTowardsTimeScaleWithRoundingMargin = [] (const MediaTime& time, uint32_t timeScale, const MediaTime& roundingMargin) {
while (true) {
MediaTime roundedTime = time.toTimeScale(timeScale);
if (abs(roundedTime - time) < roundingMargin || timeScale >= MediaTime::MaximumTimeScale)
return roundedTime;
if (!WTF::safeMultiply(timeScale, 2, timeScale) || timeScale > MediaTime::MaximumTimeScale)
timeScale = MediaTime::MaximumTimeScale;
}
};
// 1.4 If timestampOffset is not 0, then run the following steps:
if (m_timestampOffset) {
if (!trackBuffer.roundedTimestampOffset.isValid() || presentationTimestamp.timeScale() != trackBuffer.lastFrameTimescale) {
trackBuffer.lastFrameTimescale = presentationTimestamp.timeScale();
trackBuffer.roundedTimestampOffset = roundTowardsTimeScaleWithRoundingMargin(m_timestampOffset, trackBuffer.lastFrameTimescale, microsecond);
}
// 1.4.1 Add timestampOffset to the presentation timestamp.
presentationTimestamp += trackBuffer.roundedTimestampOffset;
// 1.4.2 Add timestampOffset to the decode timestamp.
decodeTimestamp += trackBuffer.roundedTimestampOffset;
}
// 1.6 ↳ If last decode timestamp for track buffer is set and decode timestamp is less than last
// decode timestamp:
// OR
// ↳ If last decode timestamp for track buffer is set and the difference between decode timestamp and
// last decode timestamp is greater than 2 times last frame duration:
MediaTime decodeDurationToCheck = trackBuffer.greatestDecodeDuration;
if (decodeDurationToCheck.isValid() && trackBuffer.lastFrameDuration.isValid()
&& (trackBuffer.lastFrameDuration > decodeDurationToCheck))
decodeDurationToCheck = trackBuffer.lastFrameDuration;
if (trackBuffer.lastDecodeTimestamp.isValid() && (decodeTimestamp < trackBuffer.lastDecodeTimestamp
|| (decodeDurationToCheck.isValid() && abs(decodeTimestamp - trackBuffer.lastDecodeTimestamp) > (decodeDurationToCheck * 2)))) {
// 1.6.1:
if (m_mode == AppendMode::Segments) {
// ↳ If mode equals "segments":
// Set group end timestamp to presentation timestamp.
m_groupEndTimestamp = presentationTimestamp;
} else {
// ↳ If mode equals "sequence":
// Set group start timestamp equal to the group end timestamp.
m_groupStartTimestamp = m_groupEndTimestamp;
}
for (auto& trackBuffer : m_trackBufferMap.values()) {
// 1.6.2 Unset the last decode timestamp on all track buffers.
trackBuffer.lastDecodeTimestamp = MediaTime::invalidTime();
// 1.6.3 Unset the last frame duration on all track buffers.
trackBuffer.greatestDecodeDuration = MediaTime::invalidTime();
trackBuffer.lastFrameDuration = MediaTime::invalidTime();
// 1.6.4 Unset the highest presentation timestamp on all track buffers.
trackBuffer.highestPresentationTimestamp = MediaTime::invalidTime();
// 1.6.5 Set the need random access point flag on all track buffers to true.
trackBuffer.needRandomAccessFlag = true;
}
// 1.6.6 Jump to the Loop Top step above to restart processing of the current coded frame.
continue;
}
if (m_mode == AppendMode::Sequence) {
// Use the generated timestamps instead of the sample's timestamps.
sample.setTimestamps(presentationTimestamp, decodeTimestamp);
} else if (trackBuffer.roundedTimestampOffset) {
// Reflect the timestamp offset into the sample.
sample.offsetTimestampsBy(trackBuffer.roundedTimestampOffset);
}
DEBUG_LOG(LOGIDENTIFIER, sample);
// 1.7 Let frame end timestamp equal the sum of presentation timestamp and frame duration.
MediaTime frameEndTimestamp = presentationTimestamp + frameDuration;
// 1.8 If presentation timestamp is less than appendWindowStart, then set the need random access
// point flag to true, drop the coded frame, and jump to the top of the loop to start processing
// the next coded frame.
// 1.9 If frame end timestamp is greater than appendWindowEnd, then set the need random access
// point flag to true, drop the coded frame, and jump to the top of the loop to start processing
// the next coded frame.
if (presentationTimestamp < m_appendWindowStart || frameEndTimestamp > m_appendWindowEnd) {
trackBuffer.needRandomAccessFlag = true;
didDropSample();
return;
}
// 1.10 If the decode timestamp is less than the presentation start time, then run the end of stream
// algorithm with the error parameter set to "decode", and abort these steps.
// NOTE: Until <https://www.w3.org/Bugs/Public/show_bug.cgi?id=27487> is resolved, we will only check
// the presentation timestamp.
MediaTime presentationStartTime = MediaTime::zeroTime();
if (presentationTimestamp < presentationStartTime) {
ERROR_LOG(LOGIDENTIFIER, "failing because presentationTimestamp (", presentationTimestamp, ") < presentationStartTime (", presentationStartTime, ")");
m_source->streamEndedWithError(MediaSource::EndOfStreamError::Decode);
return;
}
// 1.11 If the need random access point flag on track buffer equals true, then run the following steps:
if (trackBuffer.needRandomAccessFlag) {
// 1.11.1 If the coded frame is not a random access point, then drop the coded frame and jump
// to the top of the loop to start processing the next coded frame.
if (!sample.isSync()) {
didDropSample();
return;
}
// 1.11.2 Set the need random access point flag on track buffer to false.
trackBuffer.needRandomAccessFlag = false;
}
// 1.12 Let spliced audio frame be an unset variable for holding audio splice information
// 1.13 Let spliced timed text frame be an unset variable for holding timed text splice information
// FIXME: Add support for sample splicing.
SampleMap erasedSamples;
// 1.14 If last decode timestamp for track buffer is unset and presentation timestamp falls
// falls within the presentation interval of a coded frame in track buffer, then run the
// following steps:
if (trackBuffer.lastDecodeTimestamp.isInvalid()) {
auto iter = trackBuffer.samples.presentationOrder().findSampleContainingPresentationTime(presentationTimestamp);
if (iter != trackBuffer.samples.presentationOrder().end()) {
// 1.14.1 Let overlapped frame be the coded frame in track buffer that matches the condition above.
RefPtr<MediaSample> overlappedFrame = iter->second;
// 1.14.2 If track buffer contains audio coded frames:
// Run the audio splice frame algorithm and if a splice frame is returned, assign it to
// spliced audio frame.
// FIXME: Add support for sample splicing.
// If track buffer contains video coded frames:
if (trackBuffer.description && trackBuffer.description->isVideo()) {
// 1.14.2.1 Let overlapped frame presentation timestamp equal the presentation timestamp
// of overlapped frame.
MediaTime overlappedFramePresentationTimestamp = overlappedFrame->presentationTime();
// 1.14.2.2 Let remove window timestamp equal overlapped frame presentation timestamp
// plus 1 microsecond.
MediaTime removeWindowTimestamp = overlappedFramePresentationTimestamp + microsecond;
// 1.14.2.3 If the presentation timestamp is less than the remove window timestamp,
// then remove overlapped frame and any coded frames that depend on it from track buffer.
if (presentationTimestamp < removeWindowTimestamp)
erasedSamples.addSample(*iter->second);
}
// If track buffer contains timed text coded frames:
// Run the text splice frame algorithm and if a splice frame is returned, assign it to spliced timed text frame.
// FIXME: Add support for sample splicing.
}
}
// 1.15 Remove existing coded frames in track buffer:
// If highest presentation timestamp for track buffer is not set:
if (trackBuffer.highestPresentationTimestamp.isInvalid()) {
// Remove all coded frames from track buffer that have a presentation timestamp greater than or
// equal to presentation timestamp and less than frame end timestamp.
auto iter_pair = trackBuffer.samples.presentationOrder().findSamplesBetweenPresentationTimes(presentationTimestamp, frameEndTimestamp);
if (iter_pair.first != trackBuffer.samples.presentationOrder().end())
erasedSamples.addRange(iter_pair.first, iter_pair.second);
}
// There are many files out there where the frame times are not perfectly contiguous and may have small overlaps
// between the beginning of a frame and the end of the previous one; therefore a tolerance is needed whenever
// durations are considered.
// For instance, most WebM files are muxed rounded to the millisecond (the default TimecodeScale of the format)
// but their durations use a finer timescale (causing a sub-millisecond overlap). More rarely, there are also
// MP4 files with slightly off tfdt boxes, presenting a similar problem at the beginning of each fragment.
const MediaTime contiguousFrameTolerance = MediaTime(1, 1000);
// If highest presentation timestamp for track buffer is set and less than or equal to presentation timestamp
if (trackBuffer.highestPresentationTimestamp.isValid() && trackBuffer.highestPresentationTimestamp - contiguousFrameTolerance <= presentationTimestamp) {
// Remove all coded frames from track buffer that have a presentation timestamp greater than highest
// presentation timestamp and less than or equal to frame end timestamp.
do {
// NOTE: Searching from the end of the trackBuffer will be vastly more efficient if the search range is
// near the end of the buffered range. Use a linear-backwards search if the search range is within one
// frame duration of the end:
unsigned bufferedLength = trackBuffer.buffered.length();
if (!bufferedLength)
break;
MediaTime highestBufferedTime = trackBuffer.buffered.maximumBufferedTime();
MediaTime eraseBeginTime = trackBuffer.highestPresentationTimestamp - contiguousFrameTolerance;
MediaTime eraseEndTime = frameEndTimestamp - contiguousFrameTolerance;
PresentationOrderSampleMap::iterator_range range;
if (highestBufferedTime - trackBuffer.highestPresentationTimestamp < trackBuffer.lastFrameDuration)
// If the new frame is at the end of the buffered ranges, perform a sequential scan from end (O(1)).
range = trackBuffer.samples.presentationOrder().findSamplesBetweenPresentationTimesFromEnd(eraseBeginTime, eraseEndTime);
else
// In any other case, perform a binary search (O(log(n)).
range = trackBuffer.samples.presentationOrder().findSamplesBetweenPresentationTimes(eraseBeginTime, eraseEndTime);
if (range.first != trackBuffer.samples.presentationOrder().end())
erasedSamples.addRange(range.first, range.second);
} while(false);
}
// 1.16 Remove decoding dependencies of the coded frames removed in the previous step:
DecodeOrderSampleMap::MapType dependentSamples;
if (!erasedSamples.empty()) {
// If detailed information about decoding dependencies is available:
// FIXME: Add support for detailed dependency information
// Otherwise: Remove all coded frames between the coded frames removed in the previous step
// and the next random access point after those removed frames.
auto firstDecodeIter = trackBuffer.samples.decodeOrder().findSampleWithDecodeKey(erasedSamples.decodeOrder().begin()->first);
auto lastDecodeIter = trackBuffer.samples.decodeOrder().findSampleWithDecodeKey(erasedSamples.decodeOrder().rbegin()->first);
auto nextSyncIter = trackBuffer.samples.decodeOrder().findSyncSampleAfterDecodeIterator(lastDecodeIter);
dependentSamples.insert(firstDecodeIter, nextSyncIter);
// NOTE: in the case of b-frames, the previous step may leave in place samples whose presentation
// timestamp < presentationTime, but whose decode timestamp >= decodeTime. These will eventually cause
// a decode error if left in place, so remove these samples as well.
DecodeOrderSampleMap::KeyType decodeKey(sample.decodeTime(), sample.presentationTime());
auto samplesWithHigherDecodeTimes = trackBuffer.samples.decodeOrder().findSamplesBetweenDecodeKeys(decodeKey, erasedSamples.decodeOrder().begin()->first);
if (samplesWithHigherDecodeTimes.first != samplesWithHigherDecodeTimes.second)
dependentSamples.insert(samplesWithHigherDecodeTimes.first, samplesWithHigherDecodeTimes.second);
PlatformTimeRanges erasedRanges = removeSamplesFromTrackBuffer(dependentSamples, trackBuffer, this, "sourceBufferPrivateDidReceiveSample");
// Only force the TrackBuffer to re-enqueue if the removed ranges overlap with enqueued and possibly
// not yet displayed samples.
MediaTime currentMediaTime = m_source->currentTime();
if (trackBuffer.lastEnqueuedPresentationTime.isValid() && currentMediaTime < trackBuffer.lastEnqueuedPresentationTime) {
PlatformTimeRanges possiblyEnqueuedRanges(currentMediaTime, trackBuffer.lastEnqueuedPresentationTime);
possiblyEnqueuedRanges.intersectWith(erasedRanges);
if (possiblyEnqueuedRanges.length())
trackBuffer.needsReenqueueing = true;
}
erasedRanges.invert();
trackBuffer.buffered.intersectWith(erasedRanges);
setBufferedDirty(true);
}
// 1.17 If spliced audio frame is set:
// Add spliced audio frame to the track buffer.
// If spliced timed text frame is set:
// Add spliced timed text frame to the track buffer.
// FIXME: Add support for sample splicing.
// Otherwise:
// Add the coded frame with the presentation timestamp, decode timestamp, and frame duration to the track buffer.
trackBuffer.samples.addSample(sample);
// Note: The terminology here is confusing: "enqueuing" means providing a frame to the inner media framework.
// First, frames are inserted in the decode queue; later, at the end of the append all the frames in the decode
// queue are "enqueued" (sent to the inner media framework) in `provideMediaData()`.
//
// In order to check whether a frame should be added to the decode queue we check whether it starts after the
// lastEnqueuedDecodeKey.
DecodeOrderSampleMap::KeyType decodeKey(sample.decodeTime(), sample.presentationTime());
if (trackBuffer.lastEnqueuedDecodeKey.first.isInvalid() || decodeKey > trackBuffer.lastEnqueuedDecodeKey) {
trackBuffer.decodeQueue.insert(DecodeOrderSampleMap::MapType::value_type(decodeKey, &sample));
if (trackBuffer.minimumEnqueuedPresentationTime.isValid() && sample.presentationTime() < trackBuffer.minimumEnqueuedPresentationTime)
trackBuffer.needsMinimumUpcomingPresentationTimeUpdating = true;
}
// NOTE: the spec considers "Coded Frame Duration" to be the presentation duration, but this is not necessarily equal
// to the decoded duration. When comparing deltas between decode timestamps, the decode duration, not the presentation.
if (trackBuffer.lastDecodeTimestamp.isValid()) {
MediaTime lastDecodeDuration = decodeTimestamp - trackBuffer.lastDecodeTimestamp;
if (!trackBuffer.greatestDecodeDuration.isValid() || lastDecodeDuration > trackBuffer.greatestDecodeDuration)
trackBuffer.greatestDecodeDuration = lastDecodeDuration;
}
// 1.18 Set last decode timestamp for track buffer to decode timestamp.
trackBuffer.lastDecodeTimestamp = decodeTimestamp;
// 1.19 Set last frame duration for track buffer to frame duration.
trackBuffer.lastFrameDuration = frameDuration;
// 1.20 If highest presentation timestamp for track buffer is unset or frame end timestamp is greater
// than highest presentation timestamp, then set highest presentation timestamp for track buffer
// to frame end timestamp.
if (trackBuffer.highestPresentationTimestamp.isInvalid() || frameEndTimestamp > trackBuffer.highestPresentationTimestamp)
trackBuffer.highestPresentationTimestamp = frameEndTimestamp;
// 1.21 If frame end timestamp is greater than group end timestamp, then set group end timestamp equal
// to frame end timestamp.
if (m_groupEndTimestamp.isInvalid() || frameEndTimestamp > m_groupEndTimestamp)
m_groupEndTimestamp = frameEndTimestamp;
// 1.22 If generate timestamps flag equals true, then set timestampOffset equal to frame end timestamp.
if (m_shouldGenerateTimestamps) {
m_timestampOffset = frameEndTimestamp;
for (auto& trackBuffer : m_trackBufferMap.values()) {
trackBuffer.lastFrameTimescale = 0;
trackBuffer.roundedTimestampOffset = MediaTime::invalidTime();
}
}
// Eliminate small gaps between buffered ranges by coalescing
// disjoint ranges separated by less than a "fudge factor".
auto presentationEndTime = presentationTimestamp + frameDuration;
auto nearestToPresentationStartTime = trackBuffer.buffered.nearest(presentationTimestamp);
if (nearestToPresentationStartTime.isValid() && (presentationTimestamp - nearestToPresentationStartTime).isBetween(MediaTime::zeroTime(), MediaSource::currentTimeFudgeFactor()))
presentationTimestamp = nearestToPresentationStartTime;
auto nearestToPresentationEndTime = trackBuffer.buffered.nearest(presentationEndTime);
if (nearestToPresentationEndTime.isValid() && (nearestToPresentationEndTime - presentationEndTime).isBetween(MediaTime::zeroTime(), MediaSource::currentTimeFudgeFactor()))
presentationEndTime = nearestToPresentationEndTime;
trackBuffer.buffered.add(presentationTimestamp, presentationEndTime);
m_bufferedSinceLastMonitor += frameDuration.toDouble();
setBufferedDirty(true);
break;
} while (1);
// Steps 2-4 will be handled by MediaSource::monitorSourceBuffers()
// 5. If the media segment contains data beyond the current duration, then run the duration change algorithm with new
// duration set to the maximum of the current duration and the group end timestamp.
if (m_groupEndTimestamp > m_source->duration())
m_source->setDurationInternal(m_groupEndTimestamp);
}
bool SourceBuffer::hasAudio() const
{
return m_audioTracks && m_audioTracks->length();
}
bool SourceBuffer::hasVideo() const
{
return m_videoTracks && m_videoTracks->length();
}
bool SourceBuffer::sourceBufferPrivateHasAudio() const
{
return hasAudio();
}
bool SourceBuffer::sourceBufferPrivateHasVideo() const
{
return hasVideo();
}
void SourceBuffer::videoTrackSelectedChanged(VideoTrack& track)
{
// 2.4.5 Changes to selected/enabled track state
// If the selected video track changes, then run the following steps:
// 1. If the SourceBuffer associated with the previously selected video track is not associated with
// any other enabled tracks, run the following steps:
if (!track.selected()
&& (!m_videoTracks || !m_videoTracks->isAnyTrackEnabled())
&& (!m_audioTracks || !m_audioTracks->isAnyTrackEnabled())
&& (!m_textTracks || !m_textTracks->isAnyTrackEnabled())) {
// 1.1 Remove the SourceBuffer from activeSourceBuffers.
// 1.2 Queue a task to fire a simple event named removesourcebuffer at activeSourceBuffers
setActive(false);
} else if (track.selected()) {
// 2. If the SourceBuffer associated with the newly selected video track is not already in activeSourceBuffers,
// run the following steps:
// 2.1 Add the SourceBuffer to activeSourceBuffers.
// 2.2 Queue a task to fire a simple event named addsourcebuffer at activeSourceBuffers
setActive(true);
}
if (m_videoTracks && m_videoTracks->contains(track))
m_videoTracks->scheduleChangeEvent();
if (!isRemoved())
m_source->mediaElement()->videoTrackSelectedChanged(track);
}
void SourceBuffer::audioTrackEnabledChanged(AudioTrack& track)
{
// 2.4.5 Changes to selected/enabled track state
// If an audio track becomes disabled and the SourceBuffer associated with this track is not
// associated with any other enabled or selected track, then run the following steps:
if (!track.enabled()
&& (!m_videoTracks || !m_videoTracks->isAnyTrackEnabled())
&& (!m_audioTracks || !m_audioTracks->isAnyTrackEnabled())
&& (!m_textTracks || !m_textTracks->isAnyTrackEnabled())) {
// 1. Remove the SourceBuffer associated with the audio track from activeSourceBuffers
// 2. Queue a task to fire a simple event named removesourcebuffer at activeSourceBuffers
setActive(false);
} else if (track.enabled()) {
// If an audio track becomes enabled and the SourceBuffer associated with this track is
// not already in activeSourceBuffers, then run the following steps:
// 1. Add the SourceBuffer associated with the audio track to activeSourceBuffers
// 2. Queue a task to fire a simple event named addsourcebuffer at activeSourceBuffers
setActive(true);
}
if (m_audioTracks && m_audioTracks->contains(track))
m_audioTracks->scheduleChangeEvent();
if (!isRemoved())
m_source->mediaElement()->audioTrackEnabledChanged(track);
}
void SourceBuffer::textTrackModeChanged(TextTrack& track)
{
// 2.4.5 Changes to selected/enabled track state
// If a text track mode becomes "disabled" and the SourceBuffer associated with this track is not
// associated with any other enabled or selected track, then run the following steps:
if (track.mode() == TextTrack::Mode::Disabled
&& (!m_videoTracks || !m_videoTracks->isAnyTrackEnabled())
&& (!m_audioTracks || !m_audioTracks->isAnyTrackEnabled())
&& (!m_textTracks || !m_textTracks->isAnyTrackEnabled())) {
// 1. Remove the SourceBuffer associated with the audio track from activeSourceBuffers
// 2. Queue a task to fire a simple event named removesourcebuffer at activeSourceBuffers
setActive(false);
} else {
// If a text track mode becomes "showing" or "hidden" and the SourceBuffer associated with this
// track is not already in activeSourceBuffers, then run the following steps:
// 1. Add the SourceBuffer associated with the text track to activeSourceBuffers
// 2. Queue a task to fire a simple event named addsourcebuffer at activeSourceBuffers
setActive(true);
}
if (m_textTracks && m_textTracks->contains(track))
m_textTracks->scheduleChangeEvent();
if (!isRemoved())
m_source->mediaElement()->textTrackModeChanged(track);
}
void SourceBuffer::textTrackAddCue(TextTrack& track, TextTrackCue& cue)
{
if (!isRemoved())
m_source->mediaElement()->textTrackAddCue(track, cue);
}
void SourceBuffer::textTrackAddCues(TextTrack& track, const TextTrackCueList& cueList)
{
if (!isRemoved())
m_source->mediaElement()->textTrackAddCues(track, cueList);
}
void SourceBuffer::textTrackRemoveCue(TextTrack& track, TextTrackCue& cue)
{
if (!isRemoved())
m_source->mediaElement()->textTrackRemoveCue(track, cue);
}
void SourceBuffer::textTrackRemoveCues(TextTrack& track, const TextTrackCueList& cueList)
{
if (!isRemoved())
m_source->mediaElement()->textTrackRemoveCues(track, cueList);
}
void SourceBuffer::textTrackKindChanged(TextTrack& track)
{
if (!isRemoved())
m_source->mediaElement()->textTrackKindChanged(track);
}
void SourceBuffer::sourceBufferPrivateReenqueSamples(const AtomString& trackID)
{
if (isRemoved())
return;
DEBUG_LOG(LOGIDENTIFIER);
auto it = m_trackBufferMap.find(trackID);
if (it == m_trackBufferMap.end())
return;
auto& trackBuffer = it->value;
trackBuffer.needsReenqueueing = true;
reenqueueMediaForTime(trackBuffer, trackID, m_source->currentTime());
}
void SourceBuffer::sourceBufferPrivateDidBecomeReadyForMoreSamples(const AtomString& trackID)
{
if (isRemoved())
return;
DEBUG_LOG(LOGIDENTIFIER);
auto it = m_trackBufferMap.find(trackID);
if (it == m_trackBufferMap.end())
return;
auto& trackBuffer = it->value;
if (!trackBuffer.needsReenqueueing && !m_source->isSeeking())
provideMediaData(trackBuffer, trackID);
}
void SourceBuffer::provideMediaData(TrackBuffer& trackBuffer, const AtomString& trackID)
{
if (m_source->isSeeking())
return;
#if !RELEASE_LOG_DISABLED
unsigned enqueuedSamples = 0;
#endif
if (trackBuffer.needsMinimumUpcomingPresentationTimeUpdating)
resetMinimumUpcomingPresentationTime(trackBuffer, trackID);
while (!trackBuffer.decodeQueue.empty()) {
if (!m_private->isReadyForMoreSamples(trackID)) {
DEBUG_LOG(LOGIDENTIFIER, "bailing early, track id ", trackID, " is not ready for more data");
m_private->notifyClientWhenReadyForMoreSamples(trackID);
break;
}
// FIXME(rdar://problem/20635969): Remove this re-entrancy protection when the aforementioned radar is resolved; protecting
// against re-entrancy introduces a small inefficency when removing appended samples from the decode queue one at a time
// rather than when all samples have been enqueued.
auto sample = trackBuffer.decodeQueue.begin()->second;
// Do not enqueue samples spanning a significant unbuffered gap.
// NOTE: one second is somewhat arbitrary. MediaSource::monitorSourceBuffers() is run
// on the playbackTimer, which is effectively every 350ms. Allowing > 350ms gap between
// enqueued samples allows for situations where we overrun the end of a buffered range
// but don't notice for 350s of playback time, and the client can enqueue data for the
// new current time without triggering this early return.
// FIXME(135867): Make this gap detection logic less arbitrary.
MediaTime oneSecond(1, 1);
if (trackBuffer.lastEnqueuedDecodeKey.first.isValid()
&& trackBuffer.lastEnqueuedDecodeDuration.isValid()
&& sample->decodeTime() - trackBuffer.lastEnqueuedDecodeKey.first > oneSecond + trackBuffer.lastEnqueuedDecodeDuration) {
DEBUG_LOG(LOGIDENTIFIER, "bailing early because of unbuffered gap, new sample: ", sample->decodeTime(), ", last enqueued sample ends: ", trackBuffer.lastEnqueuedDecodeKey.first + trackBuffer.lastEnqueuedDecodeDuration);
break;
}
// Remove the sample from the decode queue now.
trackBuffer.decodeQueue.erase(trackBuffer.decodeQueue.begin());
trackBuffer.lastEnqueuedPresentationTime = sample->presentationTime();
trackBuffer.lastEnqueuedDecodeKey = {sample->decodeTime(), sample->presentationTime()};
trackBuffer.lastEnqueuedDecodeDuration = sample->duration();
m_private->enqueueSample(sample.releaseNonNull(), trackID);
#if !RELEASE_LOG_DISABLED
++enqueuedSamples;
#endif
}
updateMinimumUpcomingPresentationTime(trackBuffer, trackID);
#if !RELEASE_LOG_DISABLED
DEBUG_LOG(LOGIDENTIFIER, "enqueued ", enqueuedSamples, " samples, ", static_cast<size_t>(trackBuffer.decodeQueue.size()), " remaining");
#endif
trySignalAllSamplesInTrackEnqueued(trackID);
}
void SourceBuffer::updateMinimumUpcomingPresentationTime(TrackBuffer& trackBuffer, const AtomString& trackID)
{
if (!m_private->canSetMinimumUpcomingPresentationTime(trackID))
return;
if (trackBuffer.decodeQueue.empty()) {
trackBuffer.minimumEnqueuedPresentationTime = MediaTime::invalidTime();
m_private->clearMinimumUpcomingPresentationTime(trackID);
return;
}
auto minPts = std::min_element(trackBuffer.decodeQueue.begin(), trackBuffer.decodeQueue.end(), [](auto& left, auto& right) -> bool {
return left.second->outputPresentationTime() < right.second->outputPresentationTime();
});
if (minPts == trackBuffer.decodeQueue.end()) {
trackBuffer.minimumEnqueuedPresentationTime = MediaTime::invalidTime();
m_private->clearMinimumUpcomingPresentationTime(trackID);
return;
}
trackBuffer.minimumEnqueuedPresentationTime = minPts->second->outputPresentationTime();
m_private->setMinimumUpcomingPresentationTime(trackID, trackBuffer.minimumEnqueuedPresentationTime);
}
void SourceBuffer::resetMinimumUpcomingPresentationTime(TrackBuffer& trackBuffer, const AtomString& trackID)
{
if (!m_private->canSetMinimumUpcomingPresentationTime(trackID))
return;
trackBuffer.minimumEnqueuedPresentationTime = MediaTime::invalidTime();
m_private->clearMinimumUpcomingPresentationTime(trackID);
}
void SourceBuffer::trySignalAllSamplesInTrackEnqueued(const AtomString& trackID)
{
if (m_source->isEnded() && m_trackBufferMap.get(trackID).decodeQueue.empty()) {
DEBUG_LOG(LOGIDENTIFIER, "enqueued all samples from track ", trackID);
m_private->allSamplesInTrackEnqueued(trackID);
}
}
void SourceBuffer::trySignalAllSamplesEnqueued()
{
for (const AtomString& trackID : m_trackBufferMap.keys())
trySignalAllSamplesInTrackEnqueued(trackID);
}
void SourceBuffer::reenqueueMediaForTime(TrackBuffer& trackBuffer, const AtomString& trackID, const MediaTime& time)
{
m_private->flush(trackID);
trackBuffer.decodeQueue.clear();
// Find the sample which contains the current presentation time.
auto currentSamplePTSIterator = trackBuffer.samples.presentationOrder().findSampleContainingPresentationTime(time);
if (currentSamplePTSIterator == trackBuffer.samples.presentationOrder().end())
currentSamplePTSIterator = trackBuffer.samples.presentationOrder().findSampleStartingOnOrAfterPresentationTime(time);
if (currentSamplePTSIterator == trackBuffer.samples.presentationOrder().end()
|| (currentSamplePTSIterator->first - time) > MediaSource::currentTimeFudgeFactor())
return;
// Seach backward for the previous sync sample.
DecodeOrderSampleMap::KeyType decodeKey(currentSamplePTSIterator->second->decodeTime(), currentSamplePTSIterator->second->presentationTime());
auto currentSampleDTSIterator = trackBuffer.samples.decodeOrder().findSampleWithDecodeKey(decodeKey);
ASSERT(currentSampleDTSIterator != trackBuffer.samples.decodeOrder().end());
auto reverseCurrentSampleIter = --DecodeOrderSampleMap::reverse_iterator(currentSampleDTSIterator);
auto reverseLastSyncSampleIter = trackBuffer.samples.decodeOrder().findSyncSamplePriorToDecodeIterator(reverseCurrentSampleIter);
if (reverseLastSyncSampleIter == trackBuffer.samples.decodeOrder().rend())
return;
// Fill the decode queue with the non-displaying samples.
for (auto iter = reverseLastSyncSampleIter; iter != reverseCurrentSampleIter; --iter) {
auto copy = iter->second->createNonDisplayingCopy();
DecodeOrderSampleMap::KeyType decodeKey(copy->decodeTime(), copy->presentationTime());
trackBuffer.decodeQueue.insert(DecodeOrderSampleMap::MapType::value_type(decodeKey, WTFMove(copy)));
}
if (!trackBuffer.decodeQueue.empty()) {
auto lastSampleIter = trackBuffer.decodeQueue.rbegin();
auto lastSampleDecodeKey = lastSampleIter->first;
auto lastSampleDuration = lastSampleIter->second->duration();
trackBuffer.lastEnqueuedPresentationTime = lastSampleDecodeKey.second;
trackBuffer.lastEnqueuedDecodeKey = lastSampleDecodeKey;
trackBuffer.lastEnqueuedDecodeDuration = lastSampleDuration;
} else {
trackBuffer.lastEnqueuedPresentationTime = MediaTime::invalidTime();
trackBuffer.lastEnqueuedDecodeKey = {MediaTime::invalidTime(), MediaTime::invalidTime()};
trackBuffer.lastEnqueuedDecodeDuration = MediaTime::invalidTime();
}
// Fill the decode queue with the remaining samples.
for (auto iter = currentSampleDTSIterator; iter != trackBuffer.samples.decodeOrder().end(); ++iter)
trackBuffer.decodeQueue.insert(*iter);
provideMediaData(trackBuffer, trackID);
trackBuffer.needsReenqueueing = false;
}
void SourceBuffer::didDropSample()
{
if (!isRemoved())
m_source->mediaElement()->incrementDroppedFrameCount();
}
void SourceBuffer::monitorBufferingRate()
{
MonotonicTime now = MonotonicTime::now();
Seconds interval = now - m_timeOfBufferingMonitor;
double rateSinceLastMonitor = m_bufferedSinceLastMonitor / interval.seconds();
m_timeOfBufferingMonitor = now;
m_bufferedSinceLastMonitor = 0;
m_averageBufferRate += (interval.seconds() * ExponentialMovingAverageCoefficient) * (rateSinceLastMonitor - m_averageBufferRate);
DEBUG_LOG(LOGIDENTIFIER, m_averageBufferRate);
}
void SourceBuffer::updateBufferedFromTrackBuffers()
{
// 3.1 Attributes, buffered
// https://rawgit.com/w3c/media-source/45627646344eea0170dd1cbc5a3d508ca751abb8/media-source-respec.html#dom-sourcebuffer-buffered
// 2. Let highest end time be the largest track buffer ranges end time across all the track buffers managed by this SourceBuffer object.
MediaTime highestEndTime = MediaTime::negativeInfiniteTime();
for (auto& trackBuffer : m_trackBufferMap.values()) {
if (!trackBuffer.buffered.length())
continue;
highestEndTime = std::max(highestEndTime, trackBuffer.buffered.maximumBufferedTime());
}
// NOTE: Short circuit the following if none of the TrackBuffers have buffered ranges to avoid generating
// a single range of {0, 0}.
if (highestEndTime.isNegativeInfinite()) {
m_buffered->ranges() = PlatformTimeRanges();
return;
}
// 3. Let intersection ranges equal a TimeRange object containing a single range from 0 to highest end time.
PlatformTimeRanges intersectionRanges { MediaTime::zeroTime(), highestEndTime };
// 4. For each audio and video track buffer managed by this SourceBuffer, run the following steps:
for (auto& trackBuffer : m_trackBufferMap.values()) {
// 4.1 Let track ranges equal the track buffer ranges for the current track buffer.
PlatformTimeRanges trackRanges = trackBuffer.buffered;
if (!trackRanges.length())
continue;
// 4.2 If readyState is "ended", then set the end time on the last range in track ranges to highest end time.
if (m_source->isEnded())
trackRanges.add(trackRanges.maximumBufferedTime(), highestEndTime);
// 4.3 Let new intersection ranges equal the intersection between the intersection ranges and the track ranges.
// 4.4 Replace the ranges in intersection ranges with the new intersection ranges.
intersectionRanges.intersectWith(trackRanges);
}
// 5. If intersection ranges does not contain the exact same range information as the current value of this attribute,
// then update the current value of this attribute to intersection ranges.
m_buffered->ranges() = intersectionRanges;
setBufferedDirty(true);
}
bool SourceBuffer::canPlayThroughRange(PlatformTimeRanges& ranges)
{
if (isRemoved())
return false;
monitorBufferingRate();
// Assuming no fluctuations in the buffering rate, loading 1 second per second or greater
// means indefinite playback. This could be improved by taking jitter into account.
if (m_averageBufferRate > 1)
return true;
// Add up all the time yet to be buffered.
MediaTime currentTime = m_source->currentTime();
MediaTime duration = m_source->duration();
PlatformTimeRanges unbufferedRanges = ranges;
unbufferedRanges.invert();
unbufferedRanges.intersectWith(PlatformTimeRanges(currentTime, std::max(currentTime, duration)));
MediaTime unbufferedTime = unbufferedRanges.totalDuration();
if (!unbufferedTime.isValid())
return true;
MediaTime timeRemaining = duration - currentTime;
return unbufferedTime.toDouble() / m_averageBufferRate < timeRemaining.toDouble();
}
size_t SourceBuffer::extraMemoryCost() const
{
size_t extraMemoryCost = m_pendingAppendData.capacity();
for (auto& trackBuffer : m_trackBufferMap.values())
extraMemoryCost += trackBuffer.samples.sizeInBytes();
return extraMemoryCost;
}
void SourceBuffer::reportExtraMemoryAllocated()
{
size_t extraMemoryCost = this->extraMemoryCost();
if (extraMemoryCost <= m_reportedExtraMemoryCost)
return;
size_t extraMemoryCostDelta = extraMemoryCost - m_reportedExtraMemoryCost;
m_reportedExtraMemoryCost = extraMemoryCost;
JSC::JSLockHolder lock(scriptExecutionContext()->vm());
// FIXME: Adopt reportExtraMemoryVisited, and switch to reportExtraMemoryAllocated.
// https://bugs.webkit.org/show_bug.cgi?id=142595
scriptExecutionContext()->vm().heap.deprecatedReportExtraMemory(extraMemoryCostDelta);
}
Vector<String> SourceBuffer::bufferedSamplesForTrackID(const AtomString& trackID)
{
auto it = m_trackBufferMap.find(trackID);
if (it == m_trackBufferMap.end())
return Vector<String>();
TrackBuffer& trackBuffer = it->value;
Vector<String> sampleDescriptions;
for (auto& pair : trackBuffer.samples.decodeOrder())
sampleDescriptions.append(toString(*pair.second));
return sampleDescriptions;
}
Vector<String> SourceBuffer::enqueuedSamplesForTrackID(const AtomString& trackID)
{
return m_private->enqueuedSamplesForTrackID(trackID);
}
MediaTime SourceBuffer::minimumUpcomingPresentationTimeForTrackID(const AtomString& trackID)
{
return m_private->minimumUpcomingPresentationTimeForTrackID(trackID);
}
void SourceBuffer::setMaximumQueueDepthForTrackID(const AtomString& trackID, size_t maxQueueDepth)
{
m_private->setMaximumQueueDepthForTrackID(trackID, maxQueueDepth);
}
Document& SourceBuffer::document() const
{
ASSERT(scriptExecutionContext());
return downcast<Document>(*scriptExecutionContext());
}
ExceptionOr<void> SourceBuffer::setMode(AppendMode newMode)
{
// 3.1 Attributes - mode
// http://www.w3.org/TR/media-source/#widl-SourceBuffer-mode
// On setting, run the following steps:
// 1. Let new mode equal the new value being assigned to this attribute.
// 2. If generate timestamps flag equals true and new mode equals "segments", then throw an InvalidAccessError exception and abort these steps.
if (m_shouldGenerateTimestamps && newMode == AppendMode::Segments)
return Exception { InvalidAccessError };
// 3. If this object has been removed from the sourceBuffers attribute of the parent media source, then throw an InvalidStateError exception and abort these steps.
// 4. If the updating attribute equals true, then throw an InvalidStateError exception and abort these steps.
if (isRemoved() || m_updating)
return Exception { InvalidStateError };
// 5. If the readyState attribute of the parent media source is in the "ended" state then run the following steps:
if (m_source->isEnded()) {
// 5.1. Set the readyState attribute of the parent media source to "open"
// 5.2. Queue a task to fire a simple event named sourceopen at the parent media source.
m_source->openIfInEndedState();
}
// 6. If the append state equals PARSING_MEDIA_SEGMENT, then throw an InvalidStateError and abort these steps.
if (m_appendState == ParsingMediaSegment)
return Exception { InvalidStateError };
// 7. If the new mode equals "sequence", then set the group start timestamp to the group end timestamp.
if (newMode == AppendMode::Sequence)
m_groupStartTimestamp = m_groupEndTimestamp;
// 8. Update the attribute to new mode.
m_mode = newMode;
return { };
}
#if !RELEASE_LOG_DISABLED
WTFLogChannel& SourceBuffer::logChannel() const
{
return LogMediaSource;
}
#endif
} // namespace WebCore
#endif