blob: 04363258c3d5d096313b9854e1952c33247ef55e [file] [log] [blame]
/*
* Copyright (C) 2007, 2009 Apple Inc. All rights reserved.
* Copyright (C) 2007 Collabora Ltd. All rights reserved.
* Copyright (C) 2007 Alp Toker <alp@atoker.com>
* Copyright (C) 2009 Gustavo Noronha Silva <gns@gnome.org>
* Copyright (C) 2009, 2010, 2011, 2012, 2013, 2016, 2017 Igalia S.L
* Copyright (C) 2015 Sebastian Dröge <sebastian@centricular.com>
* Copyright (C) 2015, 2016, 2017 Metrological Group B.V.
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public License
* aint with this library; see the file COPYING.LIB. If not, write to
* the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
#include "config.h"
#include "MediaPlayerPrivateGStreamerMSE.h"
#if ENABLE(VIDEO) && USE(GSTREAMER) && ENABLE(MEDIA_SOURCE)
#include "AppendPipeline.h"
#include "AudioTrackPrivateGStreamer.h"
#include "GStreamerCommon.h"
#include "InbandTextTrackPrivateGStreamer.h"
#include "MIMETypeRegistry.h"
#include "MediaDescription.h"
#include "MediaPlayer.h"
#include "NotImplemented.h"
#include "SourceBufferPrivateGStreamer.h"
#include "TimeRanges.h"
#include "URL.h"
#include "VideoTrackPrivateGStreamer.h"
#include <fnmatch.h>
#include <gst/app/gstappsink.h>
#include <gst/app/gstappsrc.h>
#include <gst/gst.h>
#include <gst/pbutils/pbutils.h>
#include <gst/video/video.h>
#include <wtf/Condition.h>
#include <wtf/HashSet.h>
#include <wtf/NeverDestroyed.h>
#include <wtf/StringPrintStream.h>
#include <wtf/text/AtomicString.h>
#include <wtf/text/AtomicStringHash.h>
#if ENABLE(ENCRYPTED_MEDIA)
#include "CDMClearKey.h"
#include "SharedBuffer.h"
#endif
static const char* dumpReadyState(WebCore::MediaPlayer::ReadyState readyState)
{
switch (readyState) {
case WebCore::MediaPlayer::HaveNothing: return "HaveNothing";
case WebCore::MediaPlayer::HaveMetadata: return "HaveMetadata";
case WebCore::MediaPlayer::HaveCurrentData: return "HaveCurrentData";
case WebCore::MediaPlayer::HaveFutureData: return "HaveFutureData";
case WebCore::MediaPlayer::HaveEnoughData: return "HaveEnoughData";
default: return "(unknown)";
}
}
GST_DEBUG_CATEGORY(webkit_mse_debug);
#define GST_CAT_DEFAULT webkit_mse_debug
namespace WebCore {
void MediaPlayerPrivateGStreamerMSE::registerMediaEngine(MediaEngineRegistrar registrar)
{
if (isAvailable()) {
registrar([](MediaPlayer* player) { return std::make_unique<MediaPlayerPrivateGStreamerMSE>(player); },
getSupportedTypes, supportsType, nullptr, nullptr, nullptr, supportsKeySystem);
}
}
bool initializeGStreamerAndRegisterWebKitMSEElement()
{
registerWebKitGStreamerElements();
GST_DEBUG_CATEGORY_INIT(webkit_mse_debug, "webkitmse", 0, "WebKit MSE media player");
GRefPtr<GstElementFactory> WebKitMediaSrcFactory = adoptGRef(gst_element_factory_find("webkitmediasrc"));
if (UNLIKELY(!WebKitMediaSrcFactory))
gst_element_register(nullptr, "webkitmediasrc", GST_RANK_PRIMARY + 100, WEBKIT_TYPE_MEDIA_SRC);
return true;
}
bool MediaPlayerPrivateGStreamerMSE::isAvailable()
{
if (UNLIKELY(!initializeGStreamerAndRegisterWebKitMSEElement()))
return false;
GRefPtr<GstElementFactory> factory = adoptGRef(gst_element_factory_find("playbin"));
return factory;
}
MediaPlayerPrivateGStreamerMSE::MediaPlayerPrivateGStreamerMSE(MediaPlayer* player)
: MediaPlayerPrivateGStreamer(player)
{
GST_TRACE("creating the player (%p)", this);
}
MediaPlayerPrivateGStreamerMSE::~MediaPlayerPrivateGStreamerMSE()
{
GST_TRACE("destroying the player (%p)", this);
for (auto iterator : m_appendPipelinesMap)
iterator.value->clearPlayerPrivate();
if (m_source) {
webKitMediaSrcSetMediaPlayerPrivate(WEBKIT_MEDIA_SRC(m_source.get()), nullptr);
g_signal_handlers_disconnect_by_data(m_source.get(), this);
}
if (m_playbackPipeline)
m_playbackPipeline->setWebKitMediaSrc(nullptr);
}
void MediaPlayerPrivateGStreamerMSE::load(const String& urlString)
{
if (!urlString.startsWith("mediasource")) {
// Properly fail so the global MediaPlayer tries to fallback to the next MediaPlayerPrivate.
m_networkState = MediaPlayer::FormatError;
m_player->networkStateChanged();
return;
}
if (UNLIKELY(!initializeGStreamerAndRegisterWebKitMSEElement()))
return;
if (!m_playbackPipeline)
m_playbackPipeline = PlaybackPipeline::create();
MediaPlayerPrivateGStreamer::load(urlString);
}
void MediaPlayerPrivateGStreamerMSE::load(const String& url, MediaSourcePrivateClient* mediaSource)
{
m_mediaSource = mediaSource;
load(String::format("mediasource%s", url.utf8().data()));
}
void MediaPlayerPrivateGStreamerMSE::pause()
{
m_paused = true;
MediaPlayerPrivateGStreamer::pause();
}
MediaTime MediaPlayerPrivateGStreamerMSE::durationMediaTime() const
{
if (UNLIKELY(!m_pipeline || m_errorOccured))
return MediaTime();
return m_mediaTimeDuration;
}
void MediaPlayerPrivateGStreamerMSE::seek(const MediaTime& time)
{
if (UNLIKELY(!m_pipeline || m_errorOccured))
return;
GST_INFO("[Seek] seek attempt to %s secs", toString(time).utf8().data());
// Avoid useless seeking.
MediaTime current = currentMediaTime();
if (time == current) {
if (!m_seeking)
timeChanged();
return;
}
if (isLiveStream())
return;
if (m_seeking && m_seekIsPending) {
m_seekTime = time;
return;
}
GST_DEBUG("Seeking from %s to %s seconds", toString(current).utf8().data(), toString(time).utf8().data());
MediaTime previousSeekTime = m_seekTime;
m_seekTime = time;
if (!doSeek()) {
m_seekTime = previousSeekTime;
GST_WARNING("Seeking to %s failed", toString(time).utf8().data());
return;
}
m_isEndReached = false;
GST_DEBUG("m_seeking=%s, m_seekTime=%s", boolForPrinting(m_seeking), toString(m_seekTime).utf8().data());
}
void MediaPlayerPrivateGStreamerMSE::configurePlaySink()
{
MediaPlayerPrivateGStreamer::configurePlaySink();
GRefPtr<GstElement> playsink = adoptGRef(gst_bin_get_by_name(GST_BIN(m_pipeline.get()), "playsink"));
if (playsink) {
// The default value (0) means "send events to all the sinks", instead
// of "only to the first that returns true". This is needed for MSE seek.
g_object_set(G_OBJECT(playsink.get()), "send-event-mode", 0, nullptr);
}
}
bool MediaPlayerPrivateGStreamerMSE::changePipelineState(GstState newState)
{
if (seeking()) {
GST_DEBUG("Rejected state change to %s while seeking",
gst_element_state_get_name(newState));
return true;
}
return MediaPlayerPrivateGStreamer::changePipelineState(newState);
}
void MediaPlayerPrivateGStreamerMSE::notifySeekNeedsDataForTime(const MediaTime& seekTime)
{
// Reenqueue samples needed to resume playback in the new position.
m_mediaSource->seekToTime(seekTime);
GST_DEBUG("MSE seek to %s finished", toString(seekTime).utf8().data());
if (!m_gstSeekCompleted) {
m_gstSeekCompleted = true;
maybeFinishSeek();
}
}
bool MediaPlayerPrivateGStreamerMSE::doSeek(const MediaTime&, float, GstSeekFlags)
{
// Use doSeek() instead. If anybody is calling this version of doSeek(), something is wrong.
ASSERT_NOT_REACHED();
return false;
}
bool MediaPlayerPrivateGStreamerMSE::doSeek()
{
MediaTime seekTime = m_seekTime;
double rate = m_player->rate();
GstSeekFlags seekType = static_cast<GstSeekFlags>(GST_SEEK_FLAG_FLUSH | GST_SEEK_FLAG_ACCURATE);
// Always move to seeking state to report correct 'currentTime' while pending for actual seek to complete.
m_seeking = true;
// Check if playback pipeline is ready for seek.
GstState state, newState;
GstStateChangeReturn getStateResult = gst_element_get_state(m_pipeline.get(), &state, &newState, 0);
if (getStateResult == GST_STATE_CHANGE_FAILURE || getStateResult == GST_STATE_CHANGE_NO_PREROLL) {
GST_DEBUG("[Seek] cannot seek, current state change is %s", gst_element_state_change_return_get_name(getStateResult));
webKitMediaSrcSetReadyForSamples(WEBKIT_MEDIA_SRC(m_source.get()), true);
m_seeking = false;
return false;
}
if ((getStateResult == GST_STATE_CHANGE_ASYNC
&& !(state == GST_STATE_PLAYING && newState == GST_STATE_PAUSED))
|| state < GST_STATE_PAUSED
|| m_isEndReached
|| !m_gstSeekCompleted) {
CString reason = "Unknown reason";
if (getStateResult == GST_STATE_CHANGE_ASYNC) {
reason = String::format("In async change %s --> %s",
gst_element_state_get_name(state),
gst_element_state_get_name(newState)).utf8();
} else if (state < GST_STATE_PAUSED)
reason = "State less than PAUSED";
else if (m_isEndReached)
reason = "End reached";
else if (!m_gstSeekCompleted)
reason = "Previous seek is not finished yet";
GST_DEBUG("[Seek] Delaying the seek: %s", reason.data());
m_seekIsPending = true;
if (m_isEndReached) {
GST_DEBUG("[Seek] reset pipeline");
m_resetPipeline = true;
m_seeking = false;
if (!changePipelineState(GST_STATE_PAUSED))
loadingFailed(MediaPlayer::Empty);
else
m_seeking = true;
}
return m_seeking;
}
// Stop accepting new samples until actual seek is finished.
webKitMediaSrcSetReadyForSamples(WEBKIT_MEDIA_SRC(m_source.get()), false);
// Correct seek time if it helps to fix a small gap.
if (!isTimeBuffered(seekTime)) {
// Look if a near future time (<0.1 sec.) is buffered and change the seek target time.
if (m_mediaSource) {
const MediaTime miniGap = MediaTime(1, 10);
MediaTime nearest = m_mediaSource->buffered()->nearest(seekTime);
if (nearest.isValid() && nearest > seekTime && (nearest - seekTime) <= miniGap && isTimeBuffered(nearest + miniGap)) {
GST_DEBUG("[Seek] Changed the seek target time from %s to %s, a near point in the future", toString(seekTime).utf8().data(), toString(nearest).utf8().data());
seekTime = nearest;
}
}
}
// Check if MSE has samples for requested time and defer actual seek if needed.
if (!isTimeBuffered(seekTime)) {
GST_DEBUG("[Seek] Delaying the seek: MSE is not ready");
GstStateChangeReturn setStateResult = gst_element_set_state(m_pipeline.get(), GST_STATE_PAUSED);
if (setStateResult == GST_STATE_CHANGE_FAILURE) {
GST_DEBUG("[Seek] Cannot seek, failed to pause playback pipeline.");
webKitMediaSrcSetReadyForSamples(WEBKIT_MEDIA_SRC(m_source.get()), true);
m_seeking = false;
return false;
}
m_readyState = MediaPlayer::HaveMetadata;
notifySeekNeedsDataForTime(seekTime);
ASSERT(!m_mseSeekCompleted);
return true;
}
// Complete previous MSE seek if needed.
if (!m_mseSeekCompleted) {
m_mediaSource->monitorSourceBuffers();
ASSERT(m_mseSeekCompleted);
// Note: seekCompleted will recursively call us.
return m_seeking;
}
GST_DEBUG("We can seek now");
MediaTime startTime = seekTime, endTime = MediaTime::invalidTime();
if (rate < 0) {
startTime = MediaTime::zeroTime();
endTime = seekTime;
}
if (!rate)
rate = 1;
GST_DEBUG("Actual seek to %s, end time: %s, rate: %f", toString(startTime).utf8().data(), toString(endTime).utf8().data(), rate);
// This will call notifySeekNeedsData() after some time to tell that the pipeline is ready for sample enqueuing.
webKitMediaSrcPrepareSeek(WEBKIT_MEDIA_SRC(m_source.get()), seekTime);
m_gstSeekCompleted = false;
if (!gst_element_seek(m_pipeline.get(), rate, GST_FORMAT_TIME, seekType, GST_SEEK_TYPE_SET, toGstClockTime(startTime), GST_SEEK_TYPE_SET, toGstClockTime(endTime))) {
webKitMediaSrcSetReadyForSamples(WEBKIT_MEDIA_SRC(m_source.get()), true);
m_seeking = false;
m_gstSeekCompleted = true;
GST_DEBUG("doSeek(): gst_element_seek() failed, returning false");
return false;
}
// The samples will be enqueued in notifySeekNeedsData().
GST_DEBUG("doSeek(): gst_element_seek() succeeded, returning true");
return true;
}
void MediaPlayerPrivateGStreamerMSE::maybeFinishSeek()
{
if (!m_seeking || !m_mseSeekCompleted || !m_gstSeekCompleted)
return;
GstState state, newState;
GstStateChangeReturn getStateResult = gst_element_get_state(m_pipeline.get(), &state, &newState, 0);
if (getStateResult == GST_STATE_CHANGE_ASYNC
&& !(state == GST_STATE_PLAYING && newState == GST_STATE_PAUSED)) {
GST_DEBUG("[Seek] Delaying seek finish");
return;
}
if (m_seekIsPending) {
GST_DEBUG("[Seek] Committing pending seek to %s", toString(m_seekTime).utf8().data());
m_seekIsPending = false;
if (!doSeek()) {
GST_WARNING("[Seek] Seeking to %s failed", toString(m_seekTime).utf8().data());
m_cachedPosition = MediaTime::invalidTime();
}
return;
}
GST_DEBUG("[Seek] Seeked to %s", toString(m_seekTime).utf8().data());
webKitMediaSrcSetReadyForSamples(WEBKIT_MEDIA_SRC(m_source.get()), true);
m_seeking = false;
m_cachedPosition = MediaTime::invalidTime();
// The pipeline can still have a pending state. In this case a position query will fail.
// Right now we can use m_seekTime as a fallback.
m_canFallBackToLastFinishedSeekPosition = true;
timeChanged();
}
void MediaPlayerPrivateGStreamerMSE::updatePlaybackRate()
{
notImplemented();
}
bool MediaPlayerPrivateGStreamerMSE::seeking() const
{
return m_seeking;
}
// FIXME: MediaPlayerPrivateGStreamer manages the ReadyState on its own. We shouldn't change it manually.
void MediaPlayerPrivateGStreamerMSE::setReadyState(MediaPlayer::ReadyState readyState)
{
if (readyState == m_readyState)
return;
if (seeking()) {
GST_DEBUG("Skip ready state change(%s -> %s) due to seek\n", dumpReadyState(m_readyState), dumpReadyState(readyState));
return;
}
GST_DEBUG("Ready State Changed manually from %u to %u", m_readyState, readyState);
MediaPlayer::ReadyState oldReadyState = m_readyState;
m_readyState = readyState;
GST_DEBUG("m_readyState: %s -> %s", dumpReadyState(oldReadyState), dumpReadyState(m_readyState));
if (oldReadyState < MediaPlayer::HaveCurrentData && m_readyState >= MediaPlayer::HaveCurrentData) {
GST_DEBUG("[Seek] Reporting load state changed to trigger seek continuation");
loadStateChanged();
}
m_player->readyStateChanged();
GstState pipelineState;
GstStateChangeReturn getStateResult = gst_element_get_state(m_pipeline.get(), &pipelineState, nullptr, 250 * GST_NSECOND);
bool isPlaying = (getStateResult == GST_STATE_CHANGE_SUCCESS && pipelineState == GST_STATE_PLAYING);
if (m_readyState == MediaPlayer::HaveMetadata && oldReadyState > MediaPlayer::HaveMetadata && isPlaying) {
GST_TRACE("Changing pipeline to PAUSED...");
bool ok = changePipelineState(GST_STATE_PAUSED);
GST_TRACE("Changed pipeline to PAUSED: %s", ok ? "Success" : "Error");
}
}
void MediaPlayerPrivateGStreamerMSE::waitForSeekCompleted()
{
if (!m_seeking)
return;
GST_DEBUG("Waiting for MSE seek completed");
m_mseSeekCompleted = false;
}
void MediaPlayerPrivateGStreamerMSE::seekCompleted()
{
if (m_mseSeekCompleted)
return;
GST_DEBUG("MSE seek completed");
m_mseSeekCompleted = true;
doSeek();
if (!seeking() && m_readyState >= MediaPlayer::HaveFutureData)
changePipelineState(GST_STATE_PLAYING);
if (!seeking())
m_player->timeChanged();
}
void MediaPlayerPrivateGStreamerMSE::setRate(float)
{
notImplemented();
}
std::unique_ptr<PlatformTimeRanges> MediaPlayerPrivateGStreamerMSE::buffered() const
{
return m_mediaSource ? m_mediaSource->buffered() : std::make_unique<PlatformTimeRanges>();
}
void MediaPlayerPrivateGStreamerMSE::sourceSetup(GstElement* sourceElement)
{
m_source = sourceElement;
ASSERT(WEBKIT_IS_MEDIA_SRC(m_source.get()));
m_playbackPipeline->setWebKitMediaSrc(WEBKIT_MEDIA_SRC(m_source.get()));
MediaSourceGStreamer::open(*m_mediaSource.get(), *this);
g_signal_connect_swapped(m_source.get(), "video-changed", G_CALLBACK(videoChangedCallback), this);
g_signal_connect_swapped(m_source.get(), "audio-changed", G_CALLBACK(audioChangedCallback), this);
g_signal_connect_swapped(m_source.get(), "text-changed", G_CALLBACK(textChangedCallback), this);
webKitMediaSrcSetMediaPlayerPrivate(WEBKIT_MEDIA_SRC(m_source.get()), this);
}
void MediaPlayerPrivateGStreamerMSE::updateStates()
{
if (UNLIKELY(!m_pipeline || m_errorOccured))
return;
MediaPlayer::NetworkState oldNetworkState = m_networkState;
MediaPlayer::ReadyState oldReadyState = m_readyState;
GstState state, pending;
GstStateChangeReturn getStateResult = gst_element_get_state(m_pipeline.get(), &state, &pending, 250 * GST_NSECOND);
bool shouldUpdatePlaybackState = false;
switch (getStateResult) {
case GST_STATE_CHANGE_SUCCESS: {
GST_DEBUG("State: %s, pending: %s", gst_element_state_get_name(state), gst_element_state_get_name(pending));
// Do nothing if on EOS and state changed to READY to avoid recreating the player
// on HTMLMediaElement and properly generate the video 'ended' event.
if (m_isEndReached && state == GST_STATE_READY)
break;
m_resetPipeline = (state <= GST_STATE_READY);
if (m_resetPipeline)
m_mediaTimeDuration = MediaTime::zeroTime();
// Update ready and network states.
switch (state) {
case GST_STATE_NULL:
m_readyState = MediaPlayer::HaveNothing;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
m_networkState = MediaPlayer::Empty;
break;
case GST_STATE_READY:
m_readyState = MediaPlayer::HaveMetadata;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
m_networkState = MediaPlayer::Empty;
break;
case GST_STATE_PAUSED:
case GST_STATE_PLAYING:
if (seeking()) {
m_readyState = MediaPlayer::HaveMetadata;
// FIXME: Should we manage NetworkState too?
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
} else if (m_buffering) {
if (m_bufferingPercentage == 100) {
GST_DEBUG("[Buffering] Complete.");
m_buffering = false;
m_readyState = MediaPlayer::HaveEnoughData;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
m_networkState = m_downloadFinished ? MediaPlayer::Idle : MediaPlayer::Loading;
} else {
m_readyState = MediaPlayer::HaveCurrentData;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
m_networkState = MediaPlayer::Loading;
}
} else if (m_downloadFinished) {
m_readyState = MediaPlayer::HaveEnoughData;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
m_networkState = MediaPlayer::Loaded;
} else {
m_readyState = MediaPlayer::HaveFutureData;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
m_networkState = MediaPlayer::Loading;
}
if (m_eosMarked && state == GST_STATE_PLAYING)
m_eosPending = true;
break;
default:
ASSERT_NOT_REACHED();
break;
}
// Sync states where needed.
if (state == GST_STATE_PAUSED) {
if (!m_volumeAndMuteInitialized) {
notifyPlayerOfVolumeChange();
notifyPlayerOfMute();
m_volumeAndMuteInitialized = true;
}
if (!seeking() && !m_buffering && !m_paused && m_playbackRate) {
GST_DEBUG("[Buffering] Restarting playback.");
changePipelineState(GST_STATE_PLAYING);
}
} else if (state == GST_STATE_PLAYING) {
m_paused = false;
if ((m_buffering && !isLiveStream()) || !m_playbackRate) {
GST_DEBUG("[Buffering] Pausing stream for buffering.");
changePipelineState(GST_STATE_PAUSED);
}
} else
m_paused = true;
if (m_requestedState == GST_STATE_PAUSED && state == GST_STATE_PAUSED) {
shouldUpdatePlaybackState = true;
GST_DEBUG("Requested state change to %s was completed", gst_element_state_get_name(state));
}
break;
}
case GST_STATE_CHANGE_ASYNC:
GST_DEBUG("Async: State: %s, pending: %s", gst_element_state_get_name(state), gst_element_state_get_name(pending));
// Change in progress.
break;
case GST_STATE_CHANGE_FAILURE:
GST_WARNING("Failure: State: %s, pending: %s", gst_element_state_get_name(state), gst_element_state_get_name(pending));
// Change failed.
return;
case GST_STATE_CHANGE_NO_PREROLL:
GST_DEBUG("No preroll: State: %s, pending: %s", gst_element_state_get_name(state), gst_element_state_get_name(pending));
// Live pipelines go in PAUSED without prerolling.
m_isStreaming = true;
if (state == GST_STATE_READY) {
m_readyState = MediaPlayer::HaveNothing;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
} else if (state == GST_STATE_PAUSED) {
m_readyState = MediaPlayer::HaveEnoughData;
GST_DEBUG("m_readyState=%s", dumpReadyState(m_readyState));
m_paused = true;
} else if (state == GST_STATE_PLAYING)
m_paused = false;
if (!m_paused && m_playbackRate)
changePipelineState(GST_STATE_PLAYING);
m_networkState = MediaPlayer::Loading;
break;
default:
GST_DEBUG("Else : %d", getStateResult);
break;
}
m_requestedState = GST_STATE_VOID_PENDING;
if (shouldUpdatePlaybackState)
m_player->playbackStateChanged();
if (m_networkState != oldNetworkState) {
GST_DEBUG("Network State Changed from %u to %u", oldNetworkState, m_networkState);
m_player->networkStateChanged();
}
if (m_readyState != oldReadyState) {
GST_DEBUG("Ready State Changed from %u to %u", oldReadyState, m_readyState);
m_player->readyStateChanged();
}
if (getStateResult == GST_STATE_CHANGE_SUCCESS && state >= GST_STATE_PAUSED) {
updatePlaybackRate();
maybeFinishSeek();
}
}
void MediaPlayerPrivateGStreamerMSE::asyncStateChangeDone()
{
if (UNLIKELY(!m_pipeline || m_errorOccured))
return;
if (m_seeking)
maybeFinishSeek();
else
updateStates();
}
bool MediaPlayerPrivateGStreamerMSE::isTimeBuffered(const MediaTime &time) const
{
bool result = m_mediaSource && m_mediaSource->buffered()->contain(time);
GST_DEBUG("Time %s buffered? %s", toString(time).utf8().data(), boolForPrinting(result));
return result;
}
void MediaPlayerPrivateGStreamerMSE::setMediaSourceClient(Ref<MediaSourceClientGStreamerMSE> client)
{
m_mediaSourceClient = client.ptr();
}
RefPtr<MediaSourceClientGStreamerMSE> MediaPlayerPrivateGStreamerMSE::mediaSourceClient()
{
return m_mediaSourceClient;
}
void MediaPlayerPrivateGStreamerMSE::durationChanged()
{
if (!m_mediaSourceClient) {
GST_DEBUG("m_mediaSourceClient is null, doing nothing");
return;
}
MediaTime previousDuration = m_mediaTimeDuration;
m_mediaTimeDuration = m_mediaSourceClient->duration();
GST_TRACE("previous=%s, new=%s", toString(previousDuration).utf8().data(), toString(m_mediaTimeDuration).utf8().data());
// Avoid emiting durationchanged in the case where the previous duration was 0 because that case is already handled
// by the HTMLMediaElement.
if (m_mediaTimeDuration != previousDuration && m_mediaTimeDuration.isValid() && previousDuration.isValid()) {
m_player->durationChanged();
m_playbackPipeline->notifyDurationChanged();
m_mediaSource->durationChanged(m_mediaTimeDuration);
}
}
static HashSet<String, ASCIICaseInsensitiveHash>& mimeTypeCache()
{
static NeverDestroyed<HashSet<String, ASCIICaseInsensitiveHash>> cache = []()
{
initializeGStreamerAndRegisterWebKitMSEElement();
HashSet<String, ASCIICaseInsensitiveHash> set;
const char* mimeTypes[] = {
"video/mp4",
"audio/mp4",
"video/webm",
"audio/webm"
};
for (auto& type : mimeTypes)
set.add(type);
return set;
}();
return cache;
}
void MediaPlayerPrivateGStreamerMSE::getSupportedTypes(HashSet<String, ASCIICaseInsensitiveHash>& types)
{
types = mimeTypeCache();
}
void MediaPlayerPrivateGStreamerMSE::trackDetected(RefPtr<AppendPipeline> appendPipeline, RefPtr<WebCore::TrackPrivateBase> newTrack, bool firstTrackDetected)
{
ASSERT(appendPipeline->track() == newTrack);
GstCaps* caps = appendPipeline->appsinkCaps();
ASSERT(caps);
GST_DEBUG("track ID: %s, caps: %" GST_PTR_FORMAT, newTrack->id().string().latin1().data(), caps);
if (doCapsHaveType(caps, GST_VIDEO_CAPS_TYPE_PREFIX)) {
std::optional<FloatSize> size = getVideoResolutionFromCaps(caps);
if (size.has_value())
m_videoSize = size.value();
}
if (firstTrackDetected)
m_playbackPipeline->attachTrack(appendPipeline->sourceBufferPrivate(), newTrack, caps);
else
m_playbackPipeline->reattachTrack(appendPipeline->sourceBufferPrivate(), newTrack, caps);
}
const static HashSet<AtomicString>& codecSet()
{
static NeverDestroyed<HashSet<AtomicString>> codecTypes = []()
{
MediaPlayerPrivateGStreamerBase::initializeGStreamerAndRegisterWebKitElements();
HashSet<AtomicString> set;
GList* audioDecoderFactories = gst_element_factory_list_get_elements(GST_ELEMENT_FACTORY_TYPE_DECODER | GST_ELEMENT_FACTORY_TYPE_MEDIA_AUDIO, GST_RANK_MARGINAL);
GList* videoDecoderFactories = gst_element_factory_list_get_elements(GST_ELEMENT_FACTORY_TYPE_DECODER | GST_ELEMENT_FACTORY_TYPE_MEDIA_VIDEO, GST_RANK_MARGINAL);
enum ElementType {
AudioDecoder = 0,
VideoDecoder
};
struct GstCapsWebKitMapping {
ElementType elementType;
const char* capsString;
Vector<AtomicString> webkitCodecs;
};
GstCapsWebKitMapping mapping[] = {
{ VideoDecoder, "video/x-h264, profile=(string){ constrained-baseline, baseline }", { "x-h264", "avc*" } },
{ VideoDecoder, "video/mpeg, mpegversion=(int){1,2}, systemstream=(boolean)false", { "mpeg" } },
{ VideoDecoder, "video/x-vp8", { "vp8", "x-vp8" } },
{ VideoDecoder, "video/x-vp9", { "vp9", "x-vp9" } },
{ AudioDecoder, "audio/x-vorbis", { "vorbis", "x-vorbis" } },
{ AudioDecoder, "audio/x-opus", { "opus", "x-opus" } }
};
for (auto& current : mapping) {
GList* factories = nullptr;
switch (current.elementType) {
case AudioDecoder:
factories = audioDecoderFactories;
break;
case VideoDecoder:
factories = videoDecoderFactories;
break;
default:
g_assert_not_reached();
break;
}
g_assert_nonnull(factories);
if (gstRegistryHasElementForMediaType(factories, current.capsString) && factories != nullptr) {
if (!current.webkitCodecs.isEmpty()) {
for (const auto& mimeType : current.webkitCodecs)
set.add(mimeType);
} else
set.add(AtomicString(current.capsString));
}
}
bool audioMpegSupported = false;
if (gstRegistryHasElementForMediaType(audioDecoderFactories, "audio/mpeg, mpegversion=(int)1, layer=(int)[1, 3]")) {
audioMpegSupported = true;
set.add(AtomicString("audio/mp3"));
}
if (gstRegistryHasElementForMediaType(audioDecoderFactories, "audio/mpeg, mpegversion=(int){2, 4}")) {
audioMpegSupported = true;
set.add(AtomicString("mp4a*"));
}
if (audioMpegSupported) {
set.add(AtomicString("audio/mpeg"));
set.add(AtomicString("audio/x-mpeg"));
}
gst_plugin_feature_list_free(audioDecoderFactories);
gst_plugin_feature_list_free(videoDecoderFactories);
return set;
}();
return codecTypes;
}
bool MediaPlayerPrivateGStreamerMSE::supportsCodec(String codec)
{
// If the codec is named like a mimetype (eg: video/avc) remove the "video/" part.
size_t slashIndex = codec.find('/');
if (slashIndex != WTF::notFound)
codec = codec.substring(slashIndex+1);
for (const auto& pattern : codecSet()) {
bool codecMatchesPattern = !fnmatch(pattern.string().utf8().data(), codec.utf8().data(), 0);
if (codecMatchesPattern)
return true;
}
return false;
}
bool MediaPlayerPrivateGStreamerMSE::supportsAllCodecs(const Vector<String>& codecs)
{
for (String codec : codecs) {
if (!supportsCodec(codec))
return false;
}
return true;
}
MediaPlayer::SupportsType MediaPlayerPrivateGStreamerMSE::supportsType(const MediaEngineSupportParameters& parameters)
{
MediaPlayer::SupportsType result = MediaPlayer::IsNotSupported;
if (!parameters.isMediaSource)
return result;
auto containerType = parameters.type.containerType();
// YouTube TV provides empty types for some videos and we want to be selected as best media engine for them.
if (containerType.isEmpty()) {
result = MediaPlayer::MayBeSupported;
return result;
}
// Spec says we should not return "probably" if the codecs string is empty.
if (mimeTypeCache().contains(containerType)) {
Vector<String> codecs = parameters.type.codecs();
if (codecs.isEmpty())
result = MediaPlayer::MayBeSupported;
else
result = supportsAllCodecs(codecs) ? MediaPlayer::IsSupported : MediaPlayer::IsNotSupported;
}
return extendedSupportsType(parameters, result);
}
void MediaPlayerPrivateGStreamerMSE::markEndOfStream(MediaSourcePrivate::EndOfStreamStatus status)
{
if (status != MediaSourcePrivate::EosNoError)
return;
GST_DEBUG("Marking end of stream");
m_eosMarked = true;
updateStates();
}
MediaTime MediaPlayerPrivateGStreamerMSE::currentMediaTime() const
{
MediaTime position = MediaPlayerPrivateGStreamer::currentMediaTime();
if (m_eosPending && (paused() || (position >= durationMediaTime()))) {
if (m_networkState != MediaPlayer::Loaded) {
m_networkState = MediaPlayer::Loaded;
m_player->networkStateChanged();
}
m_eosPending = false;
m_isEndReached = true;
m_cachedPosition = m_mediaTimeDuration;
m_durationAtEOS = m_mediaTimeDuration;
m_player->timeChanged();
}
return position;
}
MediaTime MediaPlayerPrivateGStreamerMSE::maxMediaTimeSeekable() const
{
if (UNLIKELY(m_errorOccured))
return MediaTime::zeroTime();
GST_DEBUG("maxMediaTimeSeekable");
MediaTime result = durationMediaTime();
// Infinite duration means live stream.
if (result.isPositiveInfinite()) {
MediaTime maxBufferedTime = buffered()->maximumBufferedTime();
// Return the highest end time reported by the buffered attribute.
result = maxBufferedTime.isValid() ? maxBufferedTime : MediaTime::zeroTime();
}
return result;
}
#if ENABLE(ENCRYPTED_MEDIA)
void MediaPlayerPrivateGStreamerMSE::attemptToDecryptWithInstance(CDMInstance& instance)
{
if (is<CDMInstanceClearKey>(instance)) {
auto& ckInstance = downcast<CDMInstanceClearKey>(instance);
if (ckInstance.keys().isEmpty())
return;
GValue keyIDList = G_VALUE_INIT, keyValueList = G_VALUE_INIT;
g_value_init(&keyIDList, GST_TYPE_LIST);
g_value_init(&keyValueList, GST_TYPE_LIST);
auto appendBuffer =
[](GValue* valueList, const SharedBuffer& buffer)
{
GValue* bufferValue = g_new0(GValue, 1);
g_value_init(bufferValue, GST_TYPE_BUFFER);
gst_value_take_buffer(bufferValue,
gst_buffer_new_wrapped(g_memdup(buffer.data(), buffer.size()), buffer.size()));
gst_value_list_append_and_take_value(valueList, bufferValue);
};
for (auto& key : ckInstance.keys()) {
appendBuffer(&keyIDList, *key.keyIDData);
appendBuffer(&keyValueList, *key.keyValueData);
}
GUniquePtr<GstStructure> structure(gst_structure_new_empty("drm-cipher-clearkey"));
gst_structure_set_value(structure.get(), "key-ids", &keyIDList);
gst_structure_set_value(structure.get(), "key-values", &keyValueList);
gst_element_send_event(m_playbackPipeline->pipeline(), gst_event_new_custom(GST_EVENT_CUSTOM_DOWNSTREAM_OOB, gst_structure_copy(structure.get())));
}
}
#endif
} // namespace WebCore.
#endif // USE(GSTREAMER)