blob: ddea84b3377ef8299f55b445df2a905cc06dc9d2 [file] [log] [blame]
/*
* Copyright (C) 2012, 2015, 2016, 2018 Igalia S.L
* Copyright (C) 2015, 2016, 2018 Metrological Group B.V.
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "config.h"
#if USE(GSTREAMER) && USE(LIBWEBRTC)
#include "GStreamerVideoFrameLibWebRTC.h"
#include <thread>
namespace WebCore {
const GRefPtr<GstSample> GStreamerSampleFromLibWebRTCVideoFrame(const webrtc::VideoFrame& frame)
{
if (frame.video_frame_buffer()->type() == webrtc::VideoFrameBuffer::Type::kNative) {
auto framebuffer = static_cast<GStreamerVideoFrameLibWebRTC*>(frame.video_frame_buffer().get());
auto gstsample = framebuffer->getSample();
GST_LOG("Reusing native GStreamer sample: %p", gstsample.get());
return gstsample;
}
auto webrtcbuffer = frame.video_frame_buffer().get()->ToI420();
// FIXME - Check lifetime of those buffers.
const uint8_t* comps[3] = {
webrtcbuffer->DataY(),
webrtcbuffer->DataU(),
webrtcbuffer->DataV()
};
GstVideoInfo info;
gst_video_info_set_format(&info, GST_VIDEO_FORMAT_I420, frame.width(), frame.height());
auto buffer = adoptGRef(gst_buffer_new());
for (gint i = 0; i < 3; i++) {
gsize compsize = GST_VIDEO_INFO_COMP_STRIDE(&info, i) * GST_VIDEO_INFO_COMP_HEIGHT(&info, i);
GstMemory* comp = gst_memory_new_wrapped(
static_cast<GstMemoryFlags>(GST_MEMORY_FLAG_PHYSICALLY_CONTIGUOUS | GST_MEMORY_FLAG_READONLY),
const_cast<gpointer>(reinterpret_cast<const void*>(comps[i])), compsize, 0, compsize, webrtcbuffer, nullptr);
gst_buffer_append_memory(buffer.get(), comp);
}
auto caps = adoptGRef(gst_video_info_to_caps(&info));
auto sample = adoptGRef(gst_sample_new(buffer.get(), caps.get(), nullptr, nullptr));
return sample;
}
rtc::scoped_refptr<webrtc::VideoFrameBuffer> GStreamerVideoFrameLibWebRTC::create(GstSample * sample)
{
GstVideoInfo info;
if (!gst_video_info_from_caps(&info, gst_sample_get_caps(sample)))
ASSERT_NOT_REACHED();
return rtc::scoped_refptr<webrtc::VideoFrameBuffer>(new GStreamerVideoFrameLibWebRTC(sample, info));
}
std::unique_ptr<webrtc::VideoFrame> LibWebRTCVideoFrameFromGStreamerSample(GstSample* sample, webrtc::VideoRotation rotation,
int64_t timestamp, int64_t renderTimeMs)
{
auto frameBuffer(GStreamerVideoFrameLibWebRTC::create(sample));
return std::unique_ptr<webrtc::VideoFrame>(
new webrtc::VideoFrame(frameBuffer, timestamp, renderTimeMs, rotation));
}
webrtc::VideoFrameBuffer::Type GStreamerVideoFrameLibWebRTC::type() const
{
return Type::kNative;
}
GRefPtr<GstSample> GStreamerVideoFrameLibWebRTC::getSample()
{
return m_sample.get();
}
rtc::scoped_refptr<webrtc::I420BufferInterface> GStreamerVideoFrameLibWebRTC::ToI420()
{
GstMappedFrame inFrame(m_sample, GST_MAP_READ);
if (!inFrame) {
GST_WARNING("Could not map frame");
return nullptr;
}
auto newBuffer = m_bufferPool.CreateBuffer(inFrame.width(), inFrame.height());
ASSERT(newBuffer);
if (!newBuffer) {
GST_WARNING("RealtimeOutgoingVideoSourceGStreamer::videoSampleAvailable unable to allocate buffer for conversion to YUV");
return nullptr;
}
if (inFrame.format() != GST_VIDEO_FORMAT_I420) {
GstVideoInfo outInfo;
gst_video_info_set_format(&outInfo, GST_VIDEO_FORMAT_I420, inFrame.width(),
inFrame.height());
auto info = inFrame.info();
outInfo.fps_n = info->fps_n;
outInfo.fps_d = info->fps_d;
GRefPtr<GstBuffer> buffer = adoptGRef(gst_buffer_new_wrapped_full(GST_MEMORY_FLAG_NO_SHARE, newBuffer->MutableDataY(),
outInfo.size, 0, outInfo.size, nullptr, nullptr));
GstMappedFrame outFrame(buffer.get(), outInfo, GST_MAP_WRITE);
GUniquePtr<GstVideoConverter> videoConverter(gst_video_converter_new(inFrame.info(),
&outInfo, gst_structure_new("GstVideoConvertConfig",
GST_VIDEO_CONVERTER_OPT_THREADS, G_TYPE_UINT, std::max(std::thread::hardware_concurrency(), 1u), nullptr)));
ASSERT(videoConverter);
gst_video_converter_frame(videoConverter.get(), inFrame.get(), outFrame.get());
return newBuffer;
}
newBuffer->Copy(
inFrame.width(),
inFrame.height(),
inFrame.ComponentData(0),
inFrame.ComponentStride(0),
inFrame.ComponentData(1),
inFrame.ComponentStride(1),
inFrame.ComponentData(2),
inFrame.ComponentStride(2));
return newBuffer;
}
}
#endif // USE(LIBWEBRTC)