/
RealtimeIncomingVideoSourceGStreamer.cpp
122 lines (101 loc) · 4.92 KB
/
RealtimeIncomingVideoSourceGStreamer.cpp
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
/*
* Copyright (C) 2017-2022 Igalia S.L. All rights reserved.
* Copyright (C) 2022 Metrological Group B.V.
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "config.h"
#if USE(GSTREAMER_WEBRTC)
#include "RealtimeIncomingVideoSourceGStreamer.h"
#include "GStreamerCommon.h"
#include "GStreamerWebRTCUtils.h"
#include "VideoFrameGStreamer.h"
#include "VideoFrameMetadataGStreamer.h"
GST_DEBUG_CATEGORY(webkit_webrtc_incoming_video_debug);
#define GST_CAT_DEFAULT webkit_webrtc_incoming_video_debug
namespace WebCore {
RealtimeIncomingVideoSourceGStreamer::RealtimeIncomingVideoSourceGStreamer(AtomString&& videoTrackId)
: RealtimeIncomingSourceGStreamer(CaptureDevice { WTFMove(videoTrackId), CaptureDevice::DeviceType::Camera, emptyString() })
{
static std::once_flag debugRegisteredFlag;
std::call_once(debugRegisteredFlag, [] {
GST_DEBUG_CATEGORY_INIT(webkit_webrtc_incoming_video_debug, "webkitwebrtcincomingvideo", 0, "WebKit WebRTC incoming video");
});
static Atomic<uint64_t> sourceCounter = 0;
gst_element_set_name(bin(), makeString("incoming-video-source-", sourceCounter.exchangeAdd(1)).ascii().data());
GST_DEBUG_OBJECT(bin(), "New incoming video source created");
RealtimeMediaSourceSupportedConstraints constraints;
constraints.setSupportsWidth(true);
constraints.setSupportsHeight(true);
m_currentSettings = RealtimeMediaSourceSettings { };
m_currentSettings->setSupportedConstraints(WTFMove(constraints));
auto sinkPad = adoptGRef(gst_element_get_static_pad(bin(), "sink"));
gst_pad_add_probe(sinkPad.get(), static_cast<GstPadProbeType>(GST_PAD_PROBE_TYPE_BUFFER), [](GstPad*, GstPadProbeInfo* info, gpointer) -> GstPadProbeReturn {
auto videoFrameTimeMetadata = std::make_optional<VideoFrameTimeMetadata>({ });
videoFrameTimeMetadata->receiveTime = MonotonicTime::now().secondsSinceEpoch();
auto* buffer = GST_BUFFER_CAST(GST_PAD_PROBE_INFO_DATA(info));
{
GstMappedRtpBuffer rtpBuffer(buffer, GST_MAP_READ);
if (rtpBuffer)
videoFrameTimeMetadata->rtpTimestamp = gst_rtp_buffer_get_timestamp(rtpBuffer.mappedData());
}
buffer = webkitGstBufferSetVideoFrameTimeMetadata(buffer, WTFMove(videoFrameTimeMetadata));
GST_PAD_PROBE_INFO_DATA(info) = buffer;
return GST_PAD_PROBE_OK;
}, nullptr, nullptr);
start();
}
const RealtimeMediaSourceSettings& RealtimeIncomingVideoSourceGStreamer::settings()
{
if (m_currentSettings)
return m_currentSettings.value();
RealtimeMediaSourceSupportedConstraints constraints;
constraints.setSupportsWidth(true);
constraints.setSupportsHeight(true);
RealtimeMediaSourceSettings settings;
auto& size = this->size();
settings.setWidth(size.width());
settings.setHeight(size.height());
settings.setSupportedConstraints(constraints);
m_currentSettings = WTFMove(settings);
return m_currentSettings.value();
}
void RealtimeIncomingVideoSourceGStreamer::settingsDidChange(OptionSet<RealtimeMediaSourceSettings::Flag> settings)
{
if (settings.containsAny({ RealtimeMediaSourceSettings::Flag::Width, RealtimeMediaSourceSettings::Flag::Height }))
m_currentSettings = std::nullopt;
}
void RealtimeIncomingVideoSourceGStreamer::dispatchSample(GRefPtr<GstSample>&& sample)
{
auto* buffer = gst_sample_get_buffer(sample.get());
videoFrameAvailable(VideoFrameGStreamer::create(WTFMove(sample), size(), fromGstClockTime(GST_BUFFER_PTS(buffer))), { });
}
const GstStructure* RealtimeIncomingVideoSourceGStreamer::stats()
{
m_stats.reset(gst_structure_new_empty("incoming-video-stats"));
forEachVideoFrameObserver([&](auto& observer) {
auto stats = observer.queryAdditionalStats();
if (!stats)
return;
gst_structure_foreach(stats.get(), reinterpret_cast<GstStructureForeachFunc>(+[](GQuark fieldId, const GValue* value, gpointer userData) -> gboolean {
auto* source = reinterpret_cast<RealtimeIncomingVideoSourceGStreamer*>(userData);
gst_structure_set_value(source->m_stats.get(), g_quark_to_string(fieldId), value);
return TRUE;
}), this);
});
return m_stats.get();
}
} // namespace WebCore
#endif // USE(GSTREAMER_WEBRTC)