gstreamer/subprojects/gst-plugins-good/gst/rtp/gstrtpvp8depay.c
Sebastian Dröge cbc6761199 rtpvp8depay: If configured to wait for keyframes after packet loss, also do that if incomplete frames are detected
This can happen if the data inside the packets is incomplete without the
seqnums being discontinuous because of ULPFEC being used.

Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/2947>
2022-08-31 08:58:03 +00:00

579 lines
18 KiB
C

/* gstrtpvp8depay.c - Source for GstRtpVP8Depay
* Copyright (C) 2011 Sjoerd Simons <sjoerd@luon.net>
* Copyright (C) 2011 Collabora Ltd.
* Contact: Youness Alaoui <youness.alaoui@collabora.co.uk>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
*/
#ifdef HAVE_CONFIG_H
# include "config.h"
#endif
#include "gstrtpelements.h"
#include "gstrtpvp8depay.h"
#include "gstrtputils.h"
#include <gst/video/video.h>
#include <stdio.h>
GST_DEBUG_CATEGORY_STATIC (gst_rtp_vp8_depay_debug);
#define GST_CAT_DEFAULT gst_rtp_vp8_depay_debug
static void gst_rtp_vp8_depay_dispose (GObject * object);
static void gst_rtp_vp8_depay_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec);
static void gst_rtp_vp8_depay_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec);
static GstBuffer *gst_rtp_vp8_depay_process (GstRTPBaseDepayload * depayload,
GstRTPBuffer * rtp);
static GstStateChangeReturn gst_rtp_vp8_depay_change_state (GstElement *
element, GstStateChange transition);
static gboolean gst_rtp_vp8_depay_handle_event (GstRTPBaseDepayload * depay,
GstEvent * event);
static gboolean gst_rtp_vp8_depay_packet_lost (GstRTPBaseDepayload * depay,
GstEvent * event);
G_DEFINE_TYPE (GstRtpVP8Depay, gst_rtp_vp8_depay, GST_TYPE_RTP_BASE_DEPAYLOAD);
GST_ELEMENT_REGISTER_DEFINE_WITH_CODE (rtpvp8depay, "rtpvp8depay",
GST_RANK_MARGINAL, GST_TYPE_RTP_VP8_DEPAY, rtp_element_init (plugin));
static GstStaticPadTemplate gst_rtp_vp8_depay_src_template =
GST_STATIC_PAD_TEMPLATE ("src",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("video/x-vp8"));
static GstStaticPadTemplate gst_rtp_vp8_depay_sink_template =
GST_STATIC_PAD_TEMPLATE ("sink",
GST_PAD_SINK,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("application/x-rtp, "
"clock-rate = (int) 90000,"
"media = (string) \"video\","
"encoding-name = (string) { \"VP8\", \"VP8-DRAFT-IETF-01\" }"));
#define DEFAULT_WAIT_FOR_KEYFRAME FALSE
#define DEFAULT_REQUEST_KEYFRAME FALSE
enum
{
PROP_0,
PROP_WAIT_FOR_KEYFRAME,
PROP_REQUEST_KEYFRAME,
};
#define PICTURE_ID_NONE (UINT_MAX)
#define IS_PICTURE_ID_15BITS(pid) (((guint)(pid) & 0x8000) != 0)
static void
gst_rtp_vp8_depay_init (GstRtpVP8Depay * self)
{
self->adapter = gst_adapter_new ();
self->started = FALSE;
self->wait_for_keyframe = DEFAULT_WAIT_FOR_KEYFRAME;
self->request_keyframe = DEFAULT_REQUEST_KEYFRAME;
self->last_pushed_was_lost_event = FALSE;
}
static void
gst_rtp_vp8_depay_class_init (GstRtpVP8DepayClass * gst_rtp_vp8_depay_class)
{
GObjectClass *object_class = G_OBJECT_CLASS (gst_rtp_vp8_depay_class);
GstElementClass *element_class = GST_ELEMENT_CLASS (gst_rtp_vp8_depay_class);
GstRTPBaseDepayloadClass *depay_class =
(GstRTPBaseDepayloadClass *) (gst_rtp_vp8_depay_class);
gst_element_class_add_static_pad_template (element_class,
&gst_rtp_vp8_depay_sink_template);
gst_element_class_add_static_pad_template (element_class,
&gst_rtp_vp8_depay_src_template);
gst_element_class_set_static_metadata (element_class, "RTP VP8 depayloader",
"Codec/Depayloader/Network/RTP",
"Extracts VP8 video from RTP packets)",
"Sjoerd Simons <sjoerd@luon.net>");
object_class->dispose = gst_rtp_vp8_depay_dispose;
object_class->set_property = gst_rtp_vp8_depay_set_property;
object_class->get_property = gst_rtp_vp8_depay_get_property;
g_object_class_install_property (object_class, PROP_WAIT_FOR_KEYFRAME,
g_param_spec_boolean ("wait-for-keyframe", "Wait for Keyframe",
"Wait for the next keyframe after packet loss",
DEFAULT_WAIT_FOR_KEYFRAME,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
/**
* GstRtpVP8Depay:request-keyframe:
*
* Request new keyframe when packet loss is detected
*
* Since: 1.20
*/
g_object_class_install_property (object_class, PROP_REQUEST_KEYFRAME,
g_param_spec_boolean ("request-keyframe", "Request Keyframe",
"Request new keyframe when packet loss is detected",
DEFAULT_REQUEST_KEYFRAME,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
element_class->change_state = gst_rtp_vp8_depay_change_state;
depay_class->process_rtp_packet = gst_rtp_vp8_depay_process;
depay_class->handle_event = gst_rtp_vp8_depay_handle_event;
depay_class->packet_lost = gst_rtp_vp8_depay_packet_lost;
GST_DEBUG_CATEGORY_INIT (gst_rtp_vp8_depay_debug, "rtpvp8depay", 0,
"VP8 Video RTP Depayloader");
}
static void
gst_rtp_vp8_depay_dispose (GObject * object)
{
GstRtpVP8Depay *self = GST_RTP_VP8_DEPAY (object);
if (self->adapter != NULL)
g_object_unref (self->adapter);
self->adapter = NULL;
/* release any references held by the object here */
if (G_OBJECT_CLASS (gst_rtp_vp8_depay_parent_class)->dispose)
G_OBJECT_CLASS (gst_rtp_vp8_depay_parent_class)->dispose (object);
}
static void
gst_rtp_vp8_depay_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec)
{
GstRtpVP8Depay *self = GST_RTP_VP8_DEPAY (object);
switch (prop_id) {
case PROP_WAIT_FOR_KEYFRAME:
self->wait_for_keyframe = g_value_get_boolean (value);
break;
case PROP_REQUEST_KEYFRAME:
self->request_keyframe = g_value_get_boolean (value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static void
gst_rtp_vp8_depay_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec)
{
GstRtpVP8Depay *self = GST_RTP_VP8_DEPAY (object);
switch (prop_id) {
case PROP_WAIT_FOR_KEYFRAME:
g_value_set_boolean (value, self->wait_for_keyframe);
break;
case PROP_REQUEST_KEYFRAME:
g_value_set_boolean (value, self->request_keyframe);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static gint
picture_id_compare (guint16 id0, guint16 id1)
{
guint shift = 16 - (IS_PICTURE_ID_15BITS (id1) ? 15 : 7);
id0 = id0 << shift;
id1 = id1 << shift;
return ((gint16) (id1 - id0)) >> shift;
}
static void
send_last_lost_event (GstRtpVP8Depay * self)
{
if (self->last_lost_event) {
GST_ERROR_OBJECT (self,
"Sending the last stopped lost event: %" GST_PTR_FORMAT,
self->last_lost_event);
GST_RTP_BASE_DEPAYLOAD_CLASS (gst_rtp_vp8_depay_parent_class)
->packet_lost (GST_RTP_BASE_DEPAYLOAD_CAST (self),
self->last_lost_event);
gst_event_replace (&self->last_lost_event, NULL);
self->last_pushed_was_lost_event = TRUE;
}
}
static void
send_new_lost_event (GstRtpVP8Depay * self, GstClockTime timestamp,
guint new_picture_id, const gchar * reason)
{
GstEvent *event;
if (!GST_CLOCK_TIME_IS_VALID (timestamp)) {
GST_WARNING_OBJECT (self, "Can't create lost event with invalid timestmap");
return;
}
event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM,
gst_structure_new ("GstRTPPacketLost",
"timestamp", G_TYPE_UINT64, timestamp,
"duration", G_TYPE_UINT64, G_GUINT64_CONSTANT (0), NULL));
GST_DEBUG_OBJECT (self, "Pushing lost event "
"(picids 0x%x 0x%x, reason \"%s\"): %" GST_PTR_FORMAT,
self->last_picture_id, new_picture_id, reason, event);
GST_RTP_BASE_DEPAYLOAD_CLASS (gst_rtp_vp8_depay_parent_class)
->packet_lost (GST_RTP_BASE_DEPAYLOAD_CAST (self), event);
gst_event_unref (event);
}
static void
send_last_lost_event_if_needed (GstRtpVP8Depay * self, guint new_picture_id)
{
if (self->last_picture_id == PICTURE_ID_NONE)
return;
if (self->last_lost_event) {
gboolean send_lost_event = FALSE;
if (new_picture_id == PICTURE_ID_NONE) {
GST_DEBUG_OBJECT (self, "Dropping the last stopped lost event "
"(picture id does not exist): %" GST_PTR_FORMAT,
self->last_lost_event);
} else if (IS_PICTURE_ID_15BITS (self->last_picture_id) &&
!IS_PICTURE_ID_15BITS (new_picture_id)) {
GST_DEBUG_OBJECT (self, "Dropping the last stopped lost event "
"(picture id has less bits than before): %" GST_PTR_FORMAT,
self->last_lost_event);
} else if (picture_id_compare (self->last_picture_id, new_picture_id) != 1) {
GstStructure *s = gst_event_writable_structure (self->last_lost_event);
GST_DEBUG_OBJECT (self, "Sending the last stopped lost event "
"(gap in picture id %u %u): %" GST_PTR_FORMAT,
self->last_picture_id, new_picture_id, self->last_lost_event);
send_lost_event = TRUE;
/* Prevent rtpbasedepayload from dropping the event now
* that we have made sure the lost packet was not FEC */
gst_structure_remove_field (s, "might-have-been-fec");
}
if (send_lost_event)
GST_RTP_BASE_DEPAYLOAD_CLASS (gst_rtp_vp8_depay_parent_class)
->packet_lost (GST_RTP_BASE_DEPAYLOAD_CAST (self),
self->last_lost_event);
gst_event_replace (&self->last_lost_event, NULL);
}
}
static GstBuffer *
gst_rtp_vp8_depay_process (GstRTPBaseDepayload * depay, GstRTPBuffer * rtp)
{
GstRtpVP8Depay *self = GST_RTP_VP8_DEPAY (depay);
GstBuffer *payload;
guint8 *data;
guint hdrsize = 1;
guint picture_id = PICTURE_ID_NONE;
guint size = gst_rtp_buffer_get_payload_len (rtp);
guint s_bit;
guint part_id;
gboolean frame_start;
gboolean sent_lost_event = FALSE;
if (G_UNLIKELY (GST_BUFFER_IS_DISCONT (rtp->buffer))) {
GST_DEBUG_OBJECT (self, "Discontinuity, flushing adapter");
gst_adapter_clear (self->adapter);
self->started = FALSE;
if (self->wait_for_keyframe)
self->waiting_for_keyframe = TRUE;
if (self->request_keyframe)
gst_pad_push_event (GST_RTP_BASE_DEPAYLOAD_SINKPAD (depay),
gst_video_event_new_upstream_force_key_unit (GST_CLOCK_TIME_NONE,
TRUE, 0));
}
/* At least one header and one vp8 byte */
if (G_UNLIKELY (size < 2))
goto too_small;
data = gst_rtp_buffer_get_payload (rtp);
s_bit = (data[0] >> 4) & 0x1;
part_id = (data[0] >> 0) & 0x7;
/* Check X optional header */
if ((data[0] & 0x80) != 0) {
hdrsize++;
/* Check I optional header */
if ((data[1] & 0x80) != 0) {
if (G_UNLIKELY (size < 3))
goto too_small;
hdrsize++;
/* Check for 16 bits PictureID */
picture_id = data[2];
if ((data[2] & 0x80) != 0) {
if (G_UNLIKELY (size < 4))
goto too_small;
hdrsize++;
picture_id = (picture_id << 8) | data[3];
}
}
/* Check L optional header */
if ((data[1] & 0x40) != 0)
hdrsize++;
/* Check T or K optional headers */
if ((data[1] & 0x20) != 0 || (data[1] & 0x10) != 0)
hdrsize++;
}
GST_LOG_OBJECT (depay,
"hdrsize %u, size %u, picture id 0x%x, s %u, part_id %u", hdrsize, size,
picture_id, s_bit, part_id);
if (G_UNLIKELY (hdrsize >= size))
goto too_small;
frame_start = (s_bit == 1) && (part_id == 0);
if (frame_start) {
if (G_UNLIKELY (self->started)) {
GST_DEBUG_OBJECT (depay, "Incomplete frame, flushing adapter");
gst_adapter_clear (self->adapter);
self->started = FALSE;
if (self->wait_for_keyframe)
self->waiting_for_keyframe = TRUE;
if (self->request_keyframe)
gst_pad_push_event (GST_RTP_BASE_DEPAYLOAD_SINKPAD (depay),
gst_video_event_new_upstream_force_key_unit (GST_CLOCK_TIME_NONE,
TRUE, 0));
send_new_lost_event (self, GST_BUFFER_PTS (rtp->buffer), picture_id,
"Incomplete frame detected");
sent_lost_event = TRUE;
}
}
if (!self->started) {
if (G_UNLIKELY (!frame_start)) {
GST_DEBUG_OBJECT (depay,
"The frame is missing the first packet, ignoring the packet");
if (self->stop_lost_events && !sent_lost_event) {
send_last_lost_event (self);
self->stop_lost_events = FALSE;
}
if (self->wait_for_keyframe)
self->waiting_for_keyframe = TRUE;
if (self->request_keyframe)
gst_pad_push_event (GST_RTP_BASE_DEPAYLOAD_SINKPAD (depay),
gst_video_event_new_upstream_force_key_unit (GST_CLOCK_TIME_NONE,
TRUE, 0));
goto done;
}
GST_LOG_OBJECT (depay, "Found the start of the frame");
if (self->stop_lost_events && !sent_lost_event) {
send_last_lost_event_if_needed (self, picture_id);
self->stop_lost_events = FALSE;
}
self->started = TRUE;
}
payload = gst_rtp_buffer_get_payload_subbuffer (rtp, hdrsize, -1);
gst_adapter_push (self->adapter, payload);
self->last_picture_id = picture_id;
/* Marker indicates that it was the last rtp packet for this frame */
if (gst_rtp_buffer_get_marker (rtp)) {
GstBuffer *out;
guint8 header[10];
GST_LOG_OBJECT (depay,
"Found the end of the frame (%" G_GSIZE_FORMAT " bytes)",
gst_adapter_available (self->adapter));
if (gst_adapter_available (self->adapter) < 10)
goto too_small;
gst_adapter_copy (self->adapter, &header, 0, 10);
out = gst_adapter_take_buffer (self->adapter,
gst_adapter_available (self->adapter));
self->started = FALSE;
/* mark keyframes */
out = gst_buffer_make_writable (out);
/* Filter away all metas that are not sensible to copy */
gst_rtp_drop_non_video_meta (self, out);
if ((header[0] & 0x01)) {
GST_BUFFER_FLAG_SET (out, GST_BUFFER_FLAG_DELTA_UNIT);
if (self->waiting_for_keyframe) {
gst_buffer_unref (out);
out = NULL;
GST_INFO_OBJECT (self, "Dropping inter-frame before intra-frame");
gst_pad_push_event (GST_RTP_BASE_DEPAYLOAD_SINKPAD (depay),
gst_video_event_new_upstream_force_key_unit (GST_CLOCK_TIME_NONE,
TRUE, 0));
}
} else {
guint profile, width, height;
GST_BUFFER_FLAG_UNSET (out, GST_BUFFER_FLAG_DELTA_UNIT);
GST_DEBUG_OBJECT (self, "Processed keyframe");
profile = (header[0] & 0x0e) >> 1;
width = GST_READ_UINT16_LE (header + 6) & 0x3fff;
height = GST_READ_UINT16_LE (header + 8) & 0x3fff;
if (G_UNLIKELY (self->last_width != width ||
self->last_height != height || self->last_profile != profile)) {
gchar profile_str[3];
GstCaps *srccaps;
snprintf (profile_str, 3, "%u", profile);
srccaps = gst_caps_new_simple ("video/x-vp8",
"framerate", GST_TYPE_FRACTION, 0, 1,
"height", G_TYPE_INT, height,
"width", G_TYPE_INT, width,
"profile", G_TYPE_STRING, profile_str, NULL);
gst_pad_set_caps (GST_RTP_BASE_DEPAYLOAD_SRCPAD (depay), srccaps);
gst_caps_unref (srccaps);
self->last_width = width;
self->last_height = height;
self->last_profile = profile;
}
self->waiting_for_keyframe = FALSE;
}
if (picture_id != PICTURE_ID_NONE)
self->stop_lost_events = TRUE;
self->last_pushed_was_lost_event = FALSE;
return out;
}
done:
return NULL;
too_small:
GST_DEBUG_OBJECT (self, "Invalid rtp packet (too small), ignoring");
gst_adapter_clear (self->adapter);
self->started = FALSE;
goto done;
}
static GstStateChangeReturn
gst_rtp_vp8_depay_change_state (GstElement * element, GstStateChange transition)
{
GstRtpVP8Depay *self = GST_RTP_VP8_DEPAY (element);
switch (transition) {
case GST_STATE_CHANGE_READY_TO_PAUSED:
self->last_profile = -1;
self->last_height = -1;
self->last_width = -1;
self->waiting_for_keyframe = TRUE;
self->caps_sent = FALSE;
self->last_picture_id = PICTURE_ID_NONE;
gst_event_replace (&self->last_lost_event, NULL);
self->stop_lost_events = FALSE;
break;
default:
break;
}
return
GST_ELEMENT_CLASS (gst_rtp_vp8_depay_parent_class)->change_state (element,
transition);
}
static gboolean
gst_rtp_vp8_depay_handle_event (GstRTPBaseDepayload * depay, GstEvent * event)
{
GstRtpVP8Depay *self = GST_RTP_VP8_DEPAY (depay);
switch (GST_EVENT_TYPE (event)) {
case GST_EVENT_FLUSH_STOP:
self->last_profile = -1;
self->last_height = -1;
self->last_width = -1;
self->last_picture_id = PICTURE_ID_NONE;
gst_event_replace (&self->last_lost_event, NULL);
self->stop_lost_events = FALSE;
break;
default:
break;
}
return
GST_RTP_BASE_DEPAYLOAD_CLASS
(gst_rtp_vp8_depay_parent_class)->handle_event (depay, event);
}
static gboolean
gst_rtp_vp8_depay_packet_lost (GstRTPBaseDepayload * depay, GstEvent * event)
{
GstRtpVP8Depay *self = GST_RTP_VP8_DEPAY (depay);
const GstStructure *s;
gboolean might_have_been_fec;
gboolean unref_event = FALSE;
gboolean ret;
s = gst_event_get_structure (event);
if (self->stop_lost_events) {
if (gst_structure_get_boolean (s, "might-have-been-fec",
&might_have_been_fec)
&& might_have_been_fec) {
GST_DEBUG_OBJECT (depay, "Stopping lost event %" GST_PTR_FORMAT, event);
gst_event_replace (&self->last_lost_event, event);
return TRUE;
}
} else if (self->last_picture_id != PICTURE_ID_NONE) {
GstStructure *s;
if (!gst_event_is_writable (event)) {
event = gst_event_copy (event);
unref_event = TRUE;
}
s = gst_event_writable_structure (event);
/* We are currently processing a picture, let's make sure the
* base depayloader doesn't drop this lost event */
gst_structure_remove_field (s, "might-have-been-fec");
}
self->last_pushed_was_lost_event = TRUE;
ret =
GST_RTP_BASE_DEPAYLOAD_CLASS
(gst_rtp_vp8_depay_parent_class)->packet_lost (depay, event);
if (unref_event)
gst_event_unref (event);
return ret;
}