gstreamer/subprojects/gst-plugins-good/gst/rtp/gstrtph265pay.c
Sebastian Dröge b0afaffc5d rtp: In payloaders map the RTP marker flag to the corresponding buffer flag
This allows downstream of a payloader to know the RTP header's marker
flag without first having to map the buffer and parse the RTP header.

Especially inside RTP header extension implementations this can be
useful to decide which packet corresponds to e.g. the last packet of a
video frame.

Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/1776>
2022-02-28 10:13:11 +00:00

1830 lines
56 KiB
C

/* GStreamer
* Copyright (C) <2006> Wim Taymans <wim.taymans@gmail.com>
* Copyright (C) <2014> Jurgen Slowack <jurgenslowack@gmail.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
#ifdef HAVE_CONFIG_H
# include "config.h"
#endif
#include <string.h>
#include <stdlib.h>
#include <gst/rtp/gstrtpbuffer.h>
#include <gst/pbutils/pbutils.h>
#include <gst/video/video.h>
/* Included to not duplicate gst_rtp_h265_add_vps_sps_pps () */
#include "gstrtph265depay.h"
#include "gstrtpelements.h"
#include "gstrtph265pay.h"
#include "gstrtputils.h"
#include "gstbuffermemory.h"
#define AP_TYPE_ID 48
#define FU_TYPE_ID 49
GST_DEBUG_CATEGORY_STATIC (rtph265pay_debug);
#define GST_CAT_DEFAULT (rtph265pay_debug)
#define GST_TYPE_RTP_H265_AGGREGATE_MODE \
(gst_rtp_h265_aggregate_mode_get_type ())
static GType
gst_rtp_h265_aggregate_mode_get_type (void)
{
static GType type = 0;
static const GEnumValue values[] = {
{GST_RTP_H265_AGGREGATE_NONE, "Do not aggregate NAL units", "none"},
{GST_RTP_H265_AGGREGATE_ZERO_LATENCY,
"Aggregate NAL units until a VCL or suffix unit is included",
"zero-latency"},
{GST_RTP_H265_AGGREGATE_MAX,
"Aggregate all NAL units with the same timestamp (adds one frame of"
" latency)", "max"},
{0, NULL, NULL},
};
if (!type) {
type = g_enum_register_static ("GstRtpH265AggregateMode", values);
}
return type;
}
/* references:
*
* Internet Draft RTP Payload Format for High Efficiency Video Coding
*
* draft-ietf-payload-rtp-h265-03.txt
*
* This draft will be replaced with an RFC, so some details may change.
*
*/
static GstStaticPadTemplate gst_rtp_h265_pay_sink_template =
GST_STATIC_PAD_TEMPLATE ("sink",
GST_PAD_SINK,
GST_PAD_ALWAYS,
GST_STATIC_CAPS (
/* only hvc1 and byte-stream formats supported for now */
"video/x-h265, stream-format = (string) hvc1, alignment = (string) au; "
/* "video/x-h265, "
"stream-format = (string) hev1, alignment = (string) au; " */
"video/x-h265, stream-format = (string) byte-stream, "
"alignment = (string) { nal, au }")
);
static GstStaticPadTemplate gst_rtp_h265_pay_src_template =
GST_STATIC_PAD_TEMPLATE ("src",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS ("application/x-rtp, "
"media = (string) \"video\", "
"payload = (int) " GST_RTP_PAYLOAD_DYNAMIC_STRING ", "
"clock-rate = (int) 90000, " "encoding-name = (string) \"H265\"")
/* optional parameters */
/* "profile-space = (int) [ 0, 3 ], " */
/* "profile-id = (int) [ 0, 31 ], " */
/* "tier-flag = (int) [ 0, 1 ], " */
/* "level-id = (int) [ 0, 255 ], " */
/* "interop-constraints = (string) ANY, " */
/* "profile-compatibility-indicator = (string) ANY, " */
/* "sprop-sub-layer-id = (int) [ 0, 6 ], " */
/* "recv-sub-layer-id = (int) [ 0, 6 ], " */
/* "max-recv-level-id = (int) [ 0, 255 ], " */
/* "tx-mode = (string) {MST , SST}, " */
/* "sprop-vps = (string) ANY, " */
/* "sprop-sps = (string) ANY, " */
/* "sprop-pps = (string) ANY, " */
/* "sprop-sei = (string) ANY, " */
/* "max-lsr = (int) ANY, " *//* MUST be in the range of MaxLumaSR to 16 * MaxLumaSR, inclusive */
/* "max-lps = (int) ANY, " *//* MUST be in the range of MaxLumaPS to 16 * MaxLumaPS, inclusive */
/* "max-cpb = (int) ANY, " *//* MUST be in the range of MaxCPB to 16 * MaxCPB, inclusive */
/* "max-dpb = (int) [1, 16], " */
/* "max-br = (int) ANY, " *//* MUST be in the range of MaxBR to 16 * MaxBR, inclusive, for the highest level */
/* "max-tr = (int) ANY, " *//* MUST be in the range of MaxTileRows to 16 * MaxTileRows, inclusive, for the highest level */
/* "max-tc = (int) ANY, " *//* MUST be in the range of MaxTileCols to 16 * MaxTileCols, inclusive, for the highest level */
/* "max-fps = (int) ANY, " */
/* "sprop-max-don-diff = (int) [0, 32767], " */
/* "sprop-depack-buf-nalus = (int) [0, 32767], " */
/* "sprop-depack-buf-nalus = (int) [0, 4294967295], " */
/* "depack-buf-cap = (int) [1, 4294967295], " */
/* "sprop-segmentation-id = (int) [0, 3], " */
/* "sprop-spatial-segmentation-idc = (string) ANY, " */
/* "dec-parallel-cap = (string) ANY, " */
);
#define DEFAULT_CONFIG_INTERVAL 0
#define DEFAULT_AGGREGATE_MODE GST_RTP_H265_AGGREGATE_NONE
enum
{
PROP_0,
PROP_CONFIG_INTERVAL,
PROP_AGGREGATE_MODE,
};
static void gst_rtp_h265_pay_finalize (GObject * object);
static void gst_rtp_h265_pay_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec);
static void gst_rtp_h265_pay_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec);
static GstCaps *gst_rtp_h265_pay_getcaps (GstRTPBasePayload * payload,
GstPad * pad, GstCaps * filter);
static gboolean gst_rtp_h265_pay_setcaps (GstRTPBasePayload * basepayload,
GstCaps * caps);
static GstFlowReturn gst_rtp_h265_pay_handle_buffer (GstRTPBasePayload * pad,
GstBuffer * buffer);
static gboolean gst_rtp_h265_pay_sink_event (GstRTPBasePayload * payload,
GstEvent * event);
static GstStateChangeReturn gst_rtp_h265_pay_change_state (GstElement *
element, GstStateChange transition);
static gboolean gst_rtp_h265_pay_src_query (GstPad * pad, GstObject * parent,
GstQuery * query);
static void gst_rtp_h265_pay_reset_bundle (GstRtpH265Pay * rtph265pay);
#define gst_rtp_h265_pay_parent_class parent_class
G_DEFINE_TYPE (GstRtpH265Pay, gst_rtp_h265_pay, GST_TYPE_RTP_BASE_PAYLOAD);
GST_ELEMENT_REGISTER_DEFINE_WITH_CODE (rtph265pay, "rtph265pay",
GST_RANK_SECONDARY, GST_TYPE_RTP_H265_PAY, rtp_element_init (plugin));
static void
gst_rtp_h265_pay_class_init (GstRtpH265PayClass * klass)
{
GObjectClass *gobject_class;
GstElementClass *gstelement_class;
GstRTPBasePayloadClass *gstrtpbasepayload_class;
gobject_class = (GObjectClass *) klass;
gstelement_class = (GstElementClass *) klass;
gstrtpbasepayload_class = (GstRTPBasePayloadClass *) klass;
gobject_class->set_property = gst_rtp_h265_pay_set_property;
gobject_class->get_property = gst_rtp_h265_pay_get_property;
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_CONFIG_INTERVAL,
g_param_spec_int ("config-interval",
"VPS SPS PPS Send Interval",
"Send VPS, SPS and PPS Insertion Interval in seconds (sprop parameter sets "
"will be multiplexed in the data stream when detected.) "
"(0 = disabled, -1 = send with every IDR frame)",
-1, 3600, DEFAULT_CONFIG_INTERVAL,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)
);
/**
* GstRtpH265Pay:aggregate-mode
*
* Bundle suitable SPS/PPS NAL units into STAP-A aggregate packets.
*
* This can potentially reduce RTP packetization overhead but not all
* RTP implementations handle it correctly.
*
* For best compatibility, it is recommended to set this to "none" (the
* default) for RTSP and for WebRTC to "zero-latency".
*
* Since: 1.18
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_AGGREGATE_MODE,
g_param_spec_enum ("aggregate-mode",
"Attempt to use aggregate packets",
"Bundle suitable SPS/PPS NAL units into aggregate packets.",
GST_TYPE_RTP_H265_AGGREGATE_MODE,
DEFAULT_AGGREGATE_MODE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)
);
gobject_class->finalize = gst_rtp_h265_pay_finalize;
gst_element_class_add_static_pad_template (gstelement_class,
&gst_rtp_h265_pay_src_template);
gst_element_class_add_static_pad_template (gstelement_class,
&gst_rtp_h265_pay_sink_template);
gst_element_class_set_static_metadata (gstelement_class, "RTP H265 payloader",
"Codec/Payloader/Network/RTP",
"Payload-encode H265 video into RTP packets (RFC 7798)",
"Jurgen Slowack <jurgenslowack@gmail.com>");
gstelement_class->change_state =
GST_DEBUG_FUNCPTR (gst_rtp_h265_pay_change_state);
gstrtpbasepayload_class->get_caps = gst_rtp_h265_pay_getcaps;
gstrtpbasepayload_class->set_caps = gst_rtp_h265_pay_setcaps;
gstrtpbasepayload_class->handle_buffer = gst_rtp_h265_pay_handle_buffer;
gstrtpbasepayload_class->sink_event = gst_rtp_h265_pay_sink_event;
GST_DEBUG_CATEGORY_INIT (rtph265pay_debug, "rtph265pay", 0,
"H265 RTP Payloader");
gst_type_mark_as_plugin_api (GST_TYPE_RTP_H265_AGGREGATE_MODE, 0);
}
static void
gst_rtp_h265_pay_init (GstRtpH265Pay * rtph265pay)
{
rtph265pay->queue = g_array_new (FALSE, FALSE, sizeof (guint));
rtph265pay->sps = g_ptr_array_new_with_free_func (
(GDestroyNotify) gst_buffer_unref);
rtph265pay->pps = g_ptr_array_new_with_free_func (
(GDestroyNotify) gst_buffer_unref);
rtph265pay->vps = g_ptr_array_new_with_free_func (
(GDestroyNotify) gst_buffer_unref);
rtph265pay->last_vps_sps_pps = -1;
rtph265pay->vps_sps_pps_interval = DEFAULT_CONFIG_INTERVAL;
rtph265pay->aggregate_mode = DEFAULT_AGGREGATE_MODE;
rtph265pay->adapter = gst_adapter_new ();
gst_pad_set_query_function (GST_RTP_BASE_PAYLOAD_SRCPAD (rtph265pay),
gst_rtp_h265_pay_src_query);
}
static void
gst_rtp_h265_pay_clear_vps_sps_pps (GstRtpH265Pay * rtph265pay)
{
g_ptr_array_set_size (rtph265pay->vps, 0);
g_ptr_array_set_size (rtph265pay->sps, 0);
g_ptr_array_set_size (rtph265pay->pps, 0);
}
static void
gst_rtp_h265_pay_finalize (GObject * object)
{
GstRtpH265Pay *rtph265pay;
rtph265pay = GST_RTP_H265_PAY (object);
g_array_free (rtph265pay->queue, TRUE);
g_ptr_array_free (rtph265pay->sps, TRUE);
g_ptr_array_free (rtph265pay->pps, TRUE);
g_ptr_array_free (rtph265pay->vps, TRUE);
g_object_unref (rtph265pay->adapter);
gst_rtp_h265_pay_reset_bundle (rtph265pay);
G_OBJECT_CLASS (parent_class)->finalize (object);
}
static gboolean
gst_rtp_h265_pay_src_query (GstPad * pad, GstObject * parent, GstQuery * query)
{
GstRtpH265Pay *rtph265pay = GST_RTP_H265_PAY (parent);
if (GST_QUERY_TYPE (query) == GST_QUERY_LATENCY) {
gboolean retval;
gboolean live;
GstClockTime min_latency, max_latency;
retval = gst_pad_query_default (pad, parent, query);
if (!retval)
return retval;
if (rtph265pay->stream_format == GST_H265_STREAM_FORMAT_UNKNOWN ||
rtph265pay->alignment == GST_H265_ALIGNMENT_UNKNOWN)
return FALSE;
gst_query_parse_latency (query, &live, &min_latency, &max_latency);
if (rtph265pay->aggregate_mode == GST_RTP_H265_AGGREGATE_MAX &&
rtph265pay->alignment != GST_H265_ALIGNMENT_AU && rtph265pay->fps_num) {
GstClockTime one_frame = gst_util_uint64_scale_int (GST_SECOND,
rtph265pay->fps_denum, rtph265pay->fps_num);
min_latency += one_frame;
max_latency += one_frame;
gst_query_set_latency (query, live, min_latency, max_latency);
}
return TRUE;
}
return gst_pad_query_default (pad, parent, query);
}
static const gchar all_levels[][4] = {
"1",
"2",
"2.1",
"3",
"3.1",
"4",
"4.1",
"5",
"5.1",
"5.2",
"6",
"6.1",
"6.2"
};
static gboolean
parse_field (GstStructure * s, const gchar * field, gulong min, gulong max,
guint8 * result)
{
const gchar *str;
g_assert (result != NULL);
str = gst_structure_get_string (s, field);
if (str != NULL && *str != '\0') {
gulong value;
gchar *end;
value = strtoul (str, &end, 10);
if (*end == '\0' && value >= min && value <= max) {
*result = (guint8) value;
} else {
return FALSE;
}
} else {
return FALSE;
}
return TRUE;
}
static GstCaps *
gst_rtp_h265_pay_getcaps (GstRTPBasePayload * payload, GstPad * pad,
GstCaps * filter)
{
GstCaps *template_caps;
GstCaps *allowed_caps;
GstCaps *caps;
GstCaps *icaps;
guint i;
allowed_caps =
gst_pad_peer_query_caps (GST_RTP_BASE_PAYLOAD_SRCPAD (payload), NULL);
if (allowed_caps == NULL)
return NULL;
template_caps =
gst_static_pad_template_get_caps (&gst_rtp_h265_pay_sink_template);
if (gst_caps_is_any (allowed_caps)) {
caps = gst_caps_ref (template_caps);
goto done;
}
if (gst_caps_is_empty (allowed_caps)) {
caps = gst_caps_ref (allowed_caps);
goto done;
}
caps = gst_caps_new_empty ();
for (i = 0; i < gst_caps_get_size (allowed_caps); i++) {
GstStructure *s = gst_caps_get_structure (allowed_caps, i);
GstStructure *new_s = gst_structure_new_empty ("video/x-h265");
guint8 ptl[12] = { 0, };
guint8 value;
if (parse_field (s, "profile-id", 0, 31, &value)) {
const gchar *profile;
ptl[0] = value;
profile = gst_codec_utils_h265_get_profile (ptl, sizeof (ptl));
if (profile != NULL) {
GST_DEBUG_OBJECT (payload, "profile %s", profile);
gst_structure_set (new_s, "profile", G_TYPE_STRING, profile, NULL);
} else {
GST_WARNING_OBJECT (payload, "invalid profile-id %d in caps", value);
}
} else {
GST_DEBUG_OBJECT (payload, "no valid profile-id in caps");
}
if (parse_field (s, "tier-flag", 0, 1, &value)) {
const gchar *tier;
ptl[0] |= value << 5;
tier = gst_codec_utils_h265_get_tier (ptl, sizeof (ptl));
GST_DEBUG_OBJECT (payload, "tier %s", tier);
gst_structure_set (new_s, "tier", G_TYPE_STRING, tier, NULL);
} else {
GST_DEBUG_OBJECT (payload, "no valid tier-flag in caps");
}
if (parse_field (s, "level-id", 0, 255, &value)) {
const gchar *level;
ptl[11] = value;
level = gst_codec_utils_h265_get_level (ptl, sizeof (ptl));
if (level != NULL) {
GST_DEBUG_OBJECT (payload, "level %s", level);
if (strcmp (level, "1") == 0) {
gst_structure_set (new_s, "level", G_TYPE_STRING, level, NULL);
} else {
GValue levels = { 0, };
GValue val = { 0, };
int j;
g_value_init (&levels, GST_TYPE_LIST);
g_value_init (&val, G_TYPE_STRING);
for (j = 0; j < G_N_ELEMENTS (all_levels); j++) {
g_value_set_static_string (&val, all_levels[j]);
gst_value_list_prepend_value (&levels, &val);
if (!strcmp (level, all_levels[j]))
break;
}
gst_structure_take_value (new_s, "level", &levels);
}
} else {
GST_WARNING_OBJECT (payload, "invalid level-id %d in caps", value);
}
} else {
GST_DEBUG_OBJECT (payload, "no valid level-id in caps");
}
caps = gst_caps_merge_structure (caps, new_s);
}
icaps = gst_caps_intersect (caps, template_caps);
gst_caps_unref (caps);
caps = icaps;
done:
if (filter) {
GstCaps *tmp;
GST_DEBUG_OBJECT (payload, "Intersect %" GST_PTR_FORMAT " and filter %"
GST_PTR_FORMAT, caps, filter);
tmp = gst_caps_intersect_full (filter, caps, GST_CAPS_INTERSECT_FIRST);
gst_caps_unref (caps);
caps = tmp;
}
gst_caps_unref (template_caps);
gst_caps_unref (allowed_caps);
GST_LOG_OBJECT (payload, "returning caps %" GST_PTR_FORMAT, caps);
return caps;
}
/* take the currently configured VPS, SPS and PPS lists and set them on the
* caps */
static gboolean
gst_rtp_h265_pay_set_vps_sps_pps (GstRTPBasePayload * basepayload)
{
GstRtpH265Pay *payloader = GST_RTP_H265_PAY (basepayload);
gchar *set;
GString *vps;
GString *sps;
GString *pps;
guint count;
gboolean res;
GstMapInfo map;
guint i;
vps = g_string_new ("");
sps = g_string_new ("");
pps = g_string_new ("");
count = 0;
for (i = 0; i < payloader->vps->len; i++) {
GstBuffer *vps_buf =
GST_BUFFER_CAST (g_ptr_array_index (payloader->vps, i));
gst_buffer_map (vps_buf, &map, GST_MAP_READ);
set = g_base64_encode (map.data, map.size);
gst_buffer_unmap (vps_buf, &map);
g_string_append_printf (vps, "%s%s", i ? "," : "", set);
g_free (set);
count++;
}
for (i = 0; i < payloader->sps->len; i++) {
GstBuffer *sps_buf =
GST_BUFFER_CAST (g_ptr_array_index (payloader->sps, i));
gst_buffer_map (sps_buf, &map, GST_MAP_READ);
set = g_base64_encode (map.data, map.size);
gst_buffer_unmap (sps_buf, &map);
g_string_append_printf (sps, "%s%s", i ? "," : "", set);
g_free (set);
count++;
}
for (i = 0; i < payloader->pps->len; i++) {
GstBuffer *pps_buf =
GST_BUFFER_CAST (g_ptr_array_index (payloader->pps, i));
gst_buffer_map (pps_buf, &map, GST_MAP_READ);
set = g_base64_encode (map.data, map.size);
gst_buffer_unmap (pps_buf, &map);
g_string_append_printf (pps, "%s%s", i ? "," : "", set);
g_free (set);
count++;
}
if (G_LIKELY (count)) {
/* combine into output caps */
res = gst_rtp_base_payload_set_outcaps (basepayload,
"sprop-vps", G_TYPE_STRING, vps->str,
"sprop-sps", G_TYPE_STRING, sps->str,
"sprop-pps", G_TYPE_STRING, pps->str, NULL);
} else {
res = gst_rtp_base_payload_set_outcaps (basepayload, NULL);
}
g_string_free (vps, TRUE);
g_string_free (sps, TRUE);
g_string_free (pps, TRUE);
return res;
}
static gboolean
gst_rtp_h265_pay_setcaps (GstRTPBasePayload * basepayload, GstCaps * caps)
{
GstRtpH265Pay *rtph265pay;
GstStructure *str;
const GValue *value;
GstMapInfo map;
guint8 *data;
gsize size;
GstBuffer *buffer;
const gchar *alignment, *stream_format;
guint8 num_arrays;
rtph265pay = GST_RTP_H265_PAY (basepayload);
str = gst_caps_get_structure (caps, 0);
/* we can only set the output caps when we found the sprops and profile
* NALs */
gst_rtp_base_payload_set_options (basepayload, "video", TRUE, "H265", 90000);
rtph265pay->alignment = GST_H265_ALIGNMENT_UNKNOWN;
alignment = gst_structure_get_string (str, "alignment");
if (alignment) {
if (g_str_equal (alignment, "au"))
rtph265pay->alignment = GST_H265_ALIGNMENT_AU;
if (g_str_equal (alignment, "nal"))
rtph265pay->alignment = GST_H265_ALIGNMENT_NAL;
}
rtph265pay->stream_format = GST_H265_STREAM_FORMAT_UNKNOWN;
stream_format = gst_structure_get_string (str, "stream-format");
if (stream_format) {
if (g_str_equal (stream_format, "hvc1"))
rtph265pay->stream_format = GST_H265_STREAM_FORMAT_HVC1;
if (g_str_equal (stream_format, "hev1"))
rtph265pay->stream_format = GST_H265_STREAM_FORMAT_HEV1;
if (g_str_equal (stream_format, "byte-stream"))
rtph265pay->stream_format = GST_H265_STREAM_FORMAT_BYTESTREAM;
}
if (!gst_structure_get_fraction (str, "framerate", &rtph265pay->fps_num,
&rtph265pay->fps_denum))
rtph265pay->fps_num = rtph265pay->fps_denum = 0;
/* packetized HEVC video has a codec_data */
if ((value = gst_structure_get_value (str, "codec_data"))) {
guint num_vps, num_sps, num_pps;
gint i, j, nal_size;
GST_DEBUG_OBJECT (rtph265pay, "have packetized h265");
buffer = gst_value_get_buffer (value);
gst_buffer_map (buffer, &map, GST_MAP_READ);
data = map.data;
size = map.size;
/* parse the hevcC data */
if (size < 23)
goto hevcc_too_small;
/* HEVCDecoderConfigurationVersion (must be 1) */
if (data[0] != 1)
goto wrong_version;
/* profile_space | tier_flag | profile_idc */
GST_DEBUG_OBJECT (rtph265pay, "profile %06x", data[1]);
/* profile_compatibility_flags */
for (i = 2; i < 6; i++) {
for (j = 7; j >= 0; j--) {
GST_DEBUG_OBJECT (rtph265pay, "profile_compatibility_flag %06x",
(data[i] >> j) & 1);
}
}
GST_DEBUG_OBJECT (rtph265pay, "progressive_source_flag %06x",
(data[6] >> 7) & 1);
GST_DEBUG_OBJECT (rtph265pay, "interlaced_source_flag %06x",
(data[6] >> 6) & 1);
GST_DEBUG_OBJECT (rtph265pay, "non_packed_constraint_flag %06x",
(data[6] >> 5) & 1);
GST_DEBUG_OBJECT (rtph265pay, "frame_only_constraint_flag %06x",
(data[6] >> 4) & 1);
GST_DEBUG_OBJECT (rtph265pay, "level_idc %06x", data[12]);
GST_DEBUG_OBJECT (rtph265pay, "min_spatial_segmentation_idc %06x",
((data[13] ^ 0xf0) << 8) + data[14]);
GST_DEBUG_OBJECT (rtph265pay, "parrallelismType %06x (ignored by paloader)",
data[15]);
GST_DEBUG_OBJECT (rtph265pay, "sps_chroma_format_idc %06x",
data[16] ^ 0xfc);
GST_DEBUG_OBJECT (rtph265pay, "bit_depth_luma_minus8 %06x",
data[17] ^ 0xf8);
GST_DEBUG_OBJECT (rtph265pay, "bit_depth_chroma_minus8 %06x",
data[18] ^ 0xf8);
GST_DEBUG_OBJECT (rtph265pay, "avgFrameRate %06x", data[19]);
GST_DEBUG_OBJECT (rtph265pay, "avgFrameRate %06x", data[20]);
/* constFrameRate(2 bits): 0, stream may or may not be of constant framerate
* numTemporalLayers (3 bits): number of temporal layers, value from SPS
* TemporalIdNested (1 bit): sps_temporal_id_nesting_flag from SPS
* lengthSizeMinusOne (2 bits): plus 1 indicates the length of the NALUnitLength */
GST_DEBUG_OBJECT (rtph265pay, "constFrameRate %06x",
(data[21] >> 6) & 0x03);
GST_DEBUG_OBJECT (rtph265pay, "numTemporalLayers %06x",
(data[21] >> 3) & 0x07);
GST_DEBUG_OBJECT (rtph265pay, "temporal_id_nesting_flag %06x",
(data[21] >> 2) & 0x01);
rtph265pay->nal_length_size = (data[21] & 0x3) + 1;
GST_DEBUG_OBJECT (rtph265pay, "nal length %u", rtph265pay->nal_length_size);
num_arrays = GST_READ_UINT8 (data + 22);
data += 23;
size -= 23;
if (num_arrays > 0) {
if ((data[0] & 0x3f) == 0x20) { /* VPS */
data++;
num_vps = data[0] << 8 | data[1];
data += 2;
size -= 2;
for (i = 0; i < num_vps; i++) {
GstBuffer *vps_buf;
if (size < 2)
goto hevcc_error;
nal_size = (data[0] << 8) | data[1];
data += 2;
size -= 2;
GST_LOG_OBJECT (rtph265pay, "VPS %d size %d", i, nal_size);
if (size < nal_size)
goto hevcc_error;
/* make a buffer out of it and add to VPS list */
vps_buf = gst_buffer_new_and_alloc (nal_size);
gst_buffer_fill (vps_buf, 0, data, nal_size);
gst_rtp_h265_add_vps_sps_pps (GST_ELEMENT (rtph265pay),
rtph265pay->vps, rtph265pay->sps, rtph265pay->pps, vps_buf);
data += nal_size;
size -= nal_size;
}
}
--num_arrays;
}
if (num_arrays > 0) {
if ((data[0] & 0x3f) == 0x21) { /* SPS */
data++;
num_sps = data[0] << 8 | data[1];
data += 2;
size -= 2;
for (i = 0; i < num_sps; i++) {
GstBuffer *sps_buf;
if (size < 2)
goto hevcc_error;
nal_size = (data[0] << 8) | data[1];
data += 2;
size -= 2;
GST_LOG_OBJECT (rtph265pay, "SPS %d size %d", i, nal_size);
if (size < nal_size)
goto hevcc_error;
/* make a buffer out of it and add to SPS list */
sps_buf = gst_buffer_new_and_alloc (nal_size);
gst_buffer_fill (sps_buf, 0, data, nal_size);
gst_rtp_h265_add_vps_sps_pps (GST_ELEMENT (rtph265pay),
rtph265pay->vps, rtph265pay->sps, rtph265pay->pps, sps_buf);
data += nal_size;
size -= nal_size;
}
}
--num_arrays;
}
if (num_arrays > 0) {
if ((data[0] & 0x3f) == 0x22) { /* PPS */
data++;
num_pps = data[0] << 8 | data[1];
data += 2;
size -= 2;
for (i = 0; i < num_pps; i++) {
GstBuffer *pps_buf;
if (size < 2)
goto hevcc_error;
nal_size = (data[0] << 8) | data[1];
data += 2;
size -= 2;
GST_LOG_OBJECT (rtph265pay, "PPS %d size %d", i, nal_size);
if (size < nal_size)
goto hevcc_error;
/* make a buffer out of it and add to PPS list */
pps_buf = gst_buffer_new_and_alloc (nal_size);
gst_buffer_fill (pps_buf, 0, data, nal_size);
gst_rtp_h265_add_vps_sps_pps (GST_ELEMENT (rtph265pay),
rtph265pay->vps, rtph265pay->sps, rtph265pay->pps, pps_buf);
data += nal_size;
size -= nal_size;
}
}
--num_arrays;
}
/* and update the caps with the collected data */
if (!gst_rtp_h265_pay_set_vps_sps_pps (basepayload))
goto set_vps_sps_pps_failed;
GST_DEBUG_OBJECT (rtph265pay, "Caps have been set");
gst_buffer_unmap (buffer, &map);
} else {
GST_DEBUG_OBJECT (rtph265pay, "have bytestream h265");
}
return TRUE;
hevcc_too_small:
{
GST_ERROR_OBJECT (rtph265pay, "hevcC size %" G_GSIZE_FORMAT " < 7", size);
goto error;
}
wrong_version:
{
GST_ERROR_OBJECT (rtph265pay, "wrong hevcC version");
goto error;
}
hevcc_error:
{
GST_ERROR_OBJECT (rtph265pay, "hevcC too small ");
goto error;
}
set_vps_sps_pps_failed:
{
GST_ERROR_OBJECT (rtph265pay, "failed to set vps/sps/pps");
goto error;
}
error:
{
gst_buffer_unmap (buffer, &map);
return FALSE;
}
}
static guint
next_start_code (const guint8 * data, guint size)
{
/* Boyer-Moore string matching algorithm, in a degenerative
* sense because our search 'alphabet' is binary - 0 & 1 only.
* This allow us to simplify the general BM algorithm to a very
* simple form. */
/* assume 1 is in the 3rd byte */
guint offset = 2;
while (offset < size) {
if (1 == data[offset]) {
unsigned int shift = offset;
if (0 == data[--shift]) {
if (0 == data[--shift]) {
return shift;
}
}
/* The jump is always 3 because of the 1 previously matched.
* All the 0's must be after this '1' matched at offset */
offset += 3;
} else if (0 == data[offset]) {
/* maybe next byte is 1? */
offset++;
} else {
/* can jump 3 bytes forward */
offset += 3;
}
/* at each iteration, we rescan in a backward manner until
* we match 0.0.1 in reverse order. Since our search string
* has only 2 'alpabets' (i.e. 0 & 1), we know that any
* mismatch will force us to shift a fixed number of steps */
}
GST_DEBUG ("Cannot find next NAL start code. returning %u", size);
return size;
}
static gboolean
gst_rtp_h265_pay_decode_nal (GstRtpH265Pay * payloader,
const guint8 * data, guint size, GstClockTime dts, GstClockTime pts)
{
guint8 type;
gboolean updated;
/* default is no update */
updated = FALSE;
GST_DEBUG_OBJECT (payloader, "NAL payload size %u", size);
type = (data[0] >> 1) & 0x3f;
/* We record the timestamp of the last SPS/PPS so
* that we can insert them at regular intervals and when needed. */
if (GST_H265_NAL_VPS == type || GST_H265_NAL_SPS == type
|| GST_H265_NAL_PPS == type) {
GstBuffer *nal;
/* trailing 0x0 are not part of the VPS/SPS/PPS */
while (size > 0 && data[size - 1] == 0x0)
size--;
/* encode the entire NAL in base64 */
GST_DEBUG_OBJECT (payloader, "found %s (type 0x%x), size %u",
type == GST_H265_NAL_VPS ? "VPS" : type == GST_H265_NAL_SPS ?
"SPS" : "PPS", type, size);
nal = gst_buffer_new_allocate (NULL, size, NULL);
gst_buffer_fill (nal, 0, data, size);
updated = gst_rtp_h265_add_vps_sps_pps (GST_ELEMENT (payloader),
payloader->vps, payloader->sps, payloader->pps, nal);
/* remember when we last saw VPS */
if (pts != -1)
payloader->last_vps_sps_pps =
gst_segment_to_running_time (&GST_RTP_BASE_PAYLOAD_CAST
(payloader)->segment, GST_FORMAT_TIME, pts);
} else {
GST_DEBUG_OBJECT (payloader, "NALU type 0x%x, size %u", type, size);
}
return updated;
}
static GstFlowReturn gst_rtp_h265_pay_payload_nal (GstRTPBasePayload *
basepayload, GPtrArray * paybufs, GstClockTime dts, GstClockTime pts);
static GstFlowReturn gst_rtp_h265_pay_payload_nal_single (GstRTPBasePayload *
basepayload, GstBuffer * paybuf, GstClockTime dts, GstClockTime pts,
gboolean marker);
static GstFlowReturn gst_rtp_h265_pay_payload_nal_fragment (GstRTPBasePayload *
basepayload, GstBuffer * paybuf, GstClockTime dts, GstClockTime pts,
gboolean marker, guint mtu, guint8 nal_type, const guint8 * nal_header,
int size);
static GstFlowReturn gst_rtp_h265_pay_payload_nal_bundle (GstRTPBasePayload *
basepayload, GstBuffer * paybuf, GstClockTime dts, GstClockTime pts,
gboolean marker, guint8 nal_type, const guint8 * nal_header, int size);
static GstFlowReturn
gst_rtp_h265_pay_send_vps_sps_pps (GstRTPBasePayload * basepayload,
GstRtpH265Pay * rtph265pay, GstClockTime dts, GstClockTime pts)
{
GstFlowReturn ret = GST_FLOW_OK;
gboolean sent_all_vps_sps_pps = TRUE;
guint i;
GPtrArray *bufs;
bufs = g_ptr_array_new ();
for (i = 0; i < rtph265pay->vps->len; i++) {
GstBuffer *vps_buf =
GST_BUFFER_CAST (g_ptr_array_index (rtph265pay->vps, i));
GST_DEBUG_OBJECT (rtph265pay, "inserting VPS in the stream");
g_ptr_array_add (bufs, gst_buffer_ref (vps_buf));
}
for (i = 0; i < rtph265pay->sps->len; i++) {
GstBuffer *sps_buf =
GST_BUFFER_CAST (g_ptr_array_index (rtph265pay->sps, i));
GST_DEBUG_OBJECT (rtph265pay, "inserting SPS in the stream");
g_ptr_array_add (bufs, gst_buffer_ref (sps_buf));
}
for (i = 0; i < rtph265pay->pps->len; i++) {
GstBuffer *pps_buf =
GST_BUFFER_CAST (g_ptr_array_index (rtph265pay->pps, i));
GST_DEBUG_OBJECT (rtph265pay, "inserting PPS in the stream");
g_ptr_array_add (bufs, gst_buffer_ref (pps_buf));
}
ret = gst_rtp_h265_pay_payload_nal (basepayload, bufs, dts, pts);
if (ret != GST_FLOW_OK) {
/* not critical but warn */
GST_WARNING_OBJECT (basepayload, "failed pushing VPS/SPS/PPS");
sent_all_vps_sps_pps = FALSE;
}
if (pts != -1 && sent_all_vps_sps_pps)
rtph265pay->last_vps_sps_pps =
gst_segment_to_running_time (&basepayload->segment, GST_FORMAT_TIME,
pts);
return ret;
}
static void
gst_rtp_h265_pay_reset_bundle (GstRtpH265Pay * rtph265pay)
{
g_clear_pointer (&rtph265pay->bundle, gst_buffer_list_unref);
rtph265pay->bundle_size = 0;
rtph265pay->bundle_contains_vcl_or_suffix = FALSE;
}
static GstFlowReturn
gst_rtp_h265_pay_payload_nal (GstRTPBasePayload * basepayload,
GPtrArray * paybufs, GstClockTime dts, GstClockTime pts)
{
GstRtpH265Pay *rtph265pay;
guint mtu;
GstFlowReturn ret;
gint i;
gboolean sent_ps;
rtph265pay = GST_RTP_H265_PAY (basepayload);
mtu = GST_RTP_BASE_PAYLOAD_MTU (rtph265pay);
/* should set src caps before pushing stuff,
* and if we did not see enough VPS/SPS/PPS, that may not be the case */
if (G_UNLIKELY (!gst_pad_has_current_caps (GST_RTP_BASE_PAYLOAD_SRCPAD
(basepayload))))
gst_rtp_h265_pay_set_vps_sps_pps (basepayload);
ret = GST_FLOW_OK;
sent_ps = FALSE;
for (i = 0; i < paybufs->len; i++) {
guint8 nal_header[2];
guint8 nal_type;
GstBuffer *paybuf;
gboolean send_ps;
guint size;
gboolean marker;
paybuf = g_ptr_array_index (paybufs, i);
if (ret != GST_FLOW_OK) {
/* unref buffers that will not be payloaded after a flow error */
gst_buffer_unref (paybuf);
continue;
}
marker = GST_BUFFER_FLAG_IS_SET (paybuf, GST_BUFFER_FLAG_MARKER);
size = gst_buffer_get_size (paybuf);
gst_buffer_extract (paybuf, 0, nal_header, 2);
nal_type = (nal_header[0] >> 1) & 0x3f;
GST_DEBUG_OBJECT (rtph265pay, "payloading NAL Unit: datasize=%u type=%d"
" pts=%" GST_TIME_FORMAT, size, nal_type, GST_TIME_ARGS (pts));
send_ps = FALSE;
/* check if we need to emit an VPS/SPS/PPS now */
if ((nal_type == GST_H265_NAL_SLICE_TRAIL_N)
|| (nal_type == GST_H265_NAL_SLICE_TRAIL_R)
|| (nal_type == GST_H265_NAL_SLICE_TSA_N)
|| (nal_type == GST_H265_NAL_SLICE_TSA_R)
|| (nal_type == GST_H265_NAL_SLICE_STSA_N)
|| (nal_type == GST_H265_NAL_SLICE_STSA_R)
|| (nal_type == GST_H265_NAL_SLICE_RASL_N)
|| (nal_type == GST_H265_NAL_SLICE_RASL_R)
|| (nal_type == GST_H265_NAL_SLICE_BLA_W_LP)
|| (nal_type == GST_H265_NAL_SLICE_BLA_W_RADL)
|| (nal_type == GST_H265_NAL_SLICE_BLA_N_LP)
|| (nal_type == GST_H265_NAL_SLICE_IDR_W_RADL)
|| (nal_type == GST_H265_NAL_SLICE_IDR_N_LP)
|| (nal_type == GST_H265_NAL_SLICE_CRA_NUT)) {
if (rtph265pay->vps_sps_pps_interval > 0) {
if (rtph265pay->last_vps_sps_pps != -1) {
guint64 diff;
GstClockTime running_time =
gst_segment_to_running_time (&basepayload->segment,
GST_FORMAT_TIME, pts);
GST_LOG_OBJECT (rtph265pay,
"now %" GST_TIME_FORMAT ", last VPS/SPS/PPS %" GST_TIME_FORMAT,
GST_TIME_ARGS (running_time),
GST_TIME_ARGS (rtph265pay->last_vps_sps_pps));
/* calculate diff between last SPS/PPS in milliseconds */
if (running_time > rtph265pay->last_vps_sps_pps)
diff = running_time - rtph265pay->last_vps_sps_pps;
else
diff = 0;
GST_DEBUG_OBJECT (rtph265pay,
"interval since last VPS/SPS/PPS %" GST_TIME_FORMAT,
GST_TIME_ARGS (diff));
/* bigger than interval, queue SPS/PPS */
if (GST_TIME_AS_SECONDS (diff) >= rtph265pay->vps_sps_pps_interval) {
GST_DEBUG_OBJECT (rtph265pay, "time to send VPS/SPS/PPS");
send_ps = TRUE;
}
} else {
/* no known previous SPS/PPS time, send now */
GST_DEBUG_OBJECT (rtph265pay,
"no previous VPS/SPS/PPS time, send now");
send_ps = TRUE;
}
} else if (rtph265pay->vps_sps_pps_interval == -1
&& (nal_type == GST_H265_NAL_SLICE_IDR_W_RADL
|| nal_type == GST_H265_NAL_SLICE_IDR_N_LP)) {
/* send VPS/SPS/PPS before every IDR frame */
send_ps = TRUE;
}
}
if (!sent_ps && (send_ps || rtph265pay->send_vps_sps_pps)) {
/* we need to send SPS/PPS now first. FIXME, don't use the pts for
* checking when we need to send SPS/PPS but convert to running_time
* first */
rtph265pay->send_vps_sps_pps = FALSE;
sent_ps = TRUE;
GST_DEBUG_OBJECT (rtph265pay, "sending VPS/SPS/PPS before current frame");
ret =
gst_rtp_h265_pay_send_vps_sps_pps (basepayload, rtph265pay, dts, pts);
if (ret != GST_FLOW_OK) {
gst_buffer_unref (paybuf);
continue;
}
}
if (rtph265pay->aggregate_mode != GST_RTP_H265_AGGREGATE_NONE)
ret = gst_rtp_h265_pay_payload_nal_bundle (basepayload, paybuf, dts, pts,
marker, nal_type, nal_header, size);
else
ret = gst_rtp_h265_pay_payload_nal_fragment (basepayload, paybuf, dts,
pts, marker, mtu, nal_type, nal_header, size);
}
g_ptr_array_free (paybufs, TRUE);
return ret;
}
static GstFlowReturn
gst_rtp_h265_pay_payload_nal_single (GstRTPBasePayload * basepayload,
GstBuffer * paybuf, GstClockTime dts, GstClockTime pts, gboolean marker)
{
GstBufferList *outlist;
GstBuffer *outbuf;
GstRTPBuffer rtp = GST_RTP_BUFFER_INIT;
/* use buffer lists
* create buffer without payload containing only the RTP header
* (memory block at index 0) */
outbuf = gst_rtp_base_payload_allocate_output_buffer (basepayload, 0, 0, 0);
gst_rtp_buffer_map (outbuf, GST_MAP_WRITE, &rtp);
/* Mark the end of a frame */
gst_rtp_buffer_set_marker (&rtp, marker);
GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_MARKER);
/* timestamp the outbuffer */
GST_BUFFER_PTS (outbuf) = pts;
GST_BUFFER_DTS (outbuf) = dts;
/* insert payload memory block */
gst_rtp_copy_video_meta (basepayload, outbuf, paybuf);
outbuf = gst_buffer_append (outbuf, paybuf);
outlist = gst_buffer_list_new ();
/* add the buffer to the buffer list */
gst_buffer_list_add (outlist, outbuf);
gst_rtp_buffer_unmap (&rtp);
/* push the list to the next element in the pipe */
return gst_rtp_base_payload_push_list (basepayload, outlist);
}
static GstFlowReturn
gst_rtp_h265_pay_payload_nal_fragment (GstRTPBasePayload * basepayload,
GstBuffer * paybuf, GstClockTime dts, GstClockTime pts, gboolean marker,
guint mtu, guint8 nal_type, const guint8 * nal_header, int size)
{
GstRtpH265Pay *rtph265pay = (GstRtpH265Pay *) basepayload;
GstFlowReturn ret;
guint max_fragment_size, ii, pos;
GstBuffer *outbuf;
GstBufferList *outlist = NULL;
GstRTPBuffer rtp = GST_RTP_BUFFER_INIT;
guint8 *payload;
if (gst_rtp_buffer_calc_packet_len (size, 0, 0) < mtu) {
GST_DEBUG_OBJECT (rtph265pay,
"NAL Unit fit in one packet datasize=%d mtu=%d", size, mtu);
/* will fit in one packet */
return gst_rtp_h265_pay_payload_nal_single (basepayload, paybuf, dts, pts,
marker);
}
GST_DEBUG_OBJECT (basepayload,
"NAL Unit DOES NOT fit in one packet datasize=%d mtu=%d", size, mtu);
GST_DEBUG_OBJECT (basepayload, "Using FU fragmentation for data size=%d",
size - 2);
/* We keep 3 bytes for PayloadHdr and FU Header */
max_fragment_size = gst_rtp_buffer_calc_payload_len (mtu - 3, 0, 0);
outlist = gst_buffer_list_new ();
for (pos = 2, ii = 0; pos < size; pos += max_fragment_size, ii++) {
guint remaining, fragment_size;
gboolean first_fragment, last_fragment;
remaining = size - pos;
fragment_size = MIN (remaining, max_fragment_size);
first_fragment = (pos == 2);
last_fragment = (remaining <= max_fragment_size);
GST_DEBUG_OBJECT (basepayload,
"Inside FU fragmentation fragment_size=%u iteration=%d %s%s",
fragment_size, ii, first_fragment ? "first" : "",
last_fragment ? "last" : "");
/* use buffer lists
* create buffer without payload containing only the RTP header
* (memory block at index 0), and with space for PayloadHdr and FU header */
outbuf = gst_rtp_base_payload_allocate_output_buffer (basepayload, 3, 0, 0);
gst_rtp_buffer_map (outbuf, GST_MAP_WRITE, &rtp);
GST_BUFFER_DTS (outbuf) = dts;
GST_BUFFER_PTS (outbuf) = pts;
payload = gst_rtp_buffer_get_payload (&rtp);
/* PayloadHdr (type = FU_TYPE_ID (49)) */
payload[0] = (nal_header[0] & 0x81) | (FU_TYPE_ID << 1);
payload[1] = nal_header[1];
/* If it's the last fragment and the end of this au, mark the end of
* slice */
gst_rtp_buffer_set_marker (&rtp, last_fragment && marker);
if (last_fragment && marker)
GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_MARKER);
/* FU Header */
payload[2] = (first_fragment << 7) | (last_fragment << 6) |
(nal_type & 0x3f);
gst_rtp_buffer_unmap (&rtp);
/* insert payload memory block */
gst_rtp_copy_video_meta (rtph265pay, outbuf, paybuf);
gst_buffer_copy_into (outbuf, paybuf, GST_BUFFER_COPY_MEMORY, pos,
fragment_size);
/* add the buffer to the buffer list */
gst_buffer_list_add (outlist, outbuf);
}
ret = gst_rtp_base_payload_push_list (basepayload, outlist);
gst_buffer_unref (paybuf);
return ret;
}
static GstFlowReturn
gst_rtp_h265_pay_send_bundle (GstRtpH265Pay * rtph265pay, gboolean marker)
{
GstRTPBasePayload *basepayload;
GstBufferList *bundle;
guint length, bundle_size;
GstBuffer *first, *outbuf;
GstClockTime dts, pts;
bundle_size = rtph265pay->bundle_size;
if (bundle_size == 0) {
GST_DEBUG_OBJECT (rtph265pay, "no bundle, nothing to send");
return GST_FLOW_OK;
}
basepayload = GST_RTP_BASE_PAYLOAD (rtph265pay);
bundle = rtph265pay->bundle;
length = gst_buffer_list_length (bundle);
first = gst_buffer_list_get (bundle, 0);
dts = GST_BUFFER_DTS (first);
pts = GST_BUFFER_PTS (first);
if (length == 1) {
/* Push unaggregated NALU */
outbuf = gst_buffer_ref (first);
GST_DEBUG_OBJECT (rtph265pay,
"sending NAL Unit unaggregated: datasize=%u", bundle_size - 2);
} else {
guint8 ap_header[2];
guint i;
guint8 layer_id = 0xFF;
guint8 temporal_id = 0xFF;
outbuf = gst_buffer_new_allocate (NULL, sizeof ap_header, NULL);
for (i = 0; i < length; i++) {
GstBuffer *buf = gst_buffer_list_get (bundle, i);
guint8 nal_header[2];
GstMemory *size_header;
GstMapInfo map;
guint8 nal_layer_id;
guint8 nal_temporal_id;
gst_buffer_extract (buf, 0, &nal_header, sizeof nal_header);
/* Propagate F bit */
if ((nal_header[0] & 0x80))
ap_header[0] |= 0x80;
/* Select lowest layer_id & temporal_id */
nal_layer_id = ((nal_header[0] & 0x01) << 5) |
((nal_header[1] >> 3) & 0x1F);
nal_temporal_id = nal_header[1] & 0x7;
layer_id = MIN (layer_id, nal_layer_id);
temporal_id = MIN (temporal_id, nal_temporal_id);
/* append NALU size */
size_header = gst_allocator_alloc (NULL, 2, NULL);
gst_memory_map (size_header, &map, GST_MAP_WRITE);
GST_WRITE_UINT16_BE (map.data, gst_buffer_get_size (buf));
gst_memory_unmap (size_header, &map);
gst_buffer_append_memory (outbuf, size_header);
/* append NALU data */
outbuf = gst_buffer_append (outbuf, gst_buffer_ref (buf));
}
ap_header[0] = (AP_TYPE_ID << 1) | (layer_id & 0x20);
ap_header[1] = ((layer_id & 0x1F) << 3) | (temporal_id & 0x07);
gst_buffer_fill (outbuf, 0, &ap_header, sizeof ap_header);
GST_DEBUG_OBJECT (rtph265pay,
"sending AP bundle: n=%u header=%02x%02x datasize=%u",
length, ap_header[0], ap_header[1], bundle_size);
}
gst_rtp_h265_pay_reset_bundle (rtph265pay);
return gst_rtp_h265_pay_payload_nal_single (basepayload, outbuf, dts, pts,
marker);
}
static gboolean
gst_rtp_h265_pay_payload_nal_bundle (GstRTPBasePayload * basepayload,
GstBuffer * paybuf, GstClockTime dts, GstClockTime pts,
gboolean marker, guint8 nal_type, const guint8 * nal_header, int size)
{
GstRtpH265Pay *rtph265pay;
GstFlowReturn ret;
guint pay_size, bundle_size;
GstBufferList *bundle;
gboolean start_of_au;
guint mtu;
rtph265pay = GST_RTP_H265_PAY (basepayload);
mtu = GST_RTP_BASE_PAYLOAD_MTU (rtph265pay);
pay_size = 2 + gst_buffer_get_size (paybuf);
bundle = rtph265pay->bundle;
start_of_au = FALSE;
if (bundle) {
GstBuffer *first = gst_buffer_list_get (bundle, 0);
if (nal_type == GST_H265_NAL_AUD) {
GST_DEBUG_OBJECT (rtph265pay, "found access delimiter");
start_of_au = TRUE;
} else if (GST_BUFFER_IS_DISCONT (paybuf)) {
GST_DEBUG_OBJECT (rtph265pay, "found discont");
start_of_au = TRUE;
} else if (GST_BUFFER_PTS (first) != pts || GST_BUFFER_DTS (first) != dts) {
GST_DEBUG_OBJECT (rtph265pay, "found timestamp mismatch");
start_of_au = TRUE;
}
}
if (start_of_au) {
GST_DEBUG_OBJECT (rtph265pay, "sending bundle before start of AU");
ret = gst_rtp_h265_pay_send_bundle (rtph265pay, TRUE);
if (ret != GST_FLOW_OK)
goto out;
bundle = NULL;
}
bundle_size = 2 + pay_size;
if (gst_rtp_buffer_calc_packet_len (bundle_size, 0, 0) > mtu) {
GST_DEBUG_OBJECT (rtph265pay, "NAL Unit cannot fit in a bundle");
ret = gst_rtp_h265_pay_send_bundle (rtph265pay, FALSE);
if (ret != GST_FLOW_OK)
goto out;
return gst_rtp_h265_pay_payload_nal_fragment (basepayload, paybuf, dts, pts,
marker, mtu, nal_type, nal_header, size);
}
bundle_size = rtph265pay->bundle_size + pay_size;
if (gst_rtp_buffer_calc_packet_len (bundle_size, 0, 0) > mtu) {
GST_DEBUG_OBJECT (rtph265pay,
"bundle overflows, sending: bundlesize=%u datasize=2+%u mtu=%u",
rtph265pay->bundle_size, pay_size - 2, mtu);
ret = gst_rtp_h265_pay_send_bundle (rtph265pay, FALSE);
if (ret != GST_FLOW_OK)
goto out;
bundle = NULL;
}
if (!bundle) {
GST_DEBUG_OBJECT (rtph265pay, "creating new AP aggregate");
bundle = rtph265pay->bundle = gst_buffer_list_new ();
bundle_size = rtph265pay->bundle_size = 2;
rtph265pay->bundle_contains_vcl_or_suffix = FALSE;
}
GST_DEBUG_OBJECT (rtph265pay,
"bundling NAL Unit: bundlesize=%u datasize=2+%u mtu=%u",
rtph265pay->bundle_size, pay_size - 2, mtu);
paybuf = gst_buffer_make_writable (paybuf);
GST_BUFFER_PTS (paybuf) = pts;
GST_BUFFER_DTS (paybuf) = dts;
gst_buffer_list_add (bundle, gst_buffer_ref (paybuf));
rtph265pay->bundle_size += pay_size;
ret = GST_FLOW_OK;
/* In H.265, all VCL NAL units are < 32 */
if (nal_type < 32 || nal_type == GST_H265_NAL_EOS ||
nal_type == GST_H265_NAL_EOB || nal_type == GST_H265_NAL_SUFFIX_SEI ||
(nal_type >= 45 && nal_type <= 47) || (nal_type >= 56 && nal_type < 63))
rtph265pay->bundle_contains_vcl_or_suffix = TRUE;
if (marker) {
GST_DEBUG_OBJECT (rtph265pay, "sending bundle at marker");
ret = gst_rtp_h265_pay_send_bundle (rtph265pay, TRUE);
}
out:
gst_buffer_unref (paybuf);
return ret;
}
static GstFlowReturn
gst_rtp_h265_pay_handle_buffer (GstRTPBasePayload * basepayload,
GstBuffer * buffer)
{
GstRtpH265Pay *rtph265pay;
GstFlowReturn ret;
gsize size;
guint nal_len, i;
const guint8 *data;
GstClockTime dts, pts;
GArray *nal_queue;
gboolean hevc;
GstBuffer *paybuf = NULL;
gsize skip;
gboolean marker = FALSE;
gboolean discont = FALSE;
gboolean draining = (buffer == NULL);
rtph265pay = GST_RTP_H265_PAY (basepayload);
/* the input buffer contains one or more NAL units */
hevc = (rtph265pay->stream_format == GST_H265_STREAM_FORMAT_HEV1)
|| (rtph265pay->stream_format == GST_H265_STREAM_FORMAT_HVC1);
if (hevc) {
/* In hevc mode, there is no adapter, so nothing to drain */
if (draining)
return GST_FLOW_OK;
} else {
if (buffer) {
if (gst_adapter_available (rtph265pay->adapter) == 0)
discont = GST_BUFFER_FLAG_IS_SET (buffer, GST_BUFFER_FLAG_DISCONT);
marker = GST_BUFFER_FLAG_IS_SET (buffer, GST_BUFFER_FLAG_MARKER);
gst_adapter_push (rtph265pay->adapter, buffer);
buffer = NULL;
}
/* We want to use the first TS used to construct the following NAL */
dts = gst_adapter_prev_dts (rtph265pay->adapter, NULL);
pts = gst_adapter_prev_pts (rtph265pay->adapter, NULL);
size = gst_adapter_available (rtph265pay->adapter);
/* Nothing to do here if the adapter is empty, e.g. on EOS */
if (size == 0)
return GST_FLOW_OK;
data = gst_adapter_map (rtph265pay->adapter, size);
GST_DEBUG_OBJECT (basepayload, "got %" G_GSIZE_FORMAT " bytes", size);
}
ret = GST_FLOW_OK;
/* now loop over all NAL units and put them in a packet */
if (hevc) {
GstBufferMemoryMap memory;
gsize remaining_buffer_size;
guint nal_length_size;
gsize offset = 0;
GPtrArray *paybufs;
paybufs = g_ptr_array_new ();
nal_length_size = rtph265pay->nal_length_size;
gst_buffer_memory_map (buffer, &memory);
remaining_buffer_size = gst_buffer_get_size (buffer);
pts = GST_BUFFER_PTS (buffer);
dts = GST_BUFFER_DTS (buffer);
marker = GST_BUFFER_FLAG_IS_SET (buffer, GST_BUFFER_FLAG_MARKER);
GST_DEBUG_OBJECT (basepayload, "got %" G_GSIZE_FORMAT " bytes",
remaining_buffer_size);
while (remaining_buffer_size > nal_length_size) {
gint i;
nal_len = 0;
for (i = 0; i < nal_length_size; i++) {
nal_len = (nal_len << 8) + *memory.data;
if (!gst_buffer_memory_advance_bytes (&memory, 1))
break;
}
offset += nal_length_size;
remaining_buffer_size -= nal_length_size;
if (remaining_buffer_size >= nal_len) {
GST_DEBUG_OBJECT (basepayload, "got NAL of size %u", nal_len);
} else {
nal_len = remaining_buffer_size;
GST_DEBUG_OBJECT (basepayload, "got incomplete NAL of size %u",
nal_len);
}
paybuf = gst_buffer_copy_region (buffer, GST_BUFFER_COPY_ALL, offset,
nal_len);
g_ptr_array_add (paybufs, paybuf);
/* If we're at the end of the buffer, then we're at the end of the
* access unit
*/
GST_BUFFER_FLAG_UNSET (paybuf, GST_BUFFER_FLAG_MARKER);
if (remaining_buffer_size - nal_len <= nal_length_size) {
if (rtph265pay->alignment == GST_H265_ALIGNMENT_AU || marker)
GST_BUFFER_FLAG_SET (paybuf, GST_BUFFER_FLAG_MARKER);
}
GST_BUFFER_FLAG_UNSET (paybuf, GST_BUFFER_FLAG_DISCONT);
if (discont) {
GST_BUFFER_FLAG_SET (paybuf, GST_BUFFER_FLAG_DISCONT);
discont = FALSE;
}
/* Skip current nal. If it is split over multiple GstMemory
* advance_bytes () will switch to the correct GstMemory. The payloader
* does not access those bytes directly but uses gst_buffer_copy_region ()
* to create a sub-buffer referencing the nal instead */
if (!gst_buffer_memory_advance_bytes (&memory, nal_len))
break;
offset += nal_len;
remaining_buffer_size -= nal_len;
}
ret = gst_rtp_h265_pay_payload_nal (basepayload, paybufs, dts, pts);
gst_buffer_memory_unmap (&memory);
gst_buffer_unref (buffer);
} else {
guint next;
gboolean update = FALSE;
GPtrArray *paybufs;
/* get offset of first start code */
next = next_start_code (data, size);
/* skip to start code, if no start code is found, next will be size and we
* will not collect data. */
data += next;
size -= next;
nal_queue = rtph265pay->queue;
skip = next;
/* array must be empty when we get here */
g_assert (nal_queue->len == 0);
GST_DEBUG_OBJECT (basepayload,
"found first start at %u, bytes left %" G_GSIZE_FORMAT, next, size);
paybufs = g_ptr_array_new ();
/* first pass to locate NALs and parse VPS/SPS/PPS */
while (size > 4) {
/* skip start code */
data += 3;
size -= 3;
/* use next_start_code() to scan buffer.
* next_start_code() returns the offset in data,
* starting from zero to the first byte of 0.0.0.1
* If no start code is found, it returns the value of the
* 'size' parameter.
* data is unchanged by the call to next_start_code()
*/
next = next_start_code (data, size);
/* nal or au aligned input needs no delaying until next time */
if (next == size && !draining &&
rtph265pay->alignment == GST_H265_ALIGNMENT_UNKNOWN) {
/* Didn't find the start of next NAL and it's not EOS,
* handle it next time */
break;
}
/* nal length is distance to next start code */
nal_len = next;
GST_DEBUG_OBJECT (basepayload, "found next start at %u of size %u", next,
nal_len);
/* We know our stream is a valid H265 NAL packet,
* go parse it for VPS/SPS/PPS to enrich the caps */
/* order: make sure to check nal */
update = gst_rtp_h265_pay_decode_nal (rtph265pay, data, nal_len, dts, pts)
|| update;
/* move to next NAL packet */
data += nal_len;
size -= nal_len;
g_array_append_val (nal_queue, nal_len);
}
/* if has new VPS, SPS & PPS, update the output caps */
if (G_UNLIKELY (update))
if (!gst_rtp_h265_pay_set_vps_sps_pps (basepayload))
goto caps_rejected;
/* second pass to payload and push */
if (nal_queue->len != 0)
gst_adapter_flush (rtph265pay->adapter, skip);
for (i = 0; i < nal_queue->len; i++) {
guint size;
nal_len = g_array_index (nal_queue, guint, i);
/* skip start code */
gst_adapter_flush (rtph265pay->adapter, 3);
/* Trim the end unless we're the last NAL in the stream.
* In case we're not at the end of the buffer we know the next block
* starts with 0x000001 so all the 0x00 bytes at the end of this one are
* trailing 0x0 that can be discarded */
size = nal_len;
data = gst_adapter_map (rtph265pay->adapter, size);
if (i + 1 != nal_queue->len || !draining)
for (; size > 2 && data[size - 1] == 0x0; size--)
/* skip */ ;
paybuf = gst_adapter_take_buffer (rtph265pay->adapter, size);
g_assert (paybuf);
g_ptr_array_add (paybufs, paybuf);
/* If it's the last nal unit we have in non-bytestream mode, we can
* assume it's the end of an access-unit */
GST_BUFFER_FLAG_UNSET (paybuf, GST_BUFFER_FLAG_MARKER);
if (i == nal_queue->len - 1) {
if (rtph265pay->alignment == GST_H265_ALIGNMENT_AU ||
marker || draining)
GST_BUFFER_FLAG_SET (paybuf, GST_BUFFER_FLAG_MARKER);
}
GST_BUFFER_FLAG_UNSET (paybuf, GST_BUFFER_FLAG_DISCONT);
if (discont) {
GST_BUFFER_FLAG_SET (paybuf, GST_BUFFER_FLAG_DISCONT);
discont = FALSE;
}
/* move to next NAL packet */
/* Skips the trailing zeros */
gst_adapter_flush (rtph265pay->adapter, nal_len - size);
}
/* put the data in one or more RTP packets */
ret = gst_rtp_h265_pay_payload_nal (basepayload, paybufs, dts, pts);
g_array_set_size (nal_queue, 0);
}
if (ret == GST_FLOW_OK && rtph265pay->bundle_size > 0 &&
rtph265pay->aggregate_mode == GST_RTP_H265_AGGREGATE_ZERO_LATENCY &&
rtph265pay->bundle_contains_vcl_or_suffix) {
GST_DEBUG_OBJECT (rtph265pay, "sending bundle at end incoming packet");
ret = gst_rtp_h265_pay_send_bundle (rtph265pay, FALSE);
}
done:
if (!hevc) {
gst_adapter_unmap (rtph265pay->adapter);
}
return ret;
caps_rejected:
{
GST_WARNING_OBJECT (basepayload, "Could not set outcaps");
g_array_set_size (nal_queue, 0);
ret = GST_FLOW_NOT_NEGOTIATED;
goto done;
}
}
static gboolean
gst_rtp_h265_pay_sink_event (GstRTPBasePayload * payload, GstEvent * event)
{
gboolean res;
const GstStructure *s;
GstRtpH265Pay *rtph265pay = GST_RTP_H265_PAY (payload);
GstFlowReturn ret = GST_FLOW_OK;
switch (GST_EVENT_TYPE (event)) {
case GST_EVENT_FLUSH_STOP:
gst_adapter_clear (rtph265pay->adapter);
gst_rtp_h265_pay_reset_bundle (rtph265pay);
break;
case GST_EVENT_CUSTOM_DOWNSTREAM:
s = gst_event_get_structure (event);
if (gst_structure_has_name (s, "GstForceKeyUnit")) {
gboolean resend_codec_data;
if (gst_structure_get_boolean (s, "all-headers",
&resend_codec_data) && resend_codec_data)
rtph265pay->send_vps_sps_pps = TRUE;
}
break;
case GST_EVENT_EOS:
{
/* call handle_buffer with NULL to flush last NAL from adapter
* in byte-stream mode
*/
gst_rtp_h265_pay_handle_buffer (payload, NULL);
ret = gst_rtp_h265_pay_send_bundle (rtph265pay, TRUE);
break;
}
case GST_EVENT_STREAM_START:
GST_DEBUG_OBJECT (rtph265pay,
"New stream detected => Clear VPS, SPS and PPS");
gst_rtp_h265_pay_clear_vps_sps_pps (rtph265pay);
break;
default:
break;
}
if (ret != GST_FLOW_OK)
return FALSE;
res = GST_RTP_BASE_PAYLOAD_CLASS (parent_class)->sink_event (payload, event);
return res;
}
static GstStateChangeReturn
gst_rtp_h265_pay_change_state (GstElement * element, GstStateChange transition)
{
GstStateChangeReturn ret;
GstRtpH265Pay *rtph265pay = GST_RTP_H265_PAY (element);
switch (transition) {
case GST_STATE_CHANGE_READY_TO_PAUSED:
rtph265pay->send_vps_sps_pps = FALSE;
gst_adapter_clear (rtph265pay->adapter);
gst_rtp_h265_pay_reset_bundle (rtph265pay);
break;
default:
break;
}
ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
switch (transition) {
case GST_STATE_CHANGE_PAUSED_TO_READY:
rtph265pay->last_vps_sps_pps = -1;
gst_rtp_h265_pay_clear_vps_sps_pps (rtph265pay);
break;
default:
break;
}
return ret;
}
static void
gst_rtp_h265_pay_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec)
{
GstRtpH265Pay *rtph265pay;
rtph265pay = GST_RTP_H265_PAY (object);
switch (prop_id) {
case PROP_CONFIG_INTERVAL:
rtph265pay->vps_sps_pps_interval = g_value_get_int (value);
break;
case PROP_AGGREGATE_MODE:
rtph265pay->aggregate_mode = g_value_get_enum (value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static void
gst_rtp_h265_pay_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec)
{
GstRtpH265Pay *rtph265pay;
rtph265pay = GST_RTP_H265_PAY (object);
switch (prop_id) {
case PROP_CONFIG_INTERVAL:
g_value_set_int (value, rtph265pay->vps_sps_pps_interval);
break;
case PROP_AGGREGATE_MODE:
g_value_set_enum (value, rtph265pay->aggregate_mode);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}