/* * tsdemux.c * Copyright (C) 2009 Zaheer Abbas Merali * 2010 Edward Hervey * Copyright (C) 2011, Hewlett-Packard Development Company, L.P. * Author: Youness Alaoui , Collabora Ltd. * Author: Sebastian Dröge , Collabora Ltd. * Author: Edward Hervey , Collabora Ltd. * * Authors: * Zaheer Abbas Merali * Edward Hervey * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #include #include #include #include "mpegtsbase.h" #include "tsdemux.h" #include "gstmpegdesc.h" #include "gstmpegdefs.h" #include "mpegtspacketizer.h" #include "pesparse.h" #include #include #include #include #define _gst_log2(x) (log(x)/log(2)) /* * tsdemux * * See TODO for explanations on improvements needed */ #define CONTINUITY_UNSET 255 #define MAX_CONTINUITY 15 /* Seeking/Scanning related variables */ /* seek to SEEK_TIMESTAMP_OFFSET before the desired offset and search then * either accurately or for the next timestamp */ #define SEEK_TIMESTAMP_OFFSET (2500 * GST_MSECOND) #define GST_FLOW_REWINDING GST_FLOW_CUSTOM_ERROR /* latency in msecs */ #define DEFAULT_LATENCY (700) /* Limit PES packet collection to a maximum of 32MB * which is more than large enough to support an H264 frame at * maximum profile/level/bitrate at 30fps or above. * PES bigger than this limit will be output in buffers of * up to this size */ #define MAX_PES_PAYLOAD (32 * 1024 * 1024) GST_DEBUG_CATEGORY_STATIC (ts_demux_debug); #define GST_CAT_DEFAULT ts_demux_debug #define ABSDIFF(a,b) (((a) > (b)) ? ((a) - (b)) : ((b) - (a))) static GQuark QUARK_TSDEMUX; static GQuark QUARK_PID; static GQuark QUARK_PCR; static GQuark QUARK_OPCR; static GQuark QUARK_PTS; static GQuark QUARK_DTS; static GQuark QUARK_OFFSET; typedef enum { PENDING_PACKET_EMPTY = 0, /* No pending packet/buffer * Push incoming buffers to the array */ PENDING_PACKET_HEADER, /* PES header needs to be parsed * Push incoming buffers to the array */ PENDING_PACKET_BUFFER, /* Currently filling up output buffer * Push incoming buffers to the bufferlist */ PENDING_PACKET_DISCONT /* Discontinuity in incoming packets * Drop all incoming buffers */ } PendingPacketState; /* Pending buffer */ typedef struct { /* The fully reconstructed buffer */ GstBuffer *buffer; /* Raw PTS/DTS (in 90kHz units) */ guint64 pts, dts; } PendingBuffer; typedef struct _TSDemuxStream TSDemuxStream; typedef struct _TSDemuxH264ParsingInfos TSDemuxH264ParsingInfos; typedef struct _TSDemuxJP2KParsingInfos TSDemuxJP2KParsingInfos; typedef struct _TSDemuxADTSParsingInfos TSDemuxADTSParsingInfos; /* Returns TRUE if a keyframe was found */ typedef gboolean (*GstTsDemuxKeyFrameScanFunction) (TSDemuxStream * stream, guint8 * data, const gsize data_size, const gsize max_frame_offset); typedef struct { guint8 *data; gsize size; } SimpleBuffer; struct _TSDemuxH264ParsingInfos { /* H264 parsing data */ GstH264NalParser *parser; GstByteWriter *sps; GstByteWriter *pps; GstByteWriter *sei; SimpleBuffer framedata; }; struct _TSDemuxJP2KParsingInfos { /* J2K parsing data */ gboolean interlace; }; struct _TSDemuxADTSParsingInfos { guint mpegversion; }; struct _TSDemuxStream { MpegTSBaseStream stream; GstPad *pad; /* Whether the pad was added or not */ gboolean active; /* Whether this is a sparse stream (subtitles or metadata) */ gboolean sparse; /* TRUE if we are waiting for a valid timestamp */ gboolean pending_ts; /* Output data */ PendingPacketState state; /* Data being reconstructed (allocated) */ guint8 *data; /* Size of data being reconstructed (if known, else 0) */ guint expected_size; /* Amount of bytes in current ->data */ guint current_size; /* Size of ->data */ guint allocated_size; /* Current PTS/DTS for this stream (in running time) */ GstClockTime pts; GstClockTime dts; /* Reference PTS used to detect gaps */ GstClockTime gap_ref_pts; /* Number of outputted buffers */ guint32 nb_out_buffers; /* Reference number of buffers for gaps */ guint32 gap_ref_buffers; /* Current PTS/DTS for this stream (in 90kHz unit) */ guint64 raw_pts, raw_dts; /* Whether this stream needs to send a newsegment */ gboolean need_newsegment; /* Whether the next output buffer should be DISCONT */ gboolean discont; /* The value to use when calculating the newsegment */ GstClockTime first_pts; GstTagList *taglist; gint continuity_counter; /* List of pending buffers */ GList *pending; /* if != 0, output only PES from that substream */ guint8 target_pes_substream; gboolean needs_keyframe; GstClockTime seeked_pts, seeked_dts; GstTsDemuxKeyFrameScanFunction scan_function; TSDemuxH264ParsingInfos h264infos; TSDemuxJP2KParsingInfos jp2kInfos; TSDemuxADTSParsingInfos atdsInfos; }; #define VIDEO_CAPS \ GST_STATIC_CAPS (\ "video/mpeg, " \ "mpegversion = (int) { 1, 2, 4 }, " \ "systemstream = (boolean) FALSE; " \ "video/x-h264,stream-format=(string)byte-stream," \ "alignment=(string)nal;" \ "video/x-h265,stream-format=(string)byte-stream," \ "alignment=(string)nal;" \ "video/x-dirac;" \ "video/x-cavs;" \ "video/x-wmv," \ "wmvversion = (int) 3, " \ "format = (string) WVC1;" \ "image/x-jpc;" \ ) #define AUDIO_CAPS \ GST_STATIC_CAPS ( \ "audio/mpeg, " \ "mpegversion = (int) 1;" \ "audio/mpeg, " \ "mpegversion = (int) { 2, 4 }, " \ "stream-format = (string) adts; " \ "audio/mpeg, " \ "mpegversion = (int) 4, " \ "stream-format = (string) loas; " \ "audio/x-lpcm, " \ "width = (int) { 16, 20, 24 }, " \ "rate = (int) { 48000, 96000 }, " \ "channels = (int) [ 1, 8 ], " \ "dynamic_range = (int) [ 0, 255 ], " \ "emphasis = (boolean) { FALSE, TRUE }, " \ "mute = (boolean) { FALSE, TRUE }; " \ "audio/x-ac3; audio/x-eac3;" \ "audio/x-ac4;" \ "audio/x-dts;" \ "audio/x-opus;" \ "audio/x-private-ts-lpcm" \ ) /* Can also use the subpicture pads for text subtitles? */ #define SUBPICTURE_CAPS \ GST_STATIC_CAPS ("subpicture/x-pgs; subpicture/x-dvd; subpicture/x-dvb") static GstStaticPadTemplate video_template = GST_STATIC_PAD_TEMPLATE ("video_%01x_%05x", GST_PAD_SRC, GST_PAD_SOMETIMES, VIDEO_CAPS); static GstStaticPadTemplate audio_template = GST_STATIC_PAD_TEMPLATE ("audio_%01x_%05x", GST_PAD_SRC, GST_PAD_SOMETIMES, AUDIO_CAPS); static GstStaticPadTemplate subpicture_template = GST_STATIC_PAD_TEMPLATE ("subpicture_%01x_%05x", GST_PAD_SRC, GST_PAD_SOMETIMES, SUBPICTURE_CAPS); static GstStaticPadTemplate private_template = GST_STATIC_PAD_TEMPLATE ("private_%01x_%05x", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); enum { PROP_0, PROP_PROGRAM_NUMBER, PROP_EMIT_STATS, PROP_LATENCY, /* FILL ME */ }; /* Pad functions */ /* mpegtsbase methods */ static void gst_ts_demux_update_program (MpegTSBase * base, MpegTSBaseProgram * program); static void gst_ts_demux_program_started (MpegTSBase * base, MpegTSBaseProgram * program); static void gst_ts_demux_program_stopped (MpegTSBase * base, MpegTSBaseProgram * program); static gboolean gst_ts_demux_can_remove_program (MpegTSBase * base, MpegTSBaseProgram * program); static void gst_ts_demux_reset (MpegTSBase * base); static GstFlowReturn gst_ts_demux_push (MpegTSBase * base, MpegTSPacketizerPacket * packet, GstMpegtsSection * section); static void gst_ts_demux_flush (MpegTSBase * base, gboolean hard); static GstFlowReturn gst_ts_demux_drain (MpegTSBase * base); static gboolean gst_ts_demux_stream_added (MpegTSBase * base, MpegTSBaseStream * stream, MpegTSBaseProgram * program); static void gst_ts_demux_stream_removed (MpegTSBase * base, MpegTSBaseStream * stream); static GstFlowReturn gst_ts_demux_do_seek (MpegTSBase * base, GstEvent * event); static void gst_ts_demux_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_ts_demux_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static void gst_ts_demux_flush_streams (GstTSDemux * tsdemux, gboolean hard); static GstFlowReturn gst_ts_demux_push_pending_data (GstTSDemux * demux, TSDemuxStream * stream, MpegTSBaseProgram * program); static void gst_ts_demux_stream_flush (TSDemuxStream * stream, GstTSDemux * demux, gboolean hard); static gboolean push_event (MpegTSBase * base, GstEvent * event); static gboolean sink_query (MpegTSBase * base, GstQuery * query); static void gst_ts_demux_check_and_sync_streams (GstTSDemux * demux, GstClockTime time); static void _extra_init (void) { QUARK_TSDEMUX = g_quark_from_string ("tsdemux"); QUARK_PID = g_quark_from_string ("pid"); QUARK_PCR = g_quark_from_string ("pcr"); QUARK_OPCR = g_quark_from_string ("opcr"); QUARK_PTS = g_quark_from_string ("pts"); QUARK_DTS = g_quark_from_string ("dts"); QUARK_OFFSET = g_quark_from_string ("offset"); } #define gst_ts_demux_parent_class parent_class G_DEFINE_TYPE_WITH_CODE (GstTSDemux, gst_ts_demux, GST_TYPE_MPEGTS_BASE, _extra_init ()); static void gst_ts_demux_dispose (GObject * object) { GstTSDemux *demux = GST_TS_DEMUX_CAST (object); gst_flow_combiner_free (demux->flowcombiner); GST_CALL_PARENT (G_OBJECT_CLASS, dispose, (object)); } static void gst_ts_demux_class_init (GstTSDemuxClass * klass) { GObjectClass *gobject_class; GstElementClass *element_class; MpegTSBaseClass *ts_class; gobject_class = G_OBJECT_CLASS (klass); gobject_class->set_property = gst_ts_demux_set_property; gobject_class->get_property = gst_ts_demux_get_property; gobject_class->dispose = gst_ts_demux_dispose; g_object_class_install_property (gobject_class, PROP_PROGRAM_NUMBER, g_param_spec_int ("program-number", "Program number", "Program Number to demux for (-1 to ignore)", -1, G_MAXINT, -1, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_EMIT_STATS, g_param_spec_boolean ("emit-stats", "Emit statistics", "Emit messages for every pcr/opcr/pts/dts", FALSE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_LATENCY, g_param_spec_int ("latency", "Latency", "Latency to add for smooth demuxing (in ms)", -1, G_MAXINT, DEFAULT_LATENCY, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); element_class = GST_ELEMENT_CLASS (klass); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&video_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&audio_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&subpicture_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&private_template)); gst_element_class_set_static_metadata (element_class, "MPEG transport stream demuxer", "Codec/Demuxer", "Demuxes MPEG2 transport streams", "Zaheer Abbas Merali \n" "Edward Hervey "); ts_class = GST_MPEGTS_BASE_CLASS (klass); ts_class->reset = GST_DEBUG_FUNCPTR (gst_ts_demux_reset); ts_class->push = GST_DEBUG_FUNCPTR (gst_ts_demux_push); ts_class->push_event = GST_DEBUG_FUNCPTR (push_event); ts_class->sink_query = GST_DEBUG_FUNCPTR (sink_query); ts_class->program_started = GST_DEBUG_FUNCPTR (gst_ts_demux_program_started); ts_class->program_stopped = GST_DEBUG_FUNCPTR (gst_ts_demux_program_stopped); ts_class->update_program = GST_DEBUG_FUNCPTR (gst_ts_demux_update_program); ts_class->can_remove_program = gst_ts_demux_can_remove_program; ts_class->stream_added = gst_ts_demux_stream_added; ts_class->stream_removed = gst_ts_demux_stream_removed; ts_class->seek = GST_DEBUG_FUNCPTR (gst_ts_demux_do_seek); ts_class->flush = GST_DEBUG_FUNCPTR (gst_ts_demux_flush); ts_class->drain = GST_DEBUG_FUNCPTR (gst_ts_demux_drain); } static void gst_ts_demux_reset (MpegTSBase * base) { GstTSDemux *demux = (GstTSDemux *) base; demux->rate = 1.0; if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } if (demux->global_tags) { gst_tag_list_unref (demux->global_tags); demux->global_tags = NULL; } if (demux->previous_program) { mpegts_base_deactivate_and_free_program (base, demux->previous_program); demux->previous_program = NULL; } demux->have_group_id = FALSE; demux->group_id = G_MAXUINT; demux->last_seek_offset = -1; demux->program_generation = 0; } static void gst_ts_demux_init (GstTSDemux * demux) { MpegTSBase *base = (MpegTSBase *) demux; base->stream_size = sizeof (TSDemuxStream); base->parse_private_sections = TRUE; /* We are not interested in sections (all handled by mpegtsbase) */ base->push_section = FALSE; demux->flowcombiner = gst_flow_combiner_new (); demux->requested_program_number = -1; demux->program_number = -1; demux->latency = DEFAULT_LATENCY; gst_ts_demux_reset (base); } static void gst_ts_demux_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstTSDemux *demux = GST_TS_DEMUX (object); switch (prop_id) { case PROP_PROGRAM_NUMBER: /* FIXME: do something if program is switched as opposed to set at * beginning */ demux->requested_program_number = g_value_get_int (value); break; case PROP_EMIT_STATS: demux->emit_statistics = g_value_get_boolean (value); break; case PROP_LATENCY: demux->latency = g_value_get_int (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); } } static void gst_ts_demux_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstTSDemux *demux = GST_TS_DEMUX (object); switch (prop_id) { case PROP_PROGRAM_NUMBER: g_value_set_int (value, demux->requested_program_number); break; case PROP_EMIT_STATS: g_value_set_boolean (value, demux->emit_statistics); break; case PROP_LATENCY: g_value_set_int (value, demux->latency); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); } } static gboolean gst_ts_demux_get_duration (GstTSDemux * demux, GstClockTime * dur) { MpegTSBase *base = (MpegTSBase *) demux; gboolean res = FALSE; gint64 val; if (!demux->program) { GST_DEBUG_OBJECT (demux, "No active program yet, can't provide duration"); return FALSE; } /* Get total size in bytes */ if (gst_pad_peer_query_duration (base->sinkpad, GST_FORMAT_BYTES, &val)) { /* Convert it to duration */ *dur = mpegts_packetizer_offset_to_ts (base->packetizer, val, demux->program->pcr_pid); if (GST_CLOCK_TIME_IS_VALID (*dur)) res = TRUE; } return res; } static gboolean gst_ts_demux_srcpad_query (GstPad * pad, GstObject * parent, GstQuery * query) { gboolean res = TRUE; GstFormat format; GstTSDemux *demux; MpegTSBase *base; demux = GST_TS_DEMUX (parent); base = GST_MPEGTS_BASE (demux); switch (GST_QUERY_TYPE (query)) { case GST_QUERY_DURATION: { GST_DEBUG ("query duration"); gst_query_parse_duration (query, &format, NULL); if (format == GST_FORMAT_TIME) { if (!gst_pad_peer_query (base->sinkpad, query)) { GstClockTime dur; if (gst_ts_demux_get_duration (demux, &dur)) gst_query_set_duration (query, GST_FORMAT_TIME, dur); else res = FALSE; } } else { GST_DEBUG_OBJECT (demux, "only query duration on TIME is supported"); res = FALSE; } break; } case GST_QUERY_LATENCY: { GST_DEBUG ("query latency"); res = gst_pad_peer_query (base->sinkpad, query); if (res) { GstClockTime min_lat, max_lat; gboolean live; gint latency; /* According to H.222.0 Annex D.0.3 (System Time Clock recovery in the decoder) and D.0.2 (Audio and video presentation synchronization) We can end up with an interval of up to 700ms between valid PTS/DTS. We therefore allow a latency of 700ms for that. */ latency = demux->latency; if (latency < 0) latency = 700; gst_query_parse_latency (query, &live, &min_lat, &max_lat); min_lat += latency * GST_MSECOND; if (GST_CLOCK_TIME_IS_VALID (max_lat)) max_lat += latency * GST_MSECOND; gst_query_set_latency (query, live, min_lat, max_lat); } break; } case GST_QUERY_SEEKING: { GST_DEBUG ("query seeking"); gst_query_parse_seeking (query, &format, NULL, NULL, NULL); GST_DEBUG ("asked for format %s", gst_format_get_name (format)); if (format == GST_FORMAT_TIME) { gboolean seekable = FALSE; if (gst_pad_peer_query (base->sinkpad, query)) gst_query_parse_seeking (query, NULL, &seekable, NULL, NULL); /* If upstream is not seekable in TIME format we use * our own values here */ if (!seekable) { GstClockTime dur; if (gst_ts_demux_get_duration (demux, &dur)) { gst_query_set_seeking (query, GST_FORMAT_TIME, TRUE, 0, dur); GST_DEBUG ("Gave duration: %" GST_TIME_FORMAT, GST_TIME_ARGS (dur)); } } } else { GST_DEBUG_OBJECT (demux, "only TIME is supported for query seeking"); res = FALSE; } break; } case GST_QUERY_SEGMENT:{ GstFormat format; gint64 start, stop; format = base->out_segment.format; start = gst_segment_to_stream_time (&base->out_segment, format, base->out_segment.start); if ((stop = base->out_segment.stop) == -1) stop = base->out_segment.duration; else stop = gst_segment_to_stream_time (&base->out_segment, format, stop); gst_query_set_segment (query, base->out_segment.rate, format, start, stop); res = TRUE; break; } default: res = gst_pad_query_default (pad, parent, query); } return res; } static void clear_simple_buffer (SimpleBuffer * sbuf) { if (!sbuf->data) return; g_free (sbuf->data); sbuf->size = 0; sbuf->data = NULL; } static gboolean scan_keyframe_h264 (TSDemuxStream * stream, const guint8 * data, const gsize data_size, const gsize max_frame_offset) { gint offset = 0; GstH264NalUnit unit, frame_unit = { 0, }; GstH264ParserResult res = GST_H264_PARSER_OK; TSDemuxH264ParsingInfos *h264infos = &stream->h264infos; GstH264NalParser *parser = h264infos->parser; if (G_UNLIKELY (parser == NULL)) { parser = h264infos->parser = gst_h264_nal_parser_new (); h264infos->sps = gst_byte_writer_new (); h264infos->pps = gst_byte_writer_new (); h264infos->sei = gst_byte_writer_new (); } while (res == GST_H264_PARSER_OK) { res = gst_h264_parser_identify_nalu (parser, data, offset, data_size, &unit); if (res != GST_H264_PARSER_OK && res != GST_H264_PARSER_NO_NAL_END) { GST_INFO_OBJECT (stream->pad, "Error identifying nalu: %i", res); break; } res = gst_h264_parser_parse_nal (parser, &unit); if (res != GST_H264_PARSER_OK) { break; } switch (unit.type) { case GST_H264_NAL_SEI: if (frame_unit.size) break; if (gst_byte_writer_put_data (h264infos->sei, unit.data + unit.sc_offset, unit.size + unit.offset - unit.sc_offset)) { GST_DEBUG ("adding SEI %u", unit.size + unit.offset - unit.sc_offset); } else { GST_WARNING ("Could not write SEI"); } break; case GST_H264_NAL_PPS: if (frame_unit.size) break; if (gst_byte_writer_put_data (h264infos->pps, unit.data + unit.sc_offset, unit.size + unit.offset - unit.sc_offset)) { GST_DEBUG ("adding PPS %u", unit.size + unit.offset - unit.sc_offset); } else { GST_WARNING ("Could not write PPS"); } break; case GST_H264_NAL_SPS: if (frame_unit.size) break; if (gst_byte_writer_put_data (h264infos->sps, unit.data + unit.sc_offset, unit.size + unit.offset - unit.sc_offset)) { GST_DEBUG ("adding SPS %u", unit.size + unit.offset - unit.sc_offset); } else { GST_WARNING ("Could not write SPS"); } break; /* these units are considered keyframes in h264parse */ case GST_H264_NAL_SLICE: case GST_H264_NAL_SLICE_DPA: case GST_H264_NAL_SLICE_DPB: case GST_H264_NAL_SLICE_DPC: case GST_H264_NAL_SLICE_IDR: { GstH264SliceHdr slice; if (h264infos->framedata.size) break; res = gst_h264_parser_parse_slice_hdr (parser, &unit, &slice, FALSE, FALSE); if (GST_H264_IS_I_SLICE (&slice) || GST_H264_IS_SI_SLICE (&slice)) { if (*(unit.data + unit.offset + 1) & 0x80) { /* means first_mb_in_slice == 0 */ /* real frame data */ GST_DEBUG_OBJECT (stream->pad, "Found keyframe at: %u", unit.sc_offset); frame_unit = unit; } } break; } default: break; } if (offset == unit.sc_offset + unit.size) break; offset = unit.sc_offset + unit.size; } /* We've got all the infos we need (SPS / PPS and a keyframe, plus * and possibly SEI units. We can stop rewinding the stream */ if (gst_byte_writer_get_size (h264infos->sps) && gst_byte_writer_get_size (h264infos->pps) && (h264infos->framedata.size || frame_unit.size)) { guint8 *data = NULL; gsize tmpsize = gst_byte_writer_get_size (h264infos->pps); /* We know that the SPS is first so just put all our data in there */ data = gst_byte_writer_reset_and_get_data (h264infos->pps); gst_byte_writer_put_data (h264infos->sps, data, tmpsize); g_free (data); tmpsize = gst_byte_writer_get_size (h264infos->sei); if (tmpsize) { GST_DEBUG ("Adding SEI"); data = gst_byte_writer_reset_and_get_data (h264infos->sei); gst_byte_writer_put_data (h264infos->sps, data, tmpsize); g_free (data); } if (frame_unit.size) { /* We found the everything in one go! */ GST_DEBUG ("Adding Keyframe"); gst_byte_writer_put_data (h264infos->sps, frame_unit.data + frame_unit.sc_offset, stream->current_size - frame_unit.sc_offset); } else { GST_DEBUG ("Adding Keyframe"); gst_byte_writer_put_data (h264infos->sps, h264infos->framedata.data, h264infos->framedata.size); clear_simple_buffer (&h264infos->framedata); } g_free (stream->data); stream->current_size = gst_byte_writer_get_size (h264infos->sps); stream->data = gst_byte_writer_reset_and_get_data (h264infos->sps); gst_byte_writer_init (h264infos->sps); gst_byte_writer_init (h264infos->pps); gst_byte_writer_init (h264infos->sei); return TRUE; } if (frame_unit.size) { GST_DEBUG_OBJECT (stream->pad, "Keep the keyframe as this is the one" " we will push later"); h264infos->framedata.data = g_memdup (frame_unit.data + frame_unit.sc_offset, stream->current_size - frame_unit.sc_offset); h264infos->framedata.size = stream->current_size - frame_unit.sc_offset; } return FALSE; } /* We merge data from TS packets so that the scanning methods get a continuous chunk, however the scanning method will return keyframe offset which needs to be translated back to actual offset in file */ typedef struct { gint64 real_offset; /* offset of TS packet */ gint merged_offset; /* offset of merged data in buffer */ } OffsetInfo; static gboolean gst_ts_demux_adjust_seek_offset_for_keyframe (TSDemuxStream * stream, guint8 * data, guint64 size) { int scan_pid = -1; if (!stream->scan_function) return TRUE; scan_pid = ((MpegTSBaseStream *) stream)->pid; if (scan_pid != -1) { return stream->scan_function (stream, data, size, size); } return TRUE; } static GstFlowReturn gst_ts_demux_do_seek (MpegTSBase * base, GstEvent * event) { GList *tmp; GstTSDemux *demux = (GstTSDemux *) base; GstFlowReturn res = GST_FLOW_ERROR; gdouble rate; GstFormat format; GstSeekFlags flags; GstSeekType start_type, stop_type; gint64 start, stop; guint64 start_offset; gboolean update = FALSE; GstSegment seeksegment; GST_DEBUG ("seek event, %" GST_PTR_FORMAT, event); gst_event_parse_seek (event, &rate, &format, &flags, &start_type, &start, &stop_type, &stop); if (rate <= 0.0) { GST_WARNING ("Negative rate not supported"); goto done; } if (flags & (GST_SEEK_FLAG_SEGMENT)) { GST_WARNING ("seek flags 0x%x are not supported", (int) flags); goto done; } /* configure the segment with the seek variables */ memcpy (&seeksegment, &base->out_segment, sizeof (GstSegment)); GST_LOG_OBJECT (demux, "Before seek, output segment %" GST_SEGMENT_FORMAT, &seeksegment); /* record offset and rate */ demux->rate = rate; if (!gst_segment_do_seek (&seeksegment, rate, format, flags, start_type, start, stop_type, stop, &update)) { GST_DEBUG_OBJECT (demux, "Seek failed in gst_segment_do_seek()"); goto done; } GST_DEBUG_OBJECT (demux, "After seek, update %d output segment now %" GST_SEGMENT_FORMAT, update, &seeksegment); /* If the position actually changed, update == TRUE */ if (update) { GstClockTime target = seeksegment.start; if (target >= SEEK_TIMESTAMP_OFFSET) target -= SEEK_TIMESTAMP_OFFSET; else target = 0; start_offset = mpegts_packetizer_ts_to_offset (base->packetizer, target, demux->program->pcr_pid); if (G_UNLIKELY (start_offset == -1)) { GST_WARNING ("Couldn't convert start position to an offset"); goto done; } base->seek_offset = start_offset; demux->last_seek_offset = base->seek_offset; /* Reset segment if we're not doing an accurate seek */ demux->reset_segment = (!(flags & GST_SEEK_FLAG_ACCURATE)); /* Clear any existing segment - it will be recalculated after streaming recommences */ gst_event_replace (&demux->segment_event, NULL); for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = tmp->data; if (flags & GST_SEEK_FLAG_ACCURATE) stream->needs_keyframe = TRUE; stream->seeked_pts = GST_CLOCK_TIME_NONE; stream->seeked_dts = GST_CLOCK_TIME_NONE; stream->first_pts = GST_CLOCK_TIME_NONE; stream->need_newsegment = TRUE; } } else { /* Position didn't change, just update the output segment based on * our new one */ gst_event_replace (&demux->segment_event, NULL); demux->segment_event = gst_event_new_segment (&seeksegment); if (base->last_seek_seqnum) gst_event_set_seqnum (demux->segment_event, base->last_seek_seqnum); for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = tmp->data; stream->need_newsegment = TRUE; } } /* Commit the new segment */ memcpy (&base->out_segment, &seeksegment, sizeof (GstSegment)); res = GST_FLOW_OK; done: return res; } static gboolean gst_ts_demux_srcpad_event (GstPad * pad, GstObject * parent, GstEvent * event) { gboolean res = TRUE; GstTSDemux *demux = GST_TS_DEMUX (parent); GST_DEBUG_OBJECT (pad, "Got event %s", gst_event_type_get_name (GST_EVENT_TYPE (event))); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEEK: res = mpegts_base_handle_seek_event ((MpegTSBase *) demux, pad, event); if (!res) GST_WARNING ("seeking failed"); gst_event_unref (event); break; default: res = gst_pad_event_default (pad, parent, event); } return res; } static void clean_global_taglist (GstTagList * taglist) { gst_tag_list_remove_tag (taglist, GST_TAG_CONTAINER_FORMAT); gst_tag_list_remove_tag (taglist, GST_TAG_CODEC); } static gboolean push_event (MpegTSBase * base, GstEvent * event) { GstTSDemux *demux = (GstTSDemux *) base; GList *tmp; gboolean early_ret = FALSE; if (GST_EVENT_TYPE (event) == GST_EVENT_SEGMENT) { GST_DEBUG_OBJECT (base, "Ignoring segment event (recreated later)"); gst_event_unref (event); return TRUE; } else if (GST_EVENT_TYPE (event) == GST_EVENT_TAG) { /* In case we receive tags before data, store them to send later * If we already have the program, send it right away */ GstTagList *taglist; gst_event_parse_tag (event, &taglist); if (demux->global_tags == NULL) { demux->global_tags = gst_tag_list_copy (taglist); /* Tags that are stream specific for the container should be considered * global for the container streams */ if (gst_tag_list_get_scope (taglist) == GST_TAG_SCOPE_STREAM) { gst_tag_list_set_scope (demux->global_tags, GST_TAG_SCOPE_GLOBAL); } } else { demux->global_tags = gst_tag_list_make_writable (demux->global_tags); gst_tag_list_insert (demux->global_tags, taglist, GST_TAG_MERGE_REPLACE); } clean_global_taglist (demux->global_tags); /* tags are stored to be used after if there are no streams yet, * so we should never reject */ early_ret = TRUE; } if (G_UNLIKELY (demux->program == NULL)) { gst_event_unref (event); return early_ret; } for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; if (stream->pad) { /* If we are pushing out EOS, flush out pending data first */ if (GST_EVENT_TYPE (event) == GST_EVENT_EOS && gst_pad_is_active (stream->pad)) gst_ts_demux_push_pending_data (demux, stream, NULL); gst_event_ref (event); gst_pad_push_event (stream->pad, event); } } gst_event_unref (event); return TRUE; } static gboolean sink_query (MpegTSBase * base, GstQuery * query) { GstTSDemux *demux = (GstTSDemux *) base; gboolean res = FALSE; switch (GST_QUERY_TYPE (query)) { case GST_QUERY_BITRATE:{ gint64 size_bytes; GstClockTime duration; if (gst_pad_peer_query_duration (base->sinkpad, GST_FORMAT_BYTES, &size_bytes) && size_bytes > 0) { if (gst_ts_demux_get_duration (demux, &duration) && duration > 0 && duration != GST_CLOCK_TIME_NONE) { guint bitrate = gst_util_uint64_scale (8 * size_bytes, GST_SECOND, duration); GST_LOG_OBJECT (demux, "bitrate query byte length: %" G_GINT64_FORMAT " duration %" GST_TIME_FORMAT " resulting in a bitrate of %u", size_bytes, GST_TIME_ARGS (duration), bitrate); gst_query_set_bitrate (query, bitrate); res = TRUE; } } break; } default: res = GST_MPEGTS_BASE_CLASS (parent_class)->sink_query (base, query); break; } return res; } static inline void add_iso639_language_to_tags (TSDemuxStream * stream, gchar * lang_code) { const gchar *lc; GST_LOG ("Add language code for stream: '%s'", lang_code); if (!stream->taglist) stream->taglist = gst_tag_list_new_empty (); /* descriptor contains ISO 639-2 code, we want the ISO 639-1 code */ lc = gst_tag_get_language_code (lang_code); /* Only set tag if we have a valid one */ if (lc || (lang_code[0] && lang_code[1])) gst_tag_list_add (stream->taglist, GST_TAG_MERGE_REPLACE, GST_TAG_LANGUAGE_CODE, (lc) ? lc : lang_code, NULL); } static void gst_ts_demux_create_tags (TSDemuxStream * stream) { MpegTSBaseStream *bstream = (MpegTSBaseStream *) stream; const GstMpegtsDescriptor *desc = NULL; int i, nb; desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_ISO_639_LANGUAGE); if (desc) { gchar *lang_code; nb = gst_mpegts_descriptor_parse_iso_639_language_nb (desc); GST_DEBUG ("Found ISO 639 descriptor (%d entries)", nb); for (i = 0; i < nb; i++) if (gst_mpegts_descriptor_parse_iso_639_language_idx (desc, i, &lang_code, NULL)) { add_iso639_language_to_tags (stream, lang_code); g_free (lang_code); } return; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_SUBTITLING); if (desc) { gchar *lang_code; nb = gst_mpegts_descriptor_parse_dvb_subtitling_nb (desc); GST_DEBUG ("Found SUBTITLING descriptor (%d entries)", nb); for (i = 0; i < nb; i++) if (gst_mpegts_descriptor_parse_dvb_subtitling_idx (desc, i, &lang_code, NULL, NULL, NULL)) { add_iso639_language_to_tags (stream, lang_code); g_free (lang_code); } } } static GstPad * create_pad_for_stream (MpegTSBase * base, MpegTSBaseStream * bstream, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); TSDemuxStream *stream = (TSDemuxStream *) bstream; gchar *name = NULL; GstCaps *caps = NULL; GstPadTemplate *template = NULL; const GstMpegtsDescriptor *desc = NULL; GstPad *pad = NULL; gboolean sparse = FALSE; gboolean is_audio = FALSE, is_video = FALSE, is_subpicture = FALSE, is_private = FALSE; gst_ts_demux_create_tags (stream); GST_LOG ("Attempting to create pad for stream 0x%04x with stream_type %d", bstream->pid, bstream->stream_type); /* First handle BluRay-specific stream types since there is some overlap * between BluRay and non-BluRay streay type identifiers */ if (program->registration_id == DRF_ID_HDMV) { switch (bstream->stream_type) { case ST_BD_AUDIO_AC3: { const GstMpegtsDescriptor *ac3_desc; /* ATSC ac3 audio descriptor */ ac3_desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_AC3_AUDIO_STREAM); if (ac3_desc && DESC_AC_AUDIO_STREAM_bsid (ac3_desc->data) != 16) { GST_LOG ("ac3 audio"); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); } else { is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); } break; } case ST_BD_AUDIO_EAC3: case ST_BD_AUDIO_AC3_PLUS: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; case ST_BD_AUDIO_AC4: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac4"); break; case ST_BD_AUDIO_AC3_TRUE_HD: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-true-hd"); stream->target_pes_substream = 0x72; break; case ST_BD_AUDIO_LPCM: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-private-ts-lpcm"); break; case ST_BD_PGS_SUBPICTURE: is_subpicture = TRUE; caps = gst_caps_new_empty_simple ("subpicture/x-pgs"); sparse = TRUE; break; case ST_BD_AUDIO_DTS_HD: case ST_BD_AUDIO_DTS_HD_MASTER_AUDIO: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-dts"); stream->target_pes_substream = 0x71; break; } } if (caps) goto done; /* Handle non-BluRay stream types */ switch (bstream->stream_type) { case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG1: case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG2: case ST_PS_VIDEO_MPEG2_DCII: /* FIXME : Use DCII registration code (ETV1 ?) to handle that special * Stream type (ST_PS_VIDEO_MPEG2_DCII) */ /* FIXME : Use video descriptor (0x1) to refine caps with: * * frame_rate * * profile_and_level */ GST_LOG ("mpeg video"); is_video = TRUE; caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, bstream->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG1 ? 1 : 2, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MPEGTS_STREAM_TYPE_AUDIO_MPEG1: case GST_MPEGTS_STREAM_TYPE_AUDIO_MPEG2: GST_LOG ("mpeg audio"); is_audio = TRUE; caps = gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 1, NULL); /* HDV is always mpeg 1 audio layer 2 */ if (program->registration_id == DRF_ID_TSHV) gst_caps_set_simple (caps, "layer", G_TYPE_INT, 2, NULL); break; case GST_MPEGTS_STREAM_TYPE_PRIVATE_PES_PACKETS: GST_LOG ("private data"); /* FIXME: Move all of this into a common method (there might be other * types also, depending on registratino descriptors also */ desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_EXTENSION); if (desc != NULL && desc->tag_extension == GST_MTS_DESC_EXT_DVB_AC4) { GST_LOG ("ac4 audio"); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac4"); break; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_AC3); if (desc) { GST_LOG ("ac3 audio"); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); break; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_ENHANCED_AC3); if (desc) { GST_LOG ("ac3 audio"); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_TELETEXT); if (desc) { GST_LOG ("teletext"); is_private = TRUE; caps = gst_caps_new_empty_simple ("application/x-teletext"); sparse = TRUE; break; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_SUBTITLING); if (desc) { GST_LOG ("subtitling"); is_subpicture = TRUE; caps = gst_caps_new_empty_simple ("subpicture/x-dvb"); sparse = TRUE; break; } switch (bstream->registration_id) { case DRF_ID_DTS1: case DRF_ID_DTS2: case DRF_ID_DTS3: /* SMPTE registered DTS */ is_private = TRUE; caps = gst_caps_new_empty_simple ("audio/x-dts"); break; case DRF_ID_S302M: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-smpte-302m"); break; case DRF_ID_OPUS: desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_EXTENSION); if (desc != NULL && desc->tag_extension == 0x80 && desc->length >= 1) { /* User defined (provisional Opus) */ guint8 channel_config_code; GstByteReader br; /* skip tag, length and tag_extension */ gst_byte_reader_init (&br, desc->data + 3, desc->length - 1); channel_config_code = gst_byte_reader_get_uint8_unchecked (&br); if ((channel_config_code & 0x8f) <= 8) { static const guint8 coupled_stream_counts[9] = { 1, 0, 1, 1, 2, 2, 2, 3, 3 }; static const guint8 channel_map_a[8][8] = { {0}, {0, 1}, {0, 2, 1}, {0, 1, 2, 3}, {0, 4, 1, 2, 3}, {0, 4, 1, 2, 3, 5}, {0, 4, 1, 2, 3, 5, 6}, {0, 6, 1, 2, 3, 4, 5, 7}, }; static const guint8 channel_map_b[8][8] = { {0}, {0, 1}, {0, 1, 2}, {0, 1, 2, 3}, {0, 1, 2, 3, 4}, {0, 1, 2, 3, 4, 5}, {0, 1, 2, 3, 4, 5, 6}, {0, 1, 2, 3, 4, 5, 6, 7}, }; gint channels = -1, stream_count, coupled_count, mapping_family; guint8 *channel_mapping = NULL; channels = channel_config_code ? (channel_config_code & 0x0f) : 2; if (channel_config_code == 0 || channel_config_code == 0x80) { /* Dual Mono */ mapping_family = 255; if (channel_config_code == 0) { stream_count = 1; coupled_count = 1; } else { stream_count = 2; coupled_count = 0; } channel_mapping = g_new0 (guint8, channels); memcpy (channel_mapping, &channel_map_a[1], channels); } else if (channel_config_code <= 8) { mapping_family = (channels > 2) ? 1 : 0; stream_count = channel_config_code - coupled_stream_counts[channel_config_code]; coupled_count = coupled_stream_counts[channel_config_code]; if (mapping_family != 0) { channel_mapping = g_new0 (guint8, channels); memcpy (channel_mapping, &channel_map_a[channels - 1], channels); } } else if (channel_config_code >= 0x82 && channel_config_code <= 0x88) { mapping_family = 1; stream_count = channels; coupled_count = 0; channel_mapping = g_new0 (guint8, channels); memcpy (channel_mapping, &channel_map_b[channels - 1], channels); } else if (channel_config_code == 0x81) { if (gst_byte_reader_get_remaining (&br) < 2) { GST_WARNING_OBJECT (demux, "Invalid Opus descriptor with extended channel configuration"); channels = -1; break; } channels = gst_byte_reader_get_uint8_unchecked (&br); mapping_family = gst_byte_reader_get_uint8_unchecked (&br); /* Overwrite values from above */ if (channels == 0) { GST_WARNING_OBJECT (demux, "Invalid Opus descriptor with extended channel configuration"); channels = -1; break; } if (mapping_family == 0 && channels <= 2) { stream_count = channels - coupled_stream_counts[channels]; coupled_count = coupled_stream_counts[channels]; } else { GstBitReader breader; guint8 stream_count_minus_one, coupled_stream_count; gint stream_count_minus_one_len, coupled_stream_count_len; gint channel_mapping_len, i; gst_bit_reader_init (&breader, gst_byte_reader_get_data_unchecked (&br, gst_byte_reader_get_remaining (&br)), gst_byte_reader_get_remaining (&br)); stream_count_minus_one_len = ceil (_gst_log2 (channels)); if (!gst_bit_reader_get_bits_uint8 (&breader, &stream_count_minus_one, stream_count_minus_one_len)) { GST_WARNING_OBJECT (demux, "Invalid Opus descriptor with extended channel configuration"); channels = -1; break; } stream_count = stream_count_minus_one + 1; coupled_stream_count_len = ceil (_gst_log2 (stream_count_minus_one + 2)); if (!gst_bit_reader_get_bits_uint8 (&breader, &coupled_stream_count, coupled_stream_count_len)) { GST_WARNING_OBJECT (demux, "Invalid Opus descriptor with extended channel configuration"); channels = -1; break; } coupled_count = coupled_stream_count; channel_mapping_len = ceil (_gst_log2 (stream_count_minus_one + 1 + coupled_stream_count + 1)); channel_mapping = g_new0 (guint8, channels); for (i = 0; i < channels; i++) { if (!gst_bit_reader_get_bits_uint8 (&breader, &channel_mapping[i], channel_mapping_len)) { GST_WARNING_OBJECT (demux, "Invalid Opus descriptor with extended channel configuration"); break; } } /* error above */ if (i != channels) { channels = -1; g_free (channel_mapping); channel_mapping = NULL; break; } } } else { g_assert_not_reached (); } if (channels != -1) { is_audio = TRUE; caps = gst_codec_utils_opus_create_caps (48000, channels, mapping_family, stream_count, coupled_count, channel_mapping); g_free (channel_mapping); } } else { GST_WARNING_OBJECT (demux, "unexpected channel config code 0x%02x", channel_config_code); } } else { GST_WARNING_OBJECT (demux, "Opus, but no extension descriptor"); } break; case DRF_ID_HEVC: is_video = TRUE; caps = gst_caps_new_simple ("video/x-h265", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); break; case DRF_ID_KLVA: sparse = TRUE; is_private = TRUE; caps = gst_caps_new_simple ("meta/x-klv", "parsed", G_TYPE_BOOLEAN, TRUE, NULL); break; case DRF_ID_AC4: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac4"); break; } if (caps) break; /* hack for itv hd (sid 10510, video pid 3401 */ if (program->program_number == 10510 && bstream->pid == 3401) { is_video = TRUE; caps = gst_caps_new_simple ("video/x-h264", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); } break; case ST_HDV_AUX_V: /* FIXME : Should only be used with specific PMT registration_descriptor */ /* We don't expose those streams since they're only helper streams */ /* template = gst_static_pad_template_get (&private_template); */ /* name = g_strdup_printf ("private_%04x", bstream->pid); */ /* caps = gst_caps_new_simple ("hdv/aux-v", NULL); */ break; case ST_HDV_AUX_A: /* FIXME : Should only be used with specific PMT registration_descriptor */ /* We don't expose those streams since they're only helper streams */ /* template = gst_static_pad_template_get (&private_template); */ /* name = g_strdup_printf ("private_%04x", bstream->pid); */ /* caps = gst_caps_new_simple ("hdv/aux-a", NULL); */ break; case GST_MPEGTS_STREAM_TYPE_AUDIO_AAC_ADTS: is_audio = TRUE; /* prefer mpegversion 4 since it's more commonly supported one */ caps = gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 4, "stream-format", G_TYPE_STRING, "adts", NULL); /* we will set caps later once parsing adts header is done */ stream->atdsInfos.mpegversion = 4; break; case GST_MPEGTS_STREAM_TYPE_AUDIO_AAC_LATM: is_audio = TRUE; caps = gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 4, "stream-format", G_TYPE_STRING, "loas", NULL); break; case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG4: is_video = TRUE; caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 4, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MPEGTS_STREAM_TYPE_VIDEO_H264: is_video = TRUE; caps = gst_caps_new_simple ("video/x-h264", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); break; case GST_MPEGTS_STREAM_TYPE_VIDEO_HEVC: is_video = TRUE; caps = gst_caps_new_simple ("video/x-h265", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); break; case GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K: is_video = TRUE; desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_J2K_VIDEO); if (desc == NULL) { caps = gst_caps_new_empty_simple ("image/x-jpc"); break; } else { GstByteReader br; guint16 DEN_frame_rate = 0; guint16 NUM_frame_rate = 0; guint8 color_specification = 0; guint8 remaining_8b = 0; gboolean interlaced_video = 0; const gchar *interlace_mode = NULL; const gchar *colorspace = NULL; const gchar *colorimetry_mode = NULL; guint16 profile_and_level G_GNUC_UNUSED; guint32 horizontal_size G_GNUC_UNUSED; guint32 vertical_size G_GNUC_UNUSED; guint32 max_bit_rate G_GNUC_UNUSED; guint32 max_buffer_size G_GNUC_UNUSED; const guint desc_min_length = 24; if (desc->length < desc_min_length) { GST_ERROR ("GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K: descriptor length %d too short", desc->length); return NULL; } /* Skip the descriptor tag and length */ gst_byte_reader_init (&br, desc->data + 2, desc->length); profile_and_level = gst_byte_reader_get_uint16_be_unchecked (&br); horizontal_size = gst_byte_reader_get_uint32_be_unchecked (&br); vertical_size = gst_byte_reader_get_uint32_be_unchecked (&br); max_bit_rate = gst_byte_reader_get_uint32_be_unchecked (&br); max_buffer_size = gst_byte_reader_get_uint32_be_unchecked (&br); DEN_frame_rate = gst_byte_reader_get_uint16_be_unchecked (&br); NUM_frame_rate = gst_byte_reader_get_uint16_be_unchecked (&br); color_specification = gst_byte_reader_get_uint8_unchecked (&br); remaining_8b = gst_byte_reader_get_uint8_unchecked (&br); interlaced_video = remaining_8b & 0x40; /* we don't support demuxing interlaced at the moment */ if (interlaced_video) { GST_ERROR ("GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K: interlaced video not supported"); return NULL; } else { interlace_mode = "progressive"; stream->jp2kInfos.interlace = FALSE; } switch (color_specification) { case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_SRGB: colorspace = "sRGB"; colorimetry_mode = GST_VIDEO_COLORIMETRY_SRGB; break; case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_REC601: colorspace = "sYUV"; colorimetry_mode = GST_VIDEO_COLORIMETRY_BT601; break; case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_REC709: case GST_MPEGTSDEMUX_JPEG2000_COLORSPEC_CIELUV: colorspace = "sYUV"; colorimetry_mode = GST_VIDEO_COLORIMETRY_BT709; break; default: break; } caps = gst_caps_new_simple ("image/x-jpc", "framerate", GST_TYPE_FRACTION, NUM_frame_rate, DEN_frame_rate, "interlace-mode", G_TYPE_STRING, interlace_mode, "colorimetry", G_TYPE_STRING, colorimetry_mode, "colorspace", G_TYPE_STRING, colorspace, NULL); } break; case ST_VIDEO_DIRAC: if (bstream->registration_id == 0x64726163) { GST_LOG ("dirac"); /* dirac in hex */ is_video = TRUE; caps = gst_caps_new_empty_simple ("video/x-dirac"); } break; case ST_PRIVATE_EA: /* Try to detect a VC1 stream */ { gboolean is_vc1 = FALSE; /* Note/FIXME: RP-227 specifies that the registration descriptor * for vc1 can also contain other information, such as profile, * level, alignment, buffer_size, .... */ if (bstream->registration_id == DRF_ID_VC1) is_vc1 = TRUE; if (!is_vc1) { GST_WARNING ("0xea private stream type found but no descriptor " "for VC1. Assuming plain VC1."); } is_video = TRUE; caps = gst_caps_new_simple ("video/x-wmv", "wmvversion", G_TYPE_INT, 3, "format", G_TYPE_STRING, "WVC1", NULL); break; } case ST_PS_AUDIO_AC3: /* DVB_ENHANCED_AC3 */ desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_ENHANCED_AC3); if (desc) { is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; } /* If stream has ac3 descriptor * OR program is ATSC (GA94) * OR stream registration is AC-3 * then it's regular AC3 */ if (bstream->registration_id == DRF_ID_AC3 || program->registration_id == DRF_ID_GA94 || mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_AC3)) { is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); break; } GST_WARNING ("AC3 stream type found but no guaranteed " "way found to differentiate between AC3 and EAC3. " "Assuming plain AC3."); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); break; case ST_PS_AUDIO_EAC3: { /* ATSC_ENHANCED_AC3 */ if (bstream->registration_id == DRF_ID_EAC3 || mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_ATSC_EAC3)) { is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; } GST_ELEMENT_WARNING (demux, STREAM, DEMUX, ("Assuming ATSC E-AC3 audio stream."), ("ATSC E-AC3 stream type found but no guarantee way found to " "differentiate among other standards (DVB, ISDB and etc..)")); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; } case ST_PS_AUDIO_LPCM2: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-private2-lpcm"); break; case ST_PS_AUDIO_DTS: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-dts"); break; case ST_PS_AUDIO_LPCM: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-lpcm"); break; case ST_PS_DVD_SUBPICTURE: is_subpicture = TRUE; caps = gst_caps_new_empty_simple ("subpicture/x-dvd"); sparse = TRUE; break; case 0x42: /* hack for Chinese AVS video stream which use 0x42 as stream_id * NOTE: this is unofficial and within the ISO reserved range. */ is_video = TRUE; caps = gst_caps_new_empty_simple ("video/x-cavs"); break; default: GST_DEBUG ("Non-media stream (stream_type:0x%x). Not creating pad", bstream->stream_type); break; } done: if (caps) { if (is_audio) { template = gst_static_pad_template_get (&audio_template); name = g_strdup_printf ("audio_%01x_%04x", demux->program_generation, bstream->pid); gst_stream_set_stream_type (bstream->stream_object, GST_STREAM_TYPE_AUDIO); } else if (is_video) { template = gst_static_pad_template_get (&video_template); name = g_strdup_printf ("video_%01x_%04x", demux->program_generation, bstream->pid); gst_stream_set_stream_type (bstream->stream_object, GST_STREAM_TYPE_VIDEO); } else if (is_private) { template = gst_static_pad_template_get (&private_template); name = g_strdup_printf ("private_%01x_%04x", demux->program_generation, bstream->pid); } else if (is_subpicture) { template = gst_static_pad_template_get (&subpicture_template); name = g_strdup_printf ("subpicture_%01x_%04x", demux->program_generation, bstream->pid); gst_stream_set_stream_type (bstream->stream_object, GST_STREAM_TYPE_TEXT); } else g_assert_not_reached (); } if (template && name && caps) { GstEvent *event; const gchar *stream_id; GST_LOG ("stream:%p creating pad with name %s and caps %" GST_PTR_FORMAT, stream, name, caps); pad = gst_pad_new_from_template (template, name); gst_pad_set_active (pad, TRUE); gst_pad_use_fixed_caps (pad); stream_id = gst_stream_get_stream_id (bstream->stream_object); event = gst_pad_get_sticky_event (base->sinkpad, GST_EVENT_STREAM_START, 0); if (event) { if (gst_event_parse_group_id (event, &demux->group_id)) demux->have_group_id = TRUE; else demux->have_group_id = FALSE; gst_event_unref (event); } else if (!demux->have_group_id) { demux->have_group_id = TRUE; demux->group_id = gst_util_group_id_next (); } event = gst_event_new_stream_start (stream_id); gst_event_set_stream (event, bstream->stream_object); if (demux->have_group_id) gst_event_set_group_id (event, demux->group_id); if (sparse) { gst_event_set_stream_flags (event, GST_STREAM_FLAG_SPARSE); gst_stream_set_stream_flags (bstream->stream_object, GST_STREAM_FLAG_SPARSE); } stream->sparse = sparse; gst_stream_set_caps (bstream->stream_object, caps); if (!stream->taglist) stream->taglist = gst_tag_list_new_empty (); gst_pb_utils_add_codec_description_to_tag_list (stream->taglist, NULL, caps); gst_stream_set_tags (bstream->stream_object, stream->taglist); gst_pad_push_event (pad, event); gst_pad_set_caps (pad, caps); gst_pad_set_query_function (pad, gst_ts_demux_srcpad_query); gst_pad_set_event_function (pad, gst_ts_demux_srcpad_event); } g_free (name); if (template) gst_object_unref (template); if (caps) gst_caps_unref (caps); return pad; } static gboolean gst_ts_demux_stream_added (MpegTSBase * base, MpegTSBaseStream * bstream, MpegTSBaseProgram * program) { GstTSDemux *demux = (GstTSDemux *) base; TSDemuxStream *stream = (TSDemuxStream *) bstream; if (!stream->pad) { /* Create the pad */ if (bstream->stream_type != 0xff) { stream->pad = create_pad_for_stream (base, bstream, program); if (stream->pad) gst_flow_combiner_add_pad (demux->flowcombiner, stream->pad); } if (base->mode != BASE_MODE_PUSHING && bstream->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_H264) { stream->scan_function = (GstTsDemuxKeyFrameScanFunction) scan_keyframe_h264; } else { stream->scan_function = NULL; } stream->active = FALSE; stream->need_newsegment = TRUE; /* Reset segment if we're not doing an accurate seek */ demux->reset_segment = (!(base->out_segment.flags & GST_SEEK_FLAG_ACCURATE)); stream->needs_keyframe = FALSE; stream->discont = TRUE; stream->pts = GST_CLOCK_TIME_NONE; stream->dts = GST_CLOCK_TIME_NONE; stream->first_pts = GST_CLOCK_TIME_NONE; stream->raw_pts = -1; stream->raw_dts = -1; stream->pending_ts = TRUE; stream->nb_out_buffers = 0; stream->gap_ref_buffers = 0; stream->gap_ref_pts = GST_CLOCK_TIME_NONE; /* Only wait for a valid timestamp if we have a PCR_PID */ stream->pending_ts = program->pcr_pid < 0x1fff; stream->continuity_counter = CONTINUITY_UNSET; } return (stream->pad != NULL); } static void tsdemux_h264_parsing_info_clear (TSDemuxH264ParsingInfos * h264infos) { clear_simple_buffer (&h264infos->framedata); if (h264infos->parser) { gst_h264_nal_parser_free (h264infos->parser); gst_byte_writer_free (h264infos->sps); gst_byte_writer_free (h264infos->pps); gst_byte_writer_free (h264infos->sei); } } static void gst_ts_demux_stream_removed (MpegTSBase * base, MpegTSBaseStream * bstream) { TSDemuxStream *stream = (TSDemuxStream *) bstream; if (stream->pad) { gst_flow_combiner_remove_pad (GST_TS_DEMUX_CAST (base)->flowcombiner, stream->pad); if (stream->active) { if (gst_pad_is_active (stream->pad)) { /* Flush out all data */ GST_DEBUG_OBJECT (stream->pad, "Flushing out pending data"); gst_ts_demux_push_pending_data ((GstTSDemux *) base, stream, NULL); GST_DEBUG_OBJECT (stream->pad, "Pushing out EOS"); gst_pad_push_event (stream->pad, gst_event_new_eos ()); gst_pad_set_active (stream->pad, FALSE); } GST_DEBUG_OBJECT (stream->pad, "Removing pad"); gst_element_remove_pad (GST_ELEMENT_CAST (base), stream->pad); stream->active = FALSE; } else { gst_object_unref (stream->pad); } stream->pad = NULL; } gst_ts_demux_stream_flush (stream, GST_TS_DEMUX_CAST (base), TRUE); if (stream->taglist != NULL) { gst_tag_list_unref (stream->taglist); stream->taglist = NULL; } tsdemux_h264_parsing_info_clear (&stream->h264infos); } static void activate_pad_for_stream (GstTSDemux * tsdemux, TSDemuxStream * stream) { if (stream->pad) { GST_DEBUG_OBJECT (tsdemux, "Activating pad %s:%s for stream %p", GST_DEBUG_PAD_NAME (stream->pad), stream); gst_element_add_pad ((GstElement *) tsdemux, stream->pad); stream->active = TRUE; GST_DEBUG_OBJECT (stream->pad, "done adding pad"); } else if (((MpegTSBaseStream *) stream)->stream_type != 0xff) { GST_DEBUG_OBJECT (tsdemux, "stream %p (pid 0x%04x, type:0x%02x) has no pad", stream, ((MpegTSBaseStream *) stream)->pid, ((MpegTSBaseStream *) stream)->stream_type); } } static void gst_ts_demux_stream_flush (TSDemuxStream * stream, GstTSDemux * tsdemux, gboolean hard) { GST_DEBUG ("flushing stream %p", stream); g_free (stream->data); stream->data = NULL; stream->state = PENDING_PACKET_EMPTY; stream->expected_size = 0; stream->allocated_size = 0; stream->current_size = 0; stream->discont = TRUE; stream->pts = GST_CLOCK_TIME_NONE; stream->dts = GST_CLOCK_TIME_NONE; stream->raw_pts = -1; stream->raw_dts = -1; stream->pending_ts = TRUE; stream->nb_out_buffers = 0; stream->gap_ref_buffers = 0; stream->gap_ref_pts = GST_CLOCK_TIME_NONE; stream->continuity_counter = CONTINUITY_UNSET; if (G_UNLIKELY (stream->pending)) { GList *tmp; GST_DEBUG ("clearing pending %p", stream); for (tmp = stream->pending; tmp; tmp = tmp->next) { PendingBuffer *pend = (PendingBuffer *) tmp->data; gst_buffer_unref (pend->buffer); g_slice_free (PendingBuffer, pend); } g_list_free (stream->pending); stream->pending = NULL; } if (hard) { stream->first_pts = GST_CLOCK_TIME_NONE; stream->need_newsegment = TRUE; } } static void gst_ts_demux_flush_streams (GstTSDemux * demux, gboolean hard) { GList *walk; if (!demux->program) return; for (walk = demux->program->stream_list; walk; walk = g_list_next (walk)) gst_ts_demux_stream_flush (walk->data, demux, hard); } static gboolean gst_ts_demux_can_remove_program (MpegTSBase * base, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); /* If it's our current active program, we return FALSE, we'll deactivate it * ourselves when the next program gets activated */ if (demux->program == program) { GST_DEBUG ("Attempting to remove current program, delaying until new program gets activated"); demux->previous_program = program; demux->program_number = -1; return FALSE; } return TRUE; } static void gst_ts_demux_update_program (MpegTSBase * base, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); GList *tmp; GST_DEBUG ("Updating program %d", program->program_number); /* Emit collection message */ gst_element_post_message ((GstElement *) base, gst_message_new_stream_collection ((GstObject *) base, program->collection)); /* Add all streams, then fire no-more-pads */ for (tmp = program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; if (!stream->pad) { activate_pad_for_stream (demux, stream); if (stream->sparse) { /* force sending of pending sticky events which have been stored on the * pad already and which otherwise would only be sent on the first buffer * or serialized event (which means very late in case of subtitle streams), * and playsink waits for stream-start or another serialized event */ GST_DEBUG_OBJECT (stream->pad, "sparse stream, pushing GAP event"); gst_pad_push_event (stream->pad, gst_event_new_gap (0, 0)); } } } } static void gst_ts_demux_program_started (MpegTSBase * base, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); GST_DEBUG ("Current program %d, new program %d requested program %d", (gint) demux->program_number, program->program_number, demux->requested_program_number); if (demux->requested_program_number == program->program_number || (demux->requested_program_number == -1 && demux->program_number == -1)) { GList *tmp; gboolean have_pads = FALSE; GST_LOG ("program %d started", program->program_number); demux->program_number = program->program_number; demux->program = program; /* Increment the program_generation counter */ demux->program_generation = (demux->program_generation + 1) & 0xf; /* Emit collection message */ gst_element_post_message ((GstElement *) base, gst_message_new_stream_collection ((GstObject *) base, program->collection)); /* If this is not the initial program, we need to calculate * a new segment */ if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } /* DRAIN ALL STREAMS FIRST ! */ if (demux->previous_program) { GList *tmp; GST_DEBUG_OBJECT (demux, "Draining previous program"); for (tmp = demux->previous_program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; if (stream->pad) gst_ts_demux_push_pending_data (demux, stream, demux->previous_program); } } /* Add all streams, then fire no-more-pads */ for (tmp = program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; activate_pad_for_stream (demux, stream); if (stream->pad) have_pads = TRUE; } /* If there was a previous program, now is the time to deactivate it * and remove old pads (including pushing EOS) */ if (demux->previous_program) { GST_DEBUG ("Deactivating previous program"); mpegts_base_deactivate_and_free_program (base, demux->previous_program); demux->previous_program = NULL; } if (!have_pads) { /* If we had no pads, this stream is likely corrupted or unsupported and * there's not much we can do at this point */ GST_ELEMENT_ERROR (demux, STREAM, WRONG_TYPE, ("This stream contains no valid or supported streams."), ("activating program but got no pads")); return; } /* If any of the stream is sparse, push a GAP event before anything else * This is done here, and not in activate_pad_for_stream() because pushing * a GAP event *is* considering data, and we want to ensure the (potential) * old pads are all removed before we push any data on the new ones */ for (tmp = program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; if (stream->sparse) { /* force sending of pending sticky events which have been stored on the * pad already and which otherwise would only be sent on the first buffer * or serialized event (which means very late in case of subtitle streams), * and playsink waits for stream-start or another serialized event */ GST_DEBUG_OBJECT (stream->pad, "sparse stream, pushing GAP event"); gst_pad_push_event (stream->pad, gst_event_new_gap (0, 0)); } } gst_element_no_more_pads ((GstElement *) demux); } } static void gst_ts_demux_program_stopped (MpegTSBase * base, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); if (demux->program == program) { demux->program = NULL; demux->program_number = -1; } } static inline void gst_ts_demux_record_pts (GstTSDemux * demux, TSDemuxStream * stream, guint64 pts, guint64 offset) { MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; stream->raw_pts = pts; if (pts == -1) { stream->pts = GST_CLOCK_TIME_NONE; return; } GST_LOG ("pid 0x%04x raw pts:%" G_GUINT64_FORMAT " at offset %" G_GUINT64_FORMAT, bs->pid, pts, offset); /* Compute PTS in GstClockTime */ stream->pts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (pts), demux->program->pcr_pid); GST_LOG ("pid 0x%04x Stored PTS %" G_GUINT64_FORMAT, bs->pid, stream->pts); if (G_UNLIKELY (demux->emit_statistics)) { GstStructure *st; st = gst_structure_new_id_empty (QUARK_TSDEMUX); gst_structure_id_set (st, QUARK_PID, G_TYPE_UINT, bs->pid, QUARK_OFFSET, G_TYPE_UINT64, offset, QUARK_PTS, G_TYPE_UINT64, pts, NULL); gst_element_post_message (GST_ELEMENT_CAST (demux), gst_message_new_element (GST_OBJECT (demux), st)); } } static inline void gst_ts_demux_record_dts (GstTSDemux * demux, TSDemuxStream * stream, guint64 dts, guint64 offset) { MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; stream->raw_dts = dts; if (dts == -1) { stream->dts = GST_CLOCK_TIME_NONE; return; } GST_LOG ("pid 0x%04x raw dts:%" G_GUINT64_FORMAT " at offset %" G_GUINT64_FORMAT, bs->pid, dts, offset); /* Compute DTS in GstClockTime */ stream->dts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (dts), demux->program->pcr_pid); GST_LOG ("pid 0x%04x Stored DTS %" G_GUINT64_FORMAT, bs->pid, stream->dts); if (G_UNLIKELY (demux->emit_statistics)) { GstStructure *st; st = gst_structure_new_id_empty (QUARK_TSDEMUX); gst_structure_id_set (st, QUARK_PID, G_TYPE_UINT, bs->pid, QUARK_OFFSET, G_TYPE_UINT64, offset, QUARK_DTS, G_TYPE_UINT64, dts, NULL); gst_element_post_message (GST_ELEMENT_CAST (demux), gst_message_new_element (GST_OBJECT (demux), st)); } } /* This is called when we haven't got a valid initial PTS/DTS on all streams */ static gboolean check_pending_buffers (GstTSDemux * demux) { gboolean have_observation = FALSE; /* The biggest offset */ guint64 offset = 0; GList *tmp; gboolean have_only_sparse = TRUE; /* 0. Do we only have sparse stream */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; if (!tmpstream->sparse) { have_only_sparse = FALSE; break; } } /* 1. Go over all streams */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; /* 1.1 check if at least one stream got a valid DTS */ if (have_only_sparse || !tmpstream->sparse) { if ((tmpstream->raw_dts != -1 && tmpstream->dts != GST_CLOCK_TIME_NONE) || (tmpstream->raw_pts != -1 && tmpstream->pts != GST_CLOCK_TIME_NONE)) { have_observation = TRUE; break; } } } /* 2. If we don't have a valid value yet, break out */ if (have_observation == FALSE) return FALSE; /* 3. Go over all streams that have current/pending data */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; PendingBuffer *pend; guint64 firstval, lastval, ts; /* 3.1 Calculate the offset between current DTS and first DTS */ if (tmpstream->pending == NULL || tmpstream->state == PENDING_PACKET_EMPTY) continue; /* If we don't have any pending data, the offset is 0 for this stream */ if (tmpstream->pending == NULL) break; if (tmpstream->raw_dts != -1) lastval = tmpstream->raw_dts; else if (tmpstream->raw_pts != -1) lastval = tmpstream->raw_pts; else { GST_WARNING ("Don't have a last DTS/PTS to use for offset recalculation"); continue; } pend = tmpstream->pending->data; if (pend->dts != -1) firstval = pend->dts; else if (pend->pts != -1) firstval = pend->pts; else { GST_WARNING ("Don't have a first DTS/PTS to use for offset recalculation"); continue; } /* 3.2 Add to the offset the report TS for the current DTS */ ts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (lastval), demux->program->pcr_pid); if (ts == GST_CLOCK_TIME_NONE) { GST_WARNING ("THIS SHOULD NOT HAPPEN !"); continue; } ts += MPEGTIME_TO_GSTTIME (lastval - firstval); /* 3.3 If that offset is bigger than the current offset, store it */ if (ts > offset) offset = ts; } GST_DEBUG ("New initial pcr_offset %" GST_TIME_FORMAT, GST_TIME_ARGS (offset)); /* 4. Set the offset on the packetizer */ mpegts_packetizer_set_current_pcr_offset (MPEG_TS_BASE_PACKETIZER (demux), offset, demux->program->pcr_pid); /* 4. Go over all streams */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; stream->pending_ts = FALSE; /* 4.1 Set pending_ts for FALSE */ /* 4.2 Recalculate PTS/DTS (in running time) for pending data */ if (stream->pending) { GList *tmp2; for (tmp2 = stream->pending; tmp2; tmp2 = tmp2->next) { PendingBuffer *pend = (PendingBuffer *) tmp2->data; if (pend->pts != -1) GST_BUFFER_PTS (pend->buffer) = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (pend->pts), demux->program->pcr_pid); if (pend->dts != -1) GST_BUFFER_DTS (pend->buffer) = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (pend->dts), demux->program->pcr_pid); /* 4.2.2 Set first_pts to TS of lowest PTS (for segment) */ if (stream->first_pts == GST_CLOCK_TIME_NONE) { if (GST_BUFFER_PTS (pend->buffer) != GST_CLOCK_TIME_NONE) stream->first_pts = GST_BUFFER_PTS (pend->buffer); else if (GST_BUFFER_DTS (pend->buffer) != GST_CLOCK_TIME_NONE) stream->first_pts = GST_BUFFER_DTS (pend->buffer); } } } /* Recalculate PTS/DTS (in running time) for current data */ if (stream->state != PENDING_PACKET_EMPTY) { if (stream->raw_pts != -1) { stream->pts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (stream->raw_pts), demux->program->pcr_pid); if (stream->first_pts == GST_CLOCK_TIME_NONE) stream->first_pts = stream->pts; } if (stream->raw_dts != -1) { stream->dts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (stream->raw_dts), demux->program->pcr_pid); if (stream->first_pts == GST_CLOCK_TIME_NONE) stream->first_pts = stream->dts; } } } return TRUE; } static void gst_ts_demux_parse_pes_header (GstTSDemux * demux, TSDemuxStream * stream, guint8 * data, guint32 length, guint64 bufferoffset) { PESHeader header; PESParsingResult parseres; GST_MEMDUMP ("Header buffer", data, MIN (length, 32)); parseres = mpegts_parse_pes_header (data, length, &header); if (G_UNLIKELY (parseres == PES_PARSING_NEED_MORE)) goto discont; if (G_UNLIKELY (parseres == PES_PARSING_BAD)) { GST_WARNING ("Error parsing PES header. pid: 0x%x stream_type: 0x%x", stream->stream.pid, stream->stream.stream_type); goto discont; } if (stream->target_pes_substream != 0 && header.stream_id_extension != stream->target_pes_substream) { GST_DEBUG ("Skipping unwanted substream"); goto discont; } gst_ts_demux_record_dts (demux, stream, header.DTS, bufferoffset); gst_ts_demux_record_pts (demux, stream, header.PTS, bufferoffset); if (G_UNLIKELY (stream->pending_ts && (stream->pts != GST_CLOCK_TIME_NONE || stream->dts != GST_CLOCK_TIME_NONE))) { GST_DEBUG ("Got pts/dts update, rechecking all streams"); check_pending_buffers (demux); } else if (stream->first_pts == GST_CLOCK_TIME_NONE) { if (GST_CLOCK_TIME_IS_VALID (stream->pts)) stream->first_pts = stream->pts; else if (GST_CLOCK_TIME_IS_VALID (stream->dts)) stream->first_pts = stream->dts; } GST_DEBUG_OBJECT (demux, "stream PTS %" GST_TIME_FORMAT " DTS %" GST_TIME_FORMAT, GST_TIME_ARGS (stream->pts), GST_TIME_ARGS (stream->dts)); /* Remove PES headers */ GST_DEBUG ("Moving data forward by %d bytes (packet_size:%d, have:%d)", header.header_size, header.packet_length, length); stream->expected_size = header.packet_length; if (stream->expected_size) { if (G_LIKELY (stream->expected_size > header.header_size)) { stream->expected_size -= header.header_size; } else { /* next packet will have to complete this one */ GST_WARNING ("invalid header and packet size combination, empty packet"); stream->expected_size = 0; } } data += header.header_size; length -= header.header_size; /* Create the output buffer */ if (stream->expected_size) stream->allocated_size = MAX (stream->expected_size, length); else stream->allocated_size = MAX (8192, length); g_assert (stream->data == NULL); stream->data = g_malloc (stream->allocated_size); memcpy (stream->data, data, length); stream->current_size = length; stream->state = PENDING_PACKET_BUFFER; return; discont: stream->state = PENDING_PACKET_DISCONT; return; } /* ONLY CALL THIS: * * WITH packet->payload != NULL * * WITH pending/current flushed out if beginning of new PES packet */ static inline void gst_ts_demux_queue_data (GstTSDemux * demux, TSDemuxStream * stream, MpegTSPacketizerPacket * packet) { guint8 *data; guint size; guint8 cc = FLAGS_CONTINUITY_COUNTER (packet->scram_afc_cc); GST_LOG ("pid: 0x%04x state:%d", stream->stream.pid, stream->state); size = packet->data_end - packet->payload; data = packet->payload; if (stream->continuity_counter == CONTINUITY_UNSET) { GST_DEBUG ("CONTINUITY: Initialize to %d", cc); } else if ((cc == stream->continuity_counter + 1 || (stream->continuity_counter == MAX_CONTINUITY && cc == 0))) { GST_LOG ("CONTINUITY: Got expected %d", cc); } else { if (stream->state != PENDING_PACKET_EMPTY) { if (packet->payload_unit_start_indicator) { /* A mismatch is fatal, except if this is the beginning of a new * frame (from which we can recover) */ if (G_UNLIKELY (stream->data)) { g_free (stream->data); stream->data = NULL; } stream->state = PENDING_PACKET_HEADER; } else { GST_WARNING ("CONTINUITY: Mismatch packet %d, stream %d", cc, stream->continuity_counter); stream->state = PENDING_PACKET_DISCONT; } } } stream->continuity_counter = cc; if (stream->state == PENDING_PACKET_EMPTY) { if (G_UNLIKELY (!packet->payload_unit_start_indicator)) { stream->state = PENDING_PACKET_DISCONT; GST_DEBUG ("Didn't get the first packet of this PES"); } else { GST_LOG ("EMPTY=>HEADER"); stream->state = PENDING_PACKET_HEADER; } } switch (stream->state) { case PENDING_PACKET_HEADER: { GST_LOG ("HEADER: Parsing PES header"); /* parse the header */ gst_ts_demux_parse_pes_header (demux, stream, data, size, packet->offset); break; } case PENDING_PACKET_BUFFER: { GST_LOG ("BUFFER: appending data"); if (G_UNLIKELY (stream->current_size + size > stream->allocated_size)) { GST_LOG ("resizing buffer"); do { stream->allocated_size = MAX (8192, 2 * stream->allocated_size); } while (stream->current_size + size > stream->allocated_size); stream->data = g_realloc (stream->data, stream->allocated_size); } memcpy (stream->data + stream->current_size, data, size); stream->current_size += size; break; } case PENDING_PACKET_DISCONT: { GST_LOG ("DISCONT: not storing/pushing"); if (G_UNLIKELY (stream->data)) { g_free (stream->data); stream->data = NULL; } stream->continuity_counter = CONTINUITY_UNSET; break; } default: break; } return; } static void calculate_and_push_newsegment (GstTSDemux * demux, TSDemuxStream * stream, MpegTSBaseProgram * target_program) { MpegTSBase *base = (MpegTSBase *) demux; GstClockTime lowest_pts = GST_CLOCK_TIME_NONE; GstClockTime firstts = 0; GList *tmp; GST_DEBUG ("Creating new newsegment for stream %p", stream); if (target_program == NULL) target_program = demux->program; /* Speedup : if we don't need to calculate anything, go straight to pushing */ if (demux->segment_event) goto push_new_segment; /* Calculate the 'new_start' value, used for newsegment */ for (tmp = target_program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *pstream = (TSDemuxStream *) tmp->data; if (GST_CLOCK_TIME_IS_VALID (pstream->first_pts)) { if (!GST_CLOCK_TIME_IS_VALID (lowest_pts) || pstream->first_pts < lowest_pts) lowest_pts = pstream->first_pts; } } if (GST_CLOCK_TIME_IS_VALID (lowest_pts)) firstts = lowest_pts; GST_DEBUG ("lowest_pts %" G_GUINT64_FORMAT " => clocktime %" GST_TIME_FORMAT, lowest_pts, GST_TIME_ARGS (firstts)); if (base->out_segment.format != GST_FORMAT_TIME || demux->reset_segment) { /* It will happen only if it's first program or after flushes. */ GST_DEBUG ("Calculating actual segment"); if (base->segment.format == GST_FORMAT_TIME) { /* Try to recover segment info from base if it's in TIME format */ base->out_segment = base->segment; } else { /* Start from the first ts/pts */ GstSegment *seg = &base->out_segment; GstClockTime base = seg->base + seg->position - (seg->start + seg->offset); GstClockTime stop = seg->stop; gst_segment_init (seg, GST_FORMAT_TIME); seg->start = firstts; seg->stop = MAX (seg->start, stop); seg->position = firstts; seg->time = firstts; seg->rate = demux->rate; seg->base = base; } } else if (base->out_segment.start < firstts) { /* Take into account the offset to the first buffer timestamp */ if (base->out_segment.rate > 0) { base->out_segment.start = firstts; if (GST_CLOCK_TIME_IS_VALID (base->out_segment.stop)) base->out_segment.stop += firstts - base->out_segment.start; base->out_segment.position = firstts; } } GST_LOG_OBJECT (demux, "Output segment now %" GST_SEGMENT_FORMAT, &base->out_segment); if (!demux->segment_event) { demux->segment_event = gst_event_new_segment (&base->out_segment); if (base->last_seek_seqnum != GST_SEQNUM_INVALID) gst_event_set_seqnum (demux->segment_event, base->last_seek_seqnum); } push_new_segment: for (tmp = target_program->stream_list; tmp; tmp = tmp->next) { stream = (TSDemuxStream *) tmp->data; if (stream->pad == NULL) continue; if (demux->segment_event) { GST_DEBUG_OBJECT (stream->pad, "Pushing newsegment event"); gst_event_ref (demux->segment_event); gst_pad_push_event (stream->pad, demux->segment_event); } if (demux->global_tags) { gst_pad_push_event (stream->pad, gst_event_new_tag (gst_tag_list_ref (demux->global_tags))); } /* Push pending tags */ if (stream->taglist) { GST_DEBUG_OBJECT (stream->pad, "Sending tags %" GST_PTR_FORMAT, stream->taglist); gst_pad_push_event (stream->pad, gst_event_new_tag (stream->taglist)); stream->taglist = NULL; } stream->need_newsegment = FALSE; } } static void gst_ts_demux_check_and_sync_streams (GstTSDemux * demux, GstClockTime time) { GList *tmp; GST_DEBUG_OBJECT (demux, "Recheck streams and sync to at least: %" GST_TIME_FORMAT, GST_TIME_ARGS (time)); if (G_UNLIKELY (demux->program == NULL)) return; /* Go over each stream and update it to at least 'time' time. * For each stream, the pad stores the buffer counter the last time * a gap check occurred (gap_ref_buffers) and a gap_ref_pts timestamp * that is either the PTS from the stream or the PCR the pad was updated * to. * * We can check nb_out_buffers to see if any buffers were pushed since then. * This means we can detect buffers passing without PTSes fine and still generate * gaps. * * If there haven't been any buffers pushed on this stream since the last * gap check, push a gap event updating to the indicated input PCR time * and update the pad's tracking. * * If there have been buffers pushed, update the reference buffer count * and but don't push a gap event */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *ps = (TSDemuxStream *) tmp->data; GST_DEBUG_OBJECT (ps->pad, "0x%04x, PTS:%" GST_TIME_FORMAT " REFPTS:%" GST_TIME_FORMAT " Gap:%" GST_TIME_FORMAT " nb_buffers: %d (ref:%d)", ((MpegTSBaseStream *) ps)->pid, GST_TIME_ARGS (ps->pts), GST_TIME_ARGS (ps->gap_ref_pts), GST_TIME_ARGS (ps->pts - ps->gap_ref_pts), ps->nb_out_buffers, ps->gap_ref_buffers); if (ps->pad == NULL) continue; if (ps->nb_out_buffers == ps->gap_ref_buffers && ps->gap_ref_pts != ps->pts) { /* Do initial setup of pad if needed - segment etc */ GST_DEBUG_OBJECT (ps->pad, "Stream needs update. Pushing GAP event to TS %" GST_TIME_FORMAT, GST_TIME_ARGS (time)); if (G_UNLIKELY (ps->need_newsegment)) calculate_and_push_newsegment (demux, ps, NULL); /* Now send gap event */ gst_pad_push_event (ps->pad, gst_event_new_gap (time, 0)); } /* Update GAP tracking vars so we don't re-check this stream for a while */ ps->gap_ref_pts = time; if (ps->pts != GST_CLOCK_TIME_NONE && ps->pts > time) ps->gap_ref_pts = ps->pts; ps->gap_ref_buffers = ps->nb_out_buffers; } } static GstBufferList * parse_opus_access_unit (TSDemuxStream * stream) { GstByteReader reader; GstBufferList *buffer_list = NULL; buffer_list = gst_buffer_list_new (); gst_byte_reader_init (&reader, stream->data, stream->current_size); do { GstBuffer *buffer; guint16 id; guint au_size = 0; guint8 b; gboolean start_trim_flag, end_trim_flag, control_extension_flag; guint16 start_trim = 0, end_trim = 0; guint8 *packet_data; guint packet_size; if (!gst_byte_reader_get_uint16_be (&reader, &id)) goto error; /* No control header */ if ((id >> 5) != 0x3ff) goto error; do { if (!gst_byte_reader_get_uint8 (&reader, &b)) goto error; au_size += b; } while (b == 0xff); start_trim_flag = (id >> 4) & 0x1; end_trim_flag = (id >> 3) & 0x1; control_extension_flag = (id >> 2) & 0x1; if (start_trim_flag) { if (!gst_byte_reader_get_uint16_be (&reader, &start_trim)) goto error; } if (end_trim_flag) { if (!gst_byte_reader_get_uint16_be (&reader, &end_trim)) goto error; } if (control_extension_flag) { if (!gst_byte_reader_get_uint8 (&reader, &b)) goto error; if (!gst_byte_reader_skip (&reader, b)) goto error; } packet_size = au_size; /* FIXME: this should be * packet_size = au_size - gst_byte_reader_get_pos (&reader); * but ffmpeg and the only available sample stream from obe.tv * are not including the control header size in au_size */ if (gst_byte_reader_get_remaining (&reader) < packet_size) goto error; if (!gst_byte_reader_dup_data (&reader, packet_size, &packet_data)) goto error; buffer = gst_buffer_new_wrapped (packet_data, packet_size); if (start_trim != 0 || end_trim != 0) { gst_buffer_add_audio_clipping_meta (buffer, GST_FORMAT_DEFAULT, start_trim, end_trim); } gst_buffer_list_add (buffer_list, buffer); } while (gst_byte_reader_get_remaining (&reader) > 0); g_free (stream->data); stream->data = NULL; stream->current_size = 0; return buffer_list; error: { GST_ERROR ("Failed to parse Opus access unit"); g_free (stream->data); stream->data = NULL; stream->current_size = 0; if (buffer_list) gst_buffer_list_unref (buffer_list); return NULL; } } /* interlaced mode is disabled at the moment */ /*#define TSDEMUX_JP2K_SUPPORT_INTERLACE */ static GstBuffer * parse_jp2k_access_unit (TSDemuxStream * stream) { GstByteReader reader; /* header tag */ guint32 header_tag; /* Framerate box */ guint16 den G_GNUC_UNUSED; guint16 num G_GNUC_UNUSED; /* Maximum bitrate box */ guint32 MaxBr G_GNUC_UNUSED; guint32 AUF[2] = { 0, 0 }; #ifdef TSDEMUX_JP2K_SUPPORT_INTERLACE /* Field Coding Box */ guint8 Fic G_GNUC_UNUSED = 1; guint8 Fio G_GNUC_UNUSED = 0; /* header size equals 38 for non-interlaced, and 48 for interlaced */ guint header_size = stream->jp2kInfos.interlace ? 48 : 38; #else /* header size equals 38 for non-interlaced, and 48 for interlaced */ guint header_size = 38; #endif /* Time Code box */ guint32 HHMMSSFF G_GNUC_UNUSED; /* Broadcast color box */ guint8 CollC G_GNUC_UNUSED; guint8 b G_GNUC_UNUSED; guint data_location; GstBuffer *retbuf = NULL; if (stream->current_size < header_size) { GST_ERROR_OBJECT (stream->pad, "Not enough data for header"); goto error; } gst_byte_reader_init (&reader, stream->data, stream->current_size); /* Check for the location of the jp2k magic */ data_location = gst_byte_reader_masked_scan_uint32 (&reader, 0xffffffff, 0xff4fff51, 0, stream->current_size); GST_DEBUG_OBJECT (stream->pad, "data location %d", data_location); if (data_location == -1) { GST_ERROR_OBJECT (stream->pad, "Stream does not contain jp2k magic header"); goto error; } /* Elementary stream header box 'elsm' == 0x656c736d */ header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); if (header_tag != 0x656c736d) { GST_ERROR_OBJECT (stream->pad, "Expected ELSM box but found box %x instead", header_tag); goto error; } /* Frame rate box 'frat' == 0x66726174 */ header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); if (header_tag != 0x66726174) { GST_ERROR_OBJECT (stream->pad, "Expected frame rate box, but found box %x instead", header_tag); goto error; } den = gst_byte_reader_get_uint16_be_unchecked (&reader); num = gst_byte_reader_get_uint16_be_unchecked (&reader); /* Maximum bit rate box 'brat' == 0x62726174 */ header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); if (header_tag != 0x62726174) { GST_ERROR_OBJECT (stream->pad, "Expected brat box but read box %x instead", header_tag); goto error; } MaxBr = gst_byte_reader_get_uint32_be_unchecked (&reader); AUF[0] = gst_byte_reader_get_uint32_be_unchecked (&reader); if (stream->jp2kInfos.interlace) { #ifdef TSDEMUX_JP2K_SUPPORT_INTERLACE AUF[1] = gst_byte_reader_get_uint32_be_unchecked (&reader); /* Field Coding Box 'fiel' == 0x6669656c */ header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); if (header_tag != 0x6669656c) { GST_ERROR_OBJECT (stream->pad, "Expected Field Coding box but found box %x instead", header_tag); goto error; } Fic = gst_byte_reader_get_uint8_unchecked (&reader); Fio = gst_byte_reader_get_uint8_unchecked (&reader); #else GST_ERROR_OBJECT (stream->pad, "interlaced mode not supported"); goto error; #endif } /* Time Code Box 'tcod' == 0x74636f64 */ /* Some progressive streams might have a AUF[1] of value 0 present */ header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); if (header_tag == 0 && !stream->jp2kInfos.interlace) { AUF[1] = header_tag; header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); /* Bump up header size and recheck */ header_size += 4; if (stream->current_size < header_size) { GST_ERROR_OBJECT (stream->pad, "Not enough data for header"); goto error; } } if (header_tag != 0x74636f64) { GST_ERROR_OBJECT (stream->pad, "Expected Time code box but found %d box instead", header_tag); goto error; } HHMMSSFF = gst_byte_reader_get_uint32_be_unchecked (&reader); /* Broadcast Color Box 'bcol' == 0x6263686c */ header_tag = gst_byte_reader_get_uint32_be_unchecked (&reader); if (header_tag != 0x62636f6c) { GST_ERROR_OBJECT (stream->pad, "Expected Broadcast color box but found %x box instead", header_tag); goto error; } CollC = gst_byte_reader_get_uint8_unchecked (&reader); b = gst_byte_reader_get_uint8_unchecked (&reader); /* Check if we have enough data to create a valid buffer */ if ((stream->current_size - data_location) < (AUF[0] + AUF[1])) { GST_ERROR_OBJECT (stream->pad, "Required size (%d) greater than remaining size in buffer (%d)", AUF[0] + AUF[1], (stream->current_size - data_location)); if (stream->expected_size && stream->current_size != stream->expected_size) { /* warn if buffer is truncated due to draining */ GST_WARNING_OBJECT (stream->pad, "Truncated buffer: current size (%d) doesn't match expected size (%d)", stream->current_size, stream->expected_size); } else { /* kill pipeline if either we don't know what expected size is, or * we know the expected size, and thus are sure that the buffer is not * truncated due to draining */ goto error; } } retbuf = gst_buffer_new_wrapped_full (0, stream->data, stream->current_size, data_location, stream->current_size - data_location, stream->data, g_free); stream->data = NULL; stream->current_size = 0; return retbuf; error: GST_ERROR ("Failed to parse JP2K access unit"); g_free (stream->data); stream->data = NULL; stream->current_size = 0; return NULL; } static GstBuffer * parse_aac_adts_frame (TSDemuxStream * stream) { gint data_location = -1; guint frame_len; guint crc_size; guint mpegversion = 4; gint i; if (stream->current_size < 6) { GST_DEBUG_OBJECT (stream->pad, "Not enough data for header"); goto out; } /* check syncword */ for (i = 0; i < stream->current_size - 2; i++) { if ((stream->data[i] == 0xff) && ((stream->data[i + 1] & 0xf6) == 0xf0)) { data_location = i; break; } } GST_TRACE_OBJECT (stream->pad, "data location %d", data_location); if (data_location == -1) { GST_DEBUG_OBJECT (stream->pad, "Stream does not contain adts syncword"); goto out; } if (stream->current_size - data_location < 6) { GST_DEBUG_OBJECT (stream->pad, "Not enough data for header"); goto out; } frame_len = ((stream->data[data_location + 3] & 0x03) << 11) | (stream->data[data_location + 4] << 3) | ((stream->data[data_location + 5] & 0xe0) >> 5); crc_size = (stream->data[data_location + 1] & 0x01) ? 0 : 2; if (frame_len < 7 + crc_size) { GST_DEBUG_OBJECT (stream->pad, "Invalid frame len %d", frame_len); goto out; } /* this seems to be valid adts header, check mpeg version now * * TODO: check channels, rate, and profile and then update caps too? */ mpegversion = (stream->data[data_location + 1] & 0x08) ? 2 : 4; out: if (mpegversion != stream->atdsInfos.mpegversion) { GstCaps *caps; MpegTSBaseStream *bstream = (MpegTSBaseStream *) stream; GST_DEBUG_OBJECT (stream->pad, "Update mpegversion from %d to %d", stream->atdsInfos.mpegversion, mpegversion); stream->atdsInfos.mpegversion = mpegversion; caps = gst_stream_get_caps (bstream->stream_object); caps = gst_caps_make_writable (caps); gst_caps_set_simple (caps, "mpegversion", G_TYPE_INT, mpegversion, NULL); gst_stream_set_caps (bstream->stream_object, caps); gst_pad_set_caps (stream->pad, caps); gst_caps_unref (caps); } return gst_buffer_new_wrapped (stream->data, stream->current_size); } static GstFlowReturn gst_ts_demux_push_pending_data (GstTSDemux * demux, TSDemuxStream * stream, MpegTSBaseProgram * target_program) { MpegTSBase *base = GST_MPEGTS_BASE (demux); GstFlowReturn res = GST_FLOW_OK; MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; GstBuffer *buffer = NULL; GstBufferList *buffer_list = NULL; GST_DEBUG_OBJECT (stream->pad, "stream:%p, pid:0x%04x stream_type:%d state:%d", stream, bs->pid, bs->stream_type, stream->state); if (G_UNLIKELY (stream->data == NULL)) { GST_LOG ("stream->data == NULL"); goto beach; } if (G_UNLIKELY (stream->state == PENDING_PACKET_EMPTY)) { GST_LOG ("EMPTY: returning"); goto beach; } if (G_UNLIKELY (stream->state != PENDING_PACKET_BUFFER)) { GST_LOG ("state:%d, returning", stream->state); goto beach; } if (G_UNLIKELY (demux->program == NULL)) { GST_LOG_OBJECT (demux, "No program"); g_free (stream->data); goto beach; } if (stream->needs_keyframe) { MpegTSBase *base = (MpegTSBase *) demux; if ((gst_ts_demux_adjust_seek_offset_for_keyframe (stream, stream->data, stream->current_size)) || demux->last_seek_offset == 0) { GST_DEBUG_OBJECT (stream->pad, "Got Keyframe, ready to go at %" GST_TIME_FORMAT, GST_TIME_ARGS (stream->pts)); if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_PRIVATE_PES_PACKETS && bs->registration_id == DRF_ID_OPUS) { buffer_list = parse_opus_access_unit (stream); if (!buffer_list) { res = GST_FLOW_ERROR; goto beach; } if (gst_buffer_list_length (buffer_list) == 1) { buffer = gst_buffer_ref (gst_buffer_list_get (buffer_list, 0)); gst_buffer_list_unref (buffer_list); buffer_list = NULL; } } else if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K) { buffer = parse_jp2k_access_unit (stream); if (!buffer) { res = GST_FLOW_ERROR; goto beach; } } else { buffer = gst_buffer_new_wrapped (stream->data, stream->current_size); } stream->seeked_pts = stream->pts; stream->seeked_dts = stream->dts; stream->needs_keyframe = FALSE; } else { base->seek_offset = demux->last_seek_offset - 200 * base->packetsize; if (demux->last_seek_offset < 200 * base->packetsize) base->seek_offset = 0; demux->last_seek_offset = base->seek_offset; mpegts_packetizer_flush (base->packetizer, FALSE); base->mode = BASE_MODE_SEEKING; stream->continuity_counter = CONTINUITY_UNSET; res = GST_FLOW_REWINDING; g_free (stream->data); goto beach; } } else { if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_PRIVATE_PES_PACKETS && bs->registration_id == DRF_ID_OPUS) { buffer_list = parse_opus_access_unit (stream); if (!buffer_list) { res = GST_FLOW_ERROR; goto beach; } if (gst_buffer_list_length (buffer_list) == 1) { buffer = gst_buffer_ref (gst_buffer_list_get (buffer_list, 0)); gst_buffer_list_unref (buffer_list); buffer_list = NULL; } } else if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_JP2K) { buffer = parse_jp2k_access_unit (stream); if (!buffer) { res = GST_FLOW_ERROR; goto beach; } } else if (bs->stream_type == GST_MPEGTS_STREAM_TYPE_AUDIO_AAC_ADTS) { buffer = parse_aac_adts_frame (stream); if (!buffer) { res = GST_FLOW_ERROR; goto beach; } } else { buffer = gst_buffer_new_wrapped (stream->data, stream->current_size); } if (G_UNLIKELY (stream->pending_ts && !check_pending_buffers (demux))) { if (buffer) { PendingBuffer *pend; pend = g_slice_new0 (PendingBuffer); pend->buffer = buffer; pend->pts = stream->raw_pts; pend->dts = stream->raw_dts; stream->pending = g_list_append (stream->pending, pend); } else { guint i, n; n = gst_buffer_list_length (buffer_list); for (i = 0; i < n; i++) { PendingBuffer *pend; pend = g_slice_new0 (PendingBuffer); pend->buffer = gst_buffer_ref (gst_buffer_list_get (buffer_list, i)); pend->pts = i == 0 ? stream->raw_pts : -1; pend->dts = i == 0 ? stream->raw_dts : -1; stream->pending = g_list_append (stream->pending, pend); } gst_buffer_list_unref (buffer_list); } GST_DEBUG ("Not enough information to push buffers yet, storing buffer"); goto beach; } } if (G_UNLIKELY (stream->need_newsegment)) calculate_and_push_newsegment (demux, stream, target_program); /* FIXME : Push pending buffers if any */ if (G_UNLIKELY (stream->pending)) { GList *tmp; for (tmp = stream->pending; tmp; tmp = tmp->next) { PendingBuffer *pend = (PendingBuffer *) tmp->data; GST_DEBUG_OBJECT (stream->pad, "Pushing pending buffer PTS:%" GST_TIME_FORMAT " DTS:%" GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_PTS (pend->buffer)), GST_TIME_ARGS (GST_BUFFER_DTS (pend->buffer))); if (stream->discont) GST_BUFFER_FLAG_SET (pend->buffer, GST_BUFFER_FLAG_DISCONT); stream->discont = FALSE; res = gst_pad_push (stream->pad, pend->buffer); stream->nb_out_buffers += 1; g_slice_free (PendingBuffer, pend); } g_list_free (stream->pending); stream->pending = NULL; } if ((GST_CLOCK_TIME_IS_VALID (stream->seeked_pts) && stream->pts < stream->seeked_pts) || (GST_CLOCK_TIME_IS_VALID (stream->seeked_dts) && stream->pts < stream->seeked_dts)) { GST_INFO_OBJECT (stream->pad, "Droping with PTS: %" GST_TIME_FORMAT " DTS: %" GST_TIME_FORMAT " after seeking as other stream needed to be seeked further" "(seeked PTS: %" GST_TIME_FORMAT " DTS: %" GST_TIME_FORMAT ")", GST_TIME_ARGS (stream->pts), GST_TIME_ARGS (stream->dts), GST_TIME_ARGS (stream->seeked_pts), GST_TIME_ARGS (stream->seeked_dts)); if (buffer) gst_buffer_unref (buffer); if (buffer_list) gst_buffer_list_unref (buffer_list); goto beach; } GST_DEBUG_OBJECT (stream->pad, "stream->pts %" GST_TIME_FORMAT, GST_TIME_ARGS (stream->pts)); /* Decorate buffer or first buffer of the buffer list */ if (buffer_list) buffer = gst_buffer_list_get (buffer_list, 0); if (GST_CLOCK_TIME_IS_VALID (stream->pts)) GST_BUFFER_PTS (buffer) = GST_BUFFER_DTS (buffer) = stream->pts; /* DTS = PTS by default, we override it if there's a real DTS */ if (GST_CLOCK_TIME_IS_VALID (stream->dts)) GST_BUFFER_DTS (buffer) = stream->dts; if (stream->discont) GST_BUFFER_FLAG_SET (buffer, GST_BUFFER_FLAG_DISCONT); stream->discont = FALSE; if (buffer_list) buffer = NULL; GST_DEBUG_OBJECT (stream->pad, "Pushing buffer%s with PTS: %" GST_TIME_FORMAT " , DTS: %" GST_TIME_FORMAT, (buffer_list ? "list" : ""), GST_TIME_ARGS (stream->pts), GST_TIME_ARGS (stream->dts)); if (GST_CLOCK_TIME_IS_VALID (stream->dts)) { if (stream->dts > base->out_segment.position) base->out_segment.position = stream->dts; } else if (GST_CLOCK_TIME_IS_VALID (stream->pts)) { if (stream->pts > base->out_segment.position) base->out_segment.position = stream->pts; } if (buffer) { res = gst_pad_push (stream->pad, buffer); /* Record that a buffer was pushed */ stream->nb_out_buffers += 1; } else { guint n = gst_buffer_list_length (buffer_list); res = gst_pad_push_list (stream->pad, buffer_list); /* Record that a buffer was pushed */ stream->nb_out_buffers += n; } GST_DEBUG_OBJECT (stream->pad, "Returned %s", gst_flow_get_name (res)); res = gst_flow_combiner_update_flow (demux->flowcombiner, res); GST_DEBUG_OBJECT (stream->pad, "combined %s", gst_flow_get_name (res)); /* GAP / sparse stream tracking */ if (G_UNLIKELY (stream->gap_ref_pts == GST_CLOCK_TIME_NONE)) stream->gap_ref_pts = stream->pts; else { /* Look if the stream PTS has advanced 2 seconds since the last * gap check, and sync streams if it has. The first stream to * hit this will trigger a gap check */ if (G_UNLIKELY (stream->pts != GST_CLOCK_TIME_NONE && stream->pts > stream->gap_ref_pts + 2 * GST_SECOND)) { if (demux->program->pcr_pid != 0x1fff) { GstClockTime curpcr = mpegts_packetizer_get_current_time (MPEG_TS_BASE_PACKETIZER (demux), demux->program->pcr_pid); if (curpcr == GST_CLOCK_TIME_NONE || curpcr < 800 * GST_MSECOND) goto beach; curpcr -= 800 * GST_MSECOND; /* Use the current PCR (with a safety margin) to sync against */ gst_ts_demux_check_and_sync_streams (demux, curpcr); } else { /* If we don't have a PCR track, just use the current stream PTS */ gst_ts_demux_check_and_sync_streams (demux, stream->pts); } } } beach: /* Reset the PES payload collection, but don't clear the state, * we might want to keep collecting this PES */ GST_LOG ("Cleared PES data. returning %s", gst_flow_get_name (res)); if (stream->expected_size) { if (stream->current_size > stream->expected_size) stream->expected_size = 0; else stream->expected_size -= stream->current_size; } stream->data = NULL; stream->allocated_size = 0; stream->current_size = 0; return res; } static GstFlowReturn gst_ts_demux_handle_packet (GstTSDemux * demux, TSDemuxStream * stream, MpegTSPacketizerPacket * packet, GstMpegtsSection * section) { GstFlowReturn res = GST_FLOW_OK; GST_LOG ("pid 0x%04x pusi:%d, afc:%d, cont:%d, payload:%p", packet->pid, packet->payload_unit_start_indicator, packet->scram_afc_cc & 0x30, FLAGS_CONTINUITY_COUNTER (packet->scram_afc_cc), packet->payload); if (G_UNLIKELY (packet->payload_unit_start_indicator) && FLAGS_HAS_PAYLOAD (packet->scram_afc_cc)) { /* Flush previous data */ res = gst_ts_demux_push_pending_data (demux, stream, NULL); /* Tell the data collecting to expect this header */ stream->state = PENDING_PACKET_HEADER; } if (packet->payload && (res == GST_FLOW_OK || res == GST_FLOW_NOT_LINKED) && stream->pad) { gst_ts_demux_queue_data (demux, stream, packet); GST_LOG ("current_size:%d, expected_size:%d", stream->current_size, stream->expected_size); /* Finally check if the data we queued completes a packet, or got too * large and needs output now */ if ((stream->expected_size && stream->current_size >= stream->expected_size) || (stream->current_size >= MAX_PES_PAYLOAD)) { GST_LOG ("pushing packet of size %u", stream->current_size); res = gst_ts_demux_push_pending_data (demux, stream, NULL); } } /* We are rewinding to find a keyframe, * and didn't want the data to be queued */ if (res == GST_FLOW_REWINDING) res = GST_FLOW_OK; return res; } static void gst_ts_demux_flush (MpegTSBase * base, gboolean hard) { GstTSDemux *demux = GST_TS_DEMUX_CAST (base); gst_ts_demux_flush_streams (demux, hard); if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } if (demux->global_tags) { gst_tag_list_unref (demux->global_tags); demux->global_tags = NULL; } if (hard) { /* For pull mode seeks the current segment needs to be preserved */ demux->rate = 1.0; gst_segment_init (&base->out_segment, GST_FORMAT_UNDEFINED); } } static GstFlowReturn gst_ts_demux_drain (MpegTSBase * base) { GstTSDemux *demux = GST_TS_DEMUX_CAST (base); GList *tmp; GstFlowReturn res = GST_FLOW_OK; if (!demux->program) return res; for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; if (stream->pad) { res = gst_ts_demux_push_pending_data (demux, stream, NULL); if (G_UNLIKELY (res != GST_FLOW_OK)) break; } } return res; } static GstFlowReturn gst_ts_demux_push (MpegTSBase * base, MpegTSPacketizerPacket * packet, GstMpegtsSection * section) { GstTSDemux *demux = GST_TS_DEMUX_CAST (base); TSDemuxStream *stream = NULL; GstFlowReturn res = GST_FLOW_OK; if (G_LIKELY (demux->program)) { stream = (TSDemuxStream *) demux->program->streams[packet->pid]; if (stream) { res = gst_ts_demux_handle_packet (demux, stream, packet, section); } } return res; } gboolean gst_ts_demux_plugin_init (GstPlugin * plugin) { GST_DEBUG_CATEGORY_INIT (ts_demux_debug, "tsdemux", 0, "MPEG transport stream demuxer"); init_pes_parser (); return gst_element_register (plugin, "tsdemux", GST_RANK_PRIMARY, GST_TYPE_TS_DEMUX); }