/* * tsdemux.c * Copyright (C) 2009 Zaheer Abbas Merali * 2010 Edward Hervey * Copyright (C) 2011, Hewlett-Packard Development Company, L.P. * Author: Youness Alaoui , Collabora Ltd. * Author: Sebastian Dröge , Collabora Ltd. * Author: Edward Hervey , Collabora Ltd. * * Authors: * Zaheer Abbas Merali * Edward Hervey * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #include #include "mpegtsbase.h" #include "tsdemux.h" #include "gstmpegdesc.h" #include "gstmpegdefs.h" #include "mpegtspacketizer.h" #include "pesparse.h" #include #include #include /* * tsdemux * * See TODO for explanations on improvements needed */ #define CONTINUITY_UNSET 255 #define MAX_CONTINUITY 15 /* Seeking/Scanning related variables */ /* seek to SEEK_TIMESTAMP_OFFSET before the desired offset and search then * either accurately or for the next timestamp */ #define SEEK_TIMESTAMP_OFFSET (2500 * GST_MSECOND) #define GST_FLOW_REWINDING GST_FLOW_CUSTOM_ERROR /* latency in nsecs */ #define TS_LATENCY (700 * GST_MSECOND) GST_DEBUG_CATEGORY_STATIC (ts_demux_debug); #define GST_CAT_DEFAULT ts_demux_debug #define ABSDIFF(a,b) (((a) > (b)) ? ((a) - (b)) : ((b) - (a))) static GQuark QUARK_TSDEMUX; static GQuark QUARK_PID; static GQuark QUARK_PCR; static GQuark QUARK_OPCR; static GQuark QUARK_PTS; static GQuark QUARK_DTS; static GQuark QUARK_OFFSET; typedef enum { PENDING_PACKET_EMPTY = 0, /* No pending packet/buffer * Push incoming buffers to the array */ PENDING_PACKET_HEADER, /* PES header needs to be parsed * Push incoming buffers to the array */ PENDING_PACKET_BUFFER, /* Currently filling up output buffer * Push incoming buffers to the bufferlist */ PENDING_PACKET_DISCONT /* Discontinuity in incoming packets * Drop all incoming buffers */ } PendingPacketState; /* Pending buffer */ typedef struct { /* The fully reconstructed buffer */ GstBuffer *buffer; /* Raw PTS/DTS (in 90kHz units) */ guint64 pts, dts; } PendingBuffer; typedef struct _TSDemuxStream TSDemuxStream; typedef struct _TSDemuxH264ParsingInfos TSDemuxH264ParsingInfos; /* Returns TRUE if a keyframe was found */ typedef gboolean (*GstTsDemuxKeyFrameScanFunction) (TSDemuxStream * stream, guint8 * data, const gsize data_size, const gsize max_frame_offset); typedef struct { guint8 *data; gsize size; } SimpleBuffer; struct _TSDemuxH264ParsingInfos { /* H264 parsing data */ GstH264NalParser *parser; GstByteWriter *sps; GstByteWriter *pps; GstByteWriter *sei; SimpleBuffer framedata; }; struct _TSDemuxStream { MpegTSBaseStream stream; GstPad *pad; /* Whether the pad was added or not */ gboolean active; /* Whether this is a sparse stream (subtitles or metadata) */ gboolean sparse; /* TRUE if we are waiting for a valid timestamp */ gboolean pending_ts; /* Output data */ PendingPacketState state; /* Data being reconstructed (allocated) */ guint8 *data; /* Size of data being reconstructed (if known, else 0) */ guint expected_size; /* Amount of bytes in current ->data */ guint current_size; /* Size of ->data */ guint allocated_size; /* Current PTS/DTS for this stream (in running time) */ GstClockTime pts; GstClockTime dts; /* Reference PTS used to detect gaps */ GstClockTime gap_ref_pts; /* Number of outputted buffers */ guint32 nb_out_buffers; /* Reference number of buffers for gaps */ guint32 gap_ref_buffers; /* Current PTS/DTS for this stream (in 90kHz unit) */ guint64 raw_pts, raw_dts; /* Whether this stream needs to send a newsegment */ gboolean need_newsegment; /* Whether the next output buffer should be DISCONT */ gboolean discont; /* The value to use when calculating the newsegment */ GstClockTime first_dts; GstTagList *taglist; gint continuity_counter; /* List of pending buffers */ GList *pending; /* if != 0, output only PES from that substream */ guint8 target_pes_substream; gboolean needs_keyframe; GstClockTime seeked_pts, seeked_dts; GstTsDemuxKeyFrameScanFunction scan_function; TSDemuxH264ParsingInfos h264infos; }; #define VIDEO_CAPS \ GST_STATIC_CAPS (\ "video/mpeg, " \ "mpegversion = (int) { 1, 2, 4 }, " \ "systemstream = (boolean) FALSE; " \ "video/x-h264,stream-format=(string)byte-stream," \ "alignment=(string)nal;" \ "video/x-dirac;" \ "video/x-cavs;" \ "video/x-wmv," \ "wmvversion = (int) 3, " \ "format = (string) WVC1" \ ) #define AUDIO_CAPS \ GST_STATIC_CAPS ( \ "audio/mpeg, " \ "mpegversion = (int) 1;" \ "audio/mpeg, " \ "mpegversion = (int) 2, " \ "stream-format = (string) adts; " \ "audio/mpeg, " \ "mpegversion = (int) 4, " \ "stream-format = (string) loas; " \ "audio/x-lpcm, " \ "width = (int) { 16, 20, 24 }, " \ "rate = (int) { 48000, 96000 }, " \ "channels = (int) [ 1, 8 ], " \ "dynamic_range = (int) [ 0, 255 ], " \ "emphasis = (boolean) { FALSE, TRUE }, " \ "mute = (boolean) { FALSE, TRUE }; " \ "audio/x-ac3; audio/x-eac3;" \ "audio/x-dts;" \ "audio/x-private-ts-lpcm" \ ) /* Can also use the subpicture pads for text subtitles? */ #define SUBPICTURE_CAPS \ GST_STATIC_CAPS ("subpicture/x-pgs; subpicture/x-dvd") static GstStaticPadTemplate video_template = GST_STATIC_PAD_TEMPLATE ("video_%04x", GST_PAD_SRC, GST_PAD_SOMETIMES, VIDEO_CAPS); static GstStaticPadTemplate audio_template = GST_STATIC_PAD_TEMPLATE ("audio_%04x", GST_PAD_SRC, GST_PAD_SOMETIMES, AUDIO_CAPS); static GstStaticPadTemplate subpicture_template = GST_STATIC_PAD_TEMPLATE ("subpicture_%04x", GST_PAD_SRC, GST_PAD_SOMETIMES, SUBPICTURE_CAPS); static GstStaticPadTemplate private_template = GST_STATIC_PAD_TEMPLATE ("private_%04x", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); enum { PROP_0, PROP_PROGRAM_NUMBER, PROP_EMIT_STATS, /* FILL ME */ }; /* Pad functions */ /* mpegtsbase methods */ static void gst_ts_demux_program_started (MpegTSBase * base, MpegTSBaseProgram * program); static void gst_ts_demux_program_stopped (MpegTSBase * base, MpegTSBaseProgram * program); static void gst_ts_demux_reset (MpegTSBase * base); static GstFlowReturn gst_ts_demux_push (MpegTSBase * base, MpegTSPacketizerPacket * packet, GstMpegtsSection * section); static void gst_ts_demux_flush (MpegTSBase * base, gboolean hard); static GstFlowReturn gst_ts_demux_drain (MpegTSBase * base); static void gst_ts_demux_stream_added (MpegTSBase * base, MpegTSBaseStream * stream, MpegTSBaseProgram * program); static void gst_ts_demux_stream_removed (MpegTSBase * base, MpegTSBaseStream * stream); static GstFlowReturn gst_ts_demux_do_seek (MpegTSBase * base, GstEvent * event); static void gst_ts_demux_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_ts_demux_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static void gst_ts_demux_flush_streams (GstTSDemux * tsdemux, gboolean hard); static GstFlowReturn gst_ts_demux_push_pending_data (GstTSDemux * demux, TSDemuxStream * stream); static void gst_ts_demux_stream_flush (TSDemuxStream * stream, GstTSDemux * demux, gboolean hard); static gboolean push_event (MpegTSBase * base, GstEvent * event); static void gst_ts_demux_check_and_sync_streams (GstTSDemux * demux, GstClockTime time); static void _extra_init (void) { QUARK_TSDEMUX = g_quark_from_string ("tsdemux"); QUARK_PID = g_quark_from_string ("pid"); QUARK_PCR = g_quark_from_string ("pcr"); QUARK_OPCR = g_quark_from_string ("opcr"); QUARK_PTS = g_quark_from_string ("pts"); QUARK_DTS = g_quark_from_string ("dts"); QUARK_OFFSET = g_quark_from_string ("offset"); } #define gst_ts_demux_parent_class parent_class G_DEFINE_TYPE_WITH_CODE (GstTSDemux, gst_ts_demux, GST_TYPE_MPEGTS_BASE, _extra_init ()); static void gst_ts_demux_dispose (GObject * object) { GstTSDemux *demux = GST_TS_DEMUX_CAST (object); gst_flow_combiner_free (demux->flowcombiner); GST_CALL_PARENT (G_OBJECT_CLASS, dispose, (object)); } static void gst_ts_demux_class_init (GstTSDemuxClass * klass) { GObjectClass *gobject_class; GstElementClass *element_class; MpegTSBaseClass *ts_class; gobject_class = G_OBJECT_CLASS (klass); gobject_class->set_property = gst_ts_demux_set_property; gobject_class->get_property = gst_ts_demux_get_property; gobject_class->dispose = gst_ts_demux_dispose; g_object_class_install_property (gobject_class, PROP_PROGRAM_NUMBER, g_param_spec_int ("program-number", "Program number", "Program Number to demux for (-1 to ignore)", -1, G_MAXINT, -1, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_EMIT_STATS, g_param_spec_boolean ("emit-stats", "Emit statistics", "Emit messages for every pcr/opcr/pts/dts", FALSE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); element_class = GST_ELEMENT_CLASS (klass); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&video_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&audio_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&subpicture_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&private_template)); gst_element_class_set_static_metadata (element_class, "MPEG transport stream demuxer", "Codec/Demuxer", "Demuxes MPEG2 transport streams", "Zaheer Abbas Merali \n" "Edward Hervey "); ts_class = GST_MPEGTS_BASE_CLASS (klass); ts_class->reset = GST_DEBUG_FUNCPTR (gst_ts_demux_reset); ts_class->push = GST_DEBUG_FUNCPTR (gst_ts_demux_push); ts_class->push_event = GST_DEBUG_FUNCPTR (push_event); ts_class->program_started = GST_DEBUG_FUNCPTR (gst_ts_demux_program_started); ts_class->program_stopped = GST_DEBUG_FUNCPTR (gst_ts_demux_program_stopped); ts_class->stream_added = gst_ts_demux_stream_added; ts_class->stream_removed = gst_ts_demux_stream_removed; ts_class->seek = GST_DEBUG_FUNCPTR (gst_ts_demux_do_seek); ts_class->flush = GST_DEBUG_FUNCPTR (gst_ts_demux_flush); ts_class->drain = GST_DEBUG_FUNCPTR (gst_ts_demux_drain); } static void gst_ts_demux_reset (MpegTSBase * base) { GstTSDemux *demux = (GstTSDemux *) base; demux->rate = 1.0; gst_segment_init (&demux->segment, GST_FORMAT_UNDEFINED); if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } if (demux->global_tags) { gst_tag_list_unref (demux->global_tags); demux->global_tags = NULL; } demux->have_group_id = FALSE; demux->group_id = G_MAXUINT; demux->last_seek_offset = -1; } static void gst_ts_demux_init (GstTSDemux * demux) { MpegTSBase *base = (MpegTSBase *) demux; base->stream_size = sizeof (TSDemuxStream); base->parse_private_sections = TRUE; /* We are not interested in sections (all handled by mpegtsbase) */ base->push_section = FALSE; demux->flowcombiner = gst_flow_combiner_new (); demux->requested_program_number = -1; demux->program_number = -1; gst_ts_demux_reset (base); } static void gst_ts_demux_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstTSDemux *demux = GST_TS_DEMUX (object); switch (prop_id) { case PROP_PROGRAM_NUMBER: /* FIXME: do something if program is switched as opposed to set at * beginning */ demux->requested_program_number = g_value_get_int (value); break; case PROP_EMIT_STATS: demux->emit_statistics = g_value_get_boolean (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); } } static void gst_ts_demux_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstTSDemux *demux = GST_TS_DEMUX (object); switch (prop_id) { case PROP_PROGRAM_NUMBER: g_value_set_int (value, demux->requested_program_number); break; case PROP_EMIT_STATS: g_value_set_boolean (value, demux->emit_statistics); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); } } static gboolean gst_ts_demux_get_duration (GstTSDemux * demux, GstClockTime * dur) { MpegTSBase *base = (MpegTSBase *) demux; gboolean res = FALSE; gint64 val; /* Get total size in bytes */ if (gst_pad_peer_query_duration (base->sinkpad, GST_FORMAT_BYTES, &val)) { /* Convert it to duration */ *dur = mpegts_packetizer_offset_to_ts (base->packetizer, val, demux->program->pcr_pid); if (GST_CLOCK_TIME_IS_VALID (*dur)) res = TRUE; } return res; } static gboolean gst_ts_demux_srcpad_query (GstPad * pad, GstObject * parent, GstQuery * query) { gboolean res = TRUE; GstFormat format; GstTSDemux *demux; MpegTSBase *base; demux = GST_TS_DEMUX (parent); base = GST_MPEGTS_BASE (demux); switch (GST_QUERY_TYPE (query)) { case GST_QUERY_DURATION: { GST_DEBUG ("query duration"); gst_query_parse_duration (query, &format, NULL); if (format == GST_FORMAT_TIME) { if (!gst_pad_peer_query (base->sinkpad, query)) { GstClockTime dur; if (gst_ts_demux_get_duration (demux, &dur)) gst_query_set_duration (query, GST_FORMAT_TIME, dur); else res = FALSE; } } else { GST_DEBUG_OBJECT (demux, "only query duration on TIME is supported"); res = FALSE; } break; } case GST_QUERY_LATENCY: { GST_DEBUG ("query latency"); res = gst_pad_peer_query (base->sinkpad, query); if (res) { GstClockTime min_lat, max_lat; gboolean live; /* According to H.222.0 Annex D.0.3 (System Time Clock recovery in the decoder) and D.0.2 (Audio and video presentation synchronization) We can end up with an interval of up to 700ms between valid PTS/DTS. We therefore allow a latency of 700ms for that. */ gst_query_parse_latency (query, &live, &min_lat, &max_lat); if (min_lat) min_lat += TS_LATENCY; if (GST_CLOCK_TIME_IS_VALID (max_lat)) max_lat += TS_LATENCY; gst_query_set_latency (query, live, min_lat, max_lat); } break; } case GST_QUERY_SEEKING: { GST_DEBUG ("query seeking"); gst_query_parse_seeking (query, &format, NULL, NULL, NULL); GST_DEBUG ("asked for format %s", gst_format_get_name (format)); if (format == GST_FORMAT_TIME) { gboolean seekable = FALSE; if (gst_pad_peer_query (base->sinkpad, query)) gst_query_parse_seeking (query, NULL, &seekable, NULL, NULL); /* If upstream is not seekable in TIME format we use * our own values here */ if (!seekable) { GstClockTime dur; if (gst_ts_demux_get_duration (demux, &dur)) { gst_query_set_seeking (query, GST_FORMAT_TIME, TRUE, 0, dur); GST_DEBUG ("Gave duration: %" GST_TIME_FORMAT, GST_TIME_ARGS (dur)); } } } else { GST_DEBUG_OBJECT (demux, "only TIME is supported for query seeking"); res = FALSE; } break; } case GST_QUERY_SEGMENT:{ GstFormat format; gint64 start, stop; format = demux->segment.format; start = gst_segment_to_stream_time (&demux->segment, format, demux->segment.start); if ((stop = demux->segment.stop) == -1) stop = demux->segment.duration; else stop = gst_segment_to_stream_time (&demux->segment, format, stop); gst_query_set_segment (query, demux->segment.rate, format, start, stop); res = TRUE; break; } default: res = gst_pad_query_default (pad, parent, query); } return res; } static void clear_simple_buffer (SimpleBuffer * sbuf) { if (!sbuf->data) return; g_free (sbuf->data); sbuf->size = 0; sbuf->data = NULL; } static gboolean scan_keyframe_h264 (TSDemuxStream * stream, const guint8 * data, const gsize data_size, const gsize max_frame_offset) { gint offset = 0; GstH264NalUnit unit, frame_unit = { 0, }; GstH264ParserResult res = GST_H264_PARSER_OK; TSDemuxH264ParsingInfos *h264infos = &stream->h264infos; GstH264NalParser *parser = h264infos->parser; if (G_UNLIKELY (parser == NULL)) { parser = h264infos->parser = gst_h264_nal_parser_new (); h264infos->sps = gst_byte_writer_new (); h264infos->pps = gst_byte_writer_new (); h264infos->sei = gst_byte_writer_new (); } while (res == GST_H264_PARSER_OK) { res = gst_h264_parser_identify_nalu (parser, data, offset, data_size, &unit); if (res != GST_H264_PARSER_OK && res != GST_H264_PARSER_NO_NAL_END) { GST_INFO_OBJECT (stream->pad, "Error identifying nalu: %i", res); break; } res = gst_h264_parser_parse_nal (parser, &unit); if (res != GST_H264_PARSER_OK) { break; } switch (unit.type) { case GST_H264_NAL_SEI: if (frame_unit.size) break; if (gst_byte_writer_put_data (h264infos->sei, unit.data + unit.sc_offset, unit.size + unit.offset - unit.sc_offset)) { GST_DEBUG ("adding SEI %u", unit.size + unit.offset - unit.sc_offset); } else { GST_WARNING ("Could not write SEI"); } break; case GST_H264_NAL_PPS: if (frame_unit.size) break; if (gst_byte_writer_put_data (h264infos->pps, unit.data + unit.sc_offset, unit.size + unit.offset - unit.sc_offset)) { GST_DEBUG ("adding PPS %u", unit.size + unit.offset - unit.sc_offset); } else { GST_WARNING ("Could not write PPS"); } break; case GST_H264_NAL_SPS: if (frame_unit.size) break; if (gst_byte_writer_put_data (h264infos->sps, unit.data + unit.sc_offset, unit.size + unit.offset - unit.sc_offset)) { GST_DEBUG ("adding SPS %u", unit.size + unit.offset - unit.sc_offset); } else { GST_WARNING ("Could not write SPS"); } break; /* these units are considered keyframes in h264parse */ case GST_H264_NAL_SLICE: case GST_H264_NAL_SLICE_DPA: case GST_H264_NAL_SLICE_DPB: case GST_H264_NAL_SLICE_DPC: case GST_H264_NAL_SLICE_IDR: { GstH264SliceHdr slice; if (h264infos->framedata.size) break; res = gst_h264_parser_parse_slice_hdr (parser, &unit, &slice, FALSE, FALSE); if (GST_H264_IS_I_SLICE (&slice) || GST_H264_IS_SI_SLICE (&slice)) { if (*(unit.data + unit.offset + 1) & 0x80) { /* means first_mb_in_slice == 0 */ /* real frame data */ GST_DEBUG_OBJECT (stream->pad, "Found keyframe at: %u", unit.sc_offset); frame_unit = unit; } } break; } default: break; } if (offset == unit.sc_offset + unit.size) break; offset = unit.sc_offset + unit.size; } /* We've got all the infos we need (SPS / PPS and a keyframe, plus * and possibly SEI units. We can stop rewinding the stream */ if (gst_byte_writer_get_size (h264infos->sps) && gst_byte_writer_get_size (h264infos->pps) && (h264infos->framedata.size || frame_unit.size)) { guint8 *data = NULL; gsize tmpsize = gst_byte_writer_get_size (h264infos->pps); /* We know that the SPS is first so just put all our data in there */ data = gst_byte_writer_reset_and_get_data (h264infos->pps); gst_byte_writer_put_data (h264infos->sps, data, tmpsize); g_free (data); tmpsize = gst_byte_writer_get_size (h264infos->sei); if (tmpsize) { GST_DEBUG ("Adding SEI"); data = gst_byte_writer_reset_and_get_data (h264infos->sei); gst_byte_writer_put_data (h264infos->sps, data, tmpsize); g_free (data); } if (frame_unit.size) { /* We found the everything in one go! */ GST_DEBUG ("Adding Keyframe"); gst_byte_writer_put_data (h264infos->sps, frame_unit.data + frame_unit.sc_offset, stream->current_size - frame_unit.sc_offset); } else { GST_DEBUG ("Adding Keyframe"); gst_byte_writer_put_data (h264infos->sps, h264infos->framedata.data, h264infos->framedata.size); clear_simple_buffer (&h264infos->framedata); } g_free (stream->data); stream->current_size = gst_byte_writer_get_size (h264infos->sps); stream->data = gst_byte_writer_reset_and_get_data (h264infos->sps); gst_byte_writer_init (h264infos->sps); gst_byte_writer_init (h264infos->pps); gst_byte_writer_init (h264infos->sei); return TRUE; } if (frame_unit.size) { GST_DEBUG_OBJECT (stream->pad, "Keep the keyframe as this is the one" " we will push later"); h264infos->framedata.data = g_memdup (frame_unit.data + frame_unit.sc_offset, stream->current_size - frame_unit.sc_offset); h264infos->framedata.size = stream->current_size - frame_unit.sc_offset; } return FALSE; } /* We merge data from TS packets so that the scanning methods get a continuous chunk, however the scanning method will return keyframe offset which needs to be translated back to actual offset in file */ typedef struct { gint64 real_offset; /* offset of TS packet */ gint merged_offset; /* offset of merged data in buffer */ } OffsetInfo; static gboolean gst_ts_demux_adjust_seek_offset_for_keyframe (TSDemuxStream * stream, guint8 * data, guint64 size) { int scan_pid = -1; if (!stream->scan_function) return TRUE; scan_pid = ((MpegTSBaseStream *) stream)->pid; if (scan_pid != -1) { return stream->scan_function (stream, data, size, size); } return TRUE; } static GstFlowReturn gst_ts_demux_do_seek (MpegTSBase * base, GstEvent * event) { GList *tmp; GstTSDemux *demux = (GstTSDemux *) base; GstFlowReturn res = GST_FLOW_ERROR; gdouble rate; GstFormat format; GstSeekFlags flags; GstSeekType start_type, stop_type; gint64 start, stop; guint64 start_offset; gst_event_parse_seek (event, &rate, &format, &flags, &start_type, &start, &stop_type, &stop); GST_DEBUG ("seek event, rate: %f start: %" GST_TIME_FORMAT " stop: %" GST_TIME_FORMAT, rate, GST_TIME_ARGS (start), GST_TIME_ARGS (stop)); if (rate <= 0.0) { GST_WARNING ("Negative rate not supported"); goto done; } if (flags & (GST_SEEK_FLAG_SEGMENT)) { GST_WARNING ("seek flags 0x%x are not supported", (int) flags); goto done; } /* configure the segment with the seek variables */ GST_DEBUG_OBJECT (demux, "configuring seek"); if (start_type != GST_SEEK_TYPE_NONE) { start_offset = mpegts_packetizer_ts_to_offset (base->packetizer, MAX (0, start - SEEK_TIMESTAMP_OFFSET), demux->program->pcr_pid); if (G_UNLIKELY (start_offset == -1)) { GST_WARNING ("Couldn't convert start position to an offset"); goto done; } } else { for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = tmp->data; stream->need_newsegment = TRUE; } gst_segment_init (&demux->segment, GST_FORMAT_UNDEFINED); if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } demux->rate = rate; res = GST_FLOW_OK; goto done; } /* record offset and rate */ base->seek_offset = start_offset; demux->last_seek_offset = base->seek_offset; demux->rate = rate; res = GST_FLOW_OK; gst_segment_do_seek (&demux->segment, rate, format, flags, start_type, start, stop_type, stop, NULL); if (!(flags & GST_SEEK_FLAG_ACCURATE)) demux->reset_segment = TRUE; if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = tmp->data; if (flags & GST_SEEK_FLAG_ACCURATE) stream->needs_keyframe = TRUE; stream->seeked_pts = GST_CLOCK_TIME_NONE; stream->seeked_dts = GST_CLOCK_TIME_NONE; stream->need_newsegment = TRUE; stream->first_dts = GST_CLOCK_TIME_NONE; } done: return res; } static gboolean gst_ts_demux_srcpad_event (GstPad * pad, GstObject * parent, GstEvent * event) { gboolean res = TRUE; GstTSDemux *demux = GST_TS_DEMUX (parent); GST_DEBUG_OBJECT (pad, "Got event %s", gst_event_type_get_name (GST_EVENT_TYPE (event))); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEEK: res = mpegts_base_handle_seek_event ((MpegTSBase *) demux, pad, event); if (!res) GST_WARNING ("seeking failed"); gst_event_unref (event); break; default: res = gst_pad_event_default (pad, parent, event); } return res; } static void clean_global_taglist (GstTagList * taglist) { gst_tag_list_remove_tag (taglist, GST_TAG_CONTAINER_FORMAT); gst_tag_list_remove_tag (taglist, GST_TAG_CODEC); } static gboolean push_event (MpegTSBase * base, GstEvent * event) { GstTSDemux *demux = (GstTSDemux *) base; GList *tmp; gboolean early_ret = FALSE; if (GST_EVENT_TYPE (event) == GST_EVENT_SEGMENT) { GST_DEBUG_OBJECT (base, "Ignoring segment event (recreated later)"); gst_event_unref (event); return TRUE; } else if (GST_EVENT_TYPE (event) == GST_EVENT_TAG) { /* In case we receive tags before data, store them to send later * If we already have the program, send it right away */ GstTagList *taglist; gst_event_parse_tag (event, &taglist); if (demux->global_tags == NULL) { demux->global_tags = gst_tag_list_copy (taglist); /* Tags that are stream specific for the container should be considered * global for the container streams */ if (gst_tag_list_get_scope (taglist) == GST_TAG_SCOPE_STREAM) { gst_tag_list_set_scope (demux->global_tags, GST_TAG_SCOPE_GLOBAL); } } else { demux->global_tags = gst_tag_list_make_writable (demux->global_tags); gst_tag_list_insert (demux->global_tags, taglist, GST_TAG_MERGE_REPLACE); } clean_global_taglist (demux->global_tags); /* tags are stored to be used after if there are no streams yet, * so we should never reject */ early_ret = TRUE; } if (G_UNLIKELY (demux->program == NULL)) { gst_event_unref (event); return early_ret; } for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; if (stream->pad) { /* If we are pushing out EOS, flush out pending data first */ if (GST_EVENT_TYPE (event) == GST_EVENT_EOS && gst_pad_is_active (stream->pad)) gst_ts_demux_push_pending_data (demux, stream); gst_event_ref (event); gst_pad_push_event (stream->pad, event); } } gst_event_unref (event); return TRUE; } static inline void add_iso639_language_to_tags (TSDemuxStream * stream, gchar * lang_code) { const gchar *lc; GST_LOG ("Add language code for stream: '%s'", lang_code); if (!stream->taglist) stream->taglist = gst_tag_list_new_empty (); /* descriptor contains ISO 639-2 code, we want the ISO 639-1 code */ lc = gst_tag_get_language_code (lang_code); /* Only set tag if we have a valid one */ if (lc || (lang_code[0] && lang_code[1])) gst_tag_list_add (stream->taglist, GST_TAG_MERGE_REPLACE, GST_TAG_LANGUAGE_CODE, (lc) ? lc : lang_code, NULL); } static void gst_ts_demux_create_tags (TSDemuxStream * stream) { MpegTSBaseStream *bstream = (MpegTSBaseStream *) stream; const GstMpegtsDescriptor *desc = NULL; int i, nb; desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_ISO_639_LANGUAGE); if (desc) { gchar *lang_code; nb = gst_mpegts_descriptor_parse_iso_639_language_nb (desc); GST_DEBUG ("Found ISO 639 descriptor (%d entries)", nb); for (i = 0; i < nb; i++) if (gst_mpegts_descriptor_parse_iso_639_language_idx (desc, i, &lang_code, NULL)) { add_iso639_language_to_tags (stream, lang_code); g_free (lang_code); } return; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_SUBTITLING); if (desc) { gchar *lang_code; nb = gst_mpegts_descriptor_parse_dvb_subtitling_nb (desc); GST_DEBUG ("Found SUBTITLING descriptor (%d entries)", nb); for (i = 0; i < nb; i++) if (gst_mpegts_descriptor_parse_dvb_subtitling_idx (desc, i, &lang_code, NULL, NULL, NULL)) { add_iso639_language_to_tags (stream, lang_code); g_free (lang_code); } } } static GstPad * create_pad_for_stream (MpegTSBase * base, MpegTSBaseStream * bstream, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); TSDemuxStream *stream = (TSDemuxStream *) bstream; gchar *name = NULL; GstCaps *caps = NULL; GstPadTemplate *template = NULL; const GstMpegtsDescriptor *desc = NULL; GstPad *pad = NULL; gboolean sparse = FALSE; gboolean is_audio = FALSE, is_video = FALSE, is_subpicture = FALSE, is_private = FALSE; gst_ts_demux_create_tags (stream); GST_LOG ("Attempting to create pad for stream 0x%04x with stream_type %d", bstream->pid, bstream->stream_type); /* First handle BluRay-specific stream types since there is some overlap * between BluRay and non-BluRay streay type identifiers */ if (program->registration_id == DRF_ID_HDMV) { switch (bstream->stream_type) { case ST_BD_AUDIO_AC3: { const GstMpegtsDescriptor *ac3_desc; /* ATSC ac3 audio descriptor */ ac3_desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_AC3_AUDIO_STREAM); if (ac3_desc && DESC_AC_AUDIO_STREAM_bsid (ac3_desc->data) != 16) { GST_LOG ("ac3 audio"); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); } else { is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); } break; } case ST_BD_AUDIO_EAC3: case ST_BD_AUDIO_AC3_PLUS: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; case ST_BD_AUDIO_AC3_TRUE_HD: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-true-hd"); stream->target_pes_substream = 0x72; break; case ST_BD_AUDIO_LPCM: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-private-ts-lpcm"); break; case ST_BD_PGS_SUBPICTURE: is_subpicture = TRUE; caps = gst_caps_new_empty_simple ("subpicture/x-pgs"); sparse = TRUE; break; case ST_BD_AUDIO_DTS_HD: case ST_BD_AUDIO_DTS_HD_MASTER_AUDIO: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-dts"); stream->target_pes_substream = 0x71; break; } } if (caps) goto done; /* Handle non-BluRay stream types */ switch (bstream->stream_type) { case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG1: case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG2: case ST_PS_VIDEO_MPEG2_DCII: /* FIXME : Use DCII registration code (ETV1 ?) to handle that special * Stream type (ST_PS_VIDEO_MPEG2_DCII) */ /* FIXME : Use video decriptor (0x1) to refine caps with: * * frame_rate * * profile_and_level */ GST_LOG ("mpeg video"); is_video = TRUE; caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, bstream->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG1 ? 1 : 2, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MPEGTS_STREAM_TYPE_AUDIO_MPEG1: case GST_MPEGTS_STREAM_TYPE_AUDIO_MPEG2: GST_LOG ("mpeg audio"); is_audio = TRUE; caps = gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 1, NULL); /* HDV is always mpeg 1 audio layer 2 */ if (program->registration_id == DRF_ID_TSHV) gst_caps_set_simple (caps, "layer", G_TYPE_INT, 2, NULL); break; case GST_MPEGTS_STREAM_TYPE_PRIVATE_PES_PACKETS: GST_LOG ("private data"); /* FIXME: Move all of this into a common method (there might be other * types also, depending on registratino descriptors also */ desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_AC3); if (desc) { GST_LOG ("ac3 audio"); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); break; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_ENHANCED_AC3); if (desc) { GST_LOG ("ac3 audio"); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_TELETEXT); if (desc) { GST_LOG ("teletext"); is_private = TRUE; caps = gst_caps_new_empty_simple ("application/x-teletext"); sparse = TRUE; break; } desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_SUBTITLING); if (desc) { GST_LOG ("subtitling"); is_private = TRUE; caps = gst_caps_new_empty_simple ("subpicture/x-dvb"); sparse = TRUE; break; } switch (bstream->registration_id) { case DRF_ID_DTS1: case DRF_ID_DTS2: case DRF_ID_DTS3: /* SMPTE registered DTS */ is_private = TRUE; caps = gst_caps_new_empty_simple ("audio/x-dts"); break; case DRF_ID_S302M: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-smpte-302m"); break; case DRF_ID_HEVC: is_video = TRUE; caps = gst_caps_new_simple ("video/x-h265", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); break; } if (caps) break; /* hack for itv hd (sid 10510, video pid 3401 */ if (program->program_number == 10510 && bstream->pid == 3401) { is_video = TRUE; caps = gst_caps_new_simple ("video/x-h264", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); } break; case ST_HDV_AUX_V: /* FIXME : Should only be used with specific PMT registration_descriptor */ /* We don't expose those streams since they're only helper streams */ /* template = gst_static_pad_template_get (&private_template); */ /* name = g_strdup_printf ("private_%04x", bstream->pid); */ /* caps = gst_caps_new_simple ("hdv/aux-v", NULL); */ break; case ST_HDV_AUX_A: /* FIXME : Should only be used with specific PMT registration_descriptor */ /* We don't expose those streams since they're only helper streams */ /* template = gst_static_pad_template_get (&private_template); */ /* name = g_strdup_printf ("private_%04x", bstream->pid); */ /* caps = gst_caps_new_simple ("hdv/aux-a", NULL); */ break; case GST_MPEGTS_STREAM_TYPE_AUDIO_AAC_ADTS: is_audio = TRUE; caps = gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 2, "stream-format", G_TYPE_STRING, "adts", NULL); break; case GST_MPEGTS_STREAM_TYPE_AUDIO_AAC_LATM: is_audio = TRUE; caps = gst_caps_new_simple ("audio/mpeg", "mpegversion", G_TYPE_INT, 4, "stream-format", G_TYPE_STRING, "loas", NULL); break; case GST_MPEGTS_STREAM_TYPE_VIDEO_MPEG4: is_video = TRUE; caps = gst_caps_new_simple ("video/mpeg", "mpegversion", G_TYPE_INT, 4, "systemstream", G_TYPE_BOOLEAN, FALSE, NULL); break; case GST_MPEGTS_STREAM_TYPE_VIDEO_H264: is_video = TRUE; caps = gst_caps_new_simple ("video/x-h264", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); break; case GST_MPEGTS_STREAM_TYPE_VIDEO_HEVC: is_video = TRUE; caps = gst_caps_new_simple ("video/x-h265", "stream-format", G_TYPE_STRING, "byte-stream", "alignment", G_TYPE_STRING, "nal", NULL); break; case ST_VIDEO_DIRAC: if (bstream->registration_id == 0x64726163) { GST_LOG ("dirac"); /* dirac in hex */ is_video = TRUE; caps = gst_caps_new_empty_simple ("video/x-dirac"); } break; case ST_PRIVATE_EA: /* Try to detect a VC1 stream */ { gboolean is_vc1 = FALSE; /* Note/FIXME: RP-227 specifies that the registration descriptor * for vc1 can also contain other information, such as profile, * level, alignment, buffer_size, .... */ if (bstream->registration_id == DRF_ID_VC1) is_vc1 = TRUE; if (!is_vc1) { GST_WARNING ("0xea private stream type found but no descriptor " "for VC1. Assuming plain VC1."); } is_video = TRUE; caps = gst_caps_new_simple ("video/x-wmv", "wmvversion", G_TYPE_INT, 3, "format", G_TYPE_STRING, "WVC1", NULL); break; } case ST_PS_AUDIO_AC3: /* DVB_ENHANCED_AC3 */ desc = mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_ENHANCED_AC3); if (desc) { is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-eac3"); break; } /* If stream has ac3 descriptor * OR program is ATSC (GA94) * OR stream registration is AC-3 * then it's regular AC3 */ if (bstream->registration_id == DRF_ID_AC3 || program->registration_id == DRF_ID_GA94 || mpegts_get_descriptor_from_stream (bstream, GST_MTS_DESC_DVB_AC3)) { is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); break; } GST_WARNING ("AC3 stream type found but no guaranteed " "way found to differentiate between AC3 and EAC3. " "Assuming plain AC3."); is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-ac3"); break; case ST_PS_AUDIO_DTS: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-dts"); break; case ST_PS_AUDIO_LPCM: is_audio = TRUE; caps = gst_caps_new_empty_simple ("audio/x-lpcm"); break; case ST_PS_DVD_SUBPICTURE: is_subpicture = TRUE; caps = gst_caps_new_empty_simple ("subpicture/x-dvd"); sparse = TRUE; break; case 0x42: /* hack for Chinese AVS video stream which use 0x42 as stream_id * NOTE: this is unofficial and within the ISO reserved range. */ is_video = TRUE; caps = gst_caps_new_empty_simple ("video/x-cavs"); break; default: GST_WARNING ("Non-media stream (stream_type:0x%x). Not creating pad", bstream->stream_type); break; } done: if (caps) { if (is_audio) { template = gst_static_pad_template_get (&audio_template); name = g_strdup_printf ("audio_%04x", bstream->pid); } else if (is_video) { template = gst_static_pad_template_get (&video_template); name = g_strdup_printf ("video_%04x", bstream->pid); } else if (is_private) { template = gst_static_pad_template_get (&private_template); name = g_strdup_printf ("private_%04x", bstream->pid); } else if (is_subpicture) { template = gst_static_pad_template_get (&subpicture_template); name = g_strdup_printf ("subpicture_%04x", bstream->pid); } else g_assert_not_reached (); } if (template && name && caps) { GstEvent *event; gchar *stream_id; GST_LOG ("stream:%p creating pad with name %s and caps %" GST_PTR_FORMAT, stream, name, caps); pad = gst_pad_new_from_template (template, name); gst_pad_set_active (pad, TRUE); gst_pad_use_fixed_caps (pad); stream_id = gst_pad_create_stream_id_printf (pad, GST_ELEMENT_CAST (base), "%08x", bstream->pid); event = gst_pad_get_sticky_event (base->sinkpad, GST_EVENT_STREAM_START, 0); if (event) { if (gst_event_parse_group_id (event, &demux->group_id)) demux->have_group_id = TRUE; else demux->have_group_id = FALSE; gst_event_unref (event); } else if (!demux->have_group_id) { demux->have_group_id = TRUE; demux->group_id = gst_util_group_id_next (); } event = gst_event_new_stream_start (stream_id); if (demux->have_group_id) gst_event_set_group_id (event, demux->group_id); if (sparse) gst_event_set_stream_flags (event, GST_STREAM_FLAG_SPARSE); stream->sparse = sparse; gst_pad_push_event (pad, event); g_free (stream_id); gst_pad_set_caps (pad, caps); if (!stream->taglist) stream->taglist = gst_tag_list_new_empty (); gst_pb_utils_add_codec_description_to_tag_list (stream->taglist, NULL, caps); gst_pad_set_query_function (pad, gst_ts_demux_srcpad_query); gst_pad_set_event_function (pad, gst_ts_demux_srcpad_event); } if (name) g_free (name); if (template) gst_object_unref (template); if (caps) gst_caps_unref (caps); return pad; } static void gst_ts_demux_stream_added (MpegTSBase * base, MpegTSBaseStream * bstream, MpegTSBaseProgram * program) { GstTSDemux *demux = (GstTSDemux *) base; TSDemuxStream *stream = (TSDemuxStream *) bstream; if (!stream->pad) { /* Create the pad */ if (bstream->stream_type != 0xff) { stream->pad = create_pad_for_stream (base, bstream, program); if (stream->pad) gst_flow_combiner_add_pad (demux->flowcombiner, stream->pad); } if (base->mode != BASE_MODE_PUSHING && bstream->stream_type == GST_MPEGTS_STREAM_TYPE_VIDEO_H264) { stream->scan_function = (GstTsDemuxKeyFrameScanFunction) scan_keyframe_h264; } else { stream->scan_function = NULL; } stream->active = FALSE; stream->need_newsegment = TRUE; demux->reset_segment = TRUE; stream->needs_keyframe = FALSE; stream->discont = TRUE; stream->pts = GST_CLOCK_TIME_NONE; stream->dts = GST_CLOCK_TIME_NONE; stream->first_dts = GST_CLOCK_TIME_NONE; stream->raw_pts = -1; stream->raw_dts = -1; stream->pending_ts = TRUE; stream->nb_out_buffers = 0; stream->gap_ref_buffers = 0; stream->gap_ref_pts = GST_CLOCK_TIME_NONE; stream->continuity_counter = CONTINUITY_UNSET; } } static void tsdemux_h264_parsing_info_clear (TSDemuxH264ParsingInfos * h264infos) { clear_simple_buffer (&h264infos->framedata); if (h264infos->parser) { gst_h264_nal_parser_free (h264infos->parser); gst_byte_writer_free (h264infos->sps); gst_byte_writer_free (h264infos->pps); gst_byte_writer_free (h264infos->sei); } } static void gst_ts_demux_stream_removed (MpegTSBase * base, MpegTSBaseStream * bstream) { TSDemuxStream *stream = (TSDemuxStream *) bstream; if (stream->pad) { gst_flow_combiner_remove_pad (GST_TS_DEMUX_CAST (base)->flowcombiner, stream->pad); if (stream->active) { if (gst_pad_is_active (stream->pad)) { /* Flush out all data */ GST_DEBUG_OBJECT (stream->pad, "Flushing out pending data"); gst_ts_demux_push_pending_data ((GstTSDemux *) base, stream); GST_DEBUG_OBJECT (stream->pad, "Pushing out EOS"); gst_pad_push_event (stream->pad, gst_event_new_eos ()); gst_pad_set_active (stream->pad, FALSE); } GST_DEBUG_OBJECT (stream->pad, "Removing pad"); gst_element_remove_pad (GST_ELEMENT_CAST (base), stream->pad); stream->active = FALSE; } stream->pad = NULL; } gst_ts_demux_stream_flush (stream, GST_TS_DEMUX_CAST (base), TRUE); if (stream->taglist != NULL) { gst_tag_list_unref (stream->taglist); stream->taglist = NULL; } tsdemux_h264_parsing_info_clear (&stream->h264infos); } static void activate_pad_for_stream (GstTSDemux * tsdemux, TSDemuxStream * stream) { if (stream->pad) { GST_DEBUG_OBJECT (tsdemux, "Activating pad %s:%s for stream %p", GST_DEBUG_PAD_NAME (stream->pad), stream); gst_element_add_pad ((GstElement *) tsdemux, stream->pad); stream->active = TRUE; GST_DEBUG_OBJECT (stream->pad, "done adding pad"); /* force sending of pending sticky events which have been stored on the * pad already and which otherwise would only be sent on the first buffer * or serialized event (which means very late in case of subtitle streams), * and playsink waits for stream-start or another serialized event */ if (stream->sparse) { GST_DEBUG_OBJECT (stream->pad, "sparse stream, pushing GAP event"); gst_pad_push_event (stream->pad, gst_event_new_gap (0, 0)); } } else if (((MpegTSBaseStream *) stream)->stream_type != 0xff) { GST_WARNING_OBJECT (tsdemux, "stream %p (pid 0x%04x, type:0x%02x) has no pad", stream, ((MpegTSBaseStream *) stream)->pid, ((MpegTSBaseStream *) stream)->stream_type); } } static void gst_ts_demux_stream_flush (TSDemuxStream * stream, GstTSDemux * tsdemux, gboolean hard) { GST_DEBUG ("flushing stream %p", stream); if (stream->data) g_free (stream->data); stream->data = NULL; stream->state = PENDING_PACKET_EMPTY; stream->expected_size = 0; stream->allocated_size = 0; stream->current_size = 0; stream->discont = TRUE; stream->pts = GST_CLOCK_TIME_NONE; stream->dts = GST_CLOCK_TIME_NONE; stream->raw_pts = -1; stream->raw_dts = -1; stream->pending_ts = TRUE; stream->nb_out_buffers = 0; stream->gap_ref_buffers = 0; stream->gap_ref_pts = GST_CLOCK_TIME_NONE; stream->continuity_counter = CONTINUITY_UNSET; if (hard) { stream->first_dts = GST_CLOCK_TIME_NONE; stream->need_newsegment = TRUE; } } static void gst_ts_demux_flush_streams (GstTSDemux * demux, gboolean hard) { GList *walk; if (!demux->program) return; for (walk = demux->program->stream_list; walk; walk = g_list_next (walk)) gst_ts_demux_stream_flush (walk->data, demux, hard); } static void gst_ts_demux_program_started (MpegTSBase * base, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); GST_DEBUG ("Current program %d, new program %d requested program %d", (gint) demux->program_number, program->program_number, demux->requested_program_number); if (demux->requested_program_number == program->program_number || (demux->requested_program_number == -1 && demux->program_number == -1)) { GList *tmp; GST_LOG ("program %d started", program->program_number); demux->program_number = program->program_number; demux->program = program; /* If this is not the initial program, we need to calculate * a new segment */ if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } /* Add all streams, then fire no-more-pads */ for (tmp = program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; activate_pad_for_stream (demux, stream); } gst_element_no_more_pads ((GstElement *) demux); } } static void gst_ts_demux_program_stopped (MpegTSBase * base, MpegTSBaseProgram * program) { GstTSDemux *demux = GST_TS_DEMUX (base); if (demux->program == program) { demux->program = NULL; demux->program_number = -1; } } static inline void gst_ts_demux_record_pts (GstTSDemux * demux, TSDemuxStream * stream, guint64 pts, guint64 offset) { MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; stream->raw_pts = pts; if (pts == -1) { stream->pts = GST_CLOCK_TIME_NONE; return; } GST_LOG ("pid 0x%04x raw pts:%" G_GUINT64_FORMAT " at offset %" G_GUINT64_FORMAT, bs->pid, pts, offset); /* Compute PTS in GstClockTime */ stream->pts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (pts), demux->program->pcr_pid); GST_LOG ("pid 0x%04x Stored PTS %" G_GUINT64_FORMAT, bs->pid, stream->pts); if (G_UNLIKELY (demux->emit_statistics)) { GstStructure *st; st = gst_structure_new_id_empty (QUARK_TSDEMUX); gst_structure_id_set (st, QUARK_PID, G_TYPE_UINT, bs->pid, QUARK_OFFSET, G_TYPE_UINT64, offset, QUARK_PTS, G_TYPE_UINT64, pts, NULL); gst_element_post_message (GST_ELEMENT_CAST (demux), gst_message_new_element (GST_OBJECT (demux), st)); } } static inline void gst_ts_demux_record_dts (GstTSDemux * demux, TSDemuxStream * stream, guint64 dts, guint64 offset) { MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; stream->raw_dts = dts; if (dts == -1) { stream->dts = GST_CLOCK_TIME_NONE; return; } GST_LOG ("pid 0x%04x raw dts:%" G_GUINT64_FORMAT " at offset %" G_GUINT64_FORMAT, bs->pid, dts, offset); /* Compute DTS in GstClockTime */ stream->dts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (dts), demux->program->pcr_pid); GST_LOG ("pid 0x%04x Stored DTS %" G_GUINT64_FORMAT, bs->pid, stream->dts); if (G_UNLIKELY (demux->emit_statistics)) { GstStructure *st; st = gst_structure_new_id_empty (QUARK_TSDEMUX); gst_structure_id_set (st, QUARK_PID, G_TYPE_UINT, bs->pid, QUARK_OFFSET, G_TYPE_UINT64, offset, QUARK_DTS, G_TYPE_UINT64, dts, NULL); gst_element_post_message (GST_ELEMENT_CAST (demux), gst_message_new_element (GST_OBJECT (demux), st)); } } /* This is called when we haven't got a valid initial PTS/DTS on all streams */ static gboolean check_pending_buffers (GstTSDemux * demux) { gboolean have_observation = FALSE; /* The biggest offset */ guint64 offset = 0; GList *tmp; /* 1. Go over all streams */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; /* 1.1 check if at least one stream got a valid DTS */ if ((tmpstream->raw_dts != -1 && tmpstream->dts != GST_CLOCK_TIME_NONE) || (tmpstream->raw_pts != -1 && tmpstream->pts != GST_CLOCK_TIME_NONE)) { have_observation = TRUE; break; } } /* 2. If we don't have a valid value yet, break out */ if (have_observation == FALSE) return FALSE; /* 3. Go over all streams that have current/pending data */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *tmpstream = (TSDemuxStream *) tmp->data; PendingBuffer *pend; guint64 firstval, lastval, ts; /* 3.1 Calculate the offset between current DTS and first DTS */ if (tmpstream->pending == NULL || tmpstream->state == PENDING_PACKET_EMPTY) continue; /* If we don't have any pending data, the offset is 0 for this stream */ if (tmpstream->pending == NULL) break; if (tmpstream->raw_dts != -1) lastval = tmpstream->raw_dts; else if (tmpstream->raw_pts != -1) lastval = tmpstream->raw_pts; else { GST_WARNING ("Don't have a last DTS/PTS to use for offset recalculation"); continue; } pend = tmpstream->pending->data; if (pend->dts != -1) firstval = pend->dts; else if (pend->pts != -1) firstval = pend->pts; else { GST_WARNING ("Don't have a first DTS/PTS to use for offset recalculation"); continue; } /* 3.2 Add to the offset the report TS for the current DTS */ ts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (lastval), demux->program->pcr_pid); if (ts == GST_CLOCK_TIME_NONE) { GST_WARNING ("THIS SHOULD NOT HAPPEN !"); continue; } ts += MPEGTIME_TO_GSTTIME (lastval - firstval); /* 3.3 If that offset is bigger than the current offset, store it */ if (ts > offset) offset = ts; } GST_DEBUG ("New initial pcr_offset %" GST_TIME_FORMAT, GST_TIME_ARGS (offset)); /* 4. Set the offset on the packetizer */ mpegts_packetizer_set_current_pcr_offset (MPEG_TS_BASE_PACKETIZER (demux), offset, demux->program->pcr_pid); /* 4. Go over all streams */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; stream->pending_ts = FALSE; /* 4.1 Set pending_ts for FALSE */ /* 4.2 Recalculate PTS/DTS (in running time) for pending data */ if (stream->pending) { GList *tmp2; for (tmp2 = stream->pending; tmp2; tmp2 = tmp2->next) { PendingBuffer *pend = (PendingBuffer *) tmp2->data; if (pend->pts != -1) GST_BUFFER_PTS (pend->buffer) = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (pend->pts), demux->program->pcr_pid); if (pend->dts != -1) GST_BUFFER_DTS (pend->buffer) = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (pend->dts), demux->program->pcr_pid); /* 4.2.2 Set first_dts to TS of lowest DTS (for segment) */ if (stream->first_dts == GST_CLOCK_TIME_NONE) { if (GST_BUFFER_DTS (pend->buffer) != GST_CLOCK_TIME_NONE) stream->first_dts = GST_BUFFER_DTS (pend->buffer); else if (GST_BUFFER_PTS (pend->buffer) != GST_CLOCK_TIME_NONE) stream->first_dts = GST_BUFFER_PTS (pend->buffer); } } } /* Recalculate PTS/DTS (in running time) for current data */ if (stream->state != PENDING_PACKET_EMPTY) { if (stream->raw_dts != -1) { stream->dts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (stream->raw_dts), demux->program->pcr_pid); if (stream->first_dts == GST_CLOCK_TIME_NONE) stream->first_dts = stream->dts; } if (stream->raw_pts != -1) { stream->pts = mpegts_packetizer_pts_to_ts (MPEG_TS_BASE_PACKETIZER (demux), MPEGTIME_TO_GSTTIME (stream->raw_pts), demux->program->pcr_pid); if (stream->first_dts == GST_CLOCK_TIME_NONE) stream->first_dts = stream->pts; } } } return TRUE; } static void gst_ts_demux_parse_pes_header (GstTSDemux * demux, TSDemuxStream * stream, guint8 * data, guint32 length, guint64 bufferoffset) { PESHeader header; PESParsingResult parseres; GST_MEMDUMP ("Header buffer", data, MIN (length, 32)); parseres = mpegts_parse_pes_header (data, length, &header); if (G_UNLIKELY (parseres == PES_PARSING_NEED_MORE)) goto discont; if (G_UNLIKELY (parseres == PES_PARSING_BAD)) { GST_WARNING ("Error parsing PES header. pid: 0x%x stream_type: 0x%x", stream->stream.pid, stream->stream.stream_type); goto discont; } if (stream->target_pes_substream != 0 && header.stream_id_extension != stream->target_pes_substream) { GST_DEBUG ("Skipping unwanted substream"); goto discont; } gst_ts_demux_record_dts (demux, stream, header.DTS, bufferoffset); gst_ts_demux_record_pts (demux, stream, header.PTS, bufferoffset); if (G_UNLIKELY (stream->pending_ts && (stream->pts != GST_CLOCK_TIME_NONE || stream->dts != GST_CLOCK_TIME_NONE))) { GST_DEBUG ("Got pts/dts update, rechecking all streams"); check_pending_buffers (demux); } else if (stream->first_dts == GST_CLOCK_TIME_NONE) { if (GST_CLOCK_TIME_IS_VALID (stream->dts)) stream->first_dts = stream->dts; else if (GST_CLOCK_TIME_IS_VALID (stream->pts)) stream->first_dts = stream->pts; } GST_DEBUG_OBJECT (demux, "stream PTS %" GST_TIME_FORMAT " DTS %" GST_TIME_FORMAT, GST_TIME_ARGS (stream->pts), GST_TIME_ARGS (stream->dts)); /* Remove PES headers */ GST_DEBUG ("Moving data forward by %d bytes (packet_size:%d, have:%d)", header.header_size, header.packet_length, length); stream->expected_size = header.packet_length; if (stream->expected_size) { if (G_LIKELY (stream->expected_size > header.header_size)) { stream->expected_size -= header.header_size; } else { /* next packet will have to complete this one */ GST_ERROR ("invalid header and packet size combination"); stream->expected_size = 0; } } data += header.header_size; length -= header.header_size; /* Create the output buffer */ if (stream->expected_size) stream->allocated_size = MAX (stream->expected_size, length); else stream->allocated_size = MAX (8192, length); g_assert (stream->data == NULL); stream->data = g_malloc (stream->allocated_size); memcpy (stream->data, data, length); stream->current_size = length; stream->state = PENDING_PACKET_BUFFER; return; discont: stream->state = PENDING_PACKET_DISCONT; return; } /* ONLY CALL THIS: * * WITH packet->payload != NULL * * WITH pending/current flushed out if beginning of new PES packet */ static inline void gst_ts_demux_queue_data (GstTSDemux * demux, TSDemuxStream * stream, MpegTSPacketizerPacket * packet) { guint8 *data; guint size; guint8 cc = FLAGS_CONTINUITY_COUNTER (packet->scram_afc_cc); GST_LOG ("pid: 0x%04x state:%d", stream->stream.pid, stream->state); size = packet->data_end - packet->payload; data = packet->payload; if (stream->continuity_counter == CONTINUITY_UNSET) { GST_DEBUG ("CONTINUITY: Initialize to %d", cc); } else if ((cc == stream->continuity_counter + 1 || (stream->continuity_counter == MAX_CONTINUITY && cc == 0))) { GST_LOG ("CONTINUITY: Got expected %d", cc); } else { GST_WARNING ("CONTINUITY: Mismatch packet %d, stream %d", cc, stream->continuity_counter); if (stream->state != PENDING_PACKET_EMPTY) stream->state = PENDING_PACKET_DISCONT; } stream->continuity_counter = cc; if (stream->state == PENDING_PACKET_EMPTY) { if (G_UNLIKELY (!packet->payload_unit_start_indicator)) { stream->state = PENDING_PACKET_DISCONT; GST_DEBUG ("Didn't get the first packet of this PES"); } else { GST_LOG ("EMPTY=>HEADER"); stream->state = PENDING_PACKET_HEADER; } } switch (stream->state) { case PENDING_PACKET_HEADER: { GST_LOG ("HEADER: Parsing PES header"); /* parse the header */ gst_ts_demux_parse_pes_header (demux, stream, data, size, packet->offset); break; } case PENDING_PACKET_BUFFER: { GST_LOG ("BUFFER: appending data"); if (G_UNLIKELY (stream->current_size + size > stream->allocated_size)) { GST_LOG ("resizing buffer"); do { stream->allocated_size *= 2; } while (stream->current_size + size > stream->allocated_size); stream->data = g_realloc (stream->data, stream->allocated_size); } memcpy (stream->data + stream->current_size, data, size); stream->current_size += size; break; } case PENDING_PACKET_DISCONT: { GST_LOG ("DISCONT: not storing/pushing"); if (G_UNLIKELY (stream->data)) { g_free (stream->data); stream->data = NULL; } stream->continuity_counter = CONTINUITY_UNSET; break; } default: break; } return; } static void calculate_and_push_newsegment (GstTSDemux * demux, TSDemuxStream * stream) { MpegTSBase *base = (MpegTSBase *) demux; GstClockTime lowest_pts = GST_CLOCK_TIME_NONE; GstClockTime firstts = 0; GList *tmp; GST_DEBUG ("Creating new newsegment for stream %p", stream); /* Speedup : if we don't need to calculate anything, go straight to pushing */ if (demux->segment_event) goto push_new_segment; /* Calculate the 'new_start' value, used for newsegment */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *pstream = (TSDemuxStream *) tmp->data; if (GST_CLOCK_TIME_IS_VALID (pstream->first_dts)) { if (!GST_CLOCK_TIME_IS_VALID (lowest_pts) || pstream->first_dts < lowest_pts) lowest_pts = pstream->first_dts; } } if (GST_CLOCK_TIME_IS_VALID (lowest_pts)) firstts = lowest_pts; GST_DEBUG ("lowest_pts %" G_GUINT64_FORMAT " => clocktime %" GST_TIME_FORMAT, lowest_pts, GST_TIME_ARGS (firstts)); if (demux->segment.format != GST_FORMAT_TIME || demux->reset_segment) { /* It will happen only if it's first program or after flushes. */ GST_DEBUG ("Calculating actual segment"); if (base->segment.format == GST_FORMAT_TIME) { /* Try to recover segment info from base if it's in TIME format */ demux->segment = base->segment; } else { /* Start from the first ts/pts */ GstClockTime base = demux->segment.base + demux->segment.position - demux->segment.start; gst_segment_init (&demux->segment, GST_FORMAT_TIME); demux->segment.start = firstts; demux->segment.stop = GST_CLOCK_TIME_NONE; demux->segment.position = firstts; demux->segment.time = firstts; demux->segment.rate = demux->rate; demux->segment.base = base; } } else if (demux->segment.start < firstts) { /* Take into account the offset to the first buffer timestamp */ if (demux->segment.rate > 0) { demux->segment.start = firstts; if (GST_CLOCK_TIME_IS_VALID (demux->segment.stop)) demux->segment.stop += firstts - demux->segment.start; demux->segment.position = firstts; } } if (!demux->segment_event) { demux->segment_event = gst_event_new_segment (&demux->segment); GST_EVENT_SEQNUM (demux->segment_event) = base->last_seek_seqnum; } push_new_segment: for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { stream = (TSDemuxStream *) tmp->data; if (stream->pad == NULL) continue; if (demux->segment_event) { GST_DEBUG_OBJECT (stream->pad, "Pushing newsegment event"); gst_event_ref (demux->segment_event); gst_pad_push_event (stream->pad, demux->segment_event); } if (demux->global_tags) { gst_pad_push_event (stream->pad, gst_event_new_tag (gst_tag_list_ref (demux->global_tags))); } /* Push pending tags */ if (stream->taglist) { GST_DEBUG_OBJECT (stream->pad, "Sending tags %" GST_PTR_FORMAT, stream->taglist); gst_pad_push_event (stream->pad, gst_event_new_tag (stream->taglist)); stream->taglist = NULL; } stream->need_newsegment = FALSE; } } static void gst_ts_demux_check_and_sync_streams (GstTSDemux * demux, GstClockTime time) { GList *tmp; GST_DEBUG_OBJECT (demux, "Recheck streams and sync to at least: %" GST_TIME_FORMAT, GST_TIME_ARGS (time)); if (G_UNLIKELY (demux->program == NULL)) return; /* Go over each stream and update it to at least 'time' time. * For each stream, the pad stores the buffer counter the last time * a gap check occurred (gap_ref_buffers) and a gap_ref_pts timestamp * that is either the PTS from the stream or the PCR the pad was updated * to. * * We can check nb_out_buffers to see if any buffers were pushed since then. * This means we can detect buffers passing without PTSes fine and still generate * gaps. * * If there haven't been any buffers pushed on this stream since the last * gap check, push a gap event updating to the indicated input PCR time * and update the pad's tracking. * * If there have been buffers pushed, update the reference buffer count * and but don't push a gap event */ for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *ps = (TSDemuxStream *) tmp->data; GST_DEBUG_OBJECT (ps->pad, "0x%04x, PTS:%" GST_TIME_FORMAT " REFPTS:%" GST_TIME_FORMAT " Gap:%" GST_TIME_FORMAT " nb_buffers: %d (ref:%d)", ((MpegTSBaseStream *) ps)->pid, GST_TIME_ARGS (ps->pts), GST_TIME_ARGS (ps->gap_ref_pts), GST_TIME_ARGS (ps->pts - ps->gap_ref_pts), ps->nb_out_buffers, ps->gap_ref_buffers); if (ps->pad == NULL) continue; if (ps->nb_out_buffers == ps->gap_ref_buffers && ps->gap_ref_pts != ps->pts) { /* Do initial setup of pad if needed - segment etc */ GST_DEBUG_OBJECT (ps->pad, "Stream needs update. Pushing GAP event to TS %" GST_TIME_FORMAT, GST_TIME_ARGS (time)); if (G_UNLIKELY (ps->need_newsegment)) calculate_and_push_newsegment (demux, ps); /* Now send gap event */ gst_pad_push_event (ps->pad, gst_event_new_gap (time, 0)); } /* Update GAP tracking vars so we don't re-check this stream for a while */ ps->gap_ref_pts = time; if (ps->pts != GST_CLOCK_TIME_NONE && ps->pts > time) ps->gap_ref_pts = ps->pts; ps->gap_ref_buffers = ps->nb_out_buffers; } } static GstFlowReturn gst_ts_demux_push_pending_data (GstTSDemux * demux, TSDemuxStream * stream) { GstFlowReturn res = GST_FLOW_OK; #ifndef GST_DISABLE_GST_DEBUG MpegTSBaseStream *bs = (MpegTSBaseStream *) stream; #endif GstBuffer *buffer = NULL; GST_DEBUG_OBJECT (stream->pad, "stream:%p, pid:0x%04x stream_type:%d state:%d", stream, bs->pid, bs->stream_type, stream->state); if (G_UNLIKELY (stream->data == NULL)) { GST_LOG ("stream->data == NULL"); goto beach; } if (G_UNLIKELY (stream->state == PENDING_PACKET_EMPTY)) { GST_LOG ("EMPTY: returning"); goto beach; } if (G_UNLIKELY (stream->state != PENDING_PACKET_BUFFER)) { GST_LOG ("state:%d, returning", stream->state); goto beach; } if (G_UNLIKELY (demux->program == NULL)) { GST_LOG_OBJECT (demux, "No program"); g_free (stream->data); goto beach; } if (stream->needs_keyframe) { MpegTSBase *base = (MpegTSBase *) demux; if ((gst_ts_demux_adjust_seek_offset_for_keyframe (stream, stream->data, stream->current_size)) || demux->last_seek_offset == 0) { GST_DEBUG_OBJECT (stream->pad, "Got Keyframe, ready to go at %" GST_TIME_FORMAT, GST_TIME_ARGS (stream->pts)); buffer = gst_buffer_new_wrapped (stream->data, stream->current_size); stream->seeked_pts = stream->pts; stream->seeked_dts = stream->dts; stream->needs_keyframe = FALSE; } else { base->seek_offset = demux->last_seek_offset - 200 * base->packetsize; if (demux->last_seek_offset < 200 * base->packetsize) base->seek_offset = 0; demux->last_seek_offset = base->seek_offset; mpegts_packetizer_flush (base->packetizer, FALSE); base->mode = BASE_MODE_SEEKING; stream->continuity_counter = CONTINUITY_UNSET; res = GST_FLOW_REWINDING; g_free (stream->data); goto beach; } } else { buffer = gst_buffer_new_wrapped (stream->data, stream->current_size); if (G_UNLIKELY (stream->pending_ts && !check_pending_buffers (demux))) { PendingBuffer *pend; pend = g_slice_new0 (PendingBuffer); pend->buffer = buffer; pend->pts = stream->raw_pts; pend->dts = stream->raw_dts; stream->pending = g_list_append (stream->pending, pend); GST_DEBUG ("Not enough information to push buffers yet, storing buffer"); goto beach; } } if (G_UNLIKELY (stream->need_newsegment)) calculate_and_push_newsegment (demux, stream); /* FIXME : Push pending buffers if any */ if (G_UNLIKELY (stream->pending)) { GList *tmp; for (tmp = stream->pending; tmp; tmp = tmp->next) { PendingBuffer *pend = (PendingBuffer *) tmp->data; GST_DEBUG_OBJECT (stream->pad, "Pushing pending buffer PTS:%" GST_TIME_FORMAT " DTS:%" GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_PTS (pend->buffer)), GST_TIME_ARGS (GST_BUFFER_DTS (pend->buffer))); if (stream->discont) GST_BUFFER_FLAG_SET (pend->buffer, GST_BUFFER_FLAG_DISCONT); stream->discont = FALSE; res = gst_pad_push (stream->pad, pend->buffer); stream->nb_out_buffers += 1; g_slice_free (PendingBuffer, pend); } g_list_free (stream->pending); stream->pending = NULL; } if ((GST_CLOCK_TIME_IS_VALID (stream->seeked_pts) && stream->pts < stream->seeked_pts) || (GST_CLOCK_TIME_IS_VALID (stream->seeked_dts) && stream->pts < stream->seeked_dts)) { GST_INFO_OBJECT (stream->pad, "Droping with PTS: %" GST_TIME_FORMAT " DTS: %" GST_TIME_FORMAT " after seeking as other stream needed to be seeked further" "(seeked PTS: %" GST_TIME_FORMAT " DTS: %" GST_TIME_FORMAT ")", GST_TIME_ARGS (stream->pts), GST_TIME_ARGS (stream->dts), GST_TIME_ARGS (stream->seeked_pts), GST_TIME_ARGS (stream->seeked_dts)); gst_buffer_unref (buffer); goto beach; } GST_DEBUG_OBJECT (stream->pad, "stream->pts %" GST_TIME_FORMAT, GST_TIME_ARGS (stream->pts)); if (GST_CLOCK_TIME_IS_VALID (stream->pts)) GST_BUFFER_PTS (buffer) = stream->pts; if (GST_CLOCK_TIME_IS_VALID (stream->dts)) GST_BUFFER_DTS (buffer) = stream->dts; GST_DEBUG_OBJECT (stream->pad, "Pushing buffer with PTS: %" GST_TIME_FORMAT " , DTS: %" GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_PTS (buffer)), GST_TIME_ARGS (GST_BUFFER_DTS (buffer))); if (stream->discont) GST_BUFFER_FLAG_SET (buffer, GST_BUFFER_FLAG_DISCONT); stream->discont = FALSE; if (GST_CLOCK_TIME_IS_VALID (GST_BUFFER_DTS (buffer))) demux->segment.position = GST_BUFFER_DTS (buffer); else if (GST_CLOCK_TIME_IS_VALID (GST_BUFFER_PTS (buffer))) demux->segment.position = GST_BUFFER_PTS (buffer); res = gst_pad_push (stream->pad, buffer); /* Record that a buffer was pushed */ stream->nb_out_buffers += 1; GST_DEBUG_OBJECT (stream->pad, "Returned %s", gst_flow_get_name (res)); res = gst_flow_combiner_update_flow (demux->flowcombiner, res); GST_DEBUG_OBJECT (stream->pad, "combined %s", gst_flow_get_name (res)); /* GAP / sparse stream tracking */ if (G_UNLIKELY (stream->gap_ref_pts == GST_CLOCK_TIME_NONE)) stream->gap_ref_pts = stream->pts; else { /* Look if the stream PTS has advanced 2 seconds since the last * gap check, and sync streams if it has. The first stream to * hit this will trigger a gap check */ if (G_UNLIKELY (stream->pts != GST_CLOCK_TIME_NONE && stream->pts > stream->gap_ref_pts + 2 * GST_SECOND)) { GstClockTime curpcr = mpegts_packetizer_get_current_time (MPEG_TS_BASE_PACKETIZER (demux), demux->program->pcr_pid); if (curpcr == GST_CLOCK_TIME_NONE || curpcr < 800 * GST_MSECOND) goto beach; curpcr -= 800 * GST_MSECOND; gst_ts_demux_check_and_sync_streams (demux, curpcr); } } beach: /* Reset everything */ GST_LOG ("Resetting to EMPTY, returning %s", gst_flow_get_name (res)); stream->state = PENDING_PACKET_EMPTY; stream->data = NULL; stream->expected_size = 0; stream->current_size = 0; return res; } static GstFlowReturn gst_ts_demux_handle_packet (GstTSDemux * demux, TSDemuxStream * stream, MpegTSPacketizerPacket * packet, GstMpegtsSection * section) { GstFlowReturn res = GST_FLOW_OK; GST_LOG ("pid 0x%04x pusi:%d, afc:%d, cont:%d, payload:%p", packet->pid, packet->payload_unit_start_indicator, packet->scram_afc_cc & 0x30, FLAGS_CONTINUITY_COUNTER (packet->scram_afc_cc), packet->payload); if (G_UNLIKELY (packet->payload_unit_start_indicator) && FLAGS_HAS_PAYLOAD (packet->scram_afc_cc)) /* Flush previous data */ res = gst_ts_demux_push_pending_data (demux, stream); if (packet->payload && (res == GST_FLOW_OK || res == GST_FLOW_NOT_LINKED) && stream->pad) { gst_ts_demux_queue_data (demux, stream, packet); GST_LOG ("current_size:%d, expected_size:%d", stream->current_size, stream->expected_size); /* Finally check if the data we queued completes a packet */ if (stream->expected_size && stream->current_size == stream->expected_size) { GST_LOG ("pushing complete packet"); res = gst_ts_demux_push_pending_data (demux, stream); } } /* We are rewinding to find a keyframe, * and didn't want the data to be queued */ if (res == GST_FLOW_REWINDING) res = GST_FLOW_OK; return res; } static void gst_ts_demux_flush (MpegTSBase * base, gboolean hard) { GstTSDemux *demux = GST_TS_DEMUX_CAST (base); gst_ts_demux_flush_streams (demux, hard); if (demux->segment_event) { gst_event_unref (demux->segment_event); demux->segment_event = NULL; } if (demux->global_tags) { gst_tag_list_unref (demux->global_tags); demux->global_tags = NULL; } if (hard) { /* For pull mode seeks the current segment needs to be preserved */ demux->rate = 1.0; gst_segment_init (&demux->segment, GST_FORMAT_UNDEFINED); } } static GstFlowReturn gst_ts_demux_drain (MpegTSBase * base) { GstTSDemux *demux = GST_TS_DEMUX_CAST (base); GList *tmp; GstFlowReturn res = GST_FLOW_OK; if (!demux->program) return res; for (tmp = demux->program->stream_list; tmp; tmp = tmp->next) { TSDemuxStream *stream = (TSDemuxStream *) tmp->data; if (stream->pad) { res = gst_ts_demux_push_pending_data (demux, stream); if (G_UNLIKELY (res != GST_FLOW_OK)) break; } } return res; } static GstFlowReturn gst_ts_demux_push (MpegTSBase * base, MpegTSPacketizerPacket * packet, GstMpegtsSection * section) { GstTSDemux *demux = GST_TS_DEMUX_CAST (base); TSDemuxStream *stream = NULL; GstFlowReturn res = GST_FLOW_OK; if (G_LIKELY (demux->program)) { stream = (TSDemuxStream *) demux->program->streams[packet->pid]; if (stream) { res = gst_ts_demux_handle_packet (demux, stream, packet, section); } } return res; } gboolean gst_ts_demux_plugin_init (GstPlugin * plugin) { GST_DEBUG_CATEGORY_INIT (ts_demux_debug, "tsdemux", 0, "MPEG transport stream demuxer"); init_pes_parser (); return gst_element_register (plugin, "tsdemux", GST_RANK_PRIMARY, GST_TYPE_TS_DEMUX); }