mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-24 02:31:03 +00:00
2202 lines
63 KiB
C
2202 lines
63 KiB
C
/* GStreamer
|
|
* Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>,
|
|
* <2006> Edward Hervey <bilboed@bilboed.com>
|
|
* <2006> Wim Taymans <wim@fluendo.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include <string.h>
|
|
|
|
#include <libavformat/avformat.h>
|
|
#include <libavutil/imgutils.h>
|
|
/* #include <ffmpeg/avi.h> */
|
|
#include <gst/gst.h>
|
|
#include <gst/base/gstflowcombiner.h>
|
|
|
|
#include "gstav.h"
|
|
#include "gstavcodecmap.h"
|
|
#include "gstavutils.h"
|
|
#include "gstavprotocol.h"
|
|
|
|
#define MAX_STREAMS 20
|
|
|
|
typedef struct _GstFFMpegDemux GstFFMpegDemux;
|
|
typedef struct _GstFFStream GstFFStream;
|
|
|
|
struct _GstFFStream
|
|
{
|
|
GstPad *pad;
|
|
|
|
AVStream *avstream;
|
|
|
|
gboolean unknown;
|
|
GstClockTime last_ts;
|
|
gboolean discont;
|
|
gboolean eos;
|
|
|
|
GstTagList *tags; /* stream tags */
|
|
};
|
|
|
|
struct _GstFFMpegDemux
|
|
{
|
|
GstElement element;
|
|
|
|
/* We need to keep track of our pads, so we do so here. */
|
|
GstPad *sinkpad;
|
|
|
|
gboolean have_group_id;
|
|
guint group_id;
|
|
|
|
AVFormatContext *context;
|
|
gboolean opened;
|
|
|
|
GstFFStream *streams[MAX_STREAMS];
|
|
|
|
GstFlowCombiner *flowcombiner;
|
|
|
|
gint videopads, audiopads;
|
|
|
|
GstClockTime start_time;
|
|
GstClockTime duration;
|
|
|
|
/* TRUE if working in pull-mode */
|
|
gboolean seekable;
|
|
|
|
/* TRUE if the avformat demuxer can reliably handle streaming mode */
|
|
gboolean can_push;
|
|
|
|
gboolean flushing;
|
|
|
|
/* segment stuff */
|
|
GstSegment segment;
|
|
|
|
/* cached seek in READY */
|
|
GstEvent *seek_event;
|
|
|
|
/* cached upstream events */
|
|
GList *cached_events;
|
|
|
|
/* push mode data */
|
|
GstFFMpegPipe ffpipe;
|
|
GstTask *task;
|
|
GRecMutex task_lock;
|
|
};
|
|
|
|
typedef struct _GstFFMpegDemuxClass GstFFMpegDemuxClass;
|
|
|
|
struct _GstFFMpegDemuxClass
|
|
{
|
|
GstElementClass parent_class;
|
|
|
|
AVInputFormat *in_plugin;
|
|
GstPadTemplate *sinktempl;
|
|
GstPadTemplate *videosrctempl;
|
|
GstPadTemplate *audiosrctempl;
|
|
};
|
|
|
|
/* A number of function prototypes are given so we can refer to them later. */
|
|
static void gst_ffmpegdemux_class_init (GstFFMpegDemuxClass * klass);
|
|
static void gst_ffmpegdemux_base_init (GstFFMpegDemuxClass * klass);
|
|
static void gst_ffmpegdemux_init (GstFFMpegDemux * demux);
|
|
static void gst_ffmpegdemux_finalize (GObject * object);
|
|
|
|
static gboolean gst_ffmpegdemux_sink_event (GstPad * sinkpad,
|
|
GstObject * parent, GstEvent * event);
|
|
static GstFlowReturn gst_ffmpegdemux_chain (GstPad * sinkpad,
|
|
GstObject * parent, GstBuffer * buf);
|
|
|
|
static void gst_ffmpegdemux_loop (GstFFMpegDemux * demux);
|
|
static gboolean gst_ffmpegdemux_sink_activate (GstPad * sinkpad,
|
|
GstObject * parent);
|
|
static gboolean gst_ffmpegdemux_sink_activate_mode (GstPad * sinkpad,
|
|
GstObject * parent, GstPadMode mode, gboolean active);
|
|
static GstTagList *gst_ffmpeg_metadata_to_tag_list (AVDictionary * metadata);
|
|
|
|
#if 0
|
|
static gboolean
|
|
gst_ffmpegdemux_src_convert (GstPad * pad,
|
|
GstFormat src_fmt,
|
|
gint64 src_value, GstFormat * dest_fmt, gint64 * dest_value);
|
|
#endif
|
|
static gboolean
|
|
gst_ffmpegdemux_send_event (GstElement * element, GstEvent * event);
|
|
static GstStateChangeReturn
|
|
gst_ffmpegdemux_change_state (GstElement * element, GstStateChange transition);
|
|
|
|
#define GST_FFDEMUX_PARAMS_QDATA g_quark_from_static_string("avdemux-params")
|
|
|
|
static GstElementClass *parent_class = NULL;
|
|
|
|
static const gchar *
|
|
gst_ffmpegdemux_averror (gint av_errno)
|
|
{
|
|
const gchar *message = NULL;
|
|
|
|
switch (av_errno) {
|
|
case AVERROR (EINVAL):
|
|
message = "Unknown error";
|
|
break;
|
|
case AVERROR (EIO):
|
|
message = "Input/output error";
|
|
break;
|
|
case AVERROR (EDOM):
|
|
message = "Number syntax expected in filename";
|
|
break;
|
|
case AVERROR (ENOMEM):
|
|
message = "Not enough memory";
|
|
break;
|
|
case AVERROR (EILSEQ):
|
|
message = "Unknown format";
|
|
break;
|
|
case AVERROR (ENOSYS):
|
|
message = "Operation not supported";
|
|
break;
|
|
default:
|
|
message = "Unhandled error code received";
|
|
break;
|
|
}
|
|
|
|
return message;
|
|
}
|
|
|
|
static void
|
|
gst_ffmpegdemux_base_init (GstFFMpegDemuxClass * klass)
|
|
{
|
|
GstElementClass *element_class = GST_ELEMENT_CLASS (klass);
|
|
AVInputFormat *in_plugin;
|
|
GstCaps *sinkcaps;
|
|
GstPadTemplate *sinktempl, *audiosrctempl, *videosrctempl;
|
|
gchar *longname, *description, *name;
|
|
|
|
in_plugin = (AVInputFormat *)
|
|
g_type_get_qdata (G_OBJECT_CLASS_TYPE (klass), GST_FFDEMUX_PARAMS_QDATA);
|
|
g_assert (in_plugin != NULL);
|
|
|
|
name = g_strdup (in_plugin->name);
|
|
g_strdelimit (name, ".,|-<> ", '_');
|
|
|
|
/* construct the element details struct */
|
|
longname = g_strdup_printf ("libav %s demuxer", in_plugin->long_name);
|
|
description = g_strdup_printf ("libav %s demuxer", in_plugin->long_name);
|
|
gst_element_class_set_metadata (element_class, longname,
|
|
"Codec/Demuxer", description,
|
|
"Wim Taymans <wim@fluendo.com>, "
|
|
"Ronald Bultje <rbultje@ronald.bitfreak.net>, "
|
|
"Edward Hervey <bilboed@bilboed.com>");
|
|
g_free (longname);
|
|
g_free (description);
|
|
|
|
/* pad templates */
|
|
sinkcaps = gst_ffmpeg_formatid_to_caps (name);
|
|
sinktempl = gst_pad_template_new ("sink",
|
|
GST_PAD_SINK, GST_PAD_ALWAYS, sinkcaps);
|
|
g_free (name);
|
|
videosrctempl = gst_pad_template_new ("video_%u",
|
|
GST_PAD_SRC, GST_PAD_SOMETIMES, GST_CAPS_ANY);
|
|
audiosrctempl = gst_pad_template_new ("audio_%u",
|
|
GST_PAD_SRC, GST_PAD_SOMETIMES, GST_CAPS_ANY);
|
|
|
|
gst_element_class_add_pad_template (element_class, videosrctempl);
|
|
gst_element_class_add_pad_template (element_class, audiosrctempl);
|
|
gst_element_class_add_pad_template (element_class, sinktempl);
|
|
|
|
gst_caps_unref (sinkcaps);
|
|
|
|
klass->in_plugin = in_plugin;
|
|
klass->videosrctempl = videosrctempl;
|
|
klass->audiosrctempl = audiosrctempl;
|
|
klass->sinktempl = sinktempl;
|
|
}
|
|
|
|
static void
|
|
gst_ffmpegdemux_class_init (GstFFMpegDemuxClass * klass)
|
|
{
|
|
GObjectClass *gobject_class;
|
|
GstElementClass *gstelement_class;
|
|
|
|
gobject_class = (GObjectClass *) klass;
|
|
gstelement_class = (GstElementClass *) klass;
|
|
|
|
parent_class = g_type_class_peek_parent (klass);
|
|
|
|
gobject_class->finalize = GST_DEBUG_FUNCPTR (gst_ffmpegdemux_finalize);
|
|
|
|
gstelement_class->change_state = gst_ffmpegdemux_change_state;
|
|
gstelement_class->send_event = gst_ffmpegdemux_send_event;
|
|
}
|
|
|
|
static void
|
|
gst_ffmpegdemux_init (GstFFMpegDemux * demux)
|
|
{
|
|
GstFFMpegDemuxClass *oclass =
|
|
(GstFFMpegDemuxClass *) (G_OBJECT_GET_CLASS (demux));
|
|
gint n;
|
|
|
|
demux->sinkpad = gst_pad_new_from_template (oclass->sinktempl, "sink");
|
|
gst_pad_set_activate_function (demux->sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_ffmpegdemux_sink_activate));
|
|
gst_pad_set_activatemode_function (demux->sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_ffmpegdemux_sink_activate_mode));
|
|
gst_element_add_pad (GST_ELEMENT (demux), demux->sinkpad);
|
|
|
|
/* push based setup */
|
|
/* the following are not used in pull-based mode, so safe to set anyway */
|
|
gst_pad_set_event_function (demux->sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_ffmpegdemux_sink_event));
|
|
gst_pad_set_chain_function (demux->sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_ffmpegdemux_chain));
|
|
/* task for driving ffmpeg in loop function */
|
|
demux->task =
|
|
gst_task_new ((GstTaskFunction) gst_ffmpegdemux_loop, demux, NULL);
|
|
g_rec_mutex_init (&demux->task_lock);
|
|
gst_task_set_lock (demux->task, &demux->task_lock);
|
|
|
|
demux->have_group_id = FALSE;
|
|
demux->group_id = G_MAXUINT;
|
|
|
|
demux->opened = FALSE;
|
|
demux->context = NULL;
|
|
|
|
for (n = 0; n < MAX_STREAMS; n++) {
|
|
demux->streams[n] = NULL;
|
|
}
|
|
demux->videopads = 0;
|
|
demux->audiopads = 0;
|
|
|
|
demux->seek_event = NULL;
|
|
gst_segment_init (&demux->segment, GST_FORMAT_TIME);
|
|
|
|
demux->flowcombiner = gst_flow_combiner_new ();
|
|
|
|
/* push based data */
|
|
g_mutex_init (&demux->ffpipe.tlock);
|
|
g_cond_init (&demux->ffpipe.cond);
|
|
demux->ffpipe.adapter = gst_adapter_new ();
|
|
|
|
/* blacklist unreliable push-based demuxers */
|
|
if (strcmp (oclass->in_plugin->name, "ape"))
|
|
demux->can_push = TRUE;
|
|
else
|
|
demux->can_push = FALSE;
|
|
}
|
|
|
|
static void
|
|
gst_ffmpegdemux_finalize (GObject * object)
|
|
{
|
|
GstFFMpegDemux *demux;
|
|
|
|
demux = (GstFFMpegDemux *) object;
|
|
|
|
gst_flow_combiner_free (demux->flowcombiner);
|
|
|
|
g_mutex_clear (&demux->ffpipe.tlock);
|
|
g_cond_clear (&demux->ffpipe.cond);
|
|
gst_object_unref (demux->ffpipe.adapter);
|
|
|
|
gst_object_unref (demux->task);
|
|
g_rec_mutex_clear (&demux->task_lock);
|
|
|
|
G_OBJECT_CLASS (parent_class)->finalize (object);
|
|
}
|
|
|
|
static void
|
|
gst_ffmpegdemux_close (GstFFMpegDemux * demux)
|
|
{
|
|
gint n;
|
|
GstEvent **event_p;
|
|
|
|
if (!demux->opened)
|
|
return;
|
|
|
|
/* remove pads from ourselves */
|
|
for (n = 0; n < MAX_STREAMS; n++) {
|
|
GstFFStream *stream;
|
|
|
|
stream = demux->streams[n];
|
|
if (stream) {
|
|
if (stream->pad) {
|
|
gst_flow_combiner_remove_pad (demux->flowcombiner, stream->pad);
|
|
gst_element_remove_pad (GST_ELEMENT (demux), stream->pad);
|
|
}
|
|
if (stream->tags)
|
|
gst_tag_list_unref (stream->tags);
|
|
g_free (stream);
|
|
}
|
|
demux->streams[n] = NULL;
|
|
}
|
|
demux->videopads = 0;
|
|
demux->audiopads = 0;
|
|
|
|
/* close demuxer context from ffmpeg */
|
|
if (demux->seekable)
|
|
gst_ffmpegdata_close (demux->context->pb);
|
|
else
|
|
gst_ffmpeg_pipe_close (demux->context->pb);
|
|
demux->context->pb = NULL;
|
|
avformat_close_input (&demux->context);
|
|
if (demux->context)
|
|
avformat_free_context (demux->context);
|
|
demux->context = NULL;
|
|
|
|
GST_OBJECT_LOCK (demux);
|
|
demux->opened = FALSE;
|
|
event_p = &demux->seek_event;
|
|
gst_event_replace (event_p, NULL);
|
|
GST_OBJECT_UNLOCK (demux);
|
|
|
|
gst_segment_init (&demux->segment, GST_FORMAT_TIME);
|
|
}
|
|
|
|
/* send an event to all the source pads .
|
|
* Takes ownership of the event.
|
|
*
|
|
* Returns FALSE if none of the source pads handled the event.
|
|
*/
|
|
static gboolean
|
|
gst_ffmpegdemux_push_event (GstFFMpegDemux * demux, GstEvent * event)
|
|
{
|
|
gboolean res;
|
|
gint n;
|
|
|
|
res = TRUE;
|
|
|
|
for (n = 0; n < MAX_STREAMS; n++) {
|
|
GstFFStream *s = demux->streams[n];
|
|
|
|
if (s && s->pad) {
|
|
gst_event_ref (event);
|
|
res &= gst_pad_push_event (s->pad, event);
|
|
}
|
|
}
|
|
gst_event_unref (event);
|
|
|
|
return res;
|
|
}
|
|
|
|
/* set flags on all streams */
|
|
static void
|
|
gst_ffmpegdemux_set_flags (GstFFMpegDemux * demux, gboolean discont,
|
|
gboolean eos)
|
|
{
|
|
GstFFStream *s;
|
|
gint n;
|
|
|
|
for (n = 0; n < MAX_STREAMS; n++) {
|
|
if ((s = demux->streams[n])) {
|
|
s->discont = discont;
|
|
s->eos = eos;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* check if all streams are eos */
|
|
static gboolean
|
|
gst_ffmpegdemux_is_eos (GstFFMpegDemux * demux)
|
|
{
|
|
GstFFStream *s;
|
|
gint n;
|
|
|
|
for (n = 0; n < MAX_STREAMS; n++) {
|
|
if ((s = demux->streams[n])) {
|
|
GST_DEBUG ("stream %d %p eos:%d", n, s, s->eos);
|
|
if (!s->eos)
|
|
return FALSE;
|
|
}
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
/* Returns True if we at least outputted one buffer */
|
|
static gboolean
|
|
gst_ffmpegdemux_has_outputted (GstFFMpegDemux * demux)
|
|
{
|
|
GstFFStream *s;
|
|
gint n;
|
|
|
|
for (n = 0; n < MAX_STREAMS; n++) {
|
|
if ((s = demux->streams[n])) {
|
|
if (GST_CLOCK_TIME_IS_VALID (s->last_ts))
|
|
return TRUE;
|
|
}
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_do_seek (GstFFMpegDemux * demux, GstSegment * segment)
|
|
{
|
|
gboolean ret;
|
|
gint seekret;
|
|
gint64 target;
|
|
gint64 fftarget;
|
|
AVStream *stream;
|
|
gint index;
|
|
|
|
/* find default index and fail if none is present */
|
|
index = av_find_default_stream_index (demux->context);
|
|
GST_LOG_OBJECT (demux, "default stream index %d", index);
|
|
if (index < 0)
|
|
return FALSE;
|
|
|
|
ret = TRUE;
|
|
|
|
/* get the stream for seeking */
|
|
stream = demux->context->streams[index];
|
|
/* initial seek position */
|
|
target = segment->position + demux->start_time;
|
|
/* convert target to ffmpeg time */
|
|
fftarget = gst_ffmpeg_time_gst_to_ff (target, stream->time_base);
|
|
|
|
GST_LOG_OBJECT (demux, "do seek to time %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (target));
|
|
|
|
/* if we need to land on a keyframe, try to do so, we don't try to do a
|
|
* keyframe seek if we are not absolutely sure we have an index.*/
|
|
if (segment->flags & GST_SEEK_FLAG_KEY_UNIT) {
|
|
gint keyframeidx;
|
|
|
|
GST_LOG_OBJECT (demux, "looking for keyframe in ffmpeg for time %"
|
|
GST_TIME_FORMAT, GST_TIME_ARGS (target));
|
|
|
|
/* search in the index for the previous keyframe */
|
|
keyframeidx =
|
|
av_index_search_timestamp (stream, fftarget, AVSEEK_FLAG_BACKWARD);
|
|
|
|
GST_LOG_OBJECT (demux, "keyframeidx: %d", keyframeidx);
|
|
|
|
if (keyframeidx >= 0) {
|
|
#if LIBAVFORMAT_VERSION_INT >= AV_VERSION_INT(58,78,0)
|
|
fftarget = avformat_index_get_entry (stream, keyframeidx)->timestamp;
|
|
#else
|
|
fftarget = stream->index_entries[keyframeidx].timestamp;
|
|
#endif
|
|
target = gst_ffmpeg_time_ff_to_gst (fftarget, stream->time_base);
|
|
|
|
GST_LOG_OBJECT (demux,
|
|
"Found a keyframe at ffmpeg idx: %d timestamp :%" GST_TIME_FORMAT,
|
|
keyframeidx, GST_TIME_ARGS (target));
|
|
}
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (demux,
|
|
"About to call av_seek_frame (context, %d, %" G_GINT64_FORMAT
|
|
", 0) for time %" GST_TIME_FORMAT, index, fftarget,
|
|
GST_TIME_ARGS (target));
|
|
|
|
if ((seekret =
|
|
av_seek_frame (demux->context, index, fftarget,
|
|
AVSEEK_FLAG_BACKWARD)) < 0)
|
|
goto seek_failed;
|
|
|
|
GST_DEBUG_OBJECT (demux, "seek success, returned %d", seekret);
|
|
|
|
if (target > demux->start_time)
|
|
target -= demux->start_time;
|
|
else
|
|
target = 0;
|
|
|
|
segment->position = target;
|
|
segment->time = target;
|
|
segment->start = target;
|
|
|
|
return ret;
|
|
|
|
/* ERRORS */
|
|
seek_failed:
|
|
{
|
|
GST_WARNING_OBJECT (demux, "Call to av_seek_frame failed : %d", seekret);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_perform_seek (GstFFMpegDemux * demux, GstEvent * event)
|
|
{
|
|
gboolean res;
|
|
gdouble rate;
|
|
GstFormat format;
|
|
GstSeekFlags flags;
|
|
GstSeekType cur_type, stop_type;
|
|
gint64 cur, stop;
|
|
gboolean flush;
|
|
gboolean update;
|
|
GstSegment seeksegment;
|
|
|
|
if (!demux->seekable) {
|
|
GST_DEBUG_OBJECT (demux, "in push mode; ignoring seek");
|
|
return FALSE;
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (demux, "starting seek");
|
|
|
|
if (event) {
|
|
gst_event_parse_seek (event, &rate, &format, &flags,
|
|
&cur_type, &cur, &stop_type, &stop);
|
|
|
|
/* we have to have a format as the segment format. Try to convert
|
|
* if not. */
|
|
if (demux->segment.format != format) {
|
|
GstFormat fmt;
|
|
|
|
fmt = demux->segment.format;
|
|
res = TRUE;
|
|
/* FIXME, use source pad */
|
|
if (cur_type != GST_SEEK_TYPE_NONE && cur != -1)
|
|
res = gst_pad_query_convert (demux->sinkpad, format, cur, fmt, &cur);
|
|
if (res && stop_type != GST_SEEK_TYPE_NONE && stop != -1)
|
|
res = gst_pad_query_convert (demux->sinkpad, format, stop, fmt, &stop);
|
|
if (!res)
|
|
goto no_format;
|
|
|
|
format = fmt;
|
|
}
|
|
} else {
|
|
flags = 0;
|
|
}
|
|
|
|
flush = flags & GST_SEEK_FLAG_FLUSH;
|
|
|
|
/* send flush start */
|
|
if (flush) {
|
|
/* mark flushing so that the streaming thread can react on it */
|
|
GST_OBJECT_LOCK (demux);
|
|
demux->flushing = TRUE;
|
|
GST_OBJECT_UNLOCK (demux);
|
|
gst_pad_push_event (demux->sinkpad, gst_event_new_flush_start ());
|
|
gst_ffmpegdemux_push_event (demux, gst_event_new_flush_start ());
|
|
} else {
|
|
gst_pad_pause_task (demux->sinkpad);
|
|
}
|
|
|
|
/* grab streaming lock, this should eventually be possible, either
|
|
* because the task is paused or our streaming thread stopped
|
|
* because our peer is flushing. */
|
|
GST_PAD_STREAM_LOCK (demux->sinkpad);
|
|
|
|
/* make copy into temp structure, we can only update the main one
|
|
* when we actually could do the seek. */
|
|
memcpy (&seeksegment, &demux->segment, sizeof (GstSegment));
|
|
|
|
/* now configure the seek segment */
|
|
if (event) {
|
|
gst_segment_do_seek (&seeksegment, rate, format, flags,
|
|
cur_type, cur, stop_type, stop, &update);
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (demux, "segment configured from %" G_GINT64_FORMAT
|
|
" to %" G_GINT64_FORMAT ", position %" G_GINT64_FORMAT,
|
|
seeksegment.start, seeksegment.stop, seeksegment.position);
|
|
|
|
/* make the sinkpad available for data passing since we might need
|
|
* it when doing the seek */
|
|
if (flush) {
|
|
GST_OBJECT_LOCK (demux);
|
|
demux->flushing = FALSE;
|
|
GST_OBJECT_UNLOCK (demux);
|
|
gst_pad_push_event (demux->sinkpad, gst_event_new_flush_stop (TRUE));
|
|
}
|
|
|
|
/* do the seek, segment.position contains new position. */
|
|
res = gst_ffmpegdemux_do_seek (demux, &seeksegment);
|
|
|
|
/* and prepare to continue streaming */
|
|
if (flush) {
|
|
/* send flush stop, peer will accept data and events again. We
|
|
* are not yet providing data as we still have the STREAM_LOCK. */
|
|
gst_ffmpegdemux_push_event (demux, gst_event_new_flush_stop (TRUE));
|
|
}
|
|
/* if successfull seek, we update our real segment and push
|
|
* out the new segment. */
|
|
if (res) {
|
|
memcpy (&demux->segment, &seeksegment, sizeof (GstSegment));
|
|
|
|
if (demux->segment.flags & GST_SEEK_FLAG_SEGMENT) {
|
|
gst_element_post_message (GST_ELEMENT (demux),
|
|
gst_message_new_segment_start (GST_OBJECT (demux),
|
|
demux->segment.format, demux->segment.position));
|
|
}
|
|
|
|
/* now send the newsegment, FIXME, do this from the streaming thread */
|
|
GST_DEBUG_OBJECT (demux, "Sending newsegment %" GST_SEGMENT_FORMAT,
|
|
&demux->segment);
|
|
|
|
gst_ffmpegdemux_push_event (demux, gst_event_new_segment (&demux->segment));
|
|
}
|
|
|
|
/* Mark discont on all srcpads and remove eos */
|
|
gst_ffmpegdemux_set_flags (demux, TRUE, FALSE);
|
|
gst_flow_combiner_reset (demux->flowcombiner);
|
|
|
|
/* and restart the task in case it got paused explicitely or by
|
|
* the FLUSH_START event we pushed out. */
|
|
gst_pad_start_task (demux->sinkpad, (GstTaskFunction) gst_ffmpegdemux_loop,
|
|
demux->sinkpad, NULL);
|
|
|
|
/* and release the lock again so we can continue streaming */
|
|
GST_PAD_STREAM_UNLOCK (demux->sinkpad);
|
|
|
|
return res;
|
|
|
|
/* ERROR */
|
|
no_format:
|
|
{
|
|
GST_DEBUG_OBJECT (demux, "undefined format given, seek aborted.");
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_src_event (GstPad * pad, GstObject * parent, GstEvent * event)
|
|
{
|
|
GstFFMpegDemux *demux;
|
|
gboolean res = TRUE;
|
|
|
|
demux = (GstFFMpegDemux *) parent;
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_SEEK:
|
|
res = gst_ffmpegdemux_perform_seek (demux, event);
|
|
gst_event_unref (event);
|
|
break;
|
|
case GST_EVENT_LATENCY:
|
|
res = gst_pad_push_event (demux->sinkpad, event);
|
|
break;
|
|
case GST_EVENT_NAVIGATION:
|
|
case GST_EVENT_QOS:
|
|
default:
|
|
res = FALSE;
|
|
gst_event_unref (event);
|
|
break;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_send_event (GstElement * element, GstEvent * event)
|
|
{
|
|
GstFFMpegDemux *demux = (GstFFMpegDemux *) (element);
|
|
gboolean res;
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_SEEK:
|
|
GST_OBJECT_LOCK (demux);
|
|
if (!demux->opened) {
|
|
GstEvent **event_p;
|
|
|
|
GST_DEBUG_OBJECT (demux, "caching seek event");
|
|
event_p = &demux->seek_event;
|
|
gst_event_replace (event_p, event);
|
|
GST_OBJECT_UNLOCK (demux);
|
|
|
|
res = TRUE;
|
|
} else {
|
|
GST_OBJECT_UNLOCK (demux);
|
|
res = gst_ffmpegdemux_perform_seek (demux, event);
|
|
gst_event_unref (event);
|
|
}
|
|
break;
|
|
default:
|
|
res = FALSE;
|
|
break;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_src_query (GstPad * pad, GstObject * parent, GstQuery * query)
|
|
{
|
|
GstFFMpegDemux *demux;
|
|
GstFFStream *stream;
|
|
AVStream *avstream;
|
|
gboolean res = FALSE;
|
|
|
|
if (!(stream = gst_pad_get_element_private (pad)))
|
|
return FALSE;
|
|
|
|
avstream = stream->avstream;
|
|
|
|
demux = (GstFFMpegDemux *) parent;
|
|
|
|
switch (GST_QUERY_TYPE (query)) {
|
|
case GST_QUERY_POSITION:
|
|
{
|
|
GstFormat format;
|
|
gint64 timeposition;
|
|
|
|
gst_query_parse_position (query, &format, NULL);
|
|
|
|
timeposition = stream->last_ts;
|
|
if (!(GST_CLOCK_TIME_IS_VALID (timeposition)))
|
|
break;
|
|
|
|
switch (format) {
|
|
case GST_FORMAT_TIME:
|
|
gst_query_set_position (query, GST_FORMAT_TIME, timeposition);
|
|
res = TRUE;
|
|
break;
|
|
case GST_FORMAT_DEFAULT:
|
|
gst_query_set_position (query, GST_FORMAT_DEFAULT,
|
|
gst_util_uint64_scale (timeposition, avstream->avg_frame_rate.num,
|
|
GST_SECOND * avstream->avg_frame_rate.den));
|
|
res = TRUE;
|
|
break;
|
|
case GST_FORMAT_BYTES:
|
|
if (demux->videopads + demux->audiopads == 1 &&
|
|
GST_PAD_PEER (demux->sinkpad) != NULL)
|
|
res = gst_pad_query_default (pad, parent, query);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
break;
|
|
case GST_QUERY_DURATION:
|
|
{
|
|
GstFormat format;
|
|
gint64 timeduration;
|
|
|
|
gst_query_parse_duration (query, &format, NULL);
|
|
|
|
timeduration =
|
|
gst_ffmpeg_time_ff_to_gst (avstream->duration, avstream->time_base);
|
|
if (!(GST_CLOCK_TIME_IS_VALID (timeduration))) {
|
|
/* use duration of complete file if the stream duration is not known */
|
|
timeduration = demux->duration;
|
|
if (!(GST_CLOCK_TIME_IS_VALID (timeduration)))
|
|
break;
|
|
}
|
|
|
|
switch (format) {
|
|
case GST_FORMAT_TIME:
|
|
gst_query_set_duration (query, GST_FORMAT_TIME, timeduration);
|
|
res = TRUE;
|
|
break;
|
|
case GST_FORMAT_DEFAULT:
|
|
gst_query_set_duration (query, GST_FORMAT_DEFAULT,
|
|
gst_util_uint64_scale (timeduration, avstream->avg_frame_rate.num,
|
|
GST_SECOND * avstream->avg_frame_rate.den));
|
|
res = TRUE;
|
|
break;
|
|
case GST_FORMAT_BYTES:
|
|
if (demux->videopads + demux->audiopads == 1 &&
|
|
GST_PAD_PEER (demux->sinkpad) != NULL)
|
|
res = gst_pad_query_default (pad, parent, query);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
break;
|
|
case GST_QUERY_SEEKING:{
|
|
GstFormat format;
|
|
gboolean seekable;
|
|
gint64 dur = -1;
|
|
|
|
gst_query_parse_seeking (query, &format, NULL, NULL, NULL);
|
|
seekable = demux->seekable;
|
|
if (!gst_pad_query_duration (pad, format, &dur)) {
|
|
/* unlikely that we don't know duration but can seek */
|
|
seekable = FALSE;
|
|
dur = -1;
|
|
}
|
|
gst_query_set_seeking (query, format, seekable, 0, dur);
|
|
res = TRUE;
|
|
break;
|
|
}
|
|
case GST_QUERY_SEGMENT:{
|
|
GstFormat format;
|
|
gint64 start, stop;
|
|
|
|
format = demux->segment.format;
|
|
|
|
start =
|
|
gst_segment_to_stream_time (&demux->segment, format,
|
|
demux->segment.start);
|
|
if ((stop = demux->segment.stop) == -1)
|
|
stop = demux->segment.duration;
|
|
else
|
|
stop = gst_segment_to_stream_time (&demux->segment, format, stop);
|
|
|
|
gst_query_set_segment (query, demux->segment.rate, format, start, stop);
|
|
res = TRUE;
|
|
break;
|
|
}
|
|
default:
|
|
/* FIXME : ADD GST_QUERY_CONVERT */
|
|
res = gst_pad_query_default (pad, parent, query);
|
|
break;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
#if 0
|
|
/* FIXME, reenable me */
|
|
static gboolean
|
|
gst_ffmpegdemux_src_convert (GstPad * pad,
|
|
GstFormat src_fmt,
|
|
gint64 src_value, GstFormat * dest_fmt, gint64 * dest_value)
|
|
{
|
|
GstFFStream *stream;
|
|
gboolean res = TRUE;
|
|
AVStream *avstream;
|
|
|
|
if (!(stream = gst_pad_get_element_private (pad)))
|
|
return FALSE;
|
|
|
|
avstream = stream->avstream;
|
|
if (avstream->codec->codec_type != AVMEDIA_TYPE_VIDEO)
|
|
return FALSE;
|
|
|
|
switch (src_fmt) {
|
|
case GST_FORMAT_TIME:
|
|
switch (*dest_fmt) {
|
|
case GST_FORMAT_DEFAULT:
|
|
*dest_value = gst_util_uint64_scale (src_value,
|
|
avstream->avg_frame_rate.num,
|
|
GST_SECOND * avstream->avg_frame_rate.den);
|
|
break;
|
|
default:
|
|
res = FALSE;
|
|
break;
|
|
}
|
|
break;
|
|
case GST_FORMAT_DEFAULT:
|
|
switch (*dest_fmt) {
|
|
case GST_FORMAT_TIME:
|
|
*dest_value = gst_util_uint64_scale (src_value,
|
|
GST_SECOND * avstream->avg_frame_rate.num,
|
|
avstream->avg_frame_rate.den);
|
|
break;
|
|
default:
|
|
res = FALSE;
|
|
break;
|
|
}
|
|
break;
|
|
default:
|
|
res = FALSE;
|
|
break;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
#endif
|
|
|
|
static gchar *
|
|
gst_ffmpegdemux_create_padname (const gchar * templ, gint n)
|
|
{
|
|
GString *string;
|
|
|
|
/* FIXME, we just want to printf the number according to the template but
|
|
* then the format string is not a literal and we can't check arguments and
|
|
* this generates a compiler error */
|
|
string = g_string_new (templ);
|
|
g_string_truncate (string, string->len - 2);
|
|
g_string_append_printf (string, "%u", n);
|
|
|
|
return g_string_free (string, FALSE);
|
|
}
|
|
|
|
static GstFFStream *
|
|
gst_ffmpegdemux_get_stream (GstFFMpegDemux * demux, AVStream * avstream)
|
|
{
|
|
GstFFMpegDemuxClass *oclass;
|
|
GstPadTemplate *templ = NULL;
|
|
GstPad *pad;
|
|
GstCaps *caps;
|
|
gint num;
|
|
gchar *padname;
|
|
const gchar *codec;
|
|
AVCodecContext *ctx = NULL;
|
|
GstFFStream *stream;
|
|
GstEvent *event;
|
|
gchar *stream_id;
|
|
|
|
oclass = (GstFFMpegDemuxClass *) G_OBJECT_GET_CLASS (demux);
|
|
|
|
if (demux->streams[avstream->index] != NULL)
|
|
goto exists;
|
|
|
|
ctx = avcodec_alloc_context3 (NULL);
|
|
avcodec_parameters_to_context (ctx, avstream->codecpar);
|
|
|
|
/* create new stream */
|
|
stream = g_new0 (GstFFStream, 1);
|
|
demux->streams[avstream->index] = stream;
|
|
|
|
/* mark stream as unknown */
|
|
stream->unknown = TRUE;
|
|
stream->discont = TRUE;
|
|
stream->avstream = avstream;
|
|
stream->last_ts = GST_CLOCK_TIME_NONE;
|
|
stream->tags = NULL;
|
|
|
|
switch (ctx->codec_type) {
|
|
case AVMEDIA_TYPE_VIDEO:
|
|
templ = oclass->videosrctempl;
|
|
num = demux->videopads++;
|
|
/* These are not part of the codec parameters we built the
|
|
* context from */
|
|
ctx->framerate.num = avstream->r_frame_rate.num;
|
|
ctx->framerate.den = avstream->r_frame_rate.den;
|
|
break;
|
|
case AVMEDIA_TYPE_AUDIO:
|
|
templ = oclass->audiosrctempl;
|
|
num = demux->audiopads++;
|
|
break;
|
|
default:
|
|
goto unknown_type;
|
|
}
|
|
|
|
/* get caps that belongs to this stream */
|
|
caps = gst_ffmpeg_codecid_to_caps (ctx->codec_id, ctx, TRUE);
|
|
if (caps == NULL)
|
|
goto unknown_caps;
|
|
|
|
/* stream is known now */
|
|
stream->unknown = FALSE;
|
|
|
|
/* create new pad for this stream */
|
|
padname =
|
|
gst_ffmpegdemux_create_padname (GST_PAD_TEMPLATE_NAME_TEMPLATE (templ),
|
|
num);
|
|
pad = gst_pad_new_from_template (templ, padname);
|
|
g_free (padname);
|
|
|
|
gst_pad_use_fixed_caps (pad);
|
|
gst_pad_set_active (pad, TRUE);
|
|
|
|
gst_pad_set_query_function (pad, gst_ffmpegdemux_src_query);
|
|
gst_pad_set_event_function (pad, gst_ffmpegdemux_src_event);
|
|
|
|
/* store pad internally */
|
|
stream->pad = pad;
|
|
gst_pad_set_element_private (pad, stream);
|
|
|
|
/* transform some useful info to GstClockTime and remember */
|
|
{
|
|
GstClockTime tmp;
|
|
|
|
/* FIXME, actually use the start_time in some way */
|
|
tmp = gst_ffmpeg_time_ff_to_gst (avstream->start_time, avstream->time_base);
|
|
GST_DEBUG_OBJECT (demux, "stream %d: start time: %" GST_TIME_FORMAT,
|
|
avstream->index, GST_TIME_ARGS (tmp));
|
|
|
|
tmp = gst_ffmpeg_time_ff_to_gst (avstream->duration, avstream->time_base);
|
|
GST_DEBUG_OBJECT (demux, "stream %d: duration: %" GST_TIME_FORMAT,
|
|
avstream->index, GST_TIME_ARGS (tmp));
|
|
}
|
|
|
|
demux->streams[avstream->index] = stream;
|
|
|
|
|
|
stream_id =
|
|
gst_pad_create_stream_id_printf (pad, GST_ELEMENT_CAST (demux), "%03u",
|
|
avstream->index);
|
|
|
|
event = gst_pad_get_sticky_event (demux->sinkpad, GST_EVENT_STREAM_START, 0);
|
|
if (event) {
|
|
if (gst_event_parse_group_id (event, &demux->group_id))
|
|
demux->have_group_id = TRUE;
|
|
else
|
|
demux->have_group_id = FALSE;
|
|
gst_event_unref (event);
|
|
} else if (!demux->have_group_id) {
|
|
demux->have_group_id = TRUE;
|
|
demux->group_id = gst_util_group_id_next ();
|
|
}
|
|
event = gst_event_new_stream_start (stream_id);
|
|
if (demux->have_group_id)
|
|
gst_event_set_group_id (event, demux->group_id);
|
|
|
|
gst_pad_push_event (pad, event);
|
|
g_free (stream_id);
|
|
|
|
GST_INFO_OBJECT (pad, "adding pad with caps %" GST_PTR_FORMAT, caps);
|
|
gst_pad_set_caps (pad, caps);
|
|
gst_caps_unref (caps);
|
|
|
|
/* activate and add */
|
|
gst_element_add_pad (GST_ELEMENT (demux), pad);
|
|
gst_flow_combiner_add_pad (demux->flowcombiner, pad);
|
|
|
|
/* metadata */
|
|
if ((codec = gst_ffmpeg_get_codecid_longname (ctx->codec_id))) {
|
|
stream->tags = gst_ffmpeg_metadata_to_tag_list (avstream->metadata);
|
|
|
|
if (stream->tags == NULL)
|
|
stream->tags = gst_tag_list_new_empty ();
|
|
|
|
gst_tag_list_add (stream->tags, GST_TAG_MERGE_REPLACE,
|
|
(ctx->codec_type == AVMEDIA_TYPE_VIDEO) ?
|
|
GST_TAG_VIDEO_CODEC : GST_TAG_AUDIO_CODEC, codec, NULL);
|
|
}
|
|
|
|
done:
|
|
if (ctx)
|
|
avcodec_free_context (&ctx);
|
|
return stream;
|
|
|
|
/* ERRORS */
|
|
exists:
|
|
{
|
|
GST_DEBUG_OBJECT (demux, "Pad existed (stream %d)", avstream->index);
|
|
stream = demux->streams[avstream->index];
|
|
goto done;
|
|
}
|
|
unknown_type:
|
|
{
|
|
GST_WARNING_OBJECT (demux, "Unknown pad type %d", ctx->codec_type);
|
|
goto done;
|
|
}
|
|
unknown_caps:
|
|
{
|
|
GST_WARNING_OBJECT (demux, "Unknown caps for codec %d", ctx->codec_id);
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
static gchar *
|
|
safe_utf8_copy (gchar * input)
|
|
{
|
|
gchar *output;
|
|
|
|
if (!(g_utf8_validate (input, -1, NULL))) {
|
|
output = g_convert (input, strlen (input),
|
|
"UTF-8", "ISO-8859-1", NULL, NULL, NULL);
|
|
} else {
|
|
output = g_strdup (input);
|
|
}
|
|
|
|
return output;
|
|
}
|
|
|
|
/* This is a list of standard tag keys taken from the avformat.h
|
|
* header, without handling any variants. */
|
|
static const struct
|
|
{
|
|
const gchar *ffmpeg_tag_name;
|
|
const gchar *gst_tag_name;
|
|
} tagmapping[] = {
|
|
{
|
|
"album", GST_TAG_ALBUM}, {
|
|
"album_artist", GST_TAG_ALBUM_ARTIST}, {
|
|
"artist", GST_TAG_ARTIST}, {
|
|
"comment", GST_TAG_COMMENT}, {
|
|
"composer", GST_TAG_COMPOSER}, {
|
|
"copyright", GST_TAG_COPYRIGHT}, {
|
|
/* Need to convert ISO 8601 to GstDateTime: */
|
|
"creation_time", GST_TAG_DATE_TIME}, {
|
|
/* Need to convert ISO 8601 to GDateTime: */
|
|
"date", GST_TAG_DATE_TIME}, {
|
|
"disc", GST_TAG_ALBUM_VOLUME_NUMBER}, {
|
|
"encoder", GST_TAG_ENCODER}, {
|
|
"encoded_by", GST_TAG_ENCODED_BY}, {
|
|
"genre", GST_TAG_GENRE}, {
|
|
"language", GST_TAG_LANGUAGE_CODE}, {
|
|
"performer", GST_TAG_PERFORMER}, {
|
|
"publisher", GST_TAG_PUBLISHER}, {
|
|
"title", GST_TAG_TITLE}, {
|
|
"track", GST_TAG_TRACK_NUMBER}
|
|
};
|
|
|
|
static const gchar *
|
|
match_tag_name (gchar * ffmpeg_tag_name)
|
|
{
|
|
gint i;
|
|
for (i = 0; i < G_N_ELEMENTS (tagmapping); i++) {
|
|
if (!g_strcmp0 (tagmapping[i].ffmpeg_tag_name, ffmpeg_tag_name))
|
|
return tagmapping[i].gst_tag_name;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static GstTagList *
|
|
gst_ffmpeg_metadata_to_tag_list (AVDictionary * metadata)
|
|
{
|
|
AVDictionaryEntry *tag = NULL;
|
|
GstTagList *list;
|
|
list = gst_tag_list_new_empty ();
|
|
|
|
while ((tag = av_dict_get (metadata, "", tag, AV_DICT_IGNORE_SUFFIX))) {
|
|
const gchar *gsttag = match_tag_name (tag->key);
|
|
GType t;
|
|
GST_LOG ("mapping tag %s=%s\n", tag->key, tag->value);
|
|
if (gsttag == NULL) {
|
|
GST_LOG ("Ignoring unknown metadata tag %s", tag->key);
|
|
continue;
|
|
}
|
|
/* Special case, track and disc numbers may be x/n in libav, split
|
|
* them */
|
|
if (g_str_equal (gsttag, GST_TAG_TRACK_NUMBER)) {
|
|
guint track, trackcount;
|
|
if (sscanf (tag->value, "%u/%u", &track, &trackcount) == 2) {
|
|
gst_tag_list_add (list, GST_TAG_MERGE_REPLACE,
|
|
gsttag, track, GST_TAG_TRACK_COUNT, trackcount, NULL);
|
|
continue;
|
|
}
|
|
/* Fall through and handle as a single uint below */
|
|
} else if (g_str_equal (gsttag, GST_TAG_ALBUM_VOLUME_NUMBER)) {
|
|
guint disc, disc_count;
|
|
if (sscanf (tag->value, "%u/%u", &disc, &disc_count) == 2) {
|
|
gst_tag_list_add (list, GST_TAG_MERGE_REPLACE,
|
|
gsttag, disc, GST_TAG_ALBUM_VOLUME_COUNT, disc_count, NULL);
|
|
continue;
|
|
}
|
|
/* Fall through and handle as a single uint below */
|
|
}
|
|
|
|
t = gst_tag_get_type (gsttag);
|
|
if (t == G_TYPE_STRING) {
|
|
gchar *s = safe_utf8_copy (tag->value);
|
|
gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, gsttag, s, NULL);
|
|
g_free (s);
|
|
} else if (t == G_TYPE_UINT || t == G_TYPE_INT) {
|
|
gchar *end;
|
|
gint v = strtol (tag->value, &end, 10);
|
|
if (end == tag->value)
|
|
continue; /* Failed to parse */
|
|
gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, gsttag, v, NULL);
|
|
} else if (t == G_TYPE_DATE) {
|
|
guint year, month, day;
|
|
GDate *date = NULL;
|
|
if (sscanf (tag->value, "%04u-%02u-%02u", &year, &month, &day) == 3) {
|
|
date = g_date_new_dmy (day, month, year);
|
|
} else {
|
|
/* Try interpreting just as a year */
|
|
gchar *end;
|
|
|
|
year = strtol (tag->value, &end, 10);
|
|
if (end != tag->value)
|
|
date = g_date_new_dmy (1, 1, year);
|
|
}
|
|
if (date) {
|
|
gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, gsttag, date, NULL);
|
|
g_date_free (date);
|
|
}
|
|
} else if (t == GST_TYPE_DATE_TIME) {
|
|
gchar *s = safe_utf8_copy (tag->value);
|
|
GstDateTime *d = gst_date_time_new_from_iso8601_string (s);
|
|
|
|
g_free (s);
|
|
if (d) {
|
|
gst_tag_list_add (list, GST_TAG_MERGE_REPLACE, gsttag, d, NULL);
|
|
gst_date_time_unref (d);
|
|
}
|
|
} else {
|
|
GST_FIXME ("Unhandled tag %s", gsttag);
|
|
}
|
|
}
|
|
|
|
if (gst_tag_list_is_empty (list)) {
|
|
gst_tag_list_unref (list);
|
|
return NULL;
|
|
}
|
|
|
|
return list;
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_open (GstFFMpegDemux * demux)
|
|
{
|
|
AVIOContext *iocontext = NULL;
|
|
GstFFMpegDemuxClass *oclass =
|
|
(GstFFMpegDemuxClass *) G_OBJECT_GET_CLASS (demux);
|
|
gint res, n_streams, i;
|
|
GstTagList *tags;
|
|
GstEvent *event;
|
|
GList *cached_events;
|
|
GstQuery *query;
|
|
gchar *uri = NULL;
|
|
|
|
/* to be sure... */
|
|
gst_ffmpegdemux_close (demux);
|
|
|
|
/* open via our input protocol hack */
|
|
if (demux->seekable)
|
|
res = gst_ffmpegdata_open (demux->sinkpad, AVIO_FLAG_READ, &iocontext);
|
|
else
|
|
res = gst_ffmpeg_pipe_open (&demux->ffpipe, AVIO_FLAG_READ, &iocontext);
|
|
|
|
if (res < 0)
|
|
goto beach;
|
|
|
|
query = gst_query_new_uri ();
|
|
if (gst_pad_peer_query (demux->sinkpad, query)) {
|
|
gchar *query_uri, *redirect_uri;
|
|
gboolean permanent;
|
|
|
|
gst_query_parse_uri (query, &query_uri);
|
|
gst_query_parse_uri_redirection (query, &redirect_uri);
|
|
gst_query_parse_uri_redirection_permanent (query, &permanent);
|
|
|
|
if (permanent && redirect_uri) {
|
|
uri = redirect_uri;
|
|
g_free (query_uri);
|
|
} else {
|
|
uri = query_uri;
|
|
g_free (redirect_uri);
|
|
}
|
|
}
|
|
gst_query_unref (query);
|
|
|
|
GST_DEBUG_OBJECT (demux, "Opening context with URI %s", GST_STR_NULL (uri));
|
|
|
|
demux->context = avformat_alloc_context ();
|
|
demux->context->pb = iocontext;
|
|
res = avformat_open_input (&demux->context, uri, oclass->in_plugin, NULL);
|
|
|
|
g_free (uri);
|
|
|
|
GST_DEBUG_OBJECT (demux, "av_open_input returned %d", res);
|
|
if (res < 0)
|
|
goto beach;
|
|
|
|
res = gst_ffmpeg_av_find_stream_info (demux->context);
|
|
GST_DEBUG_OBJECT (demux, "av_find_stream_info returned %d", res);
|
|
if (res < 0)
|
|
goto beach;
|
|
|
|
n_streams = demux->context->nb_streams;
|
|
GST_DEBUG_OBJECT (demux, "we have %d streams", n_streams);
|
|
|
|
/* open_input_file() automatically reads the header. We can now map each
|
|
* created AVStream to a GstPad to make GStreamer handle it. */
|
|
for (i = 0; i < n_streams; i++) {
|
|
gst_ffmpegdemux_get_stream (demux, demux->context->streams[i]);
|
|
}
|
|
|
|
gst_element_no_more_pads (GST_ELEMENT (demux));
|
|
|
|
/* transform some useful info to GstClockTime and remember */
|
|
demux->start_time = gst_util_uint64_scale_int (demux->context->start_time,
|
|
GST_SECOND, AV_TIME_BASE);
|
|
GST_DEBUG_OBJECT (demux, "start time: %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (demux->start_time));
|
|
if (demux->context->duration > 0)
|
|
demux->duration = gst_util_uint64_scale_int (demux->context->duration,
|
|
GST_SECOND, AV_TIME_BASE);
|
|
else
|
|
demux->duration = GST_CLOCK_TIME_NONE;
|
|
|
|
GST_DEBUG_OBJECT (demux, "duration: %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (demux->duration));
|
|
|
|
/* store duration in the segment as well */
|
|
demux->segment.duration = demux->duration;
|
|
|
|
GST_OBJECT_LOCK (demux);
|
|
demux->opened = TRUE;
|
|
event = demux->seek_event;
|
|
demux->seek_event = NULL;
|
|
cached_events = demux->cached_events;
|
|
demux->cached_events = NULL;
|
|
GST_OBJECT_UNLOCK (demux);
|
|
|
|
if (event) {
|
|
gst_ffmpegdemux_perform_seek (demux, event);
|
|
gst_event_unref (event);
|
|
} else {
|
|
GST_DEBUG_OBJECT (demux, "Sending segment %" GST_SEGMENT_FORMAT,
|
|
&demux->segment);
|
|
gst_ffmpegdemux_push_event (demux, gst_event_new_segment (&demux->segment));
|
|
}
|
|
|
|
while (cached_events) {
|
|
event = cached_events->data;
|
|
GST_INFO_OBJECT (demux, "pushing cached event: %" GST_PTR_FORMAT, event);
|
|
gst_ffmpegdemux_push_event (demux, event);
|
|
cached_events = g_list_delete_link (cached_events, cached_events);
|
|
}
|
|
|
|
/* grab the global tags */
|
|
tags = gst_ffmpeg_metadata_to_tag_list (demux->context->metadata);
|
|
if (tags) {
|
|
GST_INFO_OBJECT (demux, "global tags: %" GST_PTR_FORMAT, tags);
|
|
}
|
|
|
|
/* now handle the stream tags */
|
|
for (i = 0; i < n_streams; i++) {
|
|
GstFFStream *stream;
|
|
|
|
stream = gst_ffmpegdemux_get_stream (demux, demux->context->streams[i]);
|
|
if (stream->pad != NULL) {
|
|
|
|
/* Global tags */
|
|
if (tags)
|
|
gst_pad_push_event (stream->pad,
|
|
gst_event_new_tag (gst_tag_list_ref (tags)));
|
|
|
|
/* Per-stream tags */
|
|
if (stream->tags != NULL) {
|
|
GST_INFO_OBJECT (stream->pad, "stream tags: %" GST_PTR_FORMAT,
|
|
stream->tags);
|
|
gst_pad_push_event (stream->pad,
|
|
gst_event_new_tag (gst_tag_list_ref (stream->tags)));
|
|
}
|
|
}
|
|
}
|
|
if (tags)
|
|
gst_tag_list_unref (tags);
|
|
return TRUE;
|
|
|
|
/* ERRORS */
|
|
beach:
|
|
{
|
|
GST_ELEMENT_ERROR (demux, LIBRARY, FAILED, (NULL),
|
|
("%s", gst_ffmpegdemux_averror (res)));
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
#define GST_FFMPEG_TYPE_FIND_SIZE 4096
|
|
#define GST_FFMPEG_TYPE_FIND_MIN_SIZE 256
|
|
|
|
static void
|
|
gst_ffmpegdemux_type_find (GstTypeFind * tf, gpointer priv)
|
|
{
|
|
const guint8 *data;
|
|
AVInputFormat *in_plugin = (AVInputFormat *) priv;
|
|
gint res = 0;
|
|
guint64 length;
|
|
GstCaps *sinkcaps;
|
|
|
|
/* We want GST_FFMPEG_TYPE_FIND_SIZE bytes, but if the file is shorter than
|
|
* that we'll give it a try... */
|
|
length = gst_type_find_get_length (tf);
|
|
if (length == 0 || length > GST_FFMPEG_TYPE_FIND_SIZE)
|
|
length = GST_FFMPEG_TYPE_FIND_SIZE;
|
|
|
|
/* The ffmpeg typefinders assume there's a certain minimum amount of data
|
|
* and will happily do invalid memory access if there isn't, so let's just
|
|
* skip the ffmpeg typefinders if the data available is too short
|
|
* (in which case it's unlikely to be a media file anyway) */
|
|
if (length < GST_FFMPEG_TYPE_FIND_MIN_SIZE) {
|
|
GST_LOG ("not typefinding %" G_GUINT64_FORMAT " bytes, too short", length);
|
|
return;
|
|
}
|
|
|
|
GST_LOG ("typefinding %" G_GUINT64_FORMAT " bytes", length);
|
|
if (in_plugin->read_probe &&
|
|
(data = gst_type_find_peek (tf, 0, length)) != NULL) {
|
|
AVProbeData probe_data;
|
|
|
|
probe_data.filename = "";
|
|
probe_data.buf = (guint8 *) data;
|
|
probe_data.buf_size = length;
|
|
|
|
res = in_plugin->read_probe (&probe_data);
|
|
if (res > 0) {
|
|
res = MAX (1, res * GST_TYPE_FIND_MAXIMUM / AVPROBE_SCORE_MAX);
|
|
/* Restrict the probability for MPEG-TS streams, because there is
|
|
* probably a better version in plugins-base, if the user has a recent
|
|
* plugins-base (in fact we shouldn't even get here for ffmpeg mpegts or
|
|
* mpegtsraw typefinders, since we blacklist them) */
|
|
if (g_str_has_prefix (in_plugin->name, "mpegts"))
|
|
res = MIN (res, GST_TYPE_FIND_POSSIBLE);
|
|
|
|
sinkcaps = gst_ffmpeg_formatid_to_caps (in_plugin->name);
|
|
|
|
GST_LOG ("libav typefinder '%s' suggests %" GST_PTR_FORMAT ", p=%u%%",
|
|
in_plugin->name, sinkcaps, res);
|
|
|
|
gst_type_find_suggest (tf, res, sinkcaps);
|
|
gst_caps_unref (sinkcaps);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Task */
|
|
static void
|
|
gst_ffmpegdemux_loop (GstFFMpegDemux * demux)
|
|
{
|
|
GstFlowReturn ret;
|
|
gint res = -1;
|
|
AVPacket pkt;
|
|
GstPad *srcpad;
|
|
GstFFStream *stream;
|
|
AVStream *avstream;
|
|
GstBuffer *outbuf = NULL;
|
|
GstClockTime timestamp, duration;
|
|
gint outsize;
|
|
gboolean rawvideo;
|
|
GstFlowReturn stream_last_flow;
|
|
gint64 pts;
|
|
|
|
/* open file if we didn't so already */
|
|
if (!demux->opened)
|
|
if (!gst_ffmpegdemux_open (demux))
|
|
goto open_failed;
|
|
|
|
GST_DEBUG_OBJECT (demux, "about to read a frame");
|
|
|
|
/* read a frame */
|
|
res = av_read_frame (demux->context, &pkt);
|
|
if (res < 0)
|
|
goto read_failed;
|
|
|
|
/* get the stream */
|
|
stream =
|
|
gst_ffmpegdemux_get_stream (demux,
|
|
demux->context->streams[pkt.stream_index]);
|
|
|
|
/* check if we know the stream */
|
|
if (stream->unknown)
|
|
goto done;
|
|
|
|
/* get more stuff belonging to this stream */
|
|
avstream = stream->avstream;
|
|
|
|
/* do timestamps, we do this first so that we can know when we
|
|
* stepped over the segment stop position. */
|
|
pts = pkt.pts;
|
|
if (G_UNLIKELY (pts < 0)) {
|
|
/* some streams have pts such this:
|
|
* 0
|
|
* -2
|
|
* -1
|
|
* 1
|
|
*
|
|
* we reset pts to 0 since for us timestamp are unsigned
|
|
*/
|
|
GST_WARNING_OBJECT (demux,
|
|
"negative pts detected: %" G_GINT64_FORMAT " resetting to 0", pts);
|
|
pts = 0;
|
|
}
|
|
timestamp = gst_ffmpeg_time_ff_to_gst (pts, avstream->time_base);
|
|
if (GST_CLOCK_TIME_IS_VALID (timestamp)) {
|
|
stream->last_ts = timestamp;
|
|
}
|
|
duration = gst_ffmpeg_time_ff_to_gst (pkt.duration, avstream->time_base);
|
|
if (G_UNLIKELY (!duration)) {
|
|
GST_WARNING_OBJECT (demux, "invalid buffer duration, setting to NONE");
|
|
duration = GST_CLOCK_TIME_NONE;
|
|
}
|
|
|
|
|
|
GST_DEBUG_OBJECT (demux,
|
|
"pkt pts:%" GST_TIME_FORMAT
|
|
" / size:%d / stream_index:%d / flags:%d / duration:%" GST_TIME_FORMAT
|
|
" / pos:%" G_GINT64_FORMAT, GST_TIME_ARGS (timestamp), pkt.size,
|
|
pkt.stream_index, pkt.flags, GST_TIME_ARGS (duration), (gint64) pkt.pos);
|
|
|
|
/* check start_time */
|
|
#if 0
|
|
if (demux->start_time != -1 && demux->start_time > timestamp)
|
|
goto drop;
|
|
#endif
|
|
|
|
if (GST_CLOCK_TIME_IS_VALID (timestamp)) {
|
|
/* start_time should be the ts of the first frame but it may actually be
|
|
* higher because of rounding when converting to gst ts. */
|
|
if (demux->start_time >= timestamp)
|
|
timestamp = 0;
|
|
else
|
|
timestamp -= demux->start_time;
|
|
}
|
|
|
|
/* check if we ran outside of the segment */
|
|
if (demux->segment.stop != -1 && timestamp > demux->segment.stop)
|
|
goto drop;
|
|
|
|
/* prepare to push packet to peer */
|
|
srcpad = stream->pad;
|
|
|
|
rawvideo = (avstream->codecpar->codec_type == AVMEDIA_TYPE_VIDEO &&
|
|
avstream->codecpar->codec_id == AV_CODEC_ID_RAWVIDEO);
|
|
|
|
if (rawvideo)
|
|
outsize = gst_ffmpeg_avpicture_get_size (avstream->codecpar->format,
|
|
avstream->codecpar->width, avstream->codecpar->height);
|
|
else
|
|
outsize = pkt.size;
|
|
|
|
outbuf = gst_buffer_new_and_alloc (outsize);
|
|
|
|
/* copy the data from packet into the target buffer
|
|
* and do conversions for raw video packets */
|
|
if (rawvideo) {
|
|
AVFrame src, dst;
|
|
const gchar *plugin_name =
|
|
((GstFFMpegDemuxClass *) (G_OBJECT_GET_CLASS (demux)))->in_plugin->name;
|
|
GstMapInfo map;
|
|
|
|
GST_WARNING ("Unknown demuxer %s, no idea what to do", plugin_name);
|
|
gst_ffmpeg_avpicture_fill (&src, pkt.data,
|
|
avstream->codecpar->format, avstream->codecpar->width,
|
|
avstream->codecpar->height);
|
|
|
|
gst_buffer_map (outbuf, &map, GST_MAP_WRITE);
|
|
gst_ffmpeg_avpicture_fill (&dst, map.data,
|
|
avstream->codecpar->format, avstream->codecpar->width,
|
|
avstream->codecpar->height);
|
|
|
|
av_image_copy (dst.data, dst.linesize, (const uint8_t **) src.data,
|
|
src.linesize, avstream->codecpar->format, avstream->codecpar->width,
|
|
avstream->codecpar->height);
|
|
gst_buffer_unmap (outbuf, &map);
|
|
} else {
|
|
gst_buffer_fill (outbuf, 0, pkt.data, outsize);
|
|
}
|
|
|
|
GST_BUFFER_TIMESTAMP (outbuf) = timestamp;
|
|
GST_BUFFER_DURATION (outbuf) = duration;
|
|
|
|
/* mark keyframes */
|
|
if (!(pkt.flags & AV_PKT_FLAG_KEY)) {
|
|
GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_DELTA_UNIT);
|
|
}
|
|
|
|
/* Mark discont */
|
|
if (stream->discont) {
|
|
GST_DEBUG_OBJECT (demux, "marking DISCONT");
|
|
GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_DISCONT);
|
|
stream->discont = FALSE;
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (demux,
|
|
"Sending out buffer time:%" GST_TIME_FORMAT " size:%" G_GSIZE_FORMAT,
|
|
GST_TIME_ARGS (timestamp), gst_buffer_get_size (outbuf));
|
|
|
|
ret = stream_last_flow = gst_pad_push (srcpad, outbuf);
|
|
|
|
/* if a pad is in e.g. WRONG_STATE, we want to pause to unlock the STREAM_LOCK */
|
|
if (((ret = gst_flow_combiner_update_flow (demux->flowcombiner,
|
|
ret)) != GST_FLOW_OK)) {
|
|
GST_WARNING_OBJECT (demux, "stream_movi flow: %s / %s",
|
|
gst_flow_get_name (stream_last_flow), gst_flow_get_name (ret));
|
|
goto pause;
|
|
}
|
|
|
|
done:
|
|
/* can destroy the packet now */
|
|
if (res == 0) {
|
|
av_packet_unref (&pkt);
|
|
}
|
|
|
|
return;
|
|
|
|
/* ERRORS */
|
|
pause:
|
|
{
|
|
GST_LOG_OBJECT (demux, "pausing task, reason %d (%s)", ret,
|
|
gst_flow_get_name (ret));
|
|
if (demux->seekable)
|
|
gst_pad_pause_task (demux->sinkpad);
|
|
else {
|
|
GstFFMpegPipe *ffpipe = &demux->ffpipe;
|
|
|
|
GST_FFMPEG_PIPE_MUTEX_LOCK (ffpipe);
|
|
/* pause task and make sure loop stops */
|
|
gst_task_pause (demux->task);
|
|
g_rec_mutex_lock (&demux->task_lock);
|
|
g_rec_mutex_unlock (&demux->task_lock);
|
|
demux->ffpipe.srcresult = ret;
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
}
|
|
|
|
if (ret == GST_FLOW_EOS) {
|
|
if (demux->segment.flags & GST_SEEK_FLAG_SEGMENT) {
|
|
gint64 stop;
|
|
|
|
if ((stop = demux->segment.stop) == -1)
|
|
stop = demux->segment.duration;
|
|
|
|
GST_LOG_OBJECT (demux, "posting segment done");
|
|
gst_element_post_message (GST_ELEMENT (demux),
|
|
gst_message_new_segment_done (GST_OBJECT (demux),
|
|
demux->segment.format, stop));
|
|
gst_ffmpegdemux_push_event (demux,
|
|
gst_event_new_segment_done (demux->segment.format, stop));
|
|
} else {
|
|
GST_LOG_OBJECT (demux, "pushing eos");
|
|
gst_ffmpegdemux_push_event (demux, gst_event_new_eos ());
|
|
}
|
|
} else if (ret == GST_FLOW_NOT_LINKED || ret < GST_FLOW_EOS) {
|
|
GST_ELEMENT_FLOW_ERROR (demux, ret);
|
|
gst_ffmpegdemux_push_event (demux, gst_event_new_eos ());
|
|
}
|
|
goto done;
|
|
}
|
|
open_failed:
|
|
{
|
|
ret = GST_FLOW_ERROR;
|
|
goto pause;
|
|
}
|
|
read_failed:
|
|
{
|
|
/* something went wrong... */
|
|
GST_WARNING_OBJECT (demux, "av_read_frame returned %d", res);
|
|
|
|
GST_OBJECT_LOCK (demux);
|
|
/* pause appropriatly based on if we are flushing or not */
|
|
if (demux->flushing)
|
|
ret = GST_FLOW_FLUSHING;
|
|
else if (gst_ffmpegdemux_has_outputted (demux)
|
|
|| gst_ffmpegdemux_is_eos (demux)) {
|
|
GST_DEBUG_OBJECT (demux, "We are EOS");
|
|
ret = GST_FLOW_EOS;
|
|
} else
|
|
ret = GST_FLOW_ERROR;
|
|
GST_OBJECT_UNLOCK (demux);
|
|
|
|
goto pause;
|
|
}
|
|
drop:
|
|
{
|
|
GST_DEBUG_OBJECT (demux, "dropping buffer out of segment, stream eos");
|
|
stream->eos = TRUE;
|
|
if (gst_ffmpegdemux_is_eos (demux)) {
|
|
av_packet_unref (&pkt);
|
|
GST_DEBUG_OBJECT (demux, "we are eos");
|
|
ret = GST_FLOW_EOS;
|
|
goto pause;
|
|
} else {
|
|
GST_DEBUG_OBJECT (demux, "some streams are not yet eos");
|
|
goto done;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_sink_event (GstPad * sinkpad, GstObject * parent,
|
|
GstEvent * event)
|
|
{
|
|
GstFFMpegDemux *demux;
|
|
GstFFMpegPipe *ffpipe;
|
|
gboolean result = TRUE;
|
|
|
|
demux = (GstFFMpegDemux *) parent;
|
|
ffpipe = &(demux->ffpipe);
|
|
|
|
GST_LOG_OBJECT (demux, "event: %" GST_PTR_FORMAT, event);
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_FLUSH_START:
|
|
/* forward event */
|
|
gst_pad_event_default (sinkpad, parent, event);
|
|
|
|
/* now unblock the chain function */
|
|
GST_FFMPEG_PIPE_MUTEX_LOCK (ffpipe);
|
|
ffpipe->srcresult = GST_FLOW_FLUSHING;
|
|
GST_FFMPEG_PIPE_SIGNAL (ffpipe);
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
|
|
/* loop might run into WRONG_STATE and end itself,
|
|
* but may also be waiting in a ffmpeg read
|
|
* trying to break that would make ffmpeg believe eos,
|
|
* so no harm to have the loop 'pausing' there ... */
|
|
goto done;
|
|
case GST_EVENT_FLUSH_STOP:
|
|
/* forward event */
|
|
gst_pad_event_default (sinkpad, parent, event);
|
|
|
|
GST_OBJECT_LOCK (demux);
|
|
g_list_foreach (demux->cached_events, (GFunc) gst_mini_object_unref,
|
|
NULL);
|
|
g_list_free (demux->cached_events);
|
|
GST_OBJECT_UNLOCK (demux);
|
|
GST_FFMPEG_PIPE_MUTEX_LOCK (ffpipe);
|
|
gst_adapter_clear (ffpipe->adapter);
|
|
ffpipe->srcresult = GST_FLOW_OK;
|
|
/* loop may have decided to end itself as a result of flush WRONG_STATE */
|
|
gst_task_start (demux->task);
|
|
demux->flushing = FALSE;
|
|
GST_LOG_OBJECT (demux, "loop started");
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
goto done;
|
|
case GST_EVENT_EOS:
|
|
/* inform the src task that it can stop now */
|
|
GST_FFMPEG_PIPE_MUTEX_LOCK (ffpipe);
|
|
ffpipe->eos = TRUE;
|
|
GST_FFMPEG_PIPE_SIGNAL (ffpipe);
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
|
|
/* eat this event for now, task will send eos when finished */
|
|
gst_event_unref (event);
|
|
goto done;
|
|
case GST_EVENT_STREAM_START:
|
|
case GST_EVENT_CAPS:
|
|
GST_LOG_OBJECT (demux, "dropping %s event", GST_EVENT_TYPE_NAME (event));
|
|
gst_event_unref (event);
|
|
goto done;
|
|
default:
|
|
/* for a serialized event, wait until an earlier data is gone,
|
|
* though this is no guarantee as to when task is done with it.
|
|
*
|
|
* If the demuxer isn't opened, push straight away, since we'll
|
|
* be waiting against a cond that will never be signalled. */
|
|
if (GST_EVENT_IS_SERIALIZED (event)) {
|
|
if (demux->opened) {
|
|
GST_FFMPEG_PIPE_MUTEX_LOCK (ffpipe);
|
|
while (!ffpipe->needed)
|
|
GST_FFMPEG_PIPE_WAIT (ffpipe);
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
} else {
|
|
/* queue events and send them later (esp. tag events) */
|
|
GST_OBJECT_LOCK (demux);
|
|
demux->cached_events = g_list_append (demux->cached_events, event);
|
|
GST_OBJECT_UNLOCK (demux);
|
|
goto done;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
result = gst_pad_event_default (sinkpad, parent, event);
|
|
|
|
done:
|
|
|
|
return result;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_ffmpegdemux_chain (GstPad * sinkpad, GstObject * parent, GstBuffer * buffer)
|
|
{
|
|
GstFFMpegDemux *demux;
|
|
GstFFMpegPipe *ffpipe;
|
|
|
|
demux = (GstFFMpegDemux *) parent;
|
|
ffpipe = &demux->ffpipe;
|
|
|
|
GST_FFMPEG_PIPE_MUTEX_LOCK (ffpipe);
|
|
|
|
if (G_UNLIKELY (ffpipe->eos))
|
|
goto eos;
|
|
|
|
if (G_UNLIKELY (ffpipe->srcresult != GST_FLOW_OK))
|
|
goto ignore;
|
|
|
|
GST_DEBUG ("Giving a buffer of %" G_GSIZE_FORMAT " bytes",
|
|
gst_buffer_get_size (buffer));
|
|
gst_adapter_push (ffpipe->adapter, buffer);
|
|
buffer = NULL;
|
|
while (gst_adapter_available (ffpipe->adapter) >= ffpipe->needed) {
|
|
GST_DEBUG ("Adapter has more that requested (ffpipe->needed:%d)",
|
|
ffpipe->needed);
|
|
GST_FFMPEG_PIPE_SIGNAL (ffpipe);
|
|
GST_FFMPEG_PIPE_WAIT (ffpipe);
|
|
/* may have become flushing */
|
|
if (G_UNLIKELY (ffpipe->srcresult != GST_FLOW_OK))
|
|
goto ignore;
|
|
}
|
|
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
|
|
return GST_FLOW_OK;
|
|
|
|
/* special cases */
|
|
eos:
|
|
{
|
|
GST_DEBUG_OBJECT (demux, "ignoring buffer at end-of-stream");
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
|
|
gst_buffer_unref (buffer);
|
|
return GST_FLOW_EOS;
|
|
}
|
|
ignore:
|
|
{
|
|
GST_DEBUG_OBJECT (demux, "ignoring buffer because src task encountered %s",
|
|
gst_flow_get_name (ffpipe->srcresult));
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
|
|
if (buffer)
|
|
gst_buffer_unref (buffer);
|
|
return GST_FLOW_FLUSHING;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_sink_activate (GstPad * sinkpad, GstObject * parent)
|
|
{
|
|
GstQuery *query;
|
|
gboolean pull_mode;
|
|
GstSchedulingFlags flags;
|
|
|
|
query = gst_query_new_scheduling ();
|
|
|
|
if (!gst_pad_peer_query (sinkpad, query)) {
|
|
gst_query_unref (query);
|
|
goto activate_push;
|
|
}
|
|
|
|
pull_mode = gst_query_has_scheduling_mode_with_flags (query,
|
|
GST_PAD_MODE_PULL, GST_SCHEDULING_FLAG_SEEKABLE);
|
|
|
|
gst_query_parse_scheduling (query, &flags, NULL, NULL, NULL);
|
|
if (flags & GST_SCHEDULING_FLAG_SEQUENTIAL)
|
|
pull_mode = FALSE;
|
|
|
|
gst_query_unref (query);
|
|
|
|
if (!pull_mode)
|
|
goto activate_push;
|
|
|
|
GST_DEBUG_OBJECT (sinkpad, "activating pull");
|
|
return gst_pad_activate_mode (sinkpad, GST_PAD_MODE_PULL, TRUE);
|
|
|
|
activate_push:
|
|
{
|
|
GST_DEBUG_OBJECT (sinkpad, "activating push");
|
|
return gst_pad_activate_mode (sinkpad, GST_PAD_MODE_PUSH, TRUE);
|
|
}
|
|
}
|
|
|
|
/* push mode:
|
|
* - not seekable
|
|
* - use gstpipe protocol, like ffmpeg's pipe protocol
|
|
* - (independently managed) task driving ffmpeg
|
|
*/
|
|
static gboolean
|
|
gst_ffmpegdemux_sink_activate_push (GstPad * sinkpad, GstObject * parent,
|
|
gboolean active)
|
|
{
|
|
GstFFMpegDemux *demux;
|
|
gboolean res = FALSE;
|
|
|
|
demux = (GstFFMpegDemux *) (parent);
|
|
|
|
if (active) {
|
|
if (demux->can_push == FALSE) {
|
|
GST_WARNING_OBJECT (demux, "Demuxer can't reliably operate in push-mode");
|
|
goto beach;
|
|
}
|
|
demux->ffpipe.eos = FALSE;
|
|
demux->ffpipe.srcresult = GST_FLOW_OK;
|
|
demux->ffpipe.needed = 0;
|
|
demux->seekable = FALSE;
|
|
res = gst_task_start (demux->task);
|
|
} else {
|
|
GstFFMpegPipe *ffpipe = &demux->ffpipe;
|
|
|
|
/* release chain and loop */
|
|
GST_FFMPEG_PIPE_MUTEX_LOCK (ffpipe);
|
|
demux->ffpipe.srcresult = GST_FLOW_FLUSHING;
|
|
/* end streaming by making ffmpeg believe eos */
|
|
demux->ffpipe.eos = TRUE;
|
|
GST_FFMPEG_PIPE_SIGNAL (ffpipe);
|
|
GST_FFMPEG_PIPE_MUTEX_UNLOCK (ffpipe);
|
|
|
|
/* make sure streaming ends */
|
|
gst_task_stop (demux->task);
|
|
g_rec_mutex_lock (&demux->task_lock);
|
|
g_rec_mutex_unlock (&demux->task_lock);
|
|
res = gst_task_join (demux->task);
|
|
demux->seekable = FALSE;
|
|
}
|
|
|
|
beach:
|
|
return res;
|
|
}
|
|
|
|
/* pull mode:
|
|
* - seekable
|
|
* - use gstreamer protocol, like ffmpeg's file protocol
|
|
* - task driving ffmpeg based on sink pad
|
|
*/
|
|
static gboolean
|
|
gst_ffmpegdemux_sink_activate_pull (GstPad * sinkpad, GstObject * parent,
|
|
gboolean active)
|
|
{
|
|
GstFFMpegDemux *demux;
|
|
gboolean res;
|
|
|
|
demux = (GstFFMpegDemux *) parent;
|
|
|
|
if (active) {
|
|
demux->seekable = TRUE;
|
|
res = gst_pad_start_task (sinkpad, (GstTaskFunction) gst_ffmpegdemux_loop,
|
|
demux, NULL);
|
|
} else {
|
|
res = gst_pad_stop_task (sinkpad);
|
|
demux->seekable = FALSE;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
static gboolean
|
|
gst_ffmpegdemux_sink_activate_mode (GstPad * sinkpad, GstObject * parent,
|
|
GstPadMode mode, gboolean active)
|
|
{
|
|
gboolean res;
|
|
|
|
switch (mode) {
|
|
case GST_PAD_MODE_PUSH:
|
|
res = gst_ffmpegdemux_sink_activate_push (sinkpad, parent, active);
|
|
break;
|
|
case GST_PAD_MODE_PULL:
|
|
res = gst_ffmpegdemux_sink_activate_pull (sinkpad, parent, active);
|
|
break;
|
|
default:
|
|
res = FALSE;
|
|
break;
|
|
}
|
|
return res;
|
|
}
|
|
|
|
static GstStateChangeReturn
|
|
gst_ffmpegdemux_change_state (GstElement * element, GstStateChange transition)
|
|
{
|
|
GstFFMpegDemux *demux = (GstFFMpegDemux *) (element);
|
|
GstStateChangeReturn ret;
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_READY_TO_PAUSED:
|
|
#if 0
|
|
/* test seek in READY here */
|
|
gst_element_send_event (element, gst_event_new_seek (1.0,
|
|
GST_FORMAT_TIME, GST_SEEK_FLAG_NONE,
|
|
GST_SEEK_TYPE_SET, 10 * GST_SECOND,
|
|
GST_SEEK_TYPE_SET, 13 * GST_SECOND));
|
|
#endif
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_PAUSED_TO_READY:
|
|
gst_ffmpegdemux_close (demux);
|
|
gst_adapter_clear (demux->ffpipe.adapter);
|
|
g_list_foreach (demux->cached_events, (GFunc) gst_mini_object_unref,
|
|
NULL);
|
|
g_list_free (demux->cached_events);
|
|
demux->cached_events = NULL;
|
|
demux->have_group_id = FALSE;
|
|
demux->group_id = G_MAXUINT;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
gboolean
|
|
gst_ffmpegdemux_register (GstPlugin * plugin)
|
|
{
|
|
GType type;
|
|
const AVInputFormat *in_plugin;
|
|
gchar *extensions;
|
|
GTypeInfo typeinfo = {
|
|
sizeof (GstFFMpegDemuxClass),
|
|
(GBaseInitFunc) gst_ffmpegdemux_base_init,
|
|
NULL,
|
|
(GClassInitFunc) gst_ffmpegdemux_class_init,
|
|
NULL,
|
|
NULL,
|
|
sizeof (GstFFMpegDemux),
|
|
0,
|
|
(GInstanceInitFunc) gst_ffmpegdemux_init,
|
|
};
|
|
|
|
void *i = 0;
|
|
|
|
GST_LOG ("Registering demuxers");
|
|
|
|
while ((in_plugin = av_demuxer_iterate (&i))) {
|
|
gchar *type_name, *typefind_name;
|
|
gint rank;
|
|
gboolean register_typefind_func = TRUE;
|
|
|
|
GST_LOG ("Attempting to handle libav demuxer plugin %s [%s]",
|
|
in_plugin->name, in_plugin->long_name);
|
|
|
|
/* no emulators */
|
|
if (in_plugin->long_name != NULL) {
|
|
if (!strncmp (in_plugin->long_name, "raw ", 4) ||
|
|
!strncmp (in_plugin->long_name, "pcm ", 4)
|
|
)
|
|
continue;
|
|
}
|
|
|
|
if (!strcmp (in_plugin->name, "audio_device") ||
|
|
!strncmp (in_plugin->name, "image", 5) ||
|
|
!strcmp (in_plugin->name, "mpegvideo") ||
|
|
!strcmp (in_plugin->name, "mjpeg") ||
|
|
!strcmp (in_plugin->name, "redir") ||
|
|
!strncmp (in_plugin->name, "u8", 2) ||
|
|
!strncmp (in_plugin->name, "u16", 3) ||
|
|
!strncmp (in_plugin->name, "u24", 3) ||
|
|
!strncmp (in_plugin->name, "u32", 3) ||
|
|
!strncmp (in_plugin->name, "s8", 2) ||
|
|
!strncmp (in_plugin->name, "s16", 3) ||
|
|
!strncmp (in_plugin->name, "s24", 3) ||
|
|
!strncmp (in_plugin->name, "s32", 3) ||
|
|
!strncmp (in_plugin->name, "f32", 3) ||
|
|
!strncmp (in_plugin->name, "f64", 3) ||
|
|
!strcmp (in_plugin->name, "mulaw") || !strcmp (in_plugin->name, "alaw")
|
|
)
|
|
continue;
|
|
|
|
/* no network demuxers */
|
|
if (!strcmp (in_plugin->name, "sdp") ||
|
|
!strcmp (in_plugin->name, "rtsp") ||
|
|
!strcmp (in_plugin->name, "applehttp")
|
|
)
|
|
continue;
|
|
|
|
/* these don't do what one would expect or
|
|
* are only partially functional/useful */
|
|
if (!strcmp (in_plugin->name, "aac") ||
|
|
!strcmp (in_plugin->name, "wv") ||
|
|
!strcmp (in_plugin->name, "ass") ||
|
|
!strcmp (in_plugin->name, "ffmetadata"))
|
|
continue;
|
|
|
|
/* Don't use the typefind functions of formats for which we already have
|
|
* better typefind functions */
|
|
if (!strcmp (in_plugin->name, "mov,mp4,m4a,3gp,3g2,mj2") ||
|
|
!strcmp (in_plugin->name, "ass") ||
|
|
!strcmp (in_plugin->name, "avi") ||
|
|
!strcmp (in_plugin->name, "asf") ||
|
|
!strcmp (in_plugin->name, "mpegvideo") ||
|
|
!strcmp (in_plugin->name, "mp3") ||
|
|
!strcmp (in_plugin->name, "matroska") ||
|
|
!strcmp (in_plugin->name, "matroska_webm") ||
|
|
!strcmp (in_plugin->name, "matroska,webm") ||
|
|
!strcmp (in_plugin->name, "mpeg") ||
|
|
!strcmp (in_plugin->name, "wav") ||
|
|
!strcmp (in_plugin->name, "au") ||
|
|
!strcmp (in_plugin->name, "tta") ||
|
|
!strcmp (in_plugin->name, "rm") ||
|
|
!strcmp (in_plugin->name, "amr") ||
|
|
!strcmp (in_plugin->name, "ogg") ||
|
|
!strcmp (in_plugin->name, "aiff") ||
|
|
!strcmp (in_plugin->name, "ape") ||
|
|
!strcmp (in_plugin->name, "dv") ||
|
|
!strcmp (in_plugin->name, "flv") ||
|
|
!strcmp (in_plugin->name, "mpc") ||
|
|
!strcmp (in_plugin->name, "mpc8") ||
|
|
!strcmp (in_plugin->name, "mpegts") ||
|
|
!strcmp (in_plugin->name, "mpegtsraw") ||
|
|
!strcmp (in_plugin->name, "mxf") ||
|
|
!strcmp (in_plugin->name, "nuv") ||
|
|
!strcmp (in_plugin->name, "swf") ||
|
|
!strcmp (in_plugin->name, "voc") ||
|
|
!strcmp (in_plugin->name, "pva") ||
|
|
!strcmp (in_plugin->name, "gif") ||
|
|
!strcmp (in_plugin->name, "vc1test") ||
|
|
!strcmp (in_plugin->name, "ivf"))
|
|
register_typefind_func = FALSE;
|
|
|
|
/* Set the rank of demuxers known to work to MARGINAL.
|
|
* Set demuxers for which we already have another implementation to NONE
|
|
* Set All others to NONE*/
|
|
/**
|
|
* element-avdemux_xwma
|
|
*
|
|
* Since: 1.20
|
|
*/
|
|
if (!strcmp (in_plugin->name, "wsvqa") ||
|
|
!strcmp (in_plugin->name, "wsaud") ||
|
|
!strcmp (in_plugin->name, "wc3movie") ||
|
|
!strcmp (in_plugin->name, "voc") ||
|
|
!strcmp (in_plugin->name, "tta") ||
|
|
!strcmp (in_plugin->name, "sol") ||
|
|
!strcmp (in_plugin->name, "smk") ||
|
|
!strcmp (in_plugin->name, "vmd") ||
|
|
!strcmp (in_plugin->name, "film_cpk") ||
|
|
!strcmp (in_plugin->name, "ingenient") ||
|
|
!strcmp (in_plugin->name, "psxstr") ||
|
|
!strcmp (in_plugin->name, "nuv") ||
|
|
!strcmp (in_plugin->name, "nut") ||
|
|
!strcmp (in_plugin->name, "nsv") ||
|
|
!strcmp (in_plugin->name, "mxf") ||
|
|
!strcmp (in_plugin->name, "mmf") ||
|
|
!strcmp (in_plugin->name, "mm") ||
|
|
!strcmp (in_plugin->name, "ipmovie") ||
|
|
!strcmp (in_plugin->name, "ape") ||
|
|
!strcmp (in_plugin->name, "RoQ") ||
|
|
!strcmp (in_plugin->name, "idcin") ||
|
|
!strcmp (in_plugin->name, "gxf") ||
|
|
!strcmp (in_plugin->name, "ffm") ||
|
|
!strcmp (in_plugin->name, "ea") ||
|
|
!strcmp (in_plugin->name, "daud") ||
|
|
!strcmp (in_plugin->name, "avs") ||
|
|
!strcmp (in_plugin->name, "aiff") ||
|
|
!strcmp (in_plugin->name, "xwma") ||
|
|
!strcmp (in_plugin->name, "4xm") ||
|
|
!strcmp (in_plugin->name, "yuv4mpegpipe") ||
|
|
!strcmp (in_plugin->name, "pva") ||
|
|
!strcmp (in_plugin->name, "mpc") ||
|
|
!strcmp (in_plugin->name, "mpc8") ||
|
|
!strcmp (in_plugin->name, "ivf") ||
|
|
!strcmp (in_plugin->name, "brstm") ||
|
|
!strcmp (in_plugin->name, "bfstm") ||
|
|
!strcmp (in_plugin->name, "gif") ||
|
|
!strcmp (in_plugin->name, "dsf") || !strcmp (in_plugin->name, "iff"))
|
|
rank = GST_RANK_MARGINAL;
|
|
else {
|
|
GST_DEBUG ("ignoring %s", in_plugin->name);
|
|
rank = GST_RANK_NONE;
|
|
continue;
|
|
}
|
|
|
|
/* construct the type */
|
|
type_name = g_strdup_printf ("avdemux_%s", in_plugin->name);
|
|
g_strdelimit (type_name, ".,|-<> ", '_');
|
|
|
|
/* if it's already registered, drop it */
|
|
if (g_type_from_name (type_name)) {
|
|
g_free (type_name);
|
|
continue;
|
|
}
|
|
|
|
typefind_name = g_strdup_printf ("avtype_%s", in_plugin->name);
|
|
g_strdelimit (typefind_name, ".,|-<> ", '_');
|
|
|
|
/* create the type now */
|
|
type = g_type_register_static (GST_TYPE_ELEMENT, type_name, &typeinfo, 0);
|
|
g_type_set_qdata (type, GST_FFDEMUX_PARAMS_QDATA, (gpointer) in_plugin);
|
|
|
|
if (in_plugin->extensions)
|
|
extensions = g_strdelimit (g_strdup (in_plugin->extensions), " ", ',');
|
|
else
|
|
extensions = NULL;
|
|
|
|
if (!gst_element_register (plugin, type_name, rank, type) ||
|
|
(register_typefind_func == TRUE &&
|
|
!gst_type_find_register (plugin, typefind_name, rank,
|
|
gst_ffmpegdemux_type_find, extensions, NULL,
|
|
(gpointer) in_plugin, NULL))) {
|
|
g_warning ("Registration of type %s failed", type_name);
|
|
g_free (type_name);
|
|
g_free (typefind_name);
|
|
g_free (extensions);
|
|
return FALSE;
|
|
}
|
|
|
|
g_free (type_name);
|
|
g_free (typefind_name);
|
|
g_free (extensions);
|
|
}
|
|
|
|
GST_LOG ("Finished registering demuxers");
|
|
|
|
return TRUE;
|
|
}
|