/* GStreamer * Copyright (C) 2006 Edward Hervey * Copyright (C) 2007 Jan Schmidt * Copyright (C) 2007 Wim Taymans * Copyright (C) 2011 Sebastian Dröge * * gstmultiqueue.c: * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:element-multiqueue * @see_also: #GstQueue * * * * Multiqueue is similar to a normal #GstQueue with the following additional * features: * * * Multiple streamhandling * * The element handles queueing data on more than one stream at once. To * achieve such a feature it has request sink pads (sink%u) and * 'sometimes' src pads (src%u). * * When requesting a given sinkpad with gst_element_request_pad(), * the associated srcpad for that stream will be created. * Example: requesting sink1 will generate src1. * * * * * Non-starvation on multiple streams * * If more than one stream is used with the element, the streams' queues * will be dynamically grown (up to a limit), in order to ensure that no * stream is risking data starvation. This guarantees that at any given * time there are at least N bytes queued and available for each individual * stream. * * If an EOS event comes through a srcpad, the associated queue will be * considered as 'not-empty' in the queue-size-growing algorithm. * * * * * Non-linked srcpads graceful handling * * In order to better support dynamic switching between streams, the multiqueue * (unlike the current GStreamer queue) continues to push buffers on non-linked * pads rather than shutting down. * * In addition, to prevent a non-linked stream from very quickly consuming all * available buffers and thus 'racing ahead' of the other streams, the element * must ensure that buffers and inlined events for a non-linked stream are pushed * in the same order as they were received, relative to the other streams * controlled by the element. This means that a buffer cannot be pushed to a * non-linked pad any sooner than buffers in any other stream which were received * before it. * * * * * * * Data is queued until one of the limits specified by the * #GstMultiQueue:max-size-buffers, #GstMultiQueue:max-size-bytes and/or * #GstMultiQueue:max-size-time properties has been reached. Any attempt to push * more buffers into the queue will block the pushing thread until more space * becomes available. #GstMultiQueue:extra-size-buffers, * * * #GstMultiQueue:extra-size-bytes and #GstMultiQueue:extra-size-time are * currently unused. * * * The default queue size limits are 5 buffers, 10MB of data, or * two second worth of data, whichever is reached first. Note that the number * of buffers will dynamically grow depending on the fill level of * other queues. * * * The #GstMultiQueue::underrun signal is emitted when all of the queues * are empty. The #GstMultiQueue::overrun signal is emitted when one of the * queues is filled. * Both signals are emitted from the context of the streaming thread. * * */ #ifdef HAVE_CONFIG_H # include "config.h" #endif #include #include #include "gstmultiqueue.h" #include /** * GstSingleQueue: * @sinkpad: associated sink #GstPad * @srcpad: associated source #GstPad * * Structure containing all information and properties about * a single queue. */ typedef struct _GstSingleQueue GstSingleQueue; struct _GstSingleQueue { /* unique identifier of the queue */ guint id; GstMultiQueue *mqueue; GstPad *sinkpad; GstPad *srcpad; /* flowreturn of previous srcpad push */ GstFlowReturn srcresult; /* If something was actually pushed on * this pad after flushing/pad activation * and the srcresult corresponds to something * real */ gboolean pushed; /* segments */ GstSegment sink_segment; GstSegment src_segment; gboolean has_src_segment; /* preferred over initializing the src_segment to * UNDEFINED as this doesn't requires adding ifs * in every segment usage */ /* position of src/sink */ GstClockTime sinktime, srctime; /* TRUE if either position needs to be recalculated */ gboolean sink_tainted, src_tainted; /* queue of data */ GstDataQueue *queue; GstDataQueueSize max_size, extra_size; GstClockTime cur_time; gboolean is_eos; gboolean flushing; /* Protected by global lock */ guint32 nextid; /* ID of the next object waiting to be pushed */ guint32 oldid; /* ID of the last object pushed (last in a series) */ guint32 last_oldid; /* Previously observed old_id, reset to MAXUINT32 on flush */ GstClockTime next_time; /* End running time of next buffer to be pushed */ GstClockTime last_time; /* Start running time of last pushed buffer */ GCond turn; /* SingleQueue turn waiting conditional */ /* for serialized queries */ GCond query_handled; gboolean last_query; GstQuery *last_handled_query; }; /* Extension of GstDataQueueItem structure for our usage */ typedef struct _GstMultiQueueItem GstMultiQueueItem; struct _GstMultiQueueItem { GstMiniObject *object; guint size; guint64 duration; gboolean visible; GDestroyNotify destroy; guint32 posid; gboolean is_query; }; static GstSingleQueue *gst_single_queue_new (GstMultiQueue * mqueue, guint id); static void gst_single_queue_free (GstSingleQueue * squeue); static void wake_up_next_non_linked (GstMultiQueue * mq); static void compute_high_id (GstMultiQueue * mq); static void compute_high_time (GstMultiQueue * mq); static void single_queue_overrun_cb (GstDataQueue * dq, GstSingleQueue * sq); static void single_queue_underrun_cb (GstDataQueue * dq, GstSingleQueue * sq); static void update_buffering (GstMultiQueue * mq, GstSingleQueue * sq); static void gst_multi_queue_post_buffering (GstMultiQueue * mq); static void gst_single_queue_flush_queue (GstSingleQueue * sq, gboolean full); static GstStaticPadTemplate sinktemplate = GST_STATIC_PAD_TEMPLATE ("sink_%u", GST_PAD_SINK, GST_PAD_REQUEST, GST_STATIC_CAPS_ANY); static GstStaticPadTemplate srctemplate = GST_STATIC_PAD_TEMPLATE ("src_%u", GST_PAD_SRC, GST_PAD_SOMETIMES, GST_STATIC_CAPS_ANY); GST_DEBUG_CATEGORY_STATIC (multi_queue_debug); #define GST_CAT_DEFAULT (multi_queue_debug) /* Signals and args */ enum { SIGNAL_UNDERRUN, SIGNAL_OVERRUN, LAST_SIGNAL }; /* default limits, we try to keep up to 2 seconds of data and if there is not * time, up to 10 MB. The number of buffers is dynamically scaled to make sure * there is data in the queues. Normally, the byte and time limits are not hit * in theses conditions. */ #define DEFAULT_MAX_SIZE_BYTES 10 * 1024 * 1024 /* 10 MB */ #define DEFAULT_MAX_SIZE_BUFFERS 5 #define DEFAULT_MAX_SIZE_TIME 2 * GST_SECOND /* second limits. When we hit one of the above limits we are probably dealing * with a badly muxed file and we scale the limits to these emergency values. * This is currently not yet implemented. * Since we dynamically scale the queue buffer size up to the limits but avoid * going above the max-size-buffers when we can, we don't really need this * aditional extra size. */ #define DEFAULT_EXTRA_SIZE_BYTES 10 * 1024 * 1024 /* 10 MB */ #define DEFAULT_EXTRA_SIZE_BUFFERS 5 #define DEFAULT_EXTRA_SIZE_TIME 3 * GST_SECOND #define DEFAULT_USE_BUFFERING FALSE #define DEFAULT_LOW_PERCENT 10 #define DEFAULT_HIGH_PERCENT 99 #define DEFAULT_SYNC_BY_RUNNING_TIME FALSE enum { PROP_0, PROP_EXTRA_SIZE_BYTES, PROP_EXTRA_SIZE_BUFFERS, PROP_EXTRA_SIZE_TIME, PROP_MAX_SIZE_BYTES, PROP_MAX_SIZE_BUFFERS, PROP_MAX_SIZE_TIME, PROP_USE_BUFFERING, PROP_LOW_PERCENT, PROP_HIGH_PERCENT, PROP_SYNC_BY_RUNNING_TIME, PROP_LAST }; #define GST_MULTI_QUEUE_MUTEX_LOCK(q) G_STMT_START { \ g_mutex_lock (&q->qlock); \ } G_STMT_END #define GST_MULTI_QUEUE_MUTEX_UNLOCK(q) G_STMT_START { \ g_mutex_unlock (&q->qlock); \ } G_STMT_END #define SET_PERCENT(mq, perc) G_STMT_START { \ if (perc != mq->percent) { \ mq->percent = perc; \ mq->percent_changed = TRUE; \ GST_DEBUG_OBJECT (mq, "buffering %d percent", perc); \ } \ } G_STMT_END static void gst_multi_queue_finalize (GObject * object); static void gst_multi_queue_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_multi_queue_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static GstPad *gst_multi_queue_request_new_pad (GstElement * element, GstPadTemplate * temp, const gchar * name, const GstCaps * caps); static void gst_multi_queue_release_pad (GstElement * element, GstPad * pad); static GstStateChangeReturn gst_multi_queue_change_state (GstElement * element, GstStateChange transition); static void gst_multi_queue_loop (GstPad * pad); #define _do_init \ GST_DEBUG_CATEGORY_INIT (multi_queue_debug, "multiqueue", 0, "multiqueue element"); #define gst_multi_queue_parent_class parent_class G_DEFINE_TYPE_WITH_CODE (GstMultiQueue, gst_multi_queue, GST_TYPE_ELEMENT, _do_init); static guint gst_multi_queue_signals[LAST_SIGNAL] = { 0 }; static void gst_multi_queue_class_init (GstMultiQueueClass * klass) { GObjectClass *gobject_class = G_OBJECT_CLASS (klass); GstElementClass *gstelement_class = GST_ELEMENT_CLASS (klass); gobject_class->set_property = gst_multi_queue_set_property; gobject_class->get_property = gst_multi_queue_get_property; /* SIGNALS */ /** * GstMultiQueue::underrun: * @multiqueue: the multiqueue instance * * This signal is emitted from the streaming thread when there is * no data in any of the queues inside the multiqueue instance (underrun). * * This indicates either starvation or EOS from the upstream data sources. */ gst_multi_queue_signals[SIGNAL_UNDERRUN] = g_signal_new ("underrun", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_FIRST, G_STRUCT_OFFSET (GstMultiQueueClass, underrun), NULL, NULL, g_cclosure_marshal_VOID__VOID, G_TYPE_NONE, 0); /** * GstMultiQueue::overrun: * @multiqueue: the multiqueue instance * * Reports that one of the queues in the multiqueue is full (overrun). * A queue is full if the total amount of data inside it (num-buffers, time, * size) is higher than the boundary values which can be set through the * GObject properties. * * This can be used as an indicator of pre-roll. */ gst_multi_queue_signals[SIGNAL_OVERRUN] = g_signal_new ("overrun", G_TYPE_FROM_CLASS (klass), G_SIGNAL_RUN_FIRST, G_STRUCT_OFFSET (GstMultiQueueClass, overrun), NULL, NULL, g_cclosure_marshal_VOID__VOID, G_TYPE_NONE, 0); /* PROPERTIES */ g_object_class_install_property (gobject_class, PROP_MAX_SIZE_BYTES, g_param_spec_uint ("max-size-bytes", "Max. size (kB)", "Max. amount of data in the queue (bytes, 0=disable)", 0, G_MAXUINT, DEFAULT_MAX_SIZE_BYTES, G_PARAM_READWRITE | GST_PARAM_MUTABLE_PLAYING | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_MAX_SIZE_BUFFERS, g_param_spec_uint ("max-size-buffers", "Max. size (buffers)", "Max. number of buffers in the queue (0=disable)", 0, G_MAXUINT, DEFAULT_MAX_SIZE_BUFFERS, G_PARAM_READWRITE | GST_PARAM_MUTABLE_PLAYING | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_MAX_SIZE_TIME, g_param_spec_uint64 ("max-size-time", "Max. size (ns)", "Max. amount of data in the queue (in ns, 0=disable)", 0, G_MAXUINT64, DEFAULT_MAX_SIZE_TIME, G_PARAM_READWRITE | GST_PARAM_MUTABLE_PLAYING | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_EXTRA_SIZE_BYTES, g_param_spec_uint ("extra-size-bytes", "Extra Size (kB)", "Amount of data the queues can grow if one of them is empty (bytes, 0=disable)" " (NOT IMPLEMENTED)", 0, G_MAXUINT, DEFAULT_EXTRA_SIZE_BYTES, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_EXTRA_SIZE_BUFFERS, g_param_spec_uint ("extra-size-buffers", "Extra Size (buffers)", "Amount of buffers the queues can grow if one of them is empty (0=disable)" " (NOT IMPLEMENTED)", 0, G_MAXUINT, DEFAULT_EXTRA_SIZE_BUFFERS, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_EXTRA_SIZE_TIME, g_param_spec_uint64 ("extra-size-time", "Extra Size (ns)", "Amount of time the queues can grow if one of them is empty (in ns, 0=disable)" " (NOT IMPLEMENTED)", 0, G_MAXUINT64, DEFAULT_EXTRA_SIZE_TIME, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstMultiQueue:use-buffering * * Enable the buffering option in multiqueue so that BUFFERING messages are * emitted based on low-/high-percent thresholds. */ g_object_class_install_property (gobject_class, PROP_USE_BUFFERING, g_param_spec_boolean ("use-buffering", "Use buffering", "Emit GST_MESSAGE_BUFFERING based on low-/high-percent thresholds", DEFAULT_USE_BUFFERING, G_PARAM_READWRITE | GST_PARAM_MUTABLE_PLAYING | G_PARAM_STATIC_STRINGS)); /** * GstMultiQueue:low-percent * * Low threshold percent for buffering to start. */ g_object_class_install_property (gobject_class, PROP_LOW_PERCENT, g_param_spec_int ("low-percent", "Low percent", "Low threshold for buffering to start", 0, 100, DEFAULT_LOW_PERCENT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstMultiQueue:high-percent * * High threshold percent for buffering to finish. */ g_object_class_install_property (gobject_class, PROP_HIGH_PERCENT, g_param_spec_int ("high-percent", "High percent", "High threshold for buffering to finish", 0, 100, DEFAULT_HIGH_PERCENT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstMultiQueue:sync-by-running-time * * If enabled multiqueue will synchronize deactivated or not-linked streams * to the activated and linked streams by taking the running time. * Otherwise multiqueue will synchronize the deactivated or not-linked * streams by keeping the order in which buffers and events arrived compared * to active and linked streams. */ g_object_class_install_property (gobject_class, PROP_SYNC_BY_RUNNING_TIME, g_param_spec_boolean ("sync-by-running-time", "Sync By Running Time", "Synchronize deactivated or not-linked streams by running time", DEFAULT_SYNC_BY_RUNNING_TIME, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); gobject_class->finalize = gst_multi_queue_finalize; gst_element_class_set_static_metadata (gstelement_class, "MultiQueue", "Generic", "Multiple data queue", "Edward Hervey "); gst_element_class_add_pad_template (gstelement_class, gst_static_pad_template_get (&sinktemplate)); gst_element_class_add_pad_template (gstelement_class, gst_static_pad_template_get (&srctemplate)); gstelement_class->request_new_pad = GST_DEBUG_FUNCPTR (gst_multi_queue_request_new_pad); gstelement_class->release_pad = GST_DEBUG_FUNCPTR (gst_multi_queue_release_pad); gstelement_class->change_state = GST_DEBUG_FUNCPTR (gst_multi_queue_change_state); } static void gst_multi_queue_init (GstMultiQueue * mqueue) { mqueue->nbqueues = 0; mqueue->queues = NULL; mqueue->max_size.bytes = DEFAULT_MAX_SIZE_BYTES; mqueue->max_size.visible = DEFAULT_MAX_SIZE_BUFFERS; mqueue->max_size.time = DEFAULT_MAX_SIZE_TIME; mqueue->extra_size.bytes = DEFAULT_EXTRA_SIZE_BYTES; mqueue->extra_size.visible = DEFAULT_EXTRA_SIZE_BUFFERS; mqueue->extra_size.time = DEFAULT_EXTRA_SIZE_TIME; mqueue->use_buffering = DEFAULT_USE_BUFFERING; mqueue->low_percent = DEFAULT_LOW_PERCENT; mqueue->high_percent = DEFAULT_HIGH_PERCENT; mqueue->sync_by_running_time = DEFAULT_SYNC_BY_RUNNING_TIME; mqueue->counter = 1; mqueue->highid = -1; mqueue->high_time = GST_CLOCK_TIME_NONE; g_mutex_init (&mqueue->qlock); g_mutex_init (&mqueue->buffering_post_lock); } static void gst_multi_queue_finalize (GObject * object) { GstMultiQueue *mqueue = GST_MULTI_QUEUE (object); g_list_foreach (mqueue->queues, (GFunc) gst_single_queue_free, NULL); g_list_free (mqueue->queues); mqueue->queues = NULL; mqueue->queues_cookie++; /* free/unref instance data */ g_mutex_clear (&mqueue->qlock); g_mutex_clear (&mqueue->buffering_post_lock); G_OBJECT_CLASS (parent_class)->finalize (object); } #define SET_CHILD_PROPERTY(mq,format) G_STMT_START { \ GList * tmp = mq->queues; \ while (tmp) { \ GstSingleQueue *q = (GstSingleQueue*)tmp->data; \ q->max_size.format = mq->max_size.format; \ update_buffering (mq, q); \ gst_data_queue_limits_changed (q->queue); \ tmp = g_list_next(tmp); \ }; \ } G_STMT_END static void gst_multi_queue_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstMultiQueue *mq = GST_MULTI_QUEUE (object); switch (prop_id) { case PROP_MAX_SIZE_BYTES: GST_MULTI_QUEUE_MUTEX_LOCK (mq); mq->max_size.bytes = g_value_get_uint (value); SET_CHILD_PROPERTY (mq, bytes); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_multi_queue_post_buffering (mq); break; case PROP_MAX_SIZE_BUFFERS: { GList *tmp; gint new_size = g_value_get_uint (value); GST_MULTI_QUEUE_MUTEX_LOCK (mq); mq->max_size.visible = new_size; tmp = mq->queues; while (tmp) { GstDataQueueSize size; GstSingleQueue *q = (GstSingleQueue *) tmp->data; gst_data_queue_get_level (q->queue, &size); GST_DEBUG_OBJECT (mq, "Queue %d: Requested buffers size: %d," " current: %d, current max %d", q->id, new_size, size.visible, q->max_size.visible); /* do not reduce max size below current level if the single queue * has grown because of empty queue */ if (new_size == 0) { q->max_size.visible = new_size; } else if (q->max_size.visible == 0) { q->max_size.visible = MAX (new_size, size.visible); } else if (new_size > size.visible) { q->max_size.visible = new_size; } update_buffering (mq, q); gst_data_queue_limits_changed (q->queue); tmp = g_list_next (tmp); } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_multi_queue_post_buffering (mq); break; } case PROP_MAX_SIZE_TIME: GST_MULTI_QUEUE_MUTEX_LOCK (mq); mq->max_size.time = g_value_get_uint64 (value); SET_CHILD_PROPERTY (mq, time); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_multi_queue_post_buffering (mq); break; case PROP_EXTRA_SIZE_BYTES: mq->extra_size.bytes = g_value_get_uint (value); break; case PROP_EXTRA_SIZE_BUFFERS: mq->extra_size.visible = g_value_get_uint (value); break; case PROP_EXTRA_SIZE_TIME: mq->extra_size.time = g_value_get_uint64 (value); break; case PROP_USE_BUFFERING: mq->use_buffering = g_value_get_boolean (value); if (!mq->use_buffering && mq->buffering) { GST_MULTI_QUEUE_MUTEX_LOCK (mq); mq->buffering = FALSE; GST_DEBUG_OBJECT (mq, "buffering 100 percent"); SET_PERCENT (mq, 100); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); } if (mq->use_buffering) { GList *tmp; GST_MULTI_QUEUE_MUTEX_LOCK (mq); tmp = mq->queues; while (tmp) { GstSingleQueue *q = (GstSingleQueue *) tmp->data; update_buffering (mq, q); gst_data_queue_limits_changed (q->queue); tmp = g_list_next (tmp); } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); } gst_multi_queue_post_buffering (mq); break; case PROP_LOW_PERCENT: mq->low_percent = g_value_get_int (value); break; case PROP_HIGH_PERCENT: mq->high_percent = g_value_get_int (value); break; case PROP_SYNC_BY_RUNNING_TIME: mq->sync_by_running_time = g_value_get_boolean (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static void gst_multi_queue_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstMultiQueue *mq = GST_MULTI_QUEUE (object); GST_MULTI_QUEUE_MUTEX_LOCK (mq); switch (prop_id) { case PROP_EXTRA_SIZE_BYTES: g_value_set_uint (value, mq->extra_size.bytes); break; case PROP_EXTRA_SIZE_BUFFERS: g_value_set_uint (value, mq->extra_size.visible); break; case PROP_EXTRA_SIZE_TIME: g_value_set_uint64 (value, mq->extra_size.time); break; case PROP_MAX_SIZE_BYTES: g_value_set_uint (value, mq->max_size.bytes); break; case PROP_MAX_SIZE_BUFFERS: g_value_set_uint (value, mq->max_size.visible); break; case PROP_MAX_SIZE_TIME: g_value_set_uint64 (value, mq->max_size.time); break; case PROP_USE_BUFFERING: g_value_set_boolean (value, mq->use_buffering); break; case PROP_LOW_PERCENT: g_value_set_int (value, mq->low_percent); break; case PROP_HIGH_PERCENT: g_value_set_int (value, mq->high_percent); break; case PROP_SYNC_BY_RUNNING_TIME: g_value_set_boolean (value, mq->sync_by_running_time); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); } static GstIterator * gst_multi_queue_iterate_internal_links (GstPad * pad, GstObject * parent) { GstIterator *it = NULL; GstPad *opad; GstSingleQueue *squeue; GstMultiQueue *mq = GST_MULTI_QUEUE (parent); GValue val = { 0, }; GST_MULTI_QUEUE_MUTEX_LOCK (mq); squeue = gst_pad_get_element_private (pad); if (!squeue) goto out; if (squeue->sinkpad == pad) opad = gst_object_ref (squeue->srcpad); else if (squeue->srcpad == pad) opad = gst_object_ref (squeue->sinkpad); else goto out; g_value_init (&val, GST_TYPE_PAD); g_value_set_object (&val, opad); it = gst_iterator_new_single (GST_TYPE_PAD, &val); g_value_unset (&val); gst_object_unref (opad); out: GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); return it; } /* * GstElement methods */ static GstPad * gst_multi_queue_request_new_pad (GstElement * element, GstPadTemplate * temp, const gchar * name, const GstCaps * caps) { GstMultiQueue *mqueue = GST_MULTI_QUEUE (element); GstSingleQueue *squeue; guint temp_id = -1; if (name) { sscanf (name + 4, "_%u", &temp_id); GST_LOG_OBJECT (element, "name : %s (id %d)", GST_STR_NULL (name), temp_id); } /* Create a new single queue, add the sink and source pad and return the sink pad */ squeue = gst_single_queue_new (mqueue, temp_id); GST_DEBUG_OBJECT (mqueue, "Returning pad %s:%s", GST_DEBUG_PAD_NAME (squeue->sinkpad)); return squeue ? squeue->sinkpad : NULL; } static void gst_multi_queue_release_pad (GstElement * element, GstPad * pad) { GstMultiQueue *mqueue = GST_MULTI_QUEUE (element); GstSingleQueue *sq = NULL; GList *tmp; GST_LOG_OBJECT (element, "pad %s:%s", GST_DEBUG_PAD_NAME (pad)); GST_MULTI_QUEUE_MUTEX_LOCK (mqueue); /* Find which single queue it belongs to, knowing that it should be a sinkpad */ for (tmp = mqueue->queues; tmp; tmp = g_list_next (tmp)) { sq = (GstSingleQueue *) tmp->data; if (sq->sinkpad == pad) break; } if (!tmp) { GST_WARNING_OBJECT (mqueue, "That pad doesn't belong to this element ???"); GST_MULTI_QUEUE_MUTEX_UNLOCK (mqueue); return; } /* FIXME: The removal of the singlequeue should probably not happen until it * finishes draining */ /* remove it from the list */ mqueue->queues = g_list_delete_link (mqueue->queues, tmp); mqueue->queues_cookie++; /* FIXME : recompute next-non-linked */ GST_MULTI_QUEUE_MUTEX_UNLOCK (mqueue); /* delete SingleQueue */ gst_data_queue_set_flushing (sq->queue, TRUE); gst_pad_set_active (sq->srcpad, FALSE); gst_pad_set_active (sq->sinkpad, FALSE); gst_pad_set_element_private (sq->srcpad, NULL); gst_pad_set_element_private (sq->sinkpad, NULL); gst_element_remove_pad (element, sq->srcpad); gst_element_remove_pad (element, sq->sinkpad); gst_single_queue_free (sq); } static GstStateChangeReturn gst_multi_queue_change_state (GstElement * element, GstStateChange transition) { GstMultiQueue *mqueue = GST_MULTI_QUEUE (element); GstSingleQueue *sq = NULL; GstStateChangeReturn result; switch (transition) { case GST_STATE_CHANGE_READY_TO_PAUSED:{ GList *tmp; /* Set all pads to non-flushing */ GST_MULTI_QUEUE_MUTEX_LOCK (mqueue); for (tmp = mqueue->queues; tmp; tmp = g_list_next (tmp)) { sq = (GstSingleQueue *) tmp->data; sq->flushing = FALSE; } /* the visible limit might not have been set on single queues that have grown because of other queueus were empty */ SET_CHILD_PROPERTY (mqueue, visible); GST_MULTI_QUEUE_MUTEX_UNLOCK (mqueue); gst_multi_queue_post_buffering (mqueue); break; } case GST_STATE_CHANGE_PAUSED_TO_READY:{ GList *tmp; /* Un-wait all waiting pads */ GST_MULTI_QUEUE_MUTEX_LOCK (mqueue); for (tmp = mqueue->queues; tmp; tmp = g_list_next (tmp)) { sq = (GstSingleQueue *) tmp->data; sq->flushing = TRUE; g_cond_signal (&sq->turn); sq->last_query = FALSE; g_cond_signal (&sq->query_handled); } GST_MULTI_QUEUE_MUTEX_UNLOCK (mqueue); break; } default: break; } result = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); switch (transition) { default: break; } return result; } static gboolean gst_single_queue_flush (GstMultiQueue * mq, GstSingleQueue * sq, gboolean flush, gboolean full) { gboolean result; GST_DEBUG_OBJECT (mq, "flush %s queue %d", (flush ? "start" : "stop"), sq->id); if (flush) { GST_MULTI_QUEUE_MUTEX_LOCK (mq); sq->srcresult = GST_FLOW_FLUSHING; gst_data_queue_set_flushing (sq->queue, TRUE); sq->flushing = TRUE; /* wake up non-linked task */ GST_LOG_OBJECT (mq, "SingleQueue %d : waking up eventually waiting task", sq->id); g_cond_signal (&sq->turn); sq->last_query = FALSE; g_cond_signal (&sq->query_handled); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); GST_LOG_OBJECT (mq, "SingleQueue %d : pausing task", sq->id); result = gst_pad_pause_task (sq->srcpad); sq->sink_tainted = sq->src_tainted = TRUE; } else { gst_single_queue_flush_queue (sq, full); GST_MULTI_QUEUE_MUTEX_LOCK (mq); gst_segment_init (&sq->sink_segment, GST_FORMAT_TIME); gst_segment_init (&sq->src_segment, GST_FORMAT_TIME); sq->has_src_segment = FALSE; /* All pads start off not-linked for a smooth kick-off */ sq->srcresult = GST_FLOW_OK; sq->pushed = FALSE; sq->cur_time = 0; sq->max_size.visible = mq->max_size.visible; sq->is_eos = FALSE; sq->nextid = 0; sq->oldid = 0; sq->last_oldid = G_MAXUINT32; sq->next_time = GST_CLOCK_TIME_NONE; sq->last_time = GST_CLOCK_TIME_NONE; gst_data_queue_set_flushing (sq->queue, FALSE); /* Reset high time to be recomputed next */ mq->high_time = GST_CLOCK_TIME_NONE; sq->flushing = FALSE; GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); GST_LOG_OBJECT (mq, "SingleQueue %d : starting task", sq->id); result = gst_pad_start_task (sq->srcpad, (GstTaskFunction) gst_multi_queue_loop, sq->srcpad, NULL); } return result; } /* WITH LOCK TAKEN */ static gint get_percentage (GstSingleQueue * sq) { GstDataQueueSize size; gint percent, tmp; gst_data_queue_get_level (sq->queue, &size); GST_DEBUG_OBJECT (sq->mqueue, "queue %d: visible %u/%u, bytes %u/%u, time %" G_GUINT64_FORMAT "/%" G_GUINT64_FORMAT, sq->id, size.visible, sq->max_size.visible, size.bytes, sq->max_size.bytes, sq->cur_time, sq->max_size.time); /* get bytes and time percentages and take the max */ if (sq->is_eos || sq->srcresult == GST_FLOW_NOT_LINKED) { percent = 100; } else { percent = 0; if (sq->max_size.time > 0) { tmp = (sq->cur_time * 100) / sq->max_size.time; percent = MAX (percent, tmp); } if (sq->max_size.bytes > 0) { tmp = (size.bytes * 100) / sq->max_size.bytes; percent = MAX (percent, tmp); } } return percent; } /* WITH LOCK TAKEN */ static void update_buffering (GstMultiQueue * mq, GstSingleQueue * sq) { gint percent; /* nothing to dowhen we are not in buffering mode */ if (!mq->use_buffering) return; percent = get_percentage (sq); if (mq->buffering) { if (percent >= mq->high_percent) { mq->buffering = FALSE; } /* make sure it increases */ percent = MAX (mq->percent, percent); SET_PERCENT (mq, percent); } else { GList *iter; gboolean is_buffering = TRUE; for (iter = mq->queues; iter; iter = g_list_next (iter)) { GstSingleQueue *oq = (GstSingleQueue *) iter->data; if (get_percentage (oq) >= mq->high_percent) { is_buffering = FALSE; break; } } if (is_buffering && percent < mq->low_percent) { mq->buffering = TRUE; SET_PERCENT (mq, percent); } } } static void gst_multi_queue_post_buffering (GstMultiQueue * mq) { GstMessage *msg = NULL; g_mutex_lock (&mq->buffering_post_lock); GST_MULTI_QUEUE_MUTEX_LOCK (mq); if (mq->percent_changed) { gint percent = mq->percent; mq->percent_changed = FALSE; percent = percent * 100 / mq->high_percent; /* clip */ if (percent > 100) percent = 100; GST_DEBUG_OBJECT (mq, "Going to post buffering: %d%%", percent); msg = gst_message_new_buffering (GST_OBJECT_CAST (mq), percent); } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); if (msg != NULL) gst_element_post_message (GST_ELEMENT_CAST (mq), msg); g_mutex_unlock (&mq->buffering_post_lock); } /* calculate the diff between running time on the sink and src of the queue. * This is the total amount of time in the queue. * WITH LOCK TAKEN */ static void update_time_level (GstMultiQueue * mq, GstSingleQueue * sq) { gint64 sink_time, src_time; if (sq->sink_tainted) { sink_time = sq->sinktime = gst_segment_to_running_time (&sq->sink_segment, GST_FORMAT_TIME, sq->sink_segment.position); if (G_UNLIKELY (sink_time != GST_CLOCK_TIME_NONE)) /* if we have a time, we become untainted and use the time */ sq->sink_tainted = FALSE; } else sink_time = sq->sinktime; if (sq->src_tainted) { GstSegment *segment; gint64 position; if (sq->has_src_segment) { segment = &sq->src_segment; position = sq->src_segment.position; } else { /* * If the src pad had no segment yet, use the sink segment * to avoid signalling overrun if the received sink segment has a * a position > max-size-time while the src pad time would be the default=0 * * This can happen when switching pads on chained/adaptive streams and the * new chain has a segment with a much larger position */ segment = &sq->sink_segment; position = sq->sink_segment.position; } src_time = sq->srctime = gst_segment_to_running_time (segment, GST_FORMAT_TIME, position); /* if we have a time, we become untainted and use the time */ if (G_UNLIKELY (src_time != GST_CLOCK_TIME_NONE)) sq->src_tainted = FALSE; } else src_time = sq->srctime; GST_DEBUG_OBJECT (mq, "queue %d, sink %" GST_TIME_FORMAT ", src %" GST_TIME_FORMAT, sq->id, GST_TIME_ARGS (sink_time), GST_TIME_ARGS (src_time)); /* This allows for streams with out of order timestamping - sometimes the * emerging timestamp is later than the arriving one(s) */ if (G_LIKELY (sink_time != -1 && src_time != -1 && sink_time > src_time)) sq->cur_time = sink_time - src_time; else sq->cur_time = 0; /* updating the time level can change the buffering state */ update_buffering (mq, sq); return; } /* take a SEGMENT event and apply the values to segment, updating the time * level of queue. */ static void apply_segment (GstMultiQueue * mq, GstSingleQueue * sq, GstEvent * event, GstSegment * segment) { gst_event_copy_segment (event, segment); /* now configure the values, we use these to track timestamps on the * sinkpad. */ if (segment->format != GST_FORMAT_TIME) { /* non-time format, pretent the current time segment is closed with a * 0 start and unknown stop time. */ segment->format = GST_FORMAT_TIME; segment->start = 0; segment->stop = -1; segment->time = 0; } GST_MULTI_QUEUE_MUTEX_LOCK (mq); if (segment == &sq->sink_segment) sq->sink_tainted = TRUE; else { sq->has_src_segment = TRUE; sq->src_tainted = TRUE; } GST_DEBUG_OBJECT (mq, "queue %d, configured SEGMENT %" GST_SEGMENT_FORMAT, sq->id, segment); /* segment can update the time level of the queue */ update_time_level (mq, sq); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_multi_queue_post_buffering (mq); } /* take a buffer and update segment, updating the time level of the queue. */ static void apply_buffer (GstMultiQueue * mq, GstSingleQueue * sq, GstClockTime timestamp, GstClockTime duration, GstSegment * segment) { GST_MULTI_QUEUE_MUTEX_LOCK (mq); /* if no timestamp is set, assume it's continuous with the previous * time */ if (timestamp == GST_CLOCK_TIME_NONE) timestamp = segment->position; /* add duration */ if (duration != GST_CLOCK_TIME_NONE) timestamp += duration; GST_DEBUG_OBJECT (mq, "queue %d, position updated to %" GST_TIME_FORMAT, sq->id, GST_TIME_ARGS (timestamp)); segment->position = timestamp; if (segment == &sq->sink_segment) sq->sink_tainted = TRUE; else sq->src_tainted = TRUE; /* calc diff with other end */ update_time_level (mq, sq); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_multi_queue_post_buffering (mq); } static void apply_gap (GstMultiQueue * mq, GstSingleQueue * sq, GstEvent * event, GstSegment * segment) { GstClockTime timestamp; GstClockTime duration; GST_MULTI_QUEUE_MUTEX_LOCK (mq); gst_event_parse_gap (event, ×tamp, &duration); if (GST_CLOCK_TIME_IS_VALID (timestamp)) { if (GST_CLOCK_TIME_IS_VALID (duration)) { timestamp += duration; } segment->position = timestamp; if (segment == &sq->sink_segment) sq->sink_tainted = TRUE; else sq->src_tainted = TRUE; /* calc diff with other end */ update_time_level (mq, sq); } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_multi_queue_post_buffering (mq); } static GstClockTime get_running_time (GstSegment * segment, GstMiniObject * object, gboolean end) { GstClockTime time = GST_CLOCK_TIME_NONE; if (GST_IS_BUFFER (object)) { GstBuffer *buf = GST_BUFFER_CAST (object); if (GST_BUFFER_TIMESTAMP_IS_VALID (buf)) { time = GST_BUFFER_TIMESTAMP (buf); if (end && GST_BUFFER_DURATION_IS_VALID (buf)) time += GST_BUFFER_DURATION (buf); if (time > segment->stop) time = segment->stop; time = gst_segment_to_running_time (segment, GST_FORMAT_TIME, time); } } else if (GST_IS_BUFFER_LIST (object)) { GstBufferList *list = GST_BUFFER_LIST_CAST (object); gint i, n; GstBuffer *buf; n = gst_buffer_list_length (list); for (i = 0; i < n; i++) { buf = gst_buffer_list_get (list, i); if (GST_BUFFER_TIMESTAMP_IS_VALID (buf)) { time = GST_BUFFER_TIMESTAMP (buf); if (end && GST_BUFFER_DURATION_IS_VALID (buf)) time += GST_BUFFER_DURATION (buf); if (time > segment->stop) time = segment->stop; time = gst_segment_to_running_time (segment, GST_FORMAT_TIME, time); if (!end) goto done; } else if (!end) { goto done; } } } else if (GST_IS_EVENT (object)) { GstEvent *event = GST_EVENT_CAST (object); /* For newsegment events return the running time of the start position */ if (GST_EVENT_TYPE (event) == GST_EVENT_SEGMENT) { const GstSegment *new_segment; gst_event_parse_segment (event, &new_segment); if (new_segment->format == GST_FORMAT_TIME) { time = gst_segment_to_running_time (new_segment, GST_FORMAT_TIME, new_segment->start); } } } done: return time; } static GstFlowReturn gst_single_queue_push_one (GstMultiQueue * mq, GstSingleQueue * sq, GstMiniObject * object, gboolean * allow_drop) { GstFlowReturn result = sq->srcresult; if (GST_IS_BUFFER (object)) { GstBuffer *buffer; GstClockTime timestamp, duration; buffer = GST_BUFFER_CAST (object); timestamp = GST_BUFFER_TIMESTAMP (buffer); duration = GST_BUFFER_DURATION (buffer); apply_buffer (mq, sq, timestamp, duration, &sq->src_segment); /* Applying the buffer may have made the queue non-full again, unblock it if needed */ gst_data_queue_limits_changed (sq->queue); if (G_UNLIKELY (*allow_drop)) { GST_DEBUG_OBJECT (mq, "SingleQueue %d : Dropping EOS buffer %p with ts %" GST_TIME_FORMAT, sq->id, buffer, GST_TIME_ARGS (timestamp)); gst_buffer_unref (buffer); } else { GST_DEBUG_OBJECT (mq, "SingleQueue %d : Pushing buffer %p with ts %" GST_TIME_FORMAT, sq->id, buffer, GST_TIME_ARGS (timestamp)); result = gst_pad_push (sq->srcpad, buffer); } } else if (GST_IS_EVENT (object)) { GstEvent *event; event = GST_EVENT_CAST (object); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_EOS: result = GST_FLOW_EOS; if (G_UNLIKELY (*allow_drop)) *allow_drop = FALSE; break; case GST_EVENT_SEGMENT: apply_segment (mq, sq, event, &sq->src_segment); /* Applying the segment may have made the queue non-full again, unblock it if needed */ gst_data_queue_limits_changed (sq->queue); if (G_UNLIKELY (*allow_drop)) { result = GST_FLOW_OK; *allow_drop = FALSE; } break; case GST_EVENT_GAP: apply_gap (mq, sq, event, &sq->src_segment); /* Applying the gap may have made the queue non-full again, unblock it if needed */ gst_data_queue_limits_changed (sq->queue); break; default: break; } if (G_UNLIKELY (*allow_drop)) { GST_DEBUG_OBJECT (mq, "SingleQueue %d : Dropping EOS event %p of type %s", sq->id, event, GST_EVENT_TYPE_NAME (event)); gst_event_unref (event); } else { GST_DEBUG_OBJECT (mq, "SingleQueue %d : Pushing event %p of type %s", sq->id, event, GST_EVENT_TYPE_NAME (event)); gst_pad_push_event (sq->srcpad, event); } } else if (GST_IS_QUERY (object)) { GstQuery *query; gboolean res; query = GST_QUERY_CAST (object); if (G_UNLIKELY (*allow_drop)) { GST_DEBUG_OBJECT (mq, "SingleQueue %d : Dropping EOS query %p", sq->id, query); gst_query_unref (query); res = FALSE; } else { res = gst_pad_peer_query (sq->srcpad, query); } GST_MULTI_QUEUE_MUTEX_LOCK (mq); sq->last_query = res; sq->last_handled_query = query; g_cond_signal (&sq->query_handled); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); } else { g_warning ("Unexpected object in singlequeue %u (refcounting problem?)", sq->id); } return result; /* ERRORS */ } static GstMiniObject * gst_multi_queue_item_steal_object (GstMultiQueueItem * item) { GstMiniObject *res; res = item->object; item->object = NULL; return res; } static void gst_multi_queue_item_destroy (GstMultiQueueItem * item) { if (!item->is_query && item->object) gst_mini_object_unref (item->object); g_slice_free (GstMultiQueueItem, item); } /* takes ownership of passed mini object! */ static GstMultiQueueItem * gst_multi_queue_buffer_item_new (GstMiniObject * object, guint32 curid) { GstMultiQueueItem *item; item = g_slice_new (GstMultiQueueItem); item->object = object; item->destroy = (GDestroyNotify) gst_multi_queue_item_destroy; item->posid = curid; item->is_query = GST_IS_QUERY (object); item->size = gst_buffer_get_size (GST_BUFFER_CAST (object)); item->duration = GST_BUFFER_DURATION (object); if (item->duration == GST_CLOCK_TIME_NONE) item->duration = 0; item->visible = TRUE; return item; } static GstMultiQueueItem * gst_multi_queue_mo_item_new (GstMiniObject * object, guint32 curid) { GstMultiQueueItem *item; item = g_slice_new (GstMultiQueueItem); item->object = object; item->destroy = (GDestroyNotify) gst_multi_queue_item_destroy; item->posid = curid; item->is_query = GST_IS_QUERY (object); item->size = 0; item->duration = 0; item->visible = FALSE; return item; } /* Each main loop attempts to push buffers until the return value * is not-linked. not-linked pads are not allowed to push data beyond * any linked pads, so they don't 'rush ahead of the pack'. */ static void gst_multi_queue_loop (GstPad * pad) { GstSingleQueue *sq; GstMultiQueueItem *item; GstDataQueueItem *sitem; GstMultiQueue *mq; GstMiniObject *object = NULL; guint32 newid; GstFlowReturn result; GstClockTime next_time; gboolean is_buffer; gboolean do_update_buffering = FALSE; gboolean dropping = FALSE; sq = (GstSingleQueue *) gst_pad_get_element_private (pad); mq = sq->mqueue; next: GST_DEBUG_OBJECT (mq, "SingleQueue %d : trying to pop an object", sq->id); if (sq->flushing) goto out_flushing; /* Get something from the queue, blocking until that happens, or we get * flushed */ if (!(gst_data_queue_pop (sq->queue, &sitem))) goto out_flushing; item = (GstMultiQueueItem *) sitem; newid = item->posid; /* steal the object and destroy the item */ object = gst_multi_queue_item_steal_object (item); gst_multi_queue_item_destroy (item); is_buffer = GST_IS_BUFFER (object); /* Get running time of the item. Events will have GST_CLOCK_TIME_NONE */ next_time = get_running_time (&sq->src_segment, object, TRUE); GST_LOG_OBJECT (mq, "SingleQueue %d : newid:%d , oldid:%d", sq->id, newid, sq->last_oldid); /* If we're not-linked, we do some extra work because we might need to * wait before pushing. If we're linked but there's a gap in the IDs, * or it's the first loop, or we just passed the previous highid, * we might need to wake some sleeping pad up, so there's extra work * there too */ GST_MULTI_QUEUE_MUTEX_LOCK (mq); if (sq->srcresult == GST_FLOW_NOT_LINKED || (sq->last_oldid == G_MAXUINT32) || (newid != (sq->last_oldid + 1)) || sq->last_oldid > mq->highid) { GST_LOG_OBJECT (mq, "CHECKING sq->srcresult: %s", gst_flow_get_name (sq->srcresult)); /* Check again if we're flushing after the lock is taken, * the flush flag might have been changed in the meantime */ if (sq->flushing) { GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); goto out_flushing; } /* Update the nextid so other threads know when to wake us up */ sq->nextid = newid; sq->next_time = next_time; /* Update the oldid (the last ID we output) for highid tracking */ if (sq->last_oldid != G_MAXUINT32) sq->oldid = sq->last_oldid; if (sq->srcresult == GST_FLOW_NOT_LINKED) { /* Go to sleep until it's time to push this buffer */ /* Recompute the highid */ compute_high_id (mq); /* Recompute the high time */ compute_high_time (mq); while (((mq->sync_by_running_time && next_time != GST_CLOCK_TIME_NONE && (mq->high_time == GST_CLOCK_TIME_NONE || next_time >= mq->high_time)) || (!mq->sync_by_running_time && newid > mq->highid)) && sq->srcresult == GST_FLOW_NOT_LINKED) { GST_DEBUG_OBJECT (mq, "queue %d sleeping for not-linked wakeup with " "newid %u, highid %u, next_time %" GST_TIME_FORMAT ", high_time %" GST_TIME_FORMAT, sq->id, newid, mq->highid, GST_TIME_ARGS (next_time), GST_TIME_ARGS (mq->high_time)); /* Wake up all non-linked pads before we sleep */ wake_up_next_non_linked (mq); mq->numwaiting++; g_cond_wait (&sq->turn, &mq->qlock); mq->numwaiting--; if (sq->flushing) { GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); goto out_flushing; } /* Recompute the high time */ compute_high_time (mq); GST_DEBUG_OBJECT (mq, "queue %d woken from sleeping for not-linked " "wakeup with newid %u, highid %u, next_time %" GST_TIME_FORMAT ", high_time %" GST_TIME_FORMAT, sq->id, newid, mq->highid, GST_TIME_ARGS (next_time), GST_TIME_ARGS (mq->high_time)); } /* Re-compute the high_id in case someone else pushed */ compute_high_id (mq); } else { compute_high_id (mq); /* Wake up all non-linked pads */ wake_up_next_non_linked (mq); } /* We're done waiting, we can clear the nextid and nexttime */ sq->nextid = 0; sq->next_time = GST_CLOCK_TIME_NONE; } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); if (sq->flushing) goto out_flushing; GST_LOG_OBJECT (mq, "BEFORE PUSHING sq->srcresult: %s", gst_flow_get_name (sq->srcresult)); /* Update time stats */ GST_MULTI_QUEUE_MUTEX_LOCK (mq); next_time = get_running_time (&sq->src_segment, object, FALSE); if (next_time != GST_CLOCK_TIME_NONE) { if (sq->last_time == GST_CLOCK_TIME_NONE || sq->last_time < next_time) sq->last_time = next_time; if (mq->high_time == GST_CLOCK_TIME_NONE || mq->high_time <= next_time) { /* Wake up all non-linked pads now that we advanced the high time */ mq->high_time = next_time; wake_up_next_non_linked (mq); } } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); /* Try to push out the new object */ result = gst_single_queue_push_one (mq, sq, object, &dropping); object = NULL; /* Check if we pushed something already and if this is * now a switch from an active to a non-active stream. * * If it is, we reset all the waiting streams, let them * push another buffer to see if they're now active again. * This allows faster switching between streams and prevents * deadlocks if downstream does any waiting too. */ GST_MULTI_QUEUE_MUTEX_LOCK (mq); if (sq->pushed && sq->srcresult == GST_FLOW_OK && result == GST_FLOW_NOT_LINKED) { GList *tmp; GST_LOG_OBJECT (mq, "SingleQueue %d : Changed from active to non-active", sq->id); compute_high_id (mq); do_update_buffering = TRUE; /* maybe no-one is waiting */ if (mq->numwaiting > 0) { /* Else figure out which singlequeue(s) need waking up */ for (tmp = mq->queues; tmp; tmp = g_list_next (tmp)) { GstSingleQueue *sq2 = (GstSingleQueue *) tmp->data; if (sq2->srcresult == GST_FLOW_NOT_LINKED) { GST_LOG_OBJECT (mq, "Waking up singlequeue %d", sq2->id); sq2->pushed = FALSE; sq2->srcresult = GST_FLOW_OK; g_cond_signal (&sq2->turn); } } } } if (is_buffer) sq->pushed = TRUE; /* now hold on a bit; * can not simply throw this result to upstream, because * that might already be onto another segment, so we have to make * sure we are relaying the correct info wrt proper segment */ if (result == GST_FLOW_EOS && !dropping && sq->srcresult != GST_FLOW_NOT_LINKED) { GST_DEBUG_OBJECT (mq, "starting EOS drop on sq %d", sq->id); dropping = TRUE; /* pretend we have not seen EOS yet for upstream's sake */ result = sq->srcresult; } else if (dropping && gst_data_queue_is_empty (sq->queue)) { /* queue empty, so stop dropping * we can commit the result we have now, * which is either OK after a segment, or EOS */ GST_DEBUG_OBJECT (mq, "committed EOS drop on sq %d", sq->id); dropping = FALSE; result = GST_FLOW_EOS; } sq->srcresult = result; sq->last_oldid = newid; if (do_update_buffering) update_buffering (mq, sq); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_multi_queue_post_buffering (mq); if (dropping) goto next; if (result != GST_FLOW_OK && result != GST_FLOW_NOT_LINKED && result != GST_FLOW_EOS) goto out_flushing; GST_LOG_OBJECT (mq, "AFTER PUSHING sq->srcresult: %s", gst_flow_get_name (sq->srcresult)); GST_MULTI_QUEUE_MUTEX_LOCK (mq); if (mq->numwaiting > 0 && sq->srcresult == GST_FLOW_EOS) { compute_high_time (mq); compute_high_id (mq); wake_up_next_non_linked (mq); } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); return; out_flushing: { if (object) gst_mini_object_unref (object); /* Need to make sure wake up any sleeping pads when we exit */ GST_MULTI_QUEUE_MUTEX_LOCK (mq); compute_high_time (mq); compute_high_id (mq); wake_up_next_non_linked (mq); sq->last_query = FALSE; g_cond_signal (&sq->query_handled); /* Post an error message if we got EOS while downstream * has returned an error flow return. After EOS there * will be no further buffer which could propagate the * error upstream */ if (sq->is_eos && sq->srcresult < GST_FLOW_EOS) { GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); GST_ELEMENT_ERROR (mq, STREAM, FAILED, ("Internal data stream error."), ("streaming stopped, reason %s", gst_flow_get_name (sq->srcresult))); } else { GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); } /* upstream needs to see fatal result ASAP to shut things down, * but might be stuck in one of our other full queues; * so empty this one and trigger dynamic queue growth. At * this point the srcresult is not OK, NOT_LINKED * or EOS, i.e. a real failure */ gst_single_queue_flush_queue (sq, FALSE); single_queue_underrun_cb (sq->queue, sq); gst_data_queue_set_flushing (sq->queue, TRUE); gst_pad_pause_task (sq->srcpad); GST_CAT_LOG_OBJECT (multi_queue_debug, mq, "SingleQueue[%d] task paused, reason:%s", sq->id, gst_flow_get_name (sq->srcresult)); return; } } /** * gst_multi_queue_chain: * * This is similar to GstQueue's chain function, except: * _ we don't have leak behaviours, * _ we push with a unique id (curid) */ static GstFlowReturn gst_multi_queue_chain (GstPad * pad, GstObject * parent, GstBuffer * buffer) { GstSingleQueue *sq; GstMultiQueue *mq; GstMultiQueueItem *item; guint32 curid; GstClockTime timestamp, duration; sq = gst_pad_get_element_private (pad); mq = sq->mqueue; /* if eos, we are always full, so avoid hanging incoming indefinitely */ if (sq->is_eos) goto was_eos; /* Get a unique incrementing id */ curid = g_atomic_int_add ((gint *) & mq->counter, 1); GST_LOG_OBJECT (mq, "SingleQueue %d : about to enqueue buffer %p with id %d", sq->id, buffer, curid); item = gst_multi_queue_buffer_item_new (GST_MINI_OBJECT_CAST (buffer), curid); timestamp = GST_BUFFER_TIMESTAMP (buffer); duration = GST_BUFFER_DURATION (buffer); if (!(gst_data_queue_push (sq->queue, (GstDataQueueItem *) item))) goto flushing; /* update time level, we must do this after pushing the data in the queue so * that we never end up filling the queue first. */ apply_buffer (mq, sq, timestamp, duration, &sq->sink_segment); done: return sq->srcresult; /* ERRORS */ flushing: { GST_LOG_OBJECT (mq, "SingleQueue %d : exit because task paused, reason: %s", sq->id, gst_flow_get_name (sq->srcresult)); gst_multi_queue_item_destroy (item); goto done; } was_eos: { GST_DEBUG_OBJECT (mq, "we are EOS, dropping buffer, return EOS"); gst_buffer_unref (buffer); return GST_FLOW_EOS; } } static gboolean gst_multi_queue_sink_activate_mode (GstPad * pad, GstObject * parent, GstPadMode mode, gboolean active) { gboolean res; GstSingleQueue *sq; GstMultiQueue *mq; sq = (GstSingleQueue *) gst_pad_get_element_private (pad); mq = (GstMultiQueue *) gst_pad_get_parent (pad); /* mq is NULL if the pad is activated/deactivated before being * added to the multiqueue */ if (mq) GST_MULTI_QUEUE_MUTEX_LOCK (mq); switch (mode) { case GST_PAD_MODE_PUSH: if (active) { /* All pads start off linked until they push one buffer */ sq->srcresult = GST_FLOW_OK; sq->pushed = FALSE; gst_data_queue_set_flushing (sq->queue, FALSE); } else { sq->srcresult = GST_FLOW_FLUSHING; sq->last_query = FALSE; g_cond_signal (&sq->query_handled); gst_data_queue_set_flushing (sq->queue, TRUE); /* Wait until streaming thread has finished */ if (mq) GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); GST_PAD_STREAM_LOCK (pad); if (mq) GST_MULTI_QUEUE_MUTEX_LOCK (mq); gst_data_queue_flush (sq->queue); if (mq) GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); GST_PAD_STREAM_UNLOCK (pad); if (mq) GST_MULTI_QUEUE_MUTEX_LOCK (mq); } res = TRUE; break; default: res = FALSE; break; } if (mq) { GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); gst_object_unref (mq); } return res; } static gboolean gst_multi_queue_sink_event (GstPad * pad, GstObject * parent, GstEvent * event) { GstSingleQueue *sq; GstMultiQueue *mq; guint32 curid; GstMultiQueueItem *item; gboolean res; GstEventType type; GstEvent *sref = NULL; sq = (GstSingleQueue *) gst_pad_get_element_private (pad); mq = (GstMultiQueue *) parent; type = GST_EVENT_TYPE (event); switch (type) { case GST_EVENT_FLUSH_START: GST_DEBUG_OBJECT (mq, "SingleQueue %d : received flush start event", sq->id); res = gst_pad_push_event (sq->srcpad, event); gst_single_queue_flush (mq, sq, TRUE, FALSE); goto done; case GST_EVENT_FLUSH_STOP: GST_DEBUG_OBJECT (mq, "SingleQueue %d : received flush stop event", sq->id); res = gst_pad_push_event (sq->srcpad, event); gst_single_queue_flush (mq, sq, FALSE, FALSE); goto done; case GST_EVENT_SEGMENT: case GST_EVENT_GAP: /* take ref because the queue will take ownership and we need the event * afterwards to update the segment */ sref = gst_event_ref (event); break; default: if (!(GST_EVENT_IS_SERIALIZED (event))) { res = gst_pad_push_event (sq->srcpad, event); goto done; } break; } /* if eos, we are always full, so avoid hanging incoming indefinitely */ if (sq->is_eos) goto was_eos; /* Get an unique incrementing id. */ curid = g_atomic_int_add ((gint *) & mq->counter, 1); item = gst_multi_queue_mo_item_new ((GstMiniObject *) event, curid); GST_DEBUG_OBJECT (mq, "SingleQueue %d : Enqueuing event %p of type %s with id %d", sq->id, event, GST_EVENT_TYPE_NAME (event), curid); if (!(res = gst_data_queue_push (sq->queue, (GstDataQueueItem *) item))) goto flushing; /* mark EOS when we received one, we must do that after putting the * buffer in the queue because EOS marks the buffer as filled. */ switch (type) { case GST_EVENT_EOS: GST_MULTI_QUEUE_MUTEX_LOCK (mq); sq->is_eos = TRUE; /* Post an error message if we got EOS while downstream * has returned an error flow return. After EOS there * will be no further buffer which could propagate the * error upstream */ if (sq->srcresult < GST_FLOW_EOS) { GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); GST_ELEMENT_ERROR (mq, STREAM, FAILED, ("Internal data stream error."), ("streaming stopped, reason %s", gst_flow_get_name (sq->srcresult))); } else { GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); } /* EOS affects the buffering state */ GST_MULTI_QUEUE_MUTEX_LOCK (mq); update_buffering (mq, sq); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); single_queue_overrun_cb (sq->queue, sq); gst_multi_queue_post_buffering (mq); break; case GST_EVENT_SEGMENT: apply_segment (mq, sq, sref, &sq->sink_segment); gst_event_unref (sref); /* a new segment allows us to accept more buffers if we got EOS * from downstream */ GST_MULTI_QUEUE_MUTEX_LOCK (mq); sq->srcresult = GST_FLOW_OK; GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); break; case GST_EVENT_GAP: apply_gap (mq, sq, sref, &sq->sink_segment); gst_event_unref (sref); default: break; } done: return res; flushing: { GST_LOG_OBJECT (mq, "SingleQueue %d : exit because task paused, reason: %s", sq->id, gst_flow_get_name (sq->srcresult)); if (sref) gst_event_unref (sref); gst_multi_queue_item_destroy (item); goto done; } was_eos: { GST_DEBUG_OBJECT (mq, "we are EOS, dropping event, return FALSE"); gst_event_unref (event); res = FALSE; goto done; } } static gboolean gst_multi_queue_sink_query (GstPad * pad, GstObject * parent, GstQuery * query) { gboolean res; GstSingleQueue *sq; GstMultiQueue *mq; sq = (GstSingleQueue *) gst_pad_get_element_private (pad); mq = (GstMultiQueue *) parent; switch (GST_QUERY_TYPE (query)) { default: if (GST_QUERY_IS_SERIALIZED (query)) { guint32 curid; GstMultiQueueItem *item; GST_MULTI_QUEUE_MUTEX_LOCK (mq); if (sq->srcresult != GST_FLOW_OK) goto out_flushing; /* serialized events go in the queue. We need to be certain that we * don't cause deadlocks waiting for the query return value. We check if * the queue is empty (nothing is blocking downstream and the query can * be pushed for sure) or we are not buffering. If we are buffering, * the pipeline waits to unblock downstream until our queue fills up * completely, which can not happen if we block on the query.. * Therefore we only potentially block when we are not buffering. */ if (!mq->use_buffering || gst_data_queue_is_empty (sq->queue)) { /* Get an unique incrementing id. */ curid = g_atomic_int_add ((gint *) & mq->counter, 1); item = gst_multi_queue_mo_item_new ((GstMiniObject *) query, curid); GST_DEBUG_OBJECT (mq, "SingleQueue %d : Enqueuing query %p of type %s with id %d", sq->id, query, GST_QUERY_TYPE_NAME (query), curid); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); res = gst_data_queue_push (sq->queue, (GstDataQueueItem *) item); GST_MULTI_QUEUE_MUTEX_LOCK (mq); /* it might be that the query has been taken out of the queue * while we were unlocked. So, we need to check if the last * handled query is the same one than the one we just * pushed. If it is, we don't need to wait for the condition * variable, otherwise we wait for the condition variable to * be signaled. */ if (sq->last_handled_query != query) g_cond_wait (&sq->query_handled, &mq->qlock); res = sq->last_query; sq->last_handled_query = NULL; } else { GST_DEBUG_OBJECT (mq, "refusing query, we are buffering and the " "queue is not empty"); res = FALSE; } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); } else { /* default handling */ res = gst_pad_query_default (pad, parent, query); } break; } return res; out_flushing: { GST_DEBUG_OBJECT (mq, "Flushing"); GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); return FALSE; } } static gboolean gst_multi_queue_src_activate_mode (GstPad * pad, GstObject * parent, GstPadMode mode, gboolean active) { GstMultiQueue *mq; GstSingleQueue *sq; gboolean result; sq = (GstSingleQueue *) gst_pad_get_element_private (pad); mq = sq->mqueue; GST_DEBUG_OBJECT (mq, "SingleQueue %d", sq->id); switch (mode) { case GST_PAD_MODE_PUSH: if (active) { result = gst_single_queue_flush (mq, sq, FALSE, TRUE); } else { result = gst_single_queue_flush (mq, sq, TRUE, TRUE); /* make sure streaming finishes */ result |= gst_pad_stop_task (pad); } break; default: result = FALSE; break; } return result; } static gboolean gst_multi_queue_src_event (GstPad * pad, GstObject * parent, GstEvent * event) { GstSingleQueue *sq = gst_pad_get_element_private (pad); GstMultiQueue *mq = sq->mqueue; gboolean ret; switch (GST_EVENT_TYPE (event)) { case GST_EVENT_RECONFIGURE: GST_MULTI_QUEUE_MUTEX_LOCK (mq); if (sq->srcresult == GST_FLOW_NOT_LINKED) { sq->srcresult = GST_FLOW_OK; g_cond_signal (&sq->turn); } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); ret = gst_pad_push_event (sq->sinkpad, event); break; default: ret = gst_pad_push_event (sq->sinkpad, event); break; } return ret; } static gboolean gst_multi_queue_src_query (GstPad * pad, GstObject * parent, GstQuery * query) { gboolean res; /* FIXME, Handle position offset depending on queue size */ switch (GST_QUERY_TYPE (query)) { default: /* default handling */ res = gst_pad_query_default (pad, parent, query); break; } return res; } /* * Next-non-linked functions */ /* WITH LOCK TAKEN */ static void wake_up_next_non_linked (GstMultiQueue * mq) { GList *tmp; /* maybe no-one is waiting */ if (mq->numwaiting < 1) return; /* Else figure out which singlequeue(s) need waking up */ for (tmp = mq->queues; tmp; tmp = g_list_next (tmp)) { GstSingleQueue *sq = (GstSingleQueue *) tmp->data; if (sq->srcresult == GST_FLOW_NOT_LINKED) { if ((mq->sync_by_running_time && mq->high_time != GST_CLOCK_TIME_NONE && sq->next_time != GST_CLOCK_TIME_NONE && sq->next_time >= mq->high_time) || (sq->nextid != 0 && sq->nextid <= mq->highid)) { GST_LOG_OBJECT (mq, "Waking up singlequeue %d", sq->id); g_cond_signal (&sq->turn); } } } } /* WITH LOCK TAKEN */ static void compute_high_id (GstMultiQueue * mq) { /* The high-id is either the highest id among the linked pads, or if all * pads are not-linked, it's the lowest not-linked pad */ GList *tmp; guint32 lowest = G_MAXUINT32; guint32 highid = G_MAXUINT32; for (tmp = mq->queues; tmp; tmp = g_list_next (tmp)) { GstSingleQueue *sq = (GstSingleQueue *) tmp->data; GST_LOG_OBJECT (mq, "inspecting sq:%d , nextid:%d, oldid:%d, srcresult:%s", sq->id, sq->nextid, sq->oldid, gst_flow_get_name (sq->srcresult)); if (sq->srcresult == GST_FLOW_NOT_LINKED) { /* No need to consider queues which are not waiting */ if (sq->nextid == 0) { GST_LOG_OBJECT (mq, "sq:%d is not waiting - ignoring", sq->id); continue; } if (sq->nextid < lowest) lowest = sq->nextid; } else if (sq->srcresult != GST_FLOW_EOS) { /* If we don't have a global highid, or the global highid is lower than * this single queue's last outputted id, store the queue's one, * unless the singlequeue is at EOS (srcresult = EOS) */ if ((highid == G_MAXUINT32) || (sq->oldid > highid)) highid = sq->oldid; } } if (highid == G_MAXUINT32 || lowest < highid) mq->highid = lowest; else mq->highid = highid; GST_LOG_OBJECT (mq, "Highid is now : %u, lowest non-linked %u", mq->highid, lowest); } /* WITH LOCK TAKEN */ static void compute_high_time (GstMultiQueue * mq) { /* The high-id is either the highest id among the linked pads, or if all * pads are not-linked, it's the lowest not-linked pad */ GList *tmp; GstClockTime highest = GST_CLOCK_TIME_NONE; GstClockTime lowest = GST_CLOCK_TIME_NONE; for (tmp = mq->queues; tmp; tmp = g_list_next (tmp)) { GstSingleQueue *sq = (GstSingleQueue *) tmp->data; GST_LOG_OBJECT (mq, "inspecting sq:%d , next_time:%" GST_TIME_FORMAT ", last_time:%" GST_TIME_FORMAT ", srcresult:%s", sq->id, GST_TIME_ARGS (sq->next_time), GST_TIME_ARGS (sq->last_time), gst_flow_get_name (sq->srcresult)); if (sq->srcresult == GST_FLOW_NOT_LINKED) { /* No need to consider queues which are not waiting */ if (sq->next_time == GST_CLOCK_TIME_NONE) { GST_LOG_OBJECT (mq, "sq:%d is not waiting - ignoring", sq->id); continue; } if (lowest == GST_CLOCK_TIME_NONE || sq->next_time < lowest) lowest = sq->next_time; } else if (sq->srcresult != GST_FLOW_EOS) { /* If we don't have a global highid, or the global highid is lower than * this single queue's last outputted id, store the queue's one, * unless the singlequeue is at EOS (srcresult = EOS) */ if (highest == GST_CLOCK_TIME_NONE || sq->last_time > highest) highest = sq->last_time; } } mq->high_time = highest; GST_LOG_OBJECT (mq, "High time is now : %" GST_TIME_FORMAT ", lowest non-linked %" GST_TIME_FORMAT, GST_TIME_ARGS (mq->high_time), GST_TIME_ARGS (lowest)); } #define IS_FILLED(q, format, value) (((q)->max_size.format) != 0 && \ ((q)->max_size.format) <= (value)) /* * GstSingleQueue functions */ static void single_queue_overrun_cb (GstDataQueue * dq, GstSingleQueue * sq) { GstMultiQueue *mq = sq->mqueue; GList *tmp; GstDataQueueSize size; gboolean filled = TRUE; gboolean empty_found = FALSE; gst_data_queue_get_level (sq->queue, &size); GST_LOG_OBJECT (mq, "Single Queue %d: EOS %d, visible %u/%u, bytes %u/%u, time %" G_GUINT64_FORMAT "/%" G_GUINT64_FORMAT, sq->id, sq->is_eos, size.visible, sq->max_size.visible, size.bytes, sq->max_size.bytes, sq->cur_time, sq->max_size.time); GST_MULTI_QUEUE_MUTEX_LOCK (mq); /* check if we reached the hard time/bytes limits */ if (sq->is_eos || IS_FILLED (sq, bytes, size.bytes) || IS_FILLED (sq, time, sq->cur_time)) { goto done; } /* Search for empty queues */ for (tmp = mq->queues; tmp; tmp = g_list_next (tmp)) { GstSingleQueue *oq = (GstSingleQueue *) tmp->data; if (oq == sq) continue; if (oq->srcresult == GST_FLOW_NOT_LINKED) { GST_LOG_OBJECT (mq, "Queue %d is not-linked", oq->id); continue; } GST_LOG_OBJECT (mq, "Checking Queue %d", oq->id); if (gst_data_queue_is_empty (oq->queue)) { GST_LOG_OBJECT (mq, "Queue %d is empty", oq->id); empty_found = TRUE; break; } } /* if hard limits are not reached then we allow one more buffer in the full * queue, but only if any of the other singelqueues are empty */ if (empty_found) { if (IS_FILLED (sq, visible, size.visible)) { sq->max_size.visible = size.visible + 1; GST_DEBUG_OBJECT (mq, "Bumping single queue %d max visible to %d", sq->id, sq->max_size.visible); filled = FALSE; } } done: GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); /* Overrun is always forwarded, since this is blocking the upstream element */ if (filled) { GST_DEBUG_OBJECT (mq, "Queue %d is filled, signalling overrun", sq->id); g_signal_emit (mq, gst_multi_queue_signals[SIGNAL_OVERRUN], 0); } } static void single_queue_underrun_cb (GstDataQueue * dq, GstSingleQueue * sq) { gboolean empty = TRUE; GstMultiQueue *mq = sq->mqueue; GList *tmp; if (sq->srcresult == GST_FLOW_NOT_LINKED) { GST_LOG_OBJECT (mq, "Single Queue %d is empty but not-linked", sq->id); return; } else { GST_LOG_OBJECT (mq, "Single Queue %d is empty, Checking other single queues", sq->id); } GST_MULTI_QUEUE_MUTEX_LOCK (mq); for (tmp = mq->queues; tmp; tmp = g_list_next (tmp)) { GstSingleQueue *oq = (GstSingleQueue *) tmp->data; if (gst_data_queue_is_full (oq->queue)) { GstDataQueueSize size; gst_data_queue_get_level (oq->queue, &size); if (IS_FILLED (oq, visible, size.visible)) { oq->max_size.visible = size.visible + 1; GST_DEBUG_OBJECT (mq, "queue %d is filled, bumping its max visible to %d", oq->id, oq->max_size.visible); gst_data_queue_limits_changed (oq->queue); } } if (!gst_data_queue_is_empty (oq->queue)) empty = FALSE; } GST_MULTI_QUEUE_MUTEX_UNLOCK (mq); if (empty) { GST_DEBUG_OBJECT (mq, "All queues are empty, signalling it"); g_signal_emit (mq, gst_multi_queue_signals[SIGNAL_UNDERRUN], 0); } } static gboolean single_queue_check_full (GstDataQueue * dataq, guint visible, guint bytes, guint64 time, GstSingleQueue * sq) { gboolean res; GstMultiQueue *mq = sq->mqueue; GST_DEBUG_OBJECT (mq, "queue %d: visible %u/%u, bytes %u/%u, time %" G_GUINT64_FORMAT "/%" G_GUINT64_FORMAT, sq->id, visible, sq->max_size.visible, bytes, sq->max_size.bytes, sq->cur_time, sq->max_size.time); /* we are always filled on EOS */ if (sq->is_eos) return TRUE; /* we never go past the max visible items unless we are in buffering mode */ if (!mq->use_buffering && IS_FILLED (sq, visible, visible)) return TRUE; /* check time or bytes */ res = IS_FILLED (sq, time, sq->cur_time) || IS_FILLED (sq, bytes, bytes); return res; } static void gst_single_queue_flush_queue (GstSingleQueue * sq, gboolean full) { GstDataQueueItem *sitem; GstMultiQueueItem *mitem; gboolean was_flushing = FALSE; while (!gst_data_queue_is_empty (sq->queue)) { GstMiniObject *data; /* FIXME: If this fails here although the queue is not empty, * we're flushing... but we want to rescue all sticky * events nonetheless. */ if (!gst_data_queue_pop (sq->queue, &sitem)) { was_flushing = TRUE; gst_data_queue_set_flushing (sq->queue, FALSE); continue; } mitem = (GstMultiQueueItem *) sitem; data = sitem->object; if (!full && !mitem->is_query && GST_IS_EVENT (data) && GST_EVENT_IS_STICKY (data) && GST_EVENT_TYPE (data) != GST_EVENT_SEGMENT && GST_EVENT_TYPE (data) != GST_EVENT_EOS) { gst_pad_store_sticky_event (sq->srcpad, GST_EVENT_CAST (data)); } sitem->destroy (sitem); } gst_data_queue_flush (sq->queue); if (was_flushing) gst_data_queue_set_flushing (sq->queue, TRUE); GST_MULTI_QUEUE_MUTEX_LOCK (sq->mqueue); update_buffering (sq->mqueue, sq); GST_MULTI_QUEUE_MUTEX_UNLOCK (sq->mqueue); gst_multi_queue_post_buffering (sq->mqueue); } static void gst_single_queue_free (GstSingleQueue * sq) { /* DRAIN QUEUE */ gst_data_queue_flush (sq->queue); g_object_unref (sq->queue); g_cond_clear (&sq->turn); g_cond_clear (&sq->query_handled); g_free (sq); } static GstSingleQueue * gst_single_queue_new (GstMultiQueue * mqueue, guint id) { GstSingleQueue *sq; gchar *name; GList *tmp; guint temp_id = (id == -1) ? 0 : id; GST_MULTI_QUEUE_MUTEX_LOCK (mqueue); /* Find an unused queue ID, if possible the passed one */ for (tmp = mqueue->queues; tmp; tmp = g_list_next (tmp)) { GstSingleQueue *sq2 = (GstSingleQueue *) tmp->data; /* This works because the IDs are sorted in ascending order */ if (sq2->id == temp_id) { /* If this ID was requested by the caller return NULL, * otherwise just get us the next one */ if (id == -1) temp_id = sq2->id + 1; else return NULL; } else if (sq2->id > temp_id) { break; } } sq = g_new0 (GstSingleQueue, 1); mqueue->nbqueues++; sq->id = temp_id; mqueue->queues = g_list_insert_before (mqueue->queues, tmp, sq); mqueue->queues_cookie++; /* copy over max_size and extra_size so we don't need to take the lock * any longer when checking if the queue is full. */ sq->max_size.visible = mqueue->max_size.visible; sq->max_size.bytes = mqueue->max_size.bytes; sq->max_size.time = mqueue->max_size.time; sq->extra_size.visible = mqueue->extra_size.visible; sq->extra_size.bytes = mqueue->extra_size.bytes; sq->extra_size.time = mqueue->extra_size.time; GST_DEBUG_OBJECT (mqueue, "Creating GstSingleQueue id:%d", sq->id); sq->mqueue = mqueue; sq->srcresult = GST_FLOW_FLUSHING; sq->pushed = FALSE; sq->queue = gst_data_queue_new ((GstDataQueueCheckFullFunction) single_queue_check_full, (GstDataQueueFullCallback) single_queue_overrun_cb, (GstDataQueueEmptyCallback) single_queue_underrun_cb, sq); sq->is_eos = FALSE; sq->flushing = FALSE; gst_segment_init (&sq->sink_segment, GST_FORMAT_TIME); gst_segment_init (&sq->src_segment, GST_FORMAT_TIME); sq->nextid = 0; sq->oldid = 0; sq->next_time = GST_CLOCK_TIME_NONE; sq->last_time = GST_CLOCK_TIME_NONE; g_cond_init (&sq->turn); g_cond_init (&sq->query_handled); sq->sinktime = GST_CLOCK_TIME_NONE; sq->srctime = GST_CLOCK_TIME_NONE; sq->sink_tainted = TRUE; sq->src_tainted = TRUE; name = g_strdup_printf ("sink_%u", sq->id); sq->sinkpad = gst_pad_new_from_static_template (&sinktemplate, name); g_free (name); gst_pad_set_chain_function (sq->sinkpad, GST_DEBUG_FUNCPTR (gst_multi_queue_chain)); gst_pad_set_activatemode_function (sq->sinkpad, GST_DEBUG_FUNCPTR (gst_multi_queue_sink_activate_mode)); gst_pad_set_event_function (sq->sinkpad, GST_DEBUG_FUNCPTR (gst_multi_queue_sink_event)); gst_pad_set_query_function (sq->sinkpad, GST_DEBUG_FUNCPTR (gst_multi_queue_sink_query)); gst_pad_set_iterate_internal_links_function (sq->sinkpad, GST_DEBUG_FUNCPTR (gst_multi_queue_iterate_internal_links)); GST_OBJECT_FLAG_SET (sq->sinkpad, GST_PAD_FLAG_PROXY_CAPS); name = g_strdup_printf ("src_%u", sq->id); sq->srcpad = gst_pad_new_from_static_template (&srctemplate, name); g_free (name); gst_pad_set_activatemode_function (sq->srcpad, GST_DEBUG_FUNCPTR (gst_multi_queue_src_activate_mode)); gst_pad_set_event_function (sq->srcpad, GST_DEBUG_FUNCPTR (gst_multi_queue_src_event)); gst_pad_set_query_function (sq->srcpad, GST_DEBUG_FUNCPTR (gst_multi_queue_src_query)); gst_pad_set_iterate_internal_links_function (sq->srcpad, GST_DEBUG_FUNCPTR (gst_multi_queue_iterate_internal_links)); GST_OBJECT_FLAG_SET (sq->srcpad, GST_PAD_FLAG_PROXY_CAPS); gst_pad_set_element_private (sq->sinkpad, (gpointer) sq); gst_pad_set_element_private (sq->srcpad, (gpointer) sq); GST_MULTI_QUEUE_MUTEX_UNLOCK (mqueue); /* only activate the pads when we are not in the NULL state * and add the pad under the state_lock to prevend state changes * between activating and adding */ g_rec_mutex_lock (GST_STATE_GET_LOCK (mqueue)); if (GST_STATE_TARGET (mqueue) != GST_STATE_NULL) { gst_pad_set_active (sq->srcpad, TRUE); gst_pad_set_active (sq->sinkpad, TRUE); } gst_element_add_pad (GST_ELEMENT (mqueue), sq->srcpad); gst_element_add_pad (GST_ELEMENT (mqueue), sq->sinkpad); g_rec_mutex_unlock (GST_STATE_GET_LOCK (mqueue)); GST_DEBUG_OBJECT (mqueue, "GstSingleQueue [%d] created and pads added", sq->id); return sq; }