2007-03-28 18:38:11 +00:00
|
|
|
/* GStreamer unit tests for multiqueue
|
|
|
|
*
|
|
|
|
* Copyright (C) 2007 Tim-Philipp Müller <tim centricular net>
|
|
|
|
*
|
|
|
|
* This library is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Library General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* This library is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Library General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Library General Public
|
|
|
|
* License along with this library; if not, write to the
|
2012-11-03 20:44:48 +00:00
|
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
|
|
* Boston, MA 02110-1301, USA.
|
2007-03-28 18:38:11 +00:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <unistd.h>
|
|
|
|
|
|
|
|
#include <gst/check/gstcheck.h>
|
|
|
|
|
2012-01-22 22:44:59 +00:00
|
|
|
static GMutex _check_lock;
|
2007-07-16 16:04:49 +00:00
|
|
|
|
2007-03-28 18:38:11 +00:00
|
|
|
static GstElement *
|
|
|
|
setup_multiqueue (GstElement * pipe, GstElement * inputs[],
|
|
|
|
GstElement * outputs[], guint num)
|
|
|
|
{
|
|
|
|
GstElement *mq;
|
|
|
|
guint i;
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
fail_unless (mq != NULL, "failed to create 'multiqueue' element");
|
|
|
|
|
|
|
|
gst_bin_add (GST_BIN (pipe), mq);
|
|
|
|
|
|
|
|
for (i = 0; i < num; ++i) {
|
|
|
|
GstPad *sinkpad = NULL;
|
|
|
|
GstPad *srcpad = NULL;
|
|
|
|
|
|
|
|
/* create multiqueue sink (and source) pad */
|
2011-11-03 16:49:45 +00:00
|
|
|
sinkpad = gst_element_get_request_pad (mq, "sink_%u");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (sinkpad != NULL,
|
|
|
|
"failed to create multiqueue request pad #%u", i);
|
|
|
|
|
|
|
|
/* link input element N to the N-th multiqueue sink pad we just created */
|
|
|
|
if (inputs != NULL && inputs[i] != NULL) {
|
|
|
|
gst_bin_add (GST_BIN (pipe), inputs[i]);
|
|
|
|
|
2008-05-21 15:57:52 +00:00
|
|
|
srcpad = gst_element_get_static_pad (inputs[i], "src");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (srcpad != NULL, "failed to find src pad for input #%u", i);
|
|
|
|
|
|
|
|
fail_unless_equals_int (GST_PAD_LINK_OK, gst_pad_link (srcpad, sinkpad));
|
|
|
|
|
|
|
|
gst_object_unref (srcpad);
|
|
|
|
srcpad = NULL;
|
|
|
|
}
|
|
|
|
gst_object_unref (sinkpad);
|
|
|
|
sinkpad = NULL;
|
|
|
|
|
|
|
|
/* link output element N to the N-th multiqueue src pad */
|
|
|
|
if (outputs != NULL && outputs[i] != NULL) {
|
|
|
|
gchar padname[10];
|
|
|
|
|
|
|
|
/* only the sink pads are by request, the source pads are sometimes pads,
|
|
|
|
* so this should return NULL */
|
2011-11-03 16:49:45 +00:00
|
|
|
srcpad = gst_element_get_request_pad (mq, "src_%u");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (srcpad == NULL);
|
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
g_snprintf (padname, sizeof (padname), "src_%u", i);
|
2008-05-21 15:57:52 +00:00
|
|
|
srcpad = gst_element_get_static_pad (mq, padname);
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (srcpad != NULL, "failed to get multiqueue src pad #%u", i);
|
|
|
|
fail_unless (GST_PAD_IS_SRC (srcpad),
|
|
|
|
"%s:%s is not a source pad?!", GST_DEBUG_PAD_NAME (srcpad));
|
|
|
|
|
|
|
|
gst_bin_add (GST_BIN (pipe), outputs[i]);
|
|
|
|
|
2008-05-21 15:57:52 +00:00
|
|
|
sinkpad = gst_element_get_static_pad (outputs[i], "sink");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (sinkpad != NULL, "failed to find sink pad of output #%u", i);
|
|
|
|
fail_unless (GST_PAD_IS_SINK (sinkpad));
|
|
|
|
|
|
|
|
fail_unless_equals_int (GST_PAD_LINK_OK, gst_pad_link (srcpad, sinkpad));
|
|
|
|
|
|
|
|
gst_object_unref (srcpad);
|
|
|
|
gst_object_unref (sinkpad);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return mq;
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_START_TEST (test_simple_pipeline)
|
|
|
|
{
|
2011-04-11 14:08:30 +00:00
|
|
|
GstElement *pipe;
|
2007-03-28 18:38:11 +00:00
|
|
|
GstElement *inputs[1];
|
|
|
|
GstElement *outputs[1];
|
|
|
|
GstMessage *msg;
|
|
|
|
|
|
|
|
pipe = gst_pipeline_new ("pipeline");
|
|
|
|
|
|
|
|
inputs[0] = gst_element_factory_make ("fakesrc", NULL);
|
|
|
|
fail_unless (inputs[0] != NULL, "failed to create 'fakesrc' element");
|
|
|
|
g_object_set (inputs[0], "num-buffers", 256, NULL);
|
|
|
|
|
|
|
|
outputs[0] = gst_element_factory_make ("fakesink", NULL);
|
|
|
|
fail_unless (outputs[0] != NULL, "failed to create 'fakesink' element");
|
|
|
|
|
2011-04-11 14:08:30 +00:00
|
|
|
setup_multiqueue (pipe, inputs, outputs, 1);
|
2007-03-28 18:38:11 +00:00
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PLAYING);
|
|
|
|
|
|
|
|
msg = gst_bus_poll (GST_ELEMENT_BUS (pipe),
|
|
|
|
GST_MESSAGE_EOS | GST_MESSAGE_ERROR, -1);
|
|
|
|
|
|
|
|
fail_if (GST_MESSAGE_TYPE (msg) == GST_MESSAGE_ERROR,
|
|
|
|
"Expected EOS message, got ERROR message");
|
|
|
|
gst_message_unref (msg);
|
|
|
|
|
|
|
|
GST_LOG ("Got EOS, cleaning up");
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_NULL);
|
|
|
|
gst_object_unref (pipe);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
|
|
|
GST_START_TEST (test_simple_shutdown_while_running)
|
|
|
|
{
|
2011-04-11 14:08:30 +00:00
|
|
|
GstElement *pipe;
|
2007-03-28 18:38:11 +00:00
|
|
|
GstElement *inputs[1];
|
|
|
|
GstElement *outputs[1];
|
|
|
|
GstMessage *msg;
|
|
|
|
|
|
|
|
pipe = gst_pipeline_new ("pipeline");
|
|
|
|
|
|
|
|
inputs[0] = gst_element_factory_make ("fakesrc", NULL);
|
|
|
|
fail_unless (inputs[0] != NULL, "failed to create 'fakesrc' element");
|
|
|
|
|
|
|
|
outputs[0] = gst_element_factory_make ("fakesink", NULL);
|
|
|
|
fail_unless (outputs[0] != NULL, "failed to create 'fakesink' element");
|
|
|
|
|
2011-04-11 14:08:30 +00:00
|
|
|
setup_multiqueue (pipe, inputs, outputs, 1);
|
2007-03-28 18:38:11 +00:00
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PAUSED);
|
|
|
|
|
|
|
|
/* wait until pipeline is up and running */
|
|
|
|
msg = gst_bus_poll (GST_ELEMENT_BUS (pipe),
|
|
|
|
GST_MESSAGE_ERROR | GST_MESSAGE_ASYNC_DONE, -1);
|
|
|
|
fail_if (GST_MESSAGE_TYPE (msg) == GST_MESSAGE_ERROR, "Got ERROR message");
|
|
|
|
gst_message_unref (msg);
|
|
|
|
|
|
|
|
GST_LOG ("pipeline is running now");
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PAUSED);
|
|
|
|
|
|
|
|
/* wait a bit to accumulate some buffers in the queue (while it's blocking
|
|
|
|
* in the sink) */
|
|
|
|
msg =
|
|
|
|
gst_bus_poll (GST_ELEMENT_BUS (pipe), GST_MESSAGE_ERROR, GST_SECOND / 4);
|
|
|
|
if (msg)
|
|
|
|
g_error ("Got ERROR message");
|
|
|
|
|
|
|
|
/* now shut down only the sink, so the queue gets a wrong-state flow return */
|
|
|
|
gst_element_set_state (outputs[0], GST_STATE_NULL);
|
|
|
|
msg =
|
|
|
|
gst_bus_poll (GST_ELEMENT_BUS (pipe), GST_MESSAGE_ERROR, GST_SECOND / 2);
|
|
|
|
if (msg)
|
|
|
|
g_error ("Got ERROR message");
|
|
|
|
|
|
|
|
GST_LOG ("Cleaning up");
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_NULL);
|
|
|
|
gst_object_unref (pipe);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
|
|
|
GST_START_TEST (test_simple_create_destroy)
|
|
|
|
{
|
|
|
|
GstElement *mq;
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
gst_object_unref (mq);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
|
|
|
GST_START_TEST (test_request_pads)
|
|
|
|
{
|
|
|
|
GstElement *mq;
|
|
|
|
GstPad *sink1, *sink2;
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
sink1 = gst_element_get_request_pad (mq, "foo_%u");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (sink1 == NULL,
|
2011-11-03 16:49:45 +00:00
|
|
|
"Expected NULL pad, as there is no request pad template for 'foo_%%u'");
|
2007-03-28 18:38:11 +00:00
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
sink1 = gst_element_get_request_pad (mq, "src_%u");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (sink1 == NULL,
|
2011-11-03 16:49:45 +00:00
|
|
|
"Expected NULL pad, as there is no request pad template for 'src_%%u'");
|
2007-03-28 18:38:11 +00:00
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
sink1 = gst_element_get_request_pad (mq, "sink_%u");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (sink1 != NULL);
|
|
|
|
fail_unless (GST_IS_PAD (sink1));
|
|
|
|
fail_unless (GST_PAD_IS_SINK (sink1));
|
|
|
|
GST_LOG ("Got pad %s:%s", GST_DEBUG_PAD_NAME (sink1));
|
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
sink2 = gst_element_get_request_pad (mq, "sink_%u");
|
2007-03-28 18:38:11 +00:00
|
|
|
fail_unless (sink2 != NULL);
|
|
|
|
fail_unless (GST_IS_PAD (sink2));
|
|
|
|
fail_unless (GST_PAD_IS_SINK (sink2));
|
|
|
|
GST_LOG ("Got pad %s:%s", GST_DEBUG_PAD_NAME (sink2));
|
|
|
|
|
|
|
|
fail_unless (sink1 != sink2);
|
|
|
|
|
|
|
|
GST_LOG ("Cleaning up");
|
|
|
|
gst_object_unref (sink1);
|
|
|
|
gst_object_unref (sink2);
|
|
|
|
gst_object_unref (mq);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
static GstPad *
|
|
|
|
mq_sinkpad_to_srcpad (GstElement * mq, GstPad * sink)
|
|
|
|
{
|
|
|
|
GstPad *srcpad = NULL;
|
|
|
|
|
|
|
|
gchar *mq_sinkpad_name;
|
|
|
|
gchar *mq_srcpad_name;
|
|
|
|
|
|
|
|
mq_sinkpad_name = gst_pad_get_name (sink);
|
|
|
|
fail_unless (g_str_has_prefix (mq_sinkpad_name, "sink"));
|
2011-11-03 16:49:45 +00:00
|
|
|
mq_srcpad_name = g_strdup_printf ("src_%s", mq_sinkpad_name + 5);
|
2008-05-21 15:57:52 +00:00
|
|
|
srcpad = gst_element_get_static_pad (mq, mq_srcpad_name);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
fail_unless (srcpad != NULL);
|
|
|
|
|
|
|
|
g_free (mq_sinkpad_name);
|
|
|
|
g_free (mq_srcpad_name);
|
|
|
|
|
|
|
|
return srcpad;
|
|
|
|
}
|
|
|
|
|
2011-03-30 08:48:47 +00:00
|
|
|
GST_START_TEST (test_request_pads_named)
|
|
|
|
{
|
|
|
|
GstElement *mq;
|
|
|
|
GstPad *sink1, *sink2, *sink3, *sink4;
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
sink1 = gst_element_get_request_pad (mq, "sink_1");
|
2011-03-30 08:48:47 +00:00
|
|
|
fail_unless (sink1 != NULL);
|
|
|
|
fail_unless (GST_IS_PAD (sink1));
|
|
|
|
fail_unless (GST_PAD_IS_SINK (sink1));
|
2011-11-03 16:49:45 +00:00
|
|
|
fail_unless_equals_string (GST_PAD_NAME (sink1), "sink_1");
|
2011-03-30 08:48:47 +00:00
|
|
|
GST_LOG ("Got pad %s:%s", GST_DEBUG_PAD_NAME (sink1));
|
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
sink3 = gst_element_get_request_pad (mq, "sink_3");
|
2011-03-30 08:48:47 +00:00
|
|
|
fail_unless (sink3 != NULL);
|
|
|
|
fail_unless (GST_IS_PAD (sink3));
|
|
|
|
fail_unless (GST_PAD_IS_SINK (sink3));
|
2011-11-03 16:49:45 +00:00
|
|
|
fail_unless_equals_string (GST_PAD_NAME (sink3), "sink_3");
|
2011-03-30 08:48:47 +00:00
|
|
|
GST_LOG ("Got pad %s:%s", GST_DEBUG_PAD_NAME (sink3));
|
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
sink2 = gst_element_get_request_pad (mq, "sink_2");
|
2011-03-30 08:48:47 +00:00
|
|
|
fail_unless (sink2 != NULL);
|
|
|
|
fail_unless (GST_IS_PAD (sink2));
|
|
|
|
fail_unless (GST_PAD_IS_SINK (sink2));
|
2011-11-03 16:49:45 +00:00
|
|
|
fail_unless_equals_string (GST_PAD_NAME (sink2), "sink_2");
|
2011-03-30 08:48:47 +00:00
|
|
|
GST_LOG ("Got pad %s:%s", GST_DEBUG_PAD_NAME (sink2));
|
|
|
|
|
|
|
|
/* This gets us the first unused id, sink0 */
|
2011-11-03 16:49:45 +00:00
|
|
|
sink4 = gst_element_get_request_pad (mq, "sink_%u");
|
2011-03-30 08:48:47 +00:00
|
|
|
fail_unless (sink4 != NULL);
|
|
|
|
fail_unless (GST_IS_PAD (sink4));
|
|
|
|
fail_unless (GST_PAD_IS_SINK (sink4));
|
2011-11-03 16:49:45 +00:00
|
|
|
fail_unless_equals_string (GST_PAD_NAME (sink4), "sink_0");
|
2011-03-30 08:48:47 +00:00
|
|
|
GST_LOG ("Got pad %s:%s", GST_DEBUG_PAD_NAME (sink4));
|
|
|
|
|
|
|
|
GST_LOG ("Cleaning up");
|
|
|
|
gst_object_unref (sink1);
|
|
|
|
gst_object_unref (sink2);
|
2011-04-14 07:07:25 +00:00
|
|
|
gst_object_unref (sink3);
|
|
|
|
gst_object_unref (sink4);
|
2011-03-30 08:48:47 +00:00
|
|
|
gst_object_unref (mq);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
2011-11-15 10:20:48 +00:00
|
|
|
static gboolean
|
2011-11-16 16:22:56 +00:00
|
|
|
mq_dummypad_query (GstPad * sinkpad, GstObject * parent, GstQuery * query)
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
{
|
2011-11-15 10:20:48 +00:00
|
|
|
gboolean res = TRUE;
|
|
|
|
|
|
|
|
switch (GST_QUERY_TYPE (query)) {
|
|
|
|
case GST_QUERY_CAPS:
|
|
|
|
{
|
|
|
|
GstCaps *filter, *caps;
|
|
|
|
|
|
|
|
gst_query_parse_caps (query, &filter);
|
|
|
|
caps = (filter ? gst_caps_ref (filter) : gst_caps_new_any ());
|
|
|
|
gst_query_set_caps_result (query, caps);
|
|
|
|
gst_caps_unref (caps);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
2011-11-16 16:22:56 +00:00
|
|
|
res = gst_pad_query_default (sinkpad, parent, query);
|
2011-11-15 10:20:48 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
return res;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
struct PadData
|
|
|
|
{
|
2016-03-16 14:42:55 +00:00
|
|
|
GstPad *input_pad;
|
|
|
|
GstPad *out_pad;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
guint8 pad_num;
|
|
|
|
guint32 *max_linked_id_ptr;
|
|
|
|
guint32 *eos_count_ptr;
|
|
|
|
gboolean is_linked;
|
2007-10-25 15:14:02 +00:00
|
|
|
gboolean first_buf;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
gint n_linked;
|
|
|
|
|
|
|
|
GMutex *mutex;
|
|
|
|
GCond *cond;
|
2015-04-05 15:47:26 +00:00
|
|
|
|
|
|
|
/* used by initial_events_nodelay */
|
|
|
|
gint event_count;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static GstFlowReturn
|
2011-11-17 11:40:45 +00:00
|
|
|
mq_dummypad_chain (GstPad * sinkpad, GstObject * parent, GstBuffer * buf)
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
{
|
|
|
|
guint32 cur_id;
|
|
|
|
struct PadData *pad_data;
|
2012-01-20 13:23:57 +00:00
|
|
|
GstMapInfo info;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
pad_data = gst_pad_get_element_private (sinkpad);
|
|
|
|
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_lock (&_check_lock);
|
2007-07-16 16:04:49 +00:00
|
|
|
fail_if (pad_data == NULL);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
/* Read an ID from the first 4 bytes of the buffer data and check it's
|
|
|
|
* what we expect */
|
2012-01-20 13:23:57 +00:00
|
|
|
fail_unless (gst_buffer_map (buf, &info, GST_MAP_READ));
|
|
|
|
fail_unless (info.size >= 4);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_unlock (&_check_lock);
|
2012-01-20 13:23:57 +00:00
|
|
|
cur_id = GST_READ_UINT32_BE (info.data);
|
|
|
|
gst_buffer_unmap (buf, &info);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
g_mutex_lock (pad_data->mutex);
|
|
|
|
|
|
|
|
/* For not-linked pads, ensure that we're not running ahead of the 'linked'
|
2007-10-25 15:14:02 +00:00
|
|
|
* pads. The first buffer is allowed to get ahead, because otherwise things can't
|
|
|
|
* always pre-roll correctly */
|
2016-03-16 14:42:55 +00:00
|
|
|
if (pad_data->max_linked_id_ptr) {
|
|
|
|
if (!pad_data->is_linked) {
|
|
|
|
/* If there are no linked pads, we can't track a max_id for them :) */
|
|
|
|
if (pad_data->n_linked > 0 && !pad_data->first_buf) {
|
|
|
|
g_mutex_lock (&_check_lock);
|
|
|
|
fail_unless (cur_id <= *(pad_data->max_linked_id_ptr) + 1,
|
|
|
|
"Got buffer %u on pad %u before buffer %u was seen on a "
|
|
|
|
"linked pad (max: %u)", cur_id, pad_data->pad_num, cur_id - 1,
|
|
|
|
*(pad_data->max_linked_id_ptr));
|
|
|
|
g_mutex_unlock (&_check_lock);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/* Update the max_id value */
|
|
|
|
if (cur_id > *(pad_data->max_linked_id_ptr))
|
|
|
|
*(pad_data->max_linked_id_ptr) = cur_id;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
}
|
|
|
|
}
|
2007-10-25 15:14:02 +00:00
|
|
|
pad_data->first_buf = FALSE;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
g_mutex_unlock (pad_data->mutex);
|
|
|
|
|
|
|
|
/* Unref the buffer */
|
|
|
|
gst_buffer_unref (buf);
|
|
|
|
|
|
|
|
/* Return OK or not-linked as indicated */
|
|
|
|
return pad_data->is_linked ? GST_FLOW_OK : GST_FLOW_NOT_LINKED;
|
|
|
|
}
|
|
|
|
|
|
|
|
static gboolean
|
2011-11-17 11:40:45 +00:00
|
|
|
mq_dummypad_event (GstPad * sinkpad, GstObject * parent, GstEvent * event)
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
{
|
|
|
|
struct PadData *pad_data;
|
|
|
|
|
|
|
|
pad_data = gst_pad_get_element_private (sinkpad);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_lock (&_check_lock);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
fail_if (pad_data == NULL);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_unlock (&_check_lock);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
if (GST_EVENT_TYPE (event) == GST_EVENT_EOS) {
|
|
|
|
g_mutex_lock (pad_data->mutex);
|
|
|
|
|
|
|
|
/* Accumulate that we've seen the EOS and signal the main thread */
|
2014-03-14 16:32:17 +00:00
|
|
|
if (pad_data->eos_count_ptr)
|
|
|
|
*(pad_data->eos_count_ptr) += 1;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
GST_DEBUG ("EOS on pad %u", pad_data->pad_num);
|
|
|
|
|
|
|
|
g_cond_broadcast (pad_data->cond);
|
|
|
|
g_mutex_unlock (pad_data->mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
gst_event_unref (event);
|
|
|
|
return TRUE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2016-03-16 14:42:55 +00:00
|
|
|
construct_n_pads (GstElement * mq, struct PadData *pad_data, gint n_pads,
|
|
|
|
gint n_linked)
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
{
|
|
|
|
gint i;
|
2013-05-09 11:32:07 +00:00
|
|
|
GstSegment segment;
|
|
|
|
|
|
|
|
gst_segment_init (&segment, GST_FORMAT_BYTES);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
/* Construct NPADS dummy output pads. The first 'n_linked' return FLOW_OK, the rest
|
|
|
|
* return NOT_LINKED. The not-linked ones check the expected ordering of
|
|
|
|
* output buffers */
|
2016-03-16 14:42:55 +00:00
|
|
|
for (i = 0; i < n_pads; i++) {
|
|
|
|
GstPad *mq_srcpad, *mq_sinkpad, *inpad, *outpad;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
gchar *name;
|
|
|
|
|
|
|
|
name = g_strdup_printf ("dummysrc%d", i);
|
2016-03-16 14:42:55 +00:00
|
|
|
inpad = gst_pad_new (name, GST_PAD_SRC);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
g_free (name);
|
2016-03-16 14:42:55 +00:00
|
|
|
gst_pad_set_query_function (inpad, mq_dummypad_query);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
mq_sinkpad = gst_element_get_request_pad (mq, "sink_%u");
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
fail_unless (mq_sinkpad != NULL);
|
2016-03-16 14:42:55 +00:00
|
|
|
fail_unless (gst_pad_link (inpad, mq_sinkpad) == GST_PAD_LINK_OK);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
gst_pad_set_active (inpad, TRUE);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
gst_pad_push_event (inpad, gst_event_new_stream_start ("test"));
|
|
|
|
gst_pad_push_event (inpad, gst_event_new_segment (&segment));
|
2013-05-09 11:32:07 +00:00
|
|
|
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
mq_srcpad = mq_sinkpad_to_srcpad (mq, mq_sinkpad);
|
|
|
|
|
|
|
|
name = g_strdup_printf ("dummysink%d", i);
|
2016-03-16 14:42:55 +00:00
|
|
|
outpad = gst_pad_new (name, GST_PAD_SINK);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
g_free (name);
|
2016-03-16 14:42:55 +00:00
|
|
|
gst_pad_set_chain_function (outpad, mq_dummypad_chain);
|
|
|
|
gst_pad_set_event_function (outpad, mq_dummypad_event);
|
|
|
|
gst_pad_set_query_function (outpad, mq_dummypad_query);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
pad_data[i].pad_num = i;
|
2016-03-16 14:42:55 +00:00
|
|
|
pad_data[i].input_pad = inpad;
|
|
|
|
pad_data[i].out_pad = outpad;
|
|
|
|
pad_data[i].max_linked_id_ptr = NULL;
|
|
|
|
pad_data[i].eos_count_ptr = NULL;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
pad_data[i].is_linked = (i < n_linked ? TRUE : FALSE);
|
|
|
|
pad_data[i].n_linked = n_linked;
|
2016-03-16 14:42:55 +00:00
|
|
|
pad_data[i].cond = NULL;
|
|
|
|
pad_data[i].mutex = NULL;
|
2007-10-25 15:14:02 +00:00
|
|
|
pad_data[i].first_buf = TRUE;
|
2016-03-16 14:42:55 +00:00
|
|
|
gst_pad_set_element_private (outpad, pad_data + i);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
fail_unless (gst_pad_link (mq_srcpad, outpad) == GST_PAD_LINK_OK);
|
|
|
|
gst_pad_set_active (outpad, TRUE);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
gst_object_unref (mq_sinkpad);
|
|
|
|
gst_object_unref (mq_srcpad);
|
|
|
|
}
|
2016-03-16 14:42:55 +00:00
|
|
|
}
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
static void
|
|
|
|
push_n_buffers (struct PadData *pad_data, gint num_buffers,
|
|
|
|
const guint8 * pad_pattern, guint pattern_size)
|
|
|
|
{
|
|
|
|
gint i;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
for (i = 0; i < num_buffers; i++) {
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
guint8 cur_pad;
|
|
|
|
GstBuffer *buf;
|
|
|
|
GstFlowReturn ret;
|
2012-01-20 13:23:57 +00:00
|
|
|
GstMapInfo info;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
cur_pad = pad_pattern[i % pattern_size];
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
buf = gst_buffer_new_and_alloc (4);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_lock (&_check_lock);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
fail_if (buf == NULL);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_unlock (&_check_lock);
|
2011-03-21 18:15:27 +00:00
|
|
|
|
2012-01-20 13:23:57 +00:00
|
|
|
fail_unless (gst_buffer_map (buf, &info, GST_MAP_WRITE));
|
|
|
|
GST_WRITE_UINT32_BE (info.data, i + 1);
|
|
|
|
gst_buffer_unmap (buf, &info);
|
2007-10-25 15:14:02 +00:00
|
|
|
GST_BUFFER_TIMESTAMP (buf) = (i + 1) * GST_SECOND;
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
ret = gst_pad_push (pad_data[cur_pad].input_pad, buf);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_lock (&_check_lock);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
if (pad_data[cur_pad].is_linked) {
|
|
|
|
fail_unless (ret == GST_FLOW_OK,
|
|
|
|
"Push on pad %d returned %d when FLOW_OK was expected", cur_pad, ret);
|
|
|
|
} else {
|
|
|
|
/* Expect OK initially, then NOT_LINKED when the srcpad starts pushing */
|
|
|
|
fail_unless (ret == GST_FLOW_OK || ret == GST_FLOW_NOT_LINKED,
|
|
|
|
"Push on pad %d returned %d when FLOW_OK or NOT_LINKED was expected",
|
|
|
|
cur_pad, ret);
|
|
|
|
}
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_unlock (&_check_lock);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
}
|
2016-03-16 14:42:55 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
run_output_order_test (gint n_linked)
|
|
|
|
{
|
|
|
|
/* This test creates a multiqueue with 2 linked output, and 3 outputs that
|
|
|
|
* return 'not-linked' when data is pushed, then verifies that all buffers
|
|
|
|
* are received on not-linked pads only after earlier buffers on the
|
|
|
|
* 'linked' pads are made */
|
|
|
|
GstElement *pipe;
|
|
|
|
GstElement *mq;
|
|
|
|
struct PadData pad_data[5];
|
|
|
|
guint32 max_linked_id;
|
|
|
|
guint32 eos_seen;
|
|
|
|
GMutex mutex;
|
|
|
|
GCond cond;
|
|
|
|
gint i;
|
|
|
|
const gint NPADS = 5;
|
|
|
|
const gint NBUFFERS = 1000;
|
|
|
|
|
|
|
|
g_mutex_init (&mutex);
|
|
|
|
g_cond_init (&cond);
|
|
|
|
|
|
|
|
pipe = gst_bin_new ("testbin");
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
fail_unless (mq != NULL);
|
|
|
|
gst_bin_add (GST_BIN (pipe), mq);
|
|
|
|
|
|
|
|
/* No limits */
|
|
|
|
g_object_set (mq,
|
|
|
|
"max-size-bytes", (guint) 0,
|
|
|
|
"max-size-buffers", (guint) 0,
|
|
|
|
"max-size-time", (guint64) 0,
|
|
|
|
"extra-size-bytes", (guint) 0,
|
|
|
|
"extra-size-buffers", (guint) 0, "extra-size-time", (guint64) 0, NULL);
|
|
|
|
|
|
|
|
construct_n_pads (mq, pad_data, NPADS, n_linked);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
for (i = 0; i < NPADS; i++) {
|
2016-03-16 14:42:55 +00:00
|
|
|
pad_data[i].max_linked_id_ptr = &max_linked_id;
|
|
|
|
/* Only look for EOS on the linked pads */
|
|
|
|
pad_data[i].eos_count_ptr = (i < n_linked) ? &eos_seen : NULL;
|
|
|
|
pad_data[i].cond = &cond;
|
|
|
|
pad_data[i].mutex = &mutex;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Run the test. Push 1000 buffers through the multiqueue in a pattern */
|
|
|
|
max_linked_id = 0;
|
|
|
|
eos_seen = 0;
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PLAYING);
|
|
|
|
|
|
|
|
{
|
|
|
|
const guint8 pad_pattern[] =
|
|
|
|
{ 0, 0, 0, 0, 1, 1, 2, 1, 0, 2, 3, 2, 3, 1, 4 };
|
|
|
|
const guint n = sizeof (pad_pattern) / sizeof (guint8);
|
|
|
|
push_n_buffers (pad_data, NBUFFERS, pad_pattern, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < NPADS; i++) {
|
|
|
|
gst_pad_push_event (pad_data[i].input_pad, gst_event_new_eos ());
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Wait while the buffers are processed */
|
2012-09-12 09:52:25 +00:00
|
|
|
g_mutex_lock (&mutex);
|
2010-02-10 13:40:17 +00:00
|
|
|
/* We wait until EOS has been pushed on all linked pads */
|
|
|
|
while (eos_seen < n_linked) {
|
2012-09-12 09:52:25 +00:00
|
|
|
g_cond_wait (&cond, &mutex);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
}
|
2012-09-12 09:52:25 +00:00
|
|
|
g_mutex_unlock (&mutex);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
|
|
|
/* Clean up */
|
2010-02-10 13:40:17 +00:00
|
|
|
for (i = 0; i < NPADS; i++) {
|
2016-03-16 14:42:55 +00:00
|
|
|
GstPad *mq_input = gst_pad_get_peer (pad_data[i].input_pad);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
gst_pad_unlink (pad_data[i].input_pad, mq_input);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
gst_element_release_request_pad (mq, mq_input);
|
|
|
|
gst_object_unref (mq_input);
|
2016-03-16 14:42:55 +00:00
|
|
|
gst_object_unref (pad_data[i].input_pad);
|
|
|
|
gst_object_unref (pad_data[i].out_pad);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_NULL);
|
|
|
|
gst_object_unref (pipe);
|
|
|
|
|
2012-09-12 09:52:25 +00:00
|
|
|
g_cond_clear (&cond);
|
|
|
|
g_mutex_clear (&mutex);
|
plugins/elements/gstmultiqueue.*: Take the multiqueue lock when updating the fill level so we don't get confused.
Original commit message from CVS:
* plugins/elements/gstmultiqueue.c: (gst_multi_queue_init),
(gst_single_queue_flush), (apply_segment), (apply_buffer),
(gst_single_queue_push_one), (gst_multi_queue_loop),
(gst_multi_queue_sink_activate_push), (gst_multi_queue_sink_event),
(gst_multi_queue_src_activate_push), (wake_up_next_non_linked),
(compute_high_id), (gst_single_queue_new):
* plugins/elements/gstmultiqueue.h:
Take the multiqueue lock when updating the fill level so we don't get
confused.
After applying a buffer or event on the src pad segment, make sure to
call gst_data_queue_limits_changed() to get the data queue to unblock
and check the filled state again.
Rework the not-linked pad handling so the logic is that not-linked
pads can push as fast as they like, but only so they never get
ahead of any linked pads.
* tests/check/elements/multiqueue.c: (mq_sinkpad_to_srcpad),
(mq_dummypad_getcaps), (mq_dummypad_chain), (mq_dummypad_event),
(run_output_order_test), (GST_START_TEST), (multiqueue_suite):
Add a test to check that not-linked pads always stay behind
linked pads.
2007-06-26 14:45:15 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
GST_START_TEST (test_output_order)
|
|
|
|
{
|
|
|
|
run_output_order_test (2);
|
|
|
|
run_output_order_test (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
GST_START_TEST (test_not_linked_eos)
|
|
|
|
{
|
|
|
|
/* This test creates a multiqueue with 1 linked output and 1 not-linked
|
|
|
|
* pad. It pushes a few buffers through each, then EOS on the linked
|
|
|
|
* pad and waits until that arrives. After that, it pushes some more
|
|
|
|
* buffers on the not-linked pad and then EOS and checks that those
|
|
|
|
* are all output */
|
|
|
|
GstElement *pipe;
|
|
|
|
GstElement *mq;
|
|
|
|
struct PadData pad_data[2];
|
|
|
|
guint32 eos_seen;
|
|
|
|
GMutex mutex;
|
|
|
|
GCond cond;
|
|
|
|
gint i;
|
|
|
|
const gint NPADS = 2;
|
|
|
|
const gint NBUFFERS = 20;
|
|
|
|
GstSegment segment;
|
|
|
|
|
|
|
|
gst_segment_init (&segment, GST_FORMAT_BYTES);
|
|
|
|
|
|
|
|
g_mutex_init (&mutex);
|
|
|
|
g_cond_init (&cond);
|
|
|
|
|
|
|
|
pipe = gst_bin_new ("testbin");
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
fail_unless (mq != NULL);
|
|
|
|
gst_bin_add (GST_BIN (pipe), mq);
|
|
|
|
|
|
|
|
/* No limits */
|
|
|
|
g_object_set (mq,
|
|
|
|
"max-size-bytes", (guint) 0,
|
|
|
|
"max-size-buffers", (guint) 0,
|
|
|
|
"max-size-time", (guint64) 0,
|
|
|
|
"extra-size-bytes", (guint) 0,
|
|
|
|
"extra-size-buffers", (guint) 0, "extra-size-time", (guint64) 0, NULL);
|
|
|
|
|
|
|
|
/* Construct NPADS dummy output pads. The first 1 returns FLOW_OK, the rest
|
|
|
|
* return NOT_LINKED. */
|
|
|
|
construct_n_pads (mq, pad_data, NPADS, 1);
|
|
|
|
for (i = 0; i < NPADS; i++) {
|
|
|
|
/* Only look for EOS on the linked pads */
|
|
|
|
pad_data[i].eos_count_ptr = &eos_seen;
|
|
|
|
pad_data[i].cond = &cond;
|
|
|
|
pad_data[i].mutex = &mutex;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Run the test. Push 20 buffers through the multiqueue in a pattern */
|
|
|
|
eos_seen = 0;
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PLAYING);
|
|
|
|
|
|
|
|
{
|
|
|
|
const guint8 pad_pattern[] = { 0, 1 };
|
|
|
|
const guint n = sizeof (pad_pattern) / sizeof (guint8);
|
|
|
|
push_n_buffers (pad_data, NBUFFERS, pad_pattern, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Make the linked pad go EOS */
|
|
|
|
gst_pad_push_event (pad_data[0].input_pad, gst_event_new_eos ());
|
|
|
|
|
|
|
|
g_mutex_lock (&mutex);
|
|
|
|
/* Wait until EOS has been seen on the linked pad */
|
|
|
|
while (eos_seen == 0)
|
|
|
|
g_cond_wait (&cond, &mutex);
|
|
|
|
g_mutex_unlock (&mutex);
|
|
|
|
|
|
|
|
/* Now push some more buffers to the not-linked pad */
|
|
|
|
{
|
|
|
|
const guint8 pad_pattern[] = { 1, 1 };
|
|
|
|
const guint n = sizeof (pad_pattern) / sizeof (guint8);
|
|
|
|
push_n_buffers (pad_data, NBUFFERS, pad_pattern, n);
|
|
|
|
}
|
|
|
|
/* And EOS on the not-linked pad */
|
|
|
|
gst_pad_push_event (pad_data[1].input_pad, gst_event_new_eos ());
|
|
|
|
|
|
|
|
g_mutex_lock (&mutex);
|
|
|
|
while (eos_seen < NPADS)
|
|
|
|
g_cond_wait (&cond, &mutex);
|
|
|
|
g_mutex_unlock (&mutex);
|
|
|
|
|
|
|
|
/* Clean up */
|
|
|
|
for (i = 0; i < NPADS; i++) {
|
|
|
|
GstPad *mq_input = gst_pad_get_peer (pad_data[i].input_pad);
|
|
|
|
|
|
|
|
gst_pad_unlink (pad_data[i].input_pad, mq_input);
|
|
|
|
gst_element_release_request_pad (mq, mq_input);
|
|
|
|
gst_object_unref (mq_input);
|
|
|
|
gst_object_unref (pad_data[i].input_pad);
|
|
|
|
gst_object_unref (pad_data[i].out_pad);
|
|
|
|
}
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_NULL);
|
|
|
|
gst_object_unref (pipe);
|
|
|
|
|
|
|
|
g_cond_clear (&cond);
|
|
|
|
g_mutex_clear (&mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
2010-10-27 16:11:35 +00:00
|
|
|
GST_START_TEST (test_sparse_stream)
|
|
|
|
{
|
|
|
|
/* This test creates a multiqueue with 2 streams. One receives
|
|
|
|
* a constant flow of buffers, the other only gets one buffer, and then
|
|
|
|
* new-segment events, and returns not-linked. The multiqueue should not fill.
|
|
|
|
*/
|
|
|
|
GstElement *pipe;
|
|
|
|
GstElement *mq;
|
|
|
|
GstPad *inputpads[2];
|
|
|
|
GstPad *sinkpads[2];
|
|
|
|
GstEvent *event;
|
|
|
|
struct PadData pad_data[2];
|
|
|
|
guint32 eos_seen, max_linked_id;
|
2012-09-12 09:52:25 +00:00
|
|
|
GMutex mutex;
|
|
|
|
GCond cond;
|
2010-10-27 16:11:35 +00:00
|
|
|
gint i;
|
|
|
|
const gint NBUFFERS = 100;
|
2011-05-13 16:07:24 +00:00
|
|
|
GstSegment segment;
|
2010-10-27 16:11:35 +00:00
|
|
|
|
2012-09-12 09:52:25 +00:00
|
|
|
g_mutex_init (&mutex);
|
|
|
|
g_cond_init (&cond);
|
2010-10-27 16:11:35 +00:00
|
|
|
|
|
|
|
pipe = gst_pipeline_new ("testbin");
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
fail_unless (mq != NULL);
|
|
|
|
gst_bin_add (GST_BIN (pipe), mq);
|
|
|
|
|
|
|
|
/* 1 second limit */
|
|
|
|
g_object_set (mq,
|
|
|
|
"max-size-bytes", (guint) 0,
|
|
|
|
"max-size-buffers", (guint) 0,
|
|
|
|
"max-size-time", (guint64) GST_SECOND,
|
|
|
|
"extra-size-bytes", (guint) 0,
|
|
|
|
"extra-size-buffers", (guint) 0, "extra-size-time", (guint64) 0, NULL);
|
|
|
|
|
2013-05-09 11:32:07 +00:00
|
|
|
gst_segment_init (&segment, GST_FORMAT_TIME);
|
|
|
|
|
2010-10-27 16:11:35 +00:00
|
|
|
/* Construct 2 dummy output pads. */
|
|
|
|
for (i = 0; i < 2; i++) {
|
|
|
|
GstPad *mq_srcpad, *mq_sinkpad;
|
|
|
|
gchar *name;
|
|
|
|
|
|
|
|
name = g_strdup_printf ("dummysrc%d", i);
|
|
|
|
inputpads[i] = gst_pad_new (name, GST_PAD_SRC);
|
|
|
|
g_free (name);
|
2011-11-15 10:20:48 +00:00
|
|
|
gst_pad_set_query_function (inputpads[i], mq_dummypad_query);
|
2010-10-27 16:11:35 +00:00
|
|
|
|
2011-11-03 16:49:45 +00:00
|
|
|
mq_sinkpad = gst_element_get_request_pad (mq, "sink_%u");
|
2010-10-27 16:11:35 +00:00
|
|
|
fail_unless (mq_sinkpad != NULL);
|
2011-11-15 10:20:48 +00:00
|
|
|
fail_unless (gst_pad_link (inputpads[i], mq_sinkpad) == GST_PAD_LINK_OK);
|
2010-10-27 16:11:35 +00:00
|
|
|
|
|
|
|
gst_pad_set_active (inputpads[i], TRUE);
|
|
|
|
|
2013-05-09 11:32:07 +00:00
|
|
|
gst_pad_push_event (inputpads[i], gst_event_new_stream_start ("test"));
|
|
|
|
gst_pad_push_event (inputpads[i], gst_event_new_segment (&segment));
|
|
|
|
|
2010-10-27 16:11:35 +00:00
|
|
|
mq_srcpad = mq_sinkpad_to_srcpad (mq, mq_sinkpad);
|
|
|
|
|
|
|
|
name = g_strdup_printf ("dummysink%d", i);
|
|
|
|
sinkpads[i] = gst_pad_new (name, GST_PAD_SINK);
|
|
|
|
g_free (name);
|
|
|
|
gst_pad_set_chain_function (sinkpads[i], mq_dummypad_chain);
|
|
|
|
gst_pad_set_event_function (sinkpads[i], mq_dummypad_event);
|
2011-11-15 10:20:48 +00:00
|
|
|
gst_pad_set_query_function (sinkpads[i], mq_dummypad_query);
|
2010-10-27 16:11:35 +00:00
|
|
|
|
|
|
|
pad_data[i].pad_num = i;
|
|
|
|
pad_data[i].max_linked_id_ptr = &max_linked_id;
|
2014-03-14 16:32:17 +00:00
|
|
|
if (i == 0)
|
|
|
|
pad_data[i].eos_count_ptr = &eos_seen;
|
|
|
|
else
|
|
|
|
pad_data[i].eos_count_ptr = NULL;
|
2010-10-27 16:11:35 +00:00
|
|
|
pad_data[i].is_linked = (i == 0) ? TRUE : FALSE;
|
|
|
|
pad_data[i].n_linked = 1;
|
2012-09-12 09:52:25 +00:00
|
|
|
pad_data[i].cond = &cond;
|
|
|
|
pad_data[i].mutex = &mutex;
|
2010-10-27 16:11:35 +00:00
|
|
|
pad_data[i].first_buf = TRUE;
|
|
|
|
gst_pad_set_element_private (sinkpads[i], pad_data + i);
|
|
|
|
|
2011-11-15 10:20:48 +00:00
|
|
|
fail_unless (gst_pad_link (mq_srcpad, sinkpads[i]) == GST_PAD_LINK_OK);
|
2010-10-27 16:11:35 +00:00
|
|
|
gst_pad_set_active (sinkpads[i], TRUE);
|
|
|
|
|
|
|
|
gst_object_unref (mq_sinkpad);
|
|
|
|
gst_object_unref (mq_srcpad);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Run the test. Push 100 buffers through the multiqueue */
|
|
|
|
max_linked_id = 0;
|
|
|
|
eos_seen = 0;
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PLAYING);
|
|
|
|
|
|
|
|
for (i = 0; i < NBUFFERS; i++) {
|
|
|
|
GstBuffer *buf;
|
|
|
|
GstFlowReturn ret;
|
|
|
|
GstClockTime ts;
|
2012-01-20 13:23:57 +00:00
|
|
|
GstMapInfo info;
|
2010-10-27 16:11:35 +00:00
|
|
|
|
|
|
|
ts = gst_util_uint64_scale_int (GST_SECOND, i, 10);
|
|
|
|
|
|
|
|
buf = gst_buffer_new_and_alloc (4);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_lock (&_check_lock);
|
2010-10-27 16:11:35 +00:00
|
|
|
fail_if (buf == NULL);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_unlock (&_check_lock);
|
2011-03-21 18:15:27 +00:00
|
|
|
|
2012-01-20 13:23:57 +00:00
|
|
|
fail_unless (gst_buffer_map (buf, &info, GST_MAP_WRITE));
|
|
|
|
GST_WRITE_UINT32_BE (info.data, i + 1);
|
|
|
|
gst_buffer_unmap (buf, &info);
|
2011-03-21 18:15:27 +00:00
|
|
|
|
2010-10-27 16:11:35 +00:00
|
|
|
GST_BUFFER_TIMESTAMP (buf) = gst_util_uint64_scale_int (GST_SECOND, i, 10);
|
|
|
|
|
|
|
|
/* If i == 0, also push the buffer to the 2nd pad */
|
|
|
|
if (i == 0)
|
|
|
|
ret = gst_pad_push (inputpads[1], gst_buffer_ref (buf));
|
|
|
|
|
|
|
|
ret = gst_pad_push (inputpads[0], buf);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_lock (&_check_lock);
|
2010-10-27 16:11:35 +00:00
|
|
|
fail_unless (ret == GST_FLOW_OK,
|
|
|
|
"Push on pad %d returned %d when FLOW_OK was expected", 0, ret);
|
2012-01-22 22:44:59 +00:00
|
|
|
g_mutex_unlock (&_check_lock);
|
2010-10-27 16:11:35 +00:00
|
|
|
|
|
|
|
/* Push a new segment update on the 2nd pad */
|
2011-05-13 16:07:24 +00:00
|
|
|
gst_segment_init (&segment, GST_FORMAT_TIME);
|
|
|
|
segment.start = ts;
|
|
|
|
segment.time = ts;
|
|
|
|
event = gst_event_new_segment (&segment);
|
2010-10-27 16:11:35 +00:00
|
|
|
gst_pad_push_event (inputpads[1], event);
|
|
|
|
}
|
|
|
|
|
|
|
|
event = gst_event_new_eos ();
|
|
|
|
gst_pad_push_event (inputpads[0], gst_event_ref (event));
|
|
|
|
gst_pad_push_event (inputpads[1], event);
|
|
|
|
|
|
|
|
/* Wait while the buffers are processed */
|
2012-09-12 09:52:25 +00:00
|
|
|
g_mutex_lock (&mutex);
|
2014-03-14 16:32:17 +00:00
|
|
|
/* We wait until EOS has been pushed on pad 1 */
|
|
|
|
while (eos_seen < 1) {
|
2012-09-12 09:52:25 +00:00
|
|
|
g_cond_wait (&cond, &mutex);
|
2010-10-27 16:11:35 +00:00
|
|
|
}
|
2012-09-12 09:52:25 +00:00
|
|
|
g_mutex_unlock (&mutex);
|
2010-10-27 16:11:35 +00:00
|
|
|
|
|
|
|
/* Clean up */
|
|
|
|
for (i = 0; i < 2; i++) {
|
|
|
|
GstPad *mq_input = gst_pad_get_peer (inputpads[i]);
|
|
|
|
|
|
|
|
gst_pad_unlink (inputpads[i], mq_input);
|
|
|
|
gst_element_release_request_pad (mq, mq_input);
|
|
|
|
gst_object_unref (mq_input);
|
|
|
|
gst_object_unref (inputpads[i]);
|
|
|
|
|
|
|
|
gst_object_unref (sinkpads[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_NULL);
|
|
|
|
gst_object_unref (pipe);
|
|
|
|
|
2012-09-12 09:52:25 +00:00
|
|
|
g_cond_clear (&cond);
|
|
|
|
g_mutex_clear (&mutex);
|
2010-10-27 16:11:35 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
2014-04-09 13:42:48 +00:00
|
|
|
static gpointer
|
|
|
|
pad_push_thread (gpointer data)
|
|
|
|
{
|
|
|
|
GstPad *pad = data;
|
|
|
|
GstBuffer *buf;
|
|
|
|
|
|
|
|
buf = gst_buffer_new ();
|
|
|
|
gst_pad_push (pad, buf);
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_START_TEST (test_limit_changes)
|
|
|
|
{
|
|
|
|
/* This test creates a multiqueue with 1 stream. The limit of the queue
|
|
|
|
* is two buffers, we check if we block once this is reached. Then we
|
|
|
|
* change the limit to three buffers and check if this is waking up
|
|
|
|
* the queue and we get the third buffer.
|
|
|
|
*/
|
|
|
|
GstElement *pipe;
|
|
|
|
GstElement *mq, *fakesink;
|
|
|
|
GstPad *inputpad;
|
|
|
|
GstPad *mq_sinkpad;
|
|
|
|
GstSegment segment;
|
|
|
|
GThread *thread;
|
|
|
|
|
|
|
|
pipe = gst_pipeline_new ("testbin");
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
fail_unless (mq != NULL);
|
|
|
|
gst_bin_add (GST_BIN (pipe), mq);
|
|
|
|
|
|
|
|
fakesink = gst_element_factory_make ("fakesink", NULL);
|
|
|
|
fail_unless (fakesink != NULL);
|
|
|
|
gst_bin_add (GST_BIN (pipe), fakesink);
|
|
|
|
|
|
|
|
g_object_set (mq,
|
|
|
|
"max-size-bytes", (guint) 0,
|
|
|
|
"max-size-buffers", (guint) 2,
|
|
|
|
"max-size-time", (guint64) 0,
|
|
|
|
"extra-size-bytes", (guint) 0,
|
|
|
|
"extra-size-buffers", (guint) 0, "extra-size-time", (guint64) 0, NULL);
|
|
|
|
|
|
|
|
gst_segment_init (&segment, GST_FORMAT_TIME);
|
|
|
|
|
|
|
|
inputpad = gst_pad_new ("dummysrc", GST_PAD_SRC);
|
|
|
|
gst_pad_set_query_function (inputpad, mq_dummypad_query);
|
|
|
|
|
|
|
|
mq_sinkpad = gst_element_get_request_pad (mq, "sink_%u");
|
|
|
|
fail_unless (mq_sinkpad != NULL);
|
|
|
|
fail_unless (gst_pad_link (inputpad, mq_sinkpad) == GST_PAD_LINK_OK);
|
|
|
|
|
|
|
|
gst_pad_set_active (inputpad, TRUE);
|
|
|
|
|
|
|
|
gst_pad_push_event (inputpad, gst_event_new_stream_start ("test"));
|
|
|
|
gst_pad_push_event (inputpad, gst_event_new_segment (&segment));
|
|
|
|
|
|
|
|
gst_object_unref (mq_sinkpad);
|
|
|
|
|
|
|
|
fail_unless (gst_element_link (mq, fakesink));
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PAUSED);
|
|
|
|
|
|
|
|
thread = g_thread_new ("push1", pad_push_thread, inputpad);
|
|
|
|
g_thread_join (thread);
|
|
|
|
thread = g_thread_new ("push2", pad_push_thread, inputpad);
|
|
|
|
g_thread_join (thread);
|
|
|
|
thread = g_thread_new ("push3", pad_push_thread, inputpad);
|
|
|
|
g_thread_join (thread);
|
|
|
|
thread = g_thread_new ("push4", pad_push_thread, inputpad);
|
|
|
|
|
|
|
|
/* Wait until we are actually blocking... we unfortunately can't
|
|
|
|
* know that without sleeping */
|
|
|
|
sleep (1);
|
|
|
|
g_object_set (mq, "max-size-buffers", (guint) 3, NULL);
|
|
|
|
g_thread_join (thread);
|
|
|
|
|
|
|
|
g_object_set (mq, "max-size-buffers", (guint) 4, NULL);
|
|
|
|
thread = g_thread_new ("push5", pad_push_thread, inputpad);
|
|
|
|
g_thread_join (thread);
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_NULL);
|
|
|
|
gst_object_unref (inputpad);
|
|
|
|
gst_object_unref (pipe);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
2014-05-08 20:33:37 +00:00
|
|
|
static GMutex block_mutex;
|
|
|
|
static GCond block_cond;
|
|
|
|
static gint unblock_count;
|
|
|
|
static gboolean expect_overrun;
|
|
|
|
|
|
|
|
static GstFlowReturn
|
|
|
|
pad_chain_block (GstPad * pad, GstObject * parent, GstBuffer * buffer)
|
|
|
|
{
|
|
|
|
g_mutex_lock (&block_mutex);
|
|
|
|
while (unblock_count == 0) {
|
|
|
|
g_cond_wait (&block_cond, &block_mutex);
|
|
|
|
}
|
|
|
|
if (unblock_count > 0) {
|
|
|
|
unblock_count--;
|
|
|
|
}
|
|
|
|
g_mutex_unlock (&block_mutex);
|
|
|
|
|
|
|
|
gst_buffer_unref (buffer);
|
|
|
|
return GST_FLOW_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static gboolean
|
|
|
|
pad_event_always_ok (GstPad * pad, GstObject * parent, GstEvent * event)
|
|
|
|
{
|
|
|
|
gst_event_unref (event);
|
|
|
|
return TRUE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
mq_overrun (GstElement * mq, gpointer udata)
|
|
|
|
{
|
|
|
|
fail_unless (expect_overrun);
|
|
|
|
|
|
|
|
/* unblock always so we don't get stuck */
|
|
|
|
g_mutex_lock (&block_mutex);
|
|
|
|
unblock_count = 2; /* let the PTS=0 and PTS=none go */
|
|
|
|
g_cond_signal (&block_cond);
|
|
|
|
g_mutex_unlock (&block_mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_START_TEST (test_buffering_with_none_pts)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* This test creates a multiqueue where source pushing blocks so we can check
|
|
|
|
* how its buffering level is reacting to GST_CLOCK_TIME_NONE buffers
|
|
|
|
* mixed with properly timestamped buffers.
|
|
|
|
*
|
|
|
|
* Sequence of pushing:
|
|
|
|
* pts=0
|
|
|
|
* pts=none
|
|
|
|
* pts=1 (it gets full now)
|
|
|
|
* pts=none (overrun expected)
|
|
|
|
*/
|
|
|
|
GstElement *mq;
|
|
|
|
GstPad *inputpad;
|
|
|
|
GstPad *outputpad;
|
|
|
|
GstPad *mq_sinkpad;
|
|
|
|
GstPad *mq_srcpad;
|
|
|
|
GstSegment segment;
|
|
|
|
GstBuffer *buffer;
|
|
|
|
|
|
|
|
g_mutex_init (&block_mutex);
|
|
|
|
g_cond_init (&block_cond);
|
|
|
|
unblock_count = 0;
|
|
|
|
expect_overrun = FALSE;
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
fail_unless (mq != NULL);
|
|
|
|
|
|
|
|
g_object_set (mq,
|
|
|
|
"max-size-bytes", (guint) 0,
|
|
|
|
"max-size-buffers", (guint) 0,
|
|
|
|
"max-size-time", (guint64) GST_SECOND, NULL);
|
|
|
|
g_signal_connect (mq, "overrun", (GCallback) mq_overrun, NULL);
|
|
|
|
|
|
|
|
gst_segment_init (&segment, GST_FORMAT_TIME);
|
|
|
|
|
|
|
|
inputpad = gst_pad_new ("dummysrc", GST_PAD_SRC);
|
|
|
|
outputpad = gst_pad_new ("dummysink", GST_PAD_SINK);
|
|
|
|
gst_pad_set_chain_function (outputpad, pad_chain_block);
|
|
|
|
gst_pad_set_event_function (outputpad, pad_event_always_ok);
|
|
|
|
mq_sinkpad = gst_element_get_request_pad (mq, "sink_%u");
|
|
|
|
mq_srcpad = gst_element_get_static_pad (mq, "src_0");
|
|
|
|
fail_unless (mq_sinkpad != NULL);
|
|
|
|
fail_unless (gst_pad_link (inputpad, mq_sinkpad) == GST_PAD_LINK_OK);
|
|
|
|
fail_unless (gst_pad_link (mq_srcpad, outputpad) == GST_PAD_LINK_OK);
|
|
|
|
|
|
|
|
gst_pad_set_active (inputpad, TRUE);
|
|
|
|
gst_pad_set_active (outputpad, TRUE);
|
|
|
|
gst_pad_push_event (inputpad, gst_event_new_stream_start ("test"));
|
|
|
|
gst_pad_push_event (inputpad, gst_event_new_segment (&segment));
|
|
|
|
|
|
|
|
gst_element_set_state (mq, GST_STATE_PAUSED);
|
|
|
|
|
|
|
|
/* push a buffer with PTS = 0 */
|
|
|
|
buffer = gst_buffer_new ();
|
|
|
|
GST_BUFFER_PTS (buffer) = 0;
|
|
|
|
fail_unless (gst_pad_push (inputpad, buffer) == GST_FLOW_OK);
|
|
|
|
|
|
|
|
/* push a buffer with PTS = NONE */
|
|
|
|
buffer = gst_buffer_new ();
|
|
|
|
GST_BUFFER_PTS (buffer) = GST_CLOCK_TIME_NONE;
|
|
|
|
fail_unless (gst_pad_push (inputpad, buffer) == GST_FLOW_OK);
|
|
|
|
|
|
|
|
/* push a buffer with PTS = 1s, so we have 1s of data in multiqueue, we are
|
|
|
|
* full */
|
|
|
|
buffer = gst_buffer_new ();
|
|
|
|
GST_BUFFER_PTS (buffer) = GST_SECOND;
|
|
|
|
fail_unless (gst_pad_push (inputpad, buffer) == GST_FLOW_OK);
|
|
|
|
|
|
|
|
/* push a buffer with PTS = NONE, the queue is full so it should overrun */
|
|
|
|
expect_overrun = TRUE;
|
|
|
|
buffer = gst_buffer_new ();
|
|
|
|
GST_BUFFER_PTS (buffer) = GST_CLOCK_TIME_NONE;
|
|
|
|
fail_unless (gst_pad_push (inputpad, buffer) == GST_FLOW_OK);
|
|
|
|
|
|
|
|
g_mutex_lock (&block_mutex);
|
|
|
|
unblock_count = -1;
|
|
|
|
g_cond_signal (&block_cond);
|
|
|
|
g_mutex_unlock (&block_mutex);
|
|
|
|
|
|
|
|
gst_element_set_state (mq, GST_STATE_NULL);
|
|
|
|
gst_object_unref (inputpad);
|
|
|
|
gst_object_unref (outputpad);
|
2014-05-30 04:42:17 +00:00
|
|
|
gst_object_unref (mq_sinkpad);
|
|
|
|
gst_object_unref (mq_srcpad);
|
2014-05-08 20:33:37 +00:00
|
|
|
gst_object_unref (mq);
|
|
|
|
g_mutex_clear (&block_mutex);
|
|
|
|
g_cond_clear (&block_cond);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
2015-04-05 15:47:26 +00:00
|
|
|
static gboolean
|
|
|
|
event_func_signal (GstPad * sinkpad, GstObject * parent, GstEvent * event)
|
|
|
|
{
|
|
|
|
struct PadData *pad_data;
|
|
|
|
|
|
|
|
GST_LOG_OBJECT (sinkpad, "%s event", GST_EVENT_TYPE_NAME (event));
|
|
|
|
|
|
|
|
pad_data = gst_pad_get_element_private (sinkpad);
|
|
|
|
|
|
|
|
g_mutex_lock (pad_data->mutex);
|
|
|
|
++pad_data->event_count;
|
|
|
|
g_cond_broadcast (pad_data->cond);
|
|
|
|
g_mutex_unlock (pad_data->mutex);
|
|
|
|
|
|
|
|
gst_event_unref (event);
|
|
|
|
return TRUE;
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_START_TEST (test_initial_events_nodelay)
|
|
|
|
{
|
|
|
|
struct PadData pad_data = { 0, };
|
|
|
|
GstElement *pipe;
|
|
|
|
GstElement *mq;
|
|
|
|
GstPad *inputpad;
|
|
|
|
GstPad *sinkpad;
|
|
|
|
GstSegment segment;
|
|
|
|
GstCaps *caps;
|
|
|
|
GMutex mutex;
|
|
|
|
GCond cond;
|
|
|
|
|
|
|
|
g_mutex_init (&mutex);
|
|
|
|
g_cond_init (&cond);
|
|
|
|
|
|
|
|
pipe = gst_pipeline_new ("testbin");
|
|
|
|
|
|
|
|
mq = gst_element_factory_make ("multiqueue", NULL);
|
|
|
|
fail_unless (mq != NULL);
|
|
|
|
gst_bin_add (GST_BIN (pipe), mq);
|
|
|
|
|
|
|
|
{
|
|
|
|
GstPad *mq_srcpad, *mq_sinkpad;
|
|
|
|
|
|
|
|
inputpad = gst_pad_new ("dummysrc", GST_PAD_SRC);
|
|
|
|
|
|
|
|
mq_sinkpad = gst_element_get_request_pad (mq, "sink_%u");
|
|
|
|
fail_unless (mq_sinkpad != NULL);
|
|
|
|
fail_unless (gst_pad_link (inputpad, mq_sinkpad) == GST_PAD_LINK_OK);
|
|
|
|
|
|
|
|
gst_pad_set_active (inputpad, TRUE);
|
|
|
|
|
|
|
|
mq_srcpad = mq_sinkpad_to_srcpad (mq, mq_sinkpad);
|
|
|
|
|
|
|
|
sinkpad = gst_pad_new ("dummysink", GST_PAD_SINK);
|
|
|
|
gst_pad_set_event_function (sinkpad, event_func_signal);
|
|
|
|
|
|
|
|
pad_data.event_count = 0;
|
|
|
|
pad_data.cond = &cond;
|
|
|
|
pad_data.mutex = &mutex;
|
|
|
|
gst_pad_set_element_private (sinkpad, &pad_data);
|
|
|
|
|
|
|
|
fail_unless (gst_pad_link (mq_srcpad, sinkpad) == GST_PAD_LINK_OK);
|
|
|
|
gst_pad_set_active (sinkpad, TRUE);
|
|
|
|
|
|
|
|
gst_object_unref (mq_sinkpad);
|
|
|
|
gst_object_unref (mq_srcpad);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Run the test: push events through multiqueue */
|
|
|
|
gst_element_set_state (pipe, GST_STATE_PLAYING);
|
|
|
|
|
|
|
|
gst_pad_push_event (inputpad, gst_event_new_stream_start ("test"));
|
|
|
|
|
|
|
|
caps = gst_caps_new_empty_simple ("foo/x-bar");
|
|
|
|
gst_pad_push_event (inputpad, gst_event_new_caps (caps));
|
|
|
|
gst_caps_unref (caps);
|
|
|
|
|
|
|
|
gst_segment_init (&segment, GST_FORMAT_TIME);
|
|
|
|
gst_pad_push_event (inputpad, gst_event_new_segment (&segment));
|
|
|
|
|
|
|
|
g_mutex_lock (&mutex);
|
|
|
|
while (pad_data.event_count < 3) {
|
|
|
|
GST_LOG ("%d events so far, waiting for more", pad_data.event_count);
|
|
|
|
g_cond_wait (&cond, &mutex);
|
|
|
|
}
|
|
|
|
g_mutex_unlock (&mutex);
|
|
|
|
|
|
|
|
/* Clean up */
|
|
|
|
{
|
|
|
|
GstPad *mq_input = gst_pad_get_peer (inputpad);
|
|
|
|
|
|
|
|
gst_pad_unlink (inputpad, mq_input);
|
|
|
|
gst_element_release_request_pad (mq, mq_input);
|
|
|
|
gst_object_unref (mq_input);
|
|
|
|
gst_object_unref (inputpad);
|
|
|
|
|
|
|
|
gst_object_unref (sinkpad);
|
|
|
|
}
|
|
|
|
|
|
|
|
gst_element_set_state (pipe, GST_STATE_NULL);
|
|
|
|
gst_object_unref (pipe);
|
|
|
|
|
|
|
|
g_cond_clear (&cond);
|
|
|
|
g_mutex_clear (&mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_END_TEST;
|
|
|
|
|
2007-03-28 18:38:11 +00:00
|
|
|
static Suite *
|
|
|
|
multiqueue_suite (void)
|
|
|
|
{
|
|
|
|
Suite *s = suite_create ("multiqueue");
|
|
|
|
TCase *tc_chain = tcase_create ("general");
|
|
|
|
|
|
|
|
suite_add_tcase (s, tc_chain);
|
|
|
|
tcase_add_test (tc_chain, test_simple_create_destroy);
|
|
|
|
tcase_add_test (tc_chain, test_simple_pipeline);
|
2007-06-06 18:11:10 +00:00
|
|
|
tcase_add_test (tc_chain, test_simple_shutdown_while_running);
|
2007-03-28 18:38:11 +00:00
|
|
|
|
|
|
|
tcase_add_test (tc_chain, test_request_pads);
|
2011-03-30 08:48:47 +00:00
|
|
|
tcase_add_test (tc_chain, test_request_pads_named);
|
2007-03-28 18:38:11 +00:00
|
|
|
|
2013-10-02 09:24:02 +00:00
|
|
|
/* Disabled, The test (and not multiqueue itself) is racy.
|
|
|
|
* See https://bugzilla.gnome.org/show_bug.cgi?id=708661 */
|
2013-10-02 11:30:54 +00:00
|
|
|
tcase_skip_broken_test (tc_chain, test_output_order);
|
2010-10-27 16:11:35 +00:00
|
|
|
|
2016-03-16 14:42:55 +00:00
|
|
|
tcase_add_test (tc_chain, test_not_linked_eos);
|
|
|
|
|
2010-10-27 16:11:35 +00:00
|
|
|
tcase_add_test (tc_chain, test_sparse_stream);
|
2014-04-09 13:42:48 +00:00
|
|
|
tcase_add_test (tc_chain, test_limit_changes);
|
|
|
|
|
2014-05-08 20:33:37 +00:00
|
|
|
tcase_add_test (tc_chain, test_buffering_with_none_pts);
|
2015-04-05 15:47:26 +00:00
|
|
|
tcase_add_test (tc_chain, test_initial_events_nodelay);
|
2014-05-08 20:33:37 +00:00
|
|
|
|
2007-03-28 18:38:11 +00:00
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
GST_CHECK_MAIN (multiqueue)
|