gstreamer/gst-libs/gst/video/convertframe.c
Sebastian Dröge 48460f5ace video: convertframe: Shut down pipeline asynchronously via the thread pool
If we use the main loop it might happen that the caller (e.g. our unit
test) already shut down the loop once the result was received and in
that case the pipeline would never ever be shut down (and our unit test
would hang).
2018-11-30 12:40:19 +02:00

806 lines
23 KiB
C

/* Small helper element for format conversion
* Copyright (C) 2005 Tim-Philipp Müller <tim centricular net>
* Copyright (C) 2010 Brandon Lewis <brandon.lewis@collabora.co.uk>
* Copyright (C) 2010 Edward Hervey <edward.hervey@collabora.co.uk>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include <string.h>
#include "video.h"
static gboolean
caps_are_raw (const GstCaps * caps)
{
guint i, len;
len = gst_caps_get_size (caps);
for (i = 0; i < len; i++) {
GstStructure *st = gst_caps_get_structure (caps, i);
if (gst_structure_has_name (st, "video/x-raw"))
return TRUE;
}
return FALSE;
}
static gboolean
create_element (const gchar * factory_name, GstElement ** element,
GError ** err)
{
*element = gst_element_factory_make (factory_name, NULL);
if (*element)
return TRUE;
if (err && *err == NULL) {
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN,
"cannot create element '%s' - please check your GStreamer installation",
factory_name);
}
return FALSE;
}
static GstElement *
get_encoder (const GstCaps * caps, GError ** err)
{
GList *encoders = NULL;
GList *filtered = NULL;
GstElementFactory *factory = NULL;
GstElement *encoder = NULL;
encoders =
gst_element_factory_list_get_elements (GST_ELEMENT_FACTORY_TYPE_ENCODER |
GST_ELEMENT_FACTORY_TYPE_MEDIA_IMAGE, GST_RANK_NONE);
if (encoders == NULL) {
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN,
"Cannot find any image encoder");
goto fail;
}
GST_INFO ("got factory list %p", encoders);
gst_plugin_feature_list_debug (encoders);
filtered =
gst_element_factory_list_filter (encoders, caps, GST_PAD_SRC, FALSE);
GST_INFO ("got filtered list %p", filtered);
if (filtered == NULL) {
gchar *tmp = gst_caps_to_string (caps);
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN,
"Cannot find any image encoder for caps %s", tmp);
g_free (tmp);
goto fail;
}
gst_plugin_feature_list_debug (filtered);
factory = (GstElementFactory *) filtered->data;
GST_INFO ("got factory %p", factory);
encoder = gst_element_factory_create (factory, NULL);
GST_INFO ("created encoder element %p, %s", encoder,
GST_ELEMENT_NAME (encoder));
fail:
if (encoders)
gst_plugin_feature_list_free (encoders);
if (filtered)
gst_plugin_feature_list_free (filtered);
return encoder;
}
static GstElement *
build_convert_frame_pipeline (GstElement ** src_element,
GstElement ** sink_element, const GstCaps * from_caps,
GstVideoCropMeta * cmeta, const GstCaps * to_caps, GError ** err)
{
GstElement *vcrop = NULL, *csp = NULL, *csp2 = NULL, *vscale = NULL;
GstElement *src = NULL, *sink = NULL, *encoder = NULL, *pipeline;
GstVideoInfo info;
GError *error = NULL;
if (cmeta) {
if (!create_element ("videocrop", &vcrop, &error)) {
g_error_free (error);
g_warning
("build_convert_frame_pipeline: Buffer has crop metadata but videocrop element is not found. Cropping will be disabled");
} else {
if (!create_element ("videoconvert", &csp2, &error))
goto no_elements;
}
}
/* videoscale is here to correct for the pixel-aspect-ratio for us */
GST_DEBUG ("creating elements");
if (!create_element ("appsrc", &src, &error) ||
!create_element ("videoconvert", &csp, &error) ||
!create_element ("videoscale", &vscale, &error) ||
!create_element ("appsink", &sink, &error))
goto no_elements;
pipeline = gst_pipeline_new ("videoconvert-pipeline");
if (pipeline == NULL)
goto no_pipeline;
/* Add black borders if necessary to keep the DAR */
g_object_set (vscale, "add-borders", TRUE, NULL);
GST_DEBUG ("adding elements");
gst_bin_add_many (GST_BIN (pipeline), src, csp, vscale, sink, NULL);
if (vcrop)
gst_bin_add_many (GST_BIN (pipeline), vcrop, csp2, NULL);
/* set caps */
g_object_set (src, "caps", from_caps, NULL);
if (vcrop) {
gst_video_info_from_caps (&info, from_caps);
g_object_set (vcrop, "left", cmeta->x, NULL);
g_object_set (vcrop, "top", cmeta->y, NULL);
g_object_set (vcrop, "right", GST_VIDEO_INFO_WIDTH (&info) - cmeta->width,
NULL);
g_object_set (vcrop, "bottom",
GST_VIDEO_INFO_HEIGHT (&info) - cmeta->height, NULL);
GST_DEBUG ("crop meta [x,y,width,height]: %d %d %d %d", cmeta->x, cmeta->y,
cmeta->width, cmeta->height);
}
g_object_set (sink, "caps", to_caps, NULL);
/* FIXME: linking is still way too expensive, profile this properly */
if (vcrop) {
GST_DEBUG ("linking src->csp2");
if (!gst_element_link_pads (src, "src", csp2, "sink"))
goto link_failed;
GST_DEBUG ("linking csp2->vcrop");
if (!gst_element_link_pads (csp2, "src", vcrop, "sink"))
goto link_failed;
GST_DEBUG ("linking vcrop->csp");
if (!gst_element_link_pads (vcrop, "src", csp, "sink"))
goto link_failed;
} else {
GST_DEBUG ("linking src->csp");
if (!gst_element_link_pads (src, "src", csp, "sink"))
goto link_failed;
}
GST_DEBUG ("linking csp->vscale");
if (!gst_element_link_pads_full (csp, "src", vscale, "sink",
GST_PAD_LINK_CHECK_NOTHING))
goto link_failed;
if (caps_are_raw (to_caps)) {
GST_DEBUG ("linking vscale->sink");
if (!gst_element_link_pads_full (vscale, "src", sink, "sink",
GST_PAD_LINK_CHECK_NOTHING))
goto link_failed;
} else {
encoder = get_encoder (to_caps, &error);
if (!encoder)
goto no_encoder;
gst_bin_add (GST_BIN (pipeline), encoder);
GST_DEBUG ("linking vscale->encoder");
if (!gst_element_link (vscale, encoder))
goto link_failed;
GST_DEBUG ("linking encoder->sink");
if (!gst_element_link_pads (encoder, "src", sink, "sink"))
goto link_failed;
}
g_object_set (src, "emit-signals", TRUE, NULL);
g_object_set (sink, "emit-signals", TRUE, NULL);
*src_element = src;
*sink_element = sink;
return pipeline;
/* ERRORS */
no_encoder:
{
gst_object_unref (pipeline);
GST_ERROR ("could not find an encoder for provided caps");
if (err)
*err = error;
else
g_error_free (error);
return NULL;
}
no_elements:
{
if (src)
gst_object_unref (src);
if (vcrop)
gst_object_unref (vcrop);
if (csp)
gst_object_unref (csp);
if (csp2)
gst_object_unref (csp2);
if (vscale)
gst_object_unref (vscale);
if (sink)
gst_object_unref (sink);
GST_ERROR ("Could not convert video frame: %s", error->message);
if (err)
*err = error;
else
g_error_free (error);
return NULL;
}
no_pipeline:
{
gst_object_unref (src);
if (vcrop)
gst_object_unref (vcrop);
gst_object_unref (csp);
if (csp2)
gst_object_unref (csp2);
gst_object_unref (vscale);
gst_object_unref (sink);
GST_ERROR ("Could not convert video frame: no pipeline (unknown error)");
if (err)
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
"Could not convert video frame: no pipeline (unknown error)");
return NULL;
}
link_failed:
{
gst_object_unref (pipeline);
GST_ERROR ("Could not convert video frame: failed to link elements");
if (err)
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_NEGOTIATION,
"Could not convert video frame: failed to link elements");
return NULL;
}
}
/**
* gst_video_convert_sample:
* @sample: a #GstSample
* @to_caps: the #GstCaps to convert to
* @timeout: the maximum amount of time allowed for the processing.
* @error: pointer to a #GError. Can be %NULL.
*
* Converts a raw video buffer into the specified output caps.
*
* The output caps can be any raw video formats or any image formats (jpeg, png, ...).
*
* The width, height and pixel-aspect-ratio can also be specified in the output caps.
*
* Returns: The converted #GstSample, or %NULL if an error happened (in which case @err
* will point to the #GError).
*/
GstSample *
gst_video_convert_sample (GstSample * sample, const GstCaps * to_caps,
GstClockTime timeout, GError ** error)
{
GstMessage *msg;
GstBuffer *buf;
GstSample *result = NULL;
GError *err = NULL;
GstBus *bus;
GstCaps *from_caps, *to_caps_copy = NULL;
GstFlowReturn ret;
GstElement *pipeline, *src, *sink;
guint i, n;
g_return_val_if_fail (sample != NULL, NULL);
g_return_val_if_fail (to_caps != NULL, NULL);
buf = gst_sample_get_buffer (sample);
g_return_val_if_fail (buf != NULL, NULL);
from_caps = gst_sample_get_caps (sample);
g_return_val_if_fail (from_caps != NULL, NULL);
to_caps_copy = gst_caps_new_empty ();
n = gst_caps_get_size (to_caps);
for (i = 0; i < n; i++) {
GstStructure *s = gst_caps_get_structure (to_caps, i);
s = gst_structure_copy (s);
gst_structure_remove_field (s, "framerate");
gst_caps_append_structure (to_caps_copy, s);
}
pipeline =
build_convert_frame_pipeline (&src, &sink, from_caps,
gst_buffer_get_video_crop_meta (buf), to_caps_copy, &err);
if (!pipeline)
goto no_pipeline;
/* now set the pipeline to the paused state, after we push the buffer into
* appsrc, this should preroll the converted buffer in appsink */
GST_DEBUG ("running conversion pipeline to caps %" GST_PTR_FORMAT,
to_caps_copy);
if (gst_element_set_state (pipeline,
GST_STATE_PAUSED) == GST_STATE_CHANGE_FAILURE)
goto state_change_failed;
/* feed buffer in appsrc */
GST_DEBUG ("feeding buffer %p, size %" G_GSIZE_FORMAT ", caps %"
GST_PTR_FORMAT, buf, gst_buffer_get_size (buf), from_caps);
g_signal_emit_by_name (src, "push-buffer", buf, &ret);
/* now see what happens. We either got an error somewhere or the pipeline
* prerolled */
bus = gst_element_get_bus (pipeline);
msg = gst_bus_timed_pop_filtered (bus,
timeout, GST_MESSAGE_ERROR | GST_MESSAGE_ASYNC_DONE);
if (msg) {
switch (GST_MESSAGE_TYPE (msg)) {
case GST_MESSAGE_ASYNC_DONE:
{
/* we're prerolled, get the frame from appsink */
g_signal_emit_by_name (sink, "pull-preroll", &result);
if (result) {
GST_DEBUG ("conversion successful: result = %p", result);
} else {
GST_ERROR ("prerolled but no result frame?!");
}
break;
}
case GST_MESSAGE_ERROR:{
gchar *dbg = NULL;
gst_message_parse_error (msg, &err, &dbg);
if (err) {
GST_ERROR ("Could not convert video frame: %s", err->message);
GST_DEBUG ("%s [debug: %s]", err->message, GST_STR_NULL (dbg));
if (error)
*error = err;
else
g_error_free (err);
}
g_free (dbg);
break;
}
default:{
g_return_val_if_reached (NULL);
}
}
gst_message_unref (msg);
} else {
GST_ERROR ("Could not convert video frame: timeout during conversion");
if (error)
*error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
"Could not convert video frame: timeout during conversion");
}
gst_element_set_state (pipeline, GST_STATE_NULL);
gst_object_unref (bus);
gst_object_unref (pipeline);
gst_caps_unref (to_caps_copy);
return result;
/* ERRORS */
no_pipeline:
state_change_failed:
{
gst_caps_unref (to_caps_copy);
if (error)
*error = err;
else
g_error_free (err);
return NULL;
}
}
typedef struct
{
gint ref_count;
GMutex mutex;
GstElement *pipeline;
GstVideoConvertSampleCallback callback;
gpointer user_data;
GDestroyNotify destroy_notify;
GMainContext *context;
GstSample *sample;
GSource *timeout_source;
gboolean finished;
/* Results */
GstSample *converted_sample;
GError *error;
} GstVideoConvertSampleContext;
static GstVideoConvertSampleContext *
gst_video_convert_frame_context_ref (GstVideoConvertSampleContext * ctx)
{
g_atomic_int_inc (&ctx->ref_count);
return ctx;
}
static void
gst_video_convert_frame_context_unref (GstVideoConvertSampleContext * ctx)
{
if (!g_atomic_int_dec_and_test (&ctx->ref_count))
return;
g_mutex_clear (&ctx->mutex);
if (ctx->timeout_source)
g_source_destroy (ctx->timeout_source);
if (ctx->sample)
gst_sample_unref (ctx->sample);
if (ctx->converted_sample)
gst_sample_unref (ctx->converted_sample);
g_clear_error (&ctx->error);
g_main_context_unref (ctx->context);
/* The pipeline was already destroyed in finish() earlier and we
* must not end up here without finish() being called */
g_warn_if_fail (ctx->pipeline == NULL);
g_slice_free (GstVideoConvertSampleContext, ctx);
}
static gboolean
convert_frame_dispatch_callback (GstVideoConvertSampleContext * ctx)
{
GstSample *sample;
GError *error;
g_return_val_if_fail (ctx->converted_sample != NULL
|| ctx->error != NULL, FALSE);
sample = ctx->converted_sample;
error = ctx->error;
ctx->converted_sample = NULL;
ctx->error = NULL;
ctx->callback (sample, error, ctx->user_data);
if (ctx->destroy_notify)
ctx->destroy_notify (ctx->user_data);
return FALSE;
}
static void
convert_frame_stop_pipeline (GstElement * element, gpointer user_data)
{
gst_element_set_state (element, GST_STATE_NULL);
}
static void
convert_frame_finish (GstVideoConvertSampleContext * context,
GstSample * sample, GError * error)
{
GSource *source;
g_return_if_fail (!context->finished);
g_return_if_fail (sample != NULL || error != NULL);
context->finished = TRUE;
context->converted_sample = sample;
context->error = error;
if (context->timeout_source)
g_source_destroy (context->timeout_source);
context->timeout_source = NULL;
source = g_timeout_source_new (0);
g_source_set_callback (source,
(GSourceFunc) convert_frame_dispatch_callback,
gst_video_convert_frame_context_ref (context),
(GDestroyNotify) gst_video_convert_frame_context_unref);
g_source_attach (source, context->context);
g_source_unref (source);
/* Asynchronously stop the pipeline here: this will set its
* state to NULL and get rid of its last reference, which in turn
* will get rid of all remaining references to our context and free
* it too. We can't do this directly here as we might be called from
* a streaming thread.
*
* We don't use the main loop here because the user might shut down it
* immediately after getting the result of the conversion above.
*/
if (context->pipeline) {
gst_element_call_async (context->pipeline, convert_frame_stop_pipeline,
NULL, NULL);
context->pipeline = NULL;
}
}
static gboolean
convert_frame_timeout_callback (GstVideoConvertSampleContext * context)
{
GError *error;
g_mutex_lock (&context->mutex);
if (context->finished)
goto done;
GST_ERROR ("Could not convert video frame: timeout");
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
"Could not convert video frame: timeout");
convert_frame_finish (context, NULL, error);
done:
g_mutex_unlock (&context->mutex);
return FALSE;
}
static gboolean
convert_frame_bus_callback (GstBus * bus, GstMessage * message,
GstVideoConvertSampleContext * context)
{
g_mutex_lock (&context->mutex);
if (context->finished)
goto done;
switch (GST_MESSAGE_TYPE (message)) {
case GST_MESSAGE_ERROR:{
GError *error;
gchar *dbg = NULL;
gst_message_parse_error (message, &error, &dbg);
GST_ERROR ("Could not convert video frame: %s", error->message);
GST_DEBUG ("%s [debug: %s]", error->message, GST_STR_NULL (dbg));
convert_frame_finish (context, NULL, error);
g_free (dbg);
break;
}
default:
break;
}
done:
g_mutex_unlock (&context->mutex);
return FALSE;
}
static void
convert_frame_need_data_callback (GstElement * src, guint size,
GstVideoConvertSampleContext * context)
{
GstFlowReturn ret = GST_FLOW_ERROR;
GError *error;
GstBuffer *buffer;
g_mutex_lock (&context->mutex);
if (context->finished)
goto done;
buffer = gst_sample_get_buffer (context->sample);
g_signal_emit_by_name (src, "push-buffer", buffer, &ret);
gst_sample_unref (context->sample);
context->sample = NULL;
if (ret != GST_FLOW_OK) {
GST_ERROR ("Could not push video frame: %s", gst_flow_get_name (ret));
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
"Could not push video frame: %s", gst_flow_get_name (ret));
convert_frame_finish (context, NULL, error);
}
done:
g_mutex_unlock (&context->mutex);
g_signal_handlers_disconnect_by_func (src, convert_frame_need_data_callback,
context);
}
static GstFlowReturn
convert_frame_new_preroll_callback (GstElement * sink,
GstVideoConvertSampleContext * context)
{
GstSample *sample = NULL;
GError *error = NULL;
g_mutex_lock (&context->mutex);
if (context->finished)
goto done;
g_signal_emit_by_name (sink, "pull-preroll", &sample);
if (!sample) {
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
"Could not get converted video sample");
}
convert_frame_finish (context, sample, error);
done:
g_mutex_unlock (&context->mutex);
g_signal_handlers_disconnect_by_func (sink, convert_frame_need_data_callback,
context);
return GST_FLOW_OK;
}
/**
* gst_video_convert_sample_async:
* @sample: a #GstSample
* @to_caps: the #GstCaps to convert to
* @timeout: the maximum amount of time allowed for the processing.
* @callback: %GstVideoConvertSampleCallback that will be called after conversion.
* @user_data: extra data that will be passed to the @callback
* @destroy_notify: %GDestroyNotify to be called after @user_data is not needed anymore
*
* Converts a raw video buffer into the specified output caps.
*
* The output caps can be any raw video formats or any image formats (jpeg, png, ...).
*
* The width, height and pixel-aspect-ratio can also be specified in the output caps.
*
* @callback will be called after conversion, when an error occured or if conversion didn't
* finish after @timeout. @callback will always be called from the thread default
* %GMainContext, see g_main_context_get_thread_default(). If GLib before 2.22 is used,
* this will always be the global default main context.
*
* @destroy_notify will be called after the callback was called and @user_data is not needed
* anymore.
*/
void
gst_video_convert_sample_async (GstSample * sample,
const GstCaps * to_caps, GstClockTime timeout,
GstVideoConvertSampleCallback callback, gpointer user_data,
GDestroyNotify destroy_notify)
{
GMainContext *context = NULL;
GError *error = NULL;
GstBus *bus;
GstBuffer *buf;
GstCaps *from_caps, *to_caps_copy = NULL;
GstElement *pipeline, *src, *sink;
guint i, n;
GSource *source;
GstVideoConvertSampleContext *ctx;
g_return_if_fail (sample != NULL);
buf = gst_sample_get_buffer (sample);
g_return_if_fail (buf != NULL);
g_return_if_fail (to_caps != NULL);
from_caps = gst_sample_get_caps (sample);
g_return_if_fail (from_caps != NULL);
g_return_if_fail (callback != NULL);
context = g_main_context_get_thread_default ();
if (!context)
context = g_main_context_default ();
to_caps_copy = gst_caps_new_empty ();
n = gst_caps_get_size (to_caps);
for (i = 0; i < n; i++) {
GstStructure *s = gst_caps_get_structure (to_caps, i);
s = gst_structure_copy (s);
gst_structure_remove_field (s, "framerate");
gst_caps_append_structure (to_caps_copy, s);
}
/* There's a reference cycle between the context and the pipeline, which is
* broken up once the finish() is called on the context. At latest when the
* timeout triggers the context will be freed */
ctx = g_slice_new0 (GstVideoConvertSampleContext);
ctx->ref_count = 1;
g_mutex_init (&ctx->mutex);
ctx->sample = gst_sample_ref (sample);
ctx->callback = callback;
ctx->user_data = user_data;
ctx->destroy_notify = destroy_notify;
ctx->context = g_main_context_ref (context);
ctx->finished = FALSE;
pipeline =
build_convert_frame_pipeline (&src, &sink, from_caps,
gst_buffer_get_video_crop_meta (buf), to_caps_copy, &error);
if (!pipeline)
goto no_pipeline;
ctx->pipeline = pipeline;
bus = gst_element_get_bus (pipeline);
if (timeout != GST_CLOCK_TIME_NONE) {
ctx->timeout_source = g_timeout_source_new (timeout / GST_MSECOND);
g_source_set_callback (ctx->timeout_source,
(GSourceFunc) convert_frame_timeout_callback,
gst_video_convert_frame_context_ref (ctx),
(GDestroyNotify) gst_video_convert_frame_context_unref);
g_source_attach (ctx->timeout_source, context);
}
g_signal_connect_data (src, "need-data",
G_CALLBACK (convert_frame_need_data_callback),
gst_video_convert_frame_context_ref (ctx),
(GClosureNotify) gst_video_convert_frame_context_unref, 0);
g_signal_connect_data (sink, "new-preroll",
G_CALLBACK (convert_frame_new_preroll_callback),
gst_video_convert_frame_context_ref (ctx),
(GClosureNotify) gst_video_convert_frame_context_unref, 0);
source = gst_bus_create_watch (bus);
g_source_set_callback (source, (GSourceFunc) convert_frame_bus_callback,
gst_video_convert_frame_context_ref (ctx),
(GDestroyNotify) gst_video_convert_frame_context_unref);
g_source_attach (source, context);
g_source_unref (source);
gst_object_unref (bus);
if (gst_element_set_state (pipeline,
GST_STATE_PAUSED) == GST_STATE_CHANGE_FAILURE)
goto state_change_failed;
gst_caps_unref (to_caps_copy);
gst_video_convert_frame_context_unref (ctx);
return;
/* ERRORS */
no_pipeline:
{
gst_caps_unref (to_caps_copy);
g_mutex_lock (&ctx->mutex);
convert_frame_finish (ctx, NULL, error);
g_mutex_unlock (&ctx->mutex);
gst_video_convert_frame_context_unref (ctx);
return;
}
state_change_failed:
{
gst_caps_unref (to_caps_copy);
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_STATE_CHANGE,
"failed to change state to PLAYING");
g_mutex_lock (&ctx->mutex);
convert_frame_finish (ctx, NULL, error);
g_mutex_unlock (&ctx->mutex);
gst_video_convert_frame_context_unref (ctx);
return;
}
}