/* Small helper element for format conversion * Copyright (C) 2005 Tim-Philipp Müller * Copyright (C) 2010 Brandon Lewis * Copyright (C) 2010 Edward Hervey * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include "video.h" static gboolean caps_are_raw (const GstCaps * caps) { guint i, len; len = gst_caps_get_size (caps); for (i = 0; i < len; i++) { GstStructure *st = gst_caps_get_structure (caps, i); if (gst_structure_has_name (st, "video/x-raw")) return TRUE; } return FALSE; } static gboolean create_element (const gchar * factory_name, GstElement ** element, GError ** err) { *element = gst_element_factory_make (factory_name, NULL); if (*element) return TRUE; if (err && *err == NULL) { *err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN, "cannot create element '%s' - please check your GStreamer installation", factory_name); } return FALSE; } static GstElement * get_encoder (const GstCaps * caps, GError ** err) { GList *encoders = NULL; GList *filtered = NULL; GstElementFactory *factory = NULL; GstElement *encoder = NULL; encoders = gst_element_factory_list_get_elements (GST_ELEMENT_FACTORY_TYPE_ENCODER | GST_ELEMENT_FACTORY_TYPE_MEDIA_IMAGE, GST_RANK_NONE); if (encoders == NULL) { *err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN, "Cannot find any image encoder"); goto fail; } GST_INFO ("got factory list %p", encoders); gst_plugin_feature_list_debug (encoders); filtered = gst_element_factory_list_filter (encoders, caps, GST_PAD_SRC, FALSE); GST_INFO ("got filtered list %p", filtered); if (filtered == NULL) { gchar *tmp = gst_caps_to_string (caps); *err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN, "Cannot find any image encoder for caps %s", tmp); g_free (tmp); goto fail; } gst_plugin_feature_list_debug (filtered); factory = (GstElementFactory *) filtered->data; GST_INFO ("got factory %p", factory); encoder = gst_element_factory_create (factory, NULL); GST_INFO ("created encoder element %p, %s", encoder, GST_ELEMENT_NAME (encoder)); fail: if (encoders) gst_plugin_feature_list_free (encoders); if (filtered) gst_plugin_feature_list_free (filtered); return encoder; } static GstElement * build_convert_frame_pipeline (GstElement ** src_element, GstElement ** sink_element, const GstCaps * from_caps, GstVideoCropMeta * cmeta, const GstCaps * to_caps, GError ** err) { GstElement *vcrop = NULL, *csp = NULL, *csp2 = NULL, *vscale = NULL; GstElement *src = NULL, *sink = NULL, *encoder = NULL, *pipeline; GstVideoInfo info; GError *error = NULL; if (cmeta) { if (!create_element ("videocrop", &vcrop, &error)) { g_error_free (error); g_warning ("build_convert_frame_pipeline: Buffer has crop metadata but videocrop element is not found. Cropping will be disabled"); } else { if (!create_element ("videoconvert", &csp2, &error)) goto no_elements; } } /* videoscale is here to correct for the pixel-aspect-ratio for us */ GST_DEBUG ("creating elements"); if (!create_element ("appsrc", &src, &error) || !create_element ("videoconvert", &csp, &error) || !create_element ("videoscale", &vscale, &error) || !create_element ("appsink", &sink, &error)) goto no_elements; pipeline = gst_pipeline_new ("videoconvert-pipeline"); if (pipeline == NULL) goto no_pipeline; /* Add black borders if necessary to keep the DAR */ g_object_set (vscale, "add-borders", TRUE, NULL); GST_DEBUG ("adding elements"); gst_bin_add_many (GST_BIN (pipeline), src, csp, vscale, sink, NULL); if (vcrop) gst_bin_add_many (GST_BIN (pipeline), vcrop, csp2, NULL); /* set caps */ g_object_set (src, "caps", from_caps, NULL); if (vcrop) { gst_video_info_from_caps (&info, from_caps); g_object_set (vcrop, "left", cmeta->x, NULL); g_object_set (vcrop, "top", cmeta->y, NULL); g_object_set (vcrop, "right", GST_VIDEO_INFO_WIDTH (&info) - cmeta->width, NULL); g_object_set (vcrop, "bottom", GST_VIDEO_INFO_HEIGHT (&info) - cmeta->height, NULL); GST_DEBUG ("crop meta [x,y,width,height]: %d %d %d %d", cmeta->x, cmeta->y, cmeta->width, cmeta->height); } g_object_set (sink, "caps", to_caps, NULL); /* FIXME: linking is still way too expensive, profile this properly */ if (vcrop) { GST_DEBUG ("linking src->csp2"); if (!gst_element_link_pads (src, "src", csp2, "sink")) goto link_failed; GST_DEBUG ("linking csp2->vcrop"); if (!gst_element_link_pads (csp2, "src", vcrop, "sink")) goto link_failed; GST_DEBUG ("linking vcrop->csp"); if (!gst_element_link_pads (vcrop, "src", csp, "sink")) goto link_failed; } else { GST_DEBUG ("linking src->csp"); if (!gst_element_link_pads (src, "src", csp, "sink")) goto link_failed; } GST_DEBUG ("linking csp->vscale"); if (!gst_element_link_pads_full (csp, "src", vscale, "sink", GST_PAD_LINK_CHECK_NOTHING)) goto link_failed; if (caps_are_raw (to_caps)) { GST_DEBUG ("linking vscale->sink"); if (!gst_element_link_pads_full (vscale, "src", sink, "sink", GST_PAD_LINK_CHECK_NOTHING)) goto link_failed; } else { encoder = get_encoder (to_caps, &error); if (!encoder) goto no_encoder; gst_bin_add (GST_BIN (pipeline), encoder); GST_DEBUG ("linking vscale->encoder"); if (!gst_element_link (vscale, encoder)) goto link_failed; GST_DEBUG ("linking encoder->sink"); if (!gst_element_link_pads (encoder, "src", sink, "sink")) goto link_failed; } g_object_set (src, "emit-signals", TRUE, NULL); g_object_set (sink, "emit-signals", TRUE, NULL); *src_element = src; *sink_element = sink; return pipeline; /* ERRORS */ no_encoder: { gst_object_unref (pipeline); GST_ERROR ("could not find an encoder for provided caps"); if (err) *err = error; else g_error_free (error); return NULL; } no_elements: { if (src) gst_object_unref (src); if (vcrop) gst_object_unref (vcrop); if (csp) gst_object_unref (csp); if (csp2) gst_object_unref (csp2); if (vscale) gst_object_unref (vscale); if (sink) gst_object_unref (sink); GST_ERROR ("Could not convert video frame: %s", error->message); if (err) *err = error; else g_error_free (error); return NULL; } no_pipeline: { gst_object_unref (src); if (vcrop) gst_object_unref (vcrop); gst_object_unref (csp); if (csp2) gst_object_unref (csp2); gst_object_unref (vscale); gst_object_unref (sink); GST_ERROR ("Could not convert video frame: no pipeline (unknown error)"); if (err) *err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED, "Could not convert video frame: no pipeline (unknown error)"); return NULL; } link_failed: { gst_object_unref (pipeline); GST_ERROR ("Could not convert video frame: failed to link elements"); if (err) *err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_NEGOTIATION, "Could not convert video frame: failed to link elements"); return NULL; } } /** * gst_video_convert_sample: * @sample: a #GstSample * @to_caps: the #GstCaps to convert to * @timeout: the maximum amount of time allowed for the processing. * @error: pointer to a #GError. Can be %NULL. * * Converts a raw video buffer into the specified output caps. * * The output caps can be any raw video formats or any image formats (jpeg, png, ...). * * The width, height and pixel-aspect-ratio can also be specified in the output caps. * * Returns: The converted #GstSample, or %NULL if an error happened (in which case @err * will point to the #GError). */ GstSample * gst_video_convert_sample (GstSample * sample, const GstCaps * to_caps, GstClockTime timeout, GError ** error) { GstMessage *msg; GstBuffer *buf; GstSample *result = NULL; GError *err = NULL; GstBus *bus; GstCaps *from_caps, *to_caps_copy = NULL; GstFlowReturn ret; GstElement *pipeline, *src, *sink; guint i, n; g_return_val_if_fail (sample != NULL, NULL); g_return_val_if_fail (to_caps != NULL, NULL); buf = gst_sample_get_buffer (sample); g_return_val_if_fail (buf != NULL, NULL); from_caps = gst_sample_get_caps (sample); g_return_val_if_fail (from_caps != NULL, NULL); to_caps_copy = gst_caps_new_empty (); n = gst_caps_get_size (to_caps); for (i = 0; i < n; i++) { GstStructure *s = gst_caps_get_structure (to_caps, i); s = gst_structure_copy (s); gst_structure_remove_field (s, "framerate"); gst_caps_append_structure (to_caps_copy, s); } pipeline = build_convert_frame_pipeline (&src, &sink, from_caps, gst_buffer_get_video_crop_meta (buf), to_caps_copy, &err); if (!pipeline) goto no_pipeline; /* now set the pipeline to the paused state, after we push the buffer into * appsrc, this should preroll the converted buffer in appsink */ GST_DEBUG ("running conversion pipeline to caps %" GST_PTR_FORMAT, to_caps_copy); gst_element_set_state (pipeline, GST_STATE_PAUSED); /* feed buffer in appsrc */ GST_DEBUG ("feeding buffer %p, size %" G_GSIZE_FORMAT ", caps %" GST_PTR_FORMAT, buf, gst_buffer_get_size (buf), from_caps); g_signal_emit_by_name (src, "push-buffer", buf, &ret); /* now see what happens. We either got an error somewhere or the pipeline * prerolled */ bus = gst_element_get_bus (pipeline); msg = gst_bus_timed_pop_filtered (bus, timeout, GST_MESSAGE_ERROR | GST_MESSAGE_ASYNC_DONE); if (msg) { switch (GST_MESSAGE_TYPE (msg)) { case GST_MESSAGE_ASYNC_DONE: { /* we're prerolled, get the frame from appsink */ g_signal_emit_by_name (sink, "pull-preroll", &result); if (result) { GST_DEBUG ("conversion successful: result = %p", result); } else { GST_ERROR ("prerolled but no result frame?!"); } break; } case GST_MESSAGE_ERROR:{ gchar *dbg = NULL; gst_message_parse_error (msg, &err, &dbg); if (err) { GST_ERROR ("Could not convert video frame: %s", err->message); GST_DEBUG ("%s [debug: %s]", err->message, GST_STR_NULL (dbg)); if (error) *error = err; else g_error_free (err); } g_free (dbg); break; } default:{ g_return_val_if_reached (NULL); } } gst_message_unref (msg); } else { GST_ERROR ("Could not convert video frame: timeout during conversion"); if (error) *error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED, "Could not convert video frame: timeout during conversion"); } gst_element_set_state (pipeline, GST_STATE_NULL); gst_object_unref (bus); gst_object_unref (pipeline); gst_caps_unref (to_caps_copy); return result; /* ERRORS */ no_pipeline: { gst_caps_unref (to_caps_copy); if (error) *error = err; else g_error_free (err); return NULL; } } typedef struct { gint ref_count; GMutex mutex; GstElement *pipeline; GstVideoConvertSampleCallback callback; gpointer user_data; GDestroyNotify destroy_notify; GMainContext *context; GstSample *sample; GSource *timeout_source; gboolean finished; /* Results */ GstSample *converted_sample; GError *error; } GstVideoConvertSampleContext; static GstVideoConvertSampleContext * gst_video_convert_frame_context_ref (GstVideoConvertSampleContext * ctx) { g_atomic_int_inc (&ctx->ref_count); return ctx; } static void gst_video_convert_frame_context_unref (GstVideoConvertSampleContext * ctx) { if (!g_atomic_int_dec_and_test (&ctx->ref_count)) return; g_mutex_clear (&ctx->mutex); if (ctx->timeout_source) g_source_destroy (ctx->timeout_source); if (ctx->sample) gst_sample_unref (ctx->sample); if (ctx->converted_sample) gst_sample_unref (ctx->converted_sample); g_clear_error (&ctx->error); g_main_context_unref (ctx->context); /* The pipeline was already destroyed in finish() earlier and we * must not end up here without finish() being called */ g_warn_if_fail (ctx->pipeline == NULL); g_slice_free (GstVideoConvertSampleContext, ctx); } static gboolean convert_frame_dispatch_callback (GstVideoConvertSampleContext * ctx) { GstSample *sample; GError *error; g_return_val_if_fail (ctx->converted_sample != NULL || ctx->error != NULL, FALSE); sample = ctx->converted_sample; error = ctx->error; ctx->converted_sample = NULL; ctx->error = NULL; ctx->callback (sample, error, ctx->user_data); if (ctx->destroy_notify) ctx->destroy_notify (ctx->user_data); return FALSE; } static gboolean convert_frame_stop_pipeline (GstElement * element) { gst_element_set_state (element, GST_STATE_NULL); return FALSE; } static void convert_frame_finish (GstVideoConvertSampleContext * context, GstSample * sample, GError * error) { GSource *source; g_return_if_fail (!context->finished); g_return_if_fail (sample != NULL || error != NULL); context->finished = TRUE; context->converted_sample = sample; context->error = error; if (context->timeout_source) g_source_destroy (context->timeout_source); context->timeout_source = NULL; source = g_timeout_source_new (0); g_source_set_callback (source, (GSourceFunc) convert_frame_dispatch_callback, gst_video_convert_frame_context_ref (context), (GDestroyNotify) gst_video_convert_frame_context_unref); g_source_attach (source, context->context); g_source_unref (source); /* Asynchronously stop the pipeline here: this will set its * state to NULL and get rid of its last reference, which in turn * will get rid of all remaining references to our context and free * it too. We can't do this directly here as we might be called from * a streaming thread. */ if (context->pipeline) { source = g_timeout_source_new (0); g_source_set_callback (source, (GSourceFunc) convert_frame_stop_pipeline, context->pipeline, (GDestroyNotify) gst_object_unref); g_source_attach (source, context->context); g_source_unref (source); context->pipeline = NULL; } } static gboolean convert_frame_timeout_callback (GstVideoConvertSampleContext * context) { GError *error; g_mutex_lock (&context->mutex); if (context->finished) goto done; GST_ERROR ("Could not convert video frame: timeout"); error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED, "Could not convert video frame: timeout"); convert_frame_finish (context, NULL, error); done: g_mutex_unlock (&context->mutex); return FALSE; } static gboolean convert_frame_bus_callback (GstBus * bus, GstMessage * message, GstVideoConvertSampleContext * context) { g_mutex_lock (&context->mutex); if (context->finished) goto done; switch (GST_MESSAGE_TYPE (message)) { case GST_MESSAGE_ERROR:{ GError *error; gchar *dbg = NULL; gst_message_parse_error (message, &error, &dbg); GST_ERROR ("Could not convert video frame: %s", error->message); GST_DEBUG ("%s [debug: %s]", error->message, GST_STR_NULL (dbg)); convert_frame_finish (context, NULL, error); g_free (dbg); break; } default: break; } done: g_mutex_unlock (&context->mutex); return FALSE; } static void convert_frame_need_data_callback (GstElement * src, guint size, GstVideoConvertSampleContext * context) { GstFlowReturn ret = GST_FLOW_ERROR; GError *error; GstBuffer *buffer; g_mutex_lock (&context->mutex); if (context->finished) goto done; buffer = gst_sample_get_buffer (context->sample); g_signal_emit_by_name (src, "push-buffer", buffer, &ret); gst_sample_unref (context->sample); context->sample = NULL; if (ret != GST_FLOW_OK) { GST_ERROR ("Could not push video frame: %s", gst_flow_get_name (ret)); error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED, "Could not push video frame: %s", gst_flow_get_name (ret)); convert_frame_finish (context, NULL, error); } done: g_mutex_unlock (&context->mutex); g_signal_handlers_disconnect_by_func (src, convert_frame_need_data_callback, context); } static GstFlowReturn convert_frame_new_preroll_callback (GstElement * sink, GstVideoConvertSampleContext * context) { GstSample *sample = NULL; GError *error = NULL; g_mutex_lock (&context->mutex); if (context->finished) goto done; g_signal_emit_by_name (sink, "pull-preroll", &sample); if (!sample) { error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED, "Could not get converted video sample"); } convert_frame_finish (context, sample, error); done: g_mutex_unlock (&context->mutex); g_signal_handlers_disconnect_by_func (sink, convert_frame_need_data_callback, context); return GST_FLOW_OK; } /** * gst_video_convert_sample_async: * @sample: a #GstSample * @to_caps: the #GstCaps to convert to * @timeout: the maximum amount of time allowed for the processing. * @callback: %GstVideoConvertSampleCallback that will be called after conversion. * @user_data: extra data that will be passed to the @callback * @destroy_notify: %GDestroyNotify to be called after @user_data is not needed anymore * * Converts a raw video buffer into the specified output caps. * * The output caps can be any raw video formats or any image formats (jpeg, png, ...). * * The width, height and pixel-aspect-ratio can also be specified in the output caps. * * @callback will be called after conversion, when an error occured or if conversion didn't * finish after @timeout. @callback will always be called from the thread default * %GMainContext, see g_main_context_get_thread_default(). If GLib before 2.22 is used, * this will always be the global default main context. * * @destroy_notify will be called after the callback was called and @user_data is not needed * anymore. */ void gst_video_convert_sample_async (GstSample * sample, const GstCaps * to_caps, GstClockTime timeout, GstVideoConvertSampleCallback callback, gpointer user_data, GDestroyNotify destroy_notify) { GMainContext *context = NULL; GError *error = NULL; GstBus *bus; GstBuffer *buf; GstCaps *from_caps, *to_caps_copy = NULL; GstElement *pipeline, *src, *sink; guint i, n; GSource *source; GstVideoConvertSampleContext *ctx; g_return_if_fail (sample != NULL); buf = gst_sample_get_buffer (sample); g_return_if_fail (buf != NULL); g_return_if_fail (to_caps != NULL); from_caps = gst_sample_get_caps (sample); g_return_if_fail (from_caps != NULL); g_return_if_fail (callback != NULL); context = g_main_context_get_thread_default (); if (!context) context = g_main_context_default (); to_caps_copy = gst_caps_new_empty (); n = gst_caps_get_size (to_caps); for (i = 0; i < n; i++) { GstStructure *s = gst_caps_get_structure (to_caps, i); s = gst_structure_copy (s); gst_structure_remove_field (s, "framerate"); gst_caps_append_structure (to_caps_copy, s); } /* There's a reference cycle between the context and the pipeline, which is * broken up once the finish() is called on the context. At latest when the * timeout triggers the context will be freed */ ctx = g_slice_new0 (GstVideoConvertSampleContext); ctx->ref_count = 1; g_mutex_init (&ctx->mutex); ctx->sample = gst_sample_ref (sample); ctx->callback = callback; ctx->user_data = user_data; ctx->destroy_notify = destroy_notify; ctx->context = g_main_context_ref (context); ctx->finished = FALSE; pipeline = build_convert_frame_pipeline (&src, &sink, from_caps, gst_buffer_get_video_crop_meta (buf), to_caps_copy, &error); if (!pipeline) goto no_pipeline; ctx->pipeline = pipeline; bus = gst_element_get_bus (pipeline); if (timeout != GST_CLOCK_TIME_NONE) { ctx->timeout_source = g_timeout_source_new (timeout / GST_MSECOND); g_source_set_callback (ctx->timeout_source, (GSourceFunc) convert_frame_timeout_callback, gst_video_convert_frame_context_ref (ctx), (GDestroyNotify) gst_video_convert_frame_context_unref); g_source_attach (ctx->timeout_source, context); } g_signal_connect_data (src, "need-data", G_CALLBACK (convert_frame_need_data_callback), gst_video_convert_frame_context_ref (ctx), (GClosureNotify) gst_video_convert_frame_context_unref, 0); g_signal_connect_data (sink, "new-preroll", G_CALLBACK (convert_frame_new_preroll_callback), gst_video_convert_frame_context_ref (ctx), (GClosureNotify) gst_video_convert_frame_context_unref, 0); source = gst_bus_create_watch (bus); g_source_set_callback (source, (GSourceFunc) convert_frame_bus_callback, gst_video_convert_frame_context_ref (ctx), (GDestroyNotify) gst_video_convert_frame_context_unref); g_source_attach (source, context); g_source_unref (source); gst_object_unref (bus); if (gst_element_set_state (pipeline, GST_STATE_PLAYING) == GST_STATE_CHANGE_FAILURE) goto state_change_failed; gst_caps_unref (to_caps_copy); gst_video_convert_frame_context_unref (ctx); return; /* ERRORS */ no_pipeline: { gst_caps_unref (to_caps_copy); g_mutex_lock (&ctx->mutex); convert_frame_finish (ctx, NULL, error); g_mutex_unlock (&ctx->mutex); gst_video_convert_frame_context_unref (ctx); return; } state_change_failed: { gst_caps_unref (to_caps_copy); error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_STATE_CHANGE, "failed to change state to PLAYING"); g_mutex_lock (&ctx->mutex); convert_frame_finish (ctx, NULL, error); g_mutex_unlock (&ctx->mutex); gst_video_convert_frame_context_unref (ctx); return; } }