mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-25 03:01:03 +00:00
806 lines
23 KiB
C
806 lines
23 KiB
C
/* Small helper element for format conversion
|
|
* Copyright (C) 2005 Tim-Philipp Müller <tim centricular net>
|
|
* Copyright (C) 2010 Brandon Lewis <brandon.lewis@collabora.co.uk>
|
|
* Copyright (C) 2010 Edward Hervey <edward.hervey@collabora.co.uk>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include <string.h>
|
|
#include "video.h"
|
|
|
|
static gboolean
|
|
caps_are_raw (const GstCaps * caps)
|
|
{
|
|
guint i, len;
|
|
|
|
len = gst_caps_get_size (caps);
|
|
|
|
for (i = 0; i < len; i++) {
|
|
GstStructure *st = gst_caps_get_structure (caps, i);
|
|
if (gst_structure_has_name (st, "video/x-raw"))
|
|
return TRUE;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static gboolean
|
|
create_element (const gchar * factory_name, GstElement ** element,
|
|
GError ** err)
|
|
{
|
|
*element = gst_element_factory_make (factory_name, NULL);
|
|
if (*element)
|
|
return TRUE;
|
|
|
|
if (err && *err == NULL) {
|
|
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN,
|
|
"cannot create element '%s' - please check your GStreamer installation",
|
|
factory_name);
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static GstElement *
|
|
get_encoder (const GstCaps * caps, GError ** err)
|
|
{
|
|
GList *encoders = NULL;
|
|
GList *filtered = NULL;
|
|
GstElementFactory *factory = NULL;
|
|
GstElement *encoder = NULL;
|
|
|
|
encoders =
|
|
gst_element_factory_list_get_elements (GST_ELEMENT_FACTORY_TYPE_ENCODER |
|
|
GST_ELEMENT_FACTORY_TYPE_MEDIA_IMAGE, GST_RANK_NONE);
|
|
|
|
if (encoders == NULL) {
|
|
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN,
|
|
"Cannot find any image encoder");
|
|
goto fail;
|
|
}
|
|
|
|
GST_INFO ("got factory list %p", encoders);
|
|
gst_plugin_feature_list_debug (encoders);
|
|
|
|
filtered =
|
|
gst_element_factory_list_filter (encoders, caps, GST_PAD_SRC, FALSE);
|
|
GST_INFO ("got filtered list %p", filtered);
|
|
|
|
if (filtered == NULL) {
|
|
gchar *tmp = gst_caps_to_string (caps);
|
|
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_MISSING_PLUGIN,
|
|
"Cannot find any image encoder for caps %s", tmp);
|
|
g_free (tmp);
|
|
goto fail;
|
|
}
|
|
|
|
gst_plugin_feature_list_debug (filtered);
|
|
|
|
factory = (GstElementFactory *) filtered->data;
|
|
|
|
GST_INFO ("got factory %p", factory);
|
|
encoder = gst_element_factory_create (factory, NULL);
|
|
|
|
GST_INFO ("created encoder element %p, %s", encoder,
|
|
GST_ELEMENT_NAME (encoder));
|
|
|
|
fail:
|
|
if (encoders)
|
|
gst_plugin_feature_list_free (encoders);
|
|
if (filtered)
|
|
gst_plugin_feature_list_free (filtered);
|
|
|
|
return encoder;
|
|
}
|
|
|
|
static GstElement *
|
|
build_convert_frame_pipeline (GstElement ** src_element,
|
|
GstElement ** sink_element, const GstCaps * from_caps,
|
|
GstVideoCropMeta * cmeta, const GstCaps * to_caps, GError ** err)
|
|
{
|
|
GstElement *vcrop = NULL, *csp = NULL, *csp2 = NULL, *vscale = NULL;
|
|
GstElement *src = NULL, *sink = NULL, *encoder = NULL, *pipeline;
|
|
GstVideoInfo info;
|
|
GError *error = NULL;
|
|
|
|
if (cmeta) {
|
|
if (!create_element ("videocrop", &vcrop, &error)) {
|
|
g_error_free (error);
|
|
g_warning
|
|
("build_convert_frame_pipeline: Buffer has crop metadata but videocrop element is not found. Cropping will be disabled");
|
|
} else {
|
|
if (!create_element ("videoconvert", &csp2, &error))
|
|
goto no_elements;
|
|
}
|
|
}
|
|
|
|
/* videoscale is here to correct for the pixel-aspect-ratio for us */
|
|
GST_DEBUG ("creating elements");
|
|
if (!create_element ("appsrc", &src, &error) ||
|
|
!create_element ("videoconvert", &csp, &error) ||
|
|
!create_element ("videoscale", &vscale, &error) ||
|
|
!create_element ("appsink", &sink, &error))
|
|
goto no_elements;
|
|
|
|
pipeline = gst_pipeline_new ("videoconvert-pipeline");
|
|
if (pipeline == NULL)
|
|
goto no_pipeline;
|
|
|
|
/* Add black borders if necessary to keep the DAR */
|
|
g_object_set (vscale, "add-borders", TRUE, NULL);
|
|
|
|
GST_DEBUG ("adding elements");
|
|
gst_bin_add_many (GST_BIN (pipeline), src, csp, vscale, sink, NULL);
|
|
if (vcrop)
|
|
gst_bin_add_many (GST_BIN (pipeline), vcrop, csp2, NULL);
|
|
|
|
/* set caps */
|
|
g_object_set (src, "caps", from_caps, NULL);
|
|
if (vcrop) {
|
|
gst_video_info_from_caps (&info, from_caps);
|
|
g_object_set (vcrop, "left", cmeta->x, NULL);
|
|
g_object_set (vcrop, "top", cmeta->y, NULL);
|
|
g_object_set (vcrop, "right", GST_VIDEO_INFO_WIDTH (&info) - cmeta->width,
|
|
NULL);
|
|
g_object_set (vcrop, "bottom",
|
|
GST_VIDEO_INFO_HEIGHT (&info) - cmeta->height, NULL);
|
|
GST_DEBUG ("crop meta [x,y,width,height]: %d %d %d %d", cmeta->x, cmeta->y,
|
|
cmeta->width, cmeta->height);
|
|
}
|
|
g_object_set (sink, "caps", to_caps, NULL);
|
|
|
|
/* FIXME: linking is still way too expensive, profile this properly */
|
|
if (vcrop) {
|
|
GST_DEBUG ("linking src->csp2");
|
|
if (!gst_element_link_pads (src, "src", csp2, "sink"))
|
|
goto link_failed;
|
|
|
|
GST_DEBUG ("linking csp2->vcrop");
|
|
if (!gst_element_link_pads (csp2, "src", vcrop, "sink"))
|
|
goto link_failed;
|
|
|
|
GST_DEBUG ("linking vcrop->csp");
|
|
if (!gst_element_link_pads (vcrop, "src", csp, "sink"))
|
|
goto link_failed;
|
|
} else {
|
|
GST_DEBUG ("linking src->csp");
|
|
if (!gst_element_link_pads (src, "src", csp, "sink"))
|
|
goto link_failed;
|
|
}
|
|
|
|
GST_DEBUG ("linking csp->vscale");
|
|
if (!gst_element_link_pads_full (csp, "src", vscale, "sink",
|
|
GST_PAD_LINK_CHECK_NOTHING))
|
|
goto link_failed;
|
|
|
|
if (caps_are_raw (to_caps)) {
|
|
GST_DEBUG ("linking vscale->sink");
|
|
|
|
if (!gst_element_link_pads_full (vscale, "src", sink, "sink",
|
|
GST_PAD_LINK_CHECK_NOTHING))
|
|
goto link_failed;
|
|
} else {
|
|
encoder = get_encoder (to_caps, &error);
|
|
if (!encoder)
|
|
goto no_encoder;
|
|
gst_bin_add (GST_BIN (pipeline), encoder);
|
|
|
|
GST_DEBUG ("linking vscale->encoder");
|
|
if (!gst_element_link (vscale, encoder))
|
|
goto link_failed;
|
|
|
|
GST_DEBUG ("linking encoder->sink");
|
|
if (!gst_element_link_pads (encoder, "src", sink, "sink"))
|
|
goto link_failed;
|
|
}
|
|
|
|
g_object_set (src, "emit-signals", TRUE, NULL);
|
|
g_object_set (sink, "emit-signals", TRUE, NULL);
|
|
|
|
*src_element = src;
|
|
*sink_element = sink;
|
|
|
|
return pipeline;
|
|
/* ERRORS */
|
|
no_encoder:
|
|
{
|
|
gst_object_unref (pipeline);
|
|
|
|
GST_ERROR ("could not find an encoder for provided caps");
|
|
if (err)
|
|
*err = error;
|
|
else
|
|
g_error_free (error);
|
|
|
|
return NULL;
|
|
}
|
|
no_elements:
|
|
{
|
|
if (src)
|
|
gst_object_unref (src);
|
|
if (vcrop)
|
|
gst_object_unref (vcrop);
|
|
if (csp)
|
|
gst_object_unref (csp);
|
|
if (csp2)
|
|
gst_object_unref (csp2);
|
|
if (vscale)
|
|
gst_object_unref (vscale);
|
|
if (sink)
|
|
gst_object_unref (sink);
|
|
GST_ERROR ("Could not convert video frame: %s", error->message);
|
|
if (err)
|
|
*err = error;
|
|
else
|
|
g_error_free (error);
|
|
return NULL;
|
|
}
|
|
no_pipeline:
|
|
{
|
|
gst_object_unref (src);
|
|
if (vcrop)
|
|
gst_object_unref (vcrop);
|
|
gst_object_unref (csp);
|
|
if (csp2)
|
|
gst_object_unref (csp2);
|
|
gst_object_unref (vscale);
|
|
gst_object_unref (sink);
|
|
|
|
GST_ERROR ("Could not convert video frame: no pipeline (unknown error)");
|
|
if (err)
|
|
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
|
|
"Could not convert video frame: no pipeline (unknown error)");
|
|
return NULL;
|
|
}
|
|
link_failed:
|
|
{
|
|
gst_object_unref (pipeline);
|
|
|
|
GST_ERROR ("Could not convert video frame: failed to link elements");
|
|
if (err)
|
|
*err = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_NEGOTIATION,
|
|
"Could not convert video frame: failed to link elements");
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* gst_video_convert_sample:
|
|
* @sample: a #GstSample
|
|
* @to_caps: the #GstCaps to convert to
|
|
* @timeout: the maximum amount of time allowed for the processing.
|
|
* @error: pointer to a #GError. Can be %NULL.
|
|
*
|
|
* Converts a raw video buffer into the specified output caps.
|
|
*
|
|
* The output caps can be any raw video formats or any image formats (jpeg, png, ...).
|
|
*
|
|
* The width, height and pixel-aspect-ratio can also be specified in the output caps.
|
|
*
|
|
* Returns: The converted #GstSample, or %NULL if an error happened (in which case @err
|
|
* will point to the #GError).
|
|
*/
|
|
GstSample *
|
|
gst_video_convert_sample (GstSample * sample, const GstCaps * to_caps,
|
|
GstClockTime timeout, GError ** error)
|
|
{
|
|
GstMessage *msg;
|
|
GstBuffer *buf;
|
|
GstSample *result = NULL;
|
|
GError *err = NULL;
|
|
GstBus *bus;
|
|
GstCaps *from_caps, *to_caps_copy = NULL;
|
|
GstFlowReturn ret;
|
|
GstElement *pipeline, *src, *sink;
|
|
guint i, n;
|
|
|
|
g_return_val_if_fail (sample != NULL, NULL);
|
|
g_return_val_if_fail (to_caps != NULL, NULL);
|
|
|
|
buf = gst_sample_get_buffer (sample);
|
|
g_return_val_if_fail (buf != NULL, NULL);
|
|
|
|
from_caps = gst_sample_get_caps (sample);
|
|
g_return_val_if_fail (from_caps != NULL, NULL);
|
|
|
|
to_caps_copy = gst_caps_new_empty ();
|
|
n = gst_caps_get_size (to_caps);
|
|
for (i = 0; i < n; i++) {
|
|
GstStructure *s = gst_caps_get_structure (to_caps, i);
|
|
|
|
s = gst_structure_copy (s);
|
|
gst_structure_remove_field (s, "framerate");
|
|
gst_caps_append_structure (to_caps_copy, s);
|
|
}
|
|
|
|
pipeline =
|
|
build_convert_frame_pipeline (&src, &sink, from_caps,
|
|
gst_buffer_get_video_crop_meta (buf), to_caps_copy, &err);
|
|
if (!pipeline)
|
|
goto no_pipeline;
|
|
|
|
/* now set the pipeline to the paused state, after we push the buffer into
|
|
* appsrc, this should preroll the converted buffer in appsink */
|
|
GST_DEBUG ("running conversion pipeline to caps %" GST_PTR_FORMAT,
|
|
to_caps_copy);
|
|
if (gst_element_set_state (pipeline,
|
|
GST_STATE_PAUSED) == GST_STATE_CHANGE_FAILURE)
|
|
goto state_change_failed;
|
|
|
|
/* feed buffer in appsrc */
|
|
GST_DEBUG ("feeding buffer %p, size %" G_GSIZE_FORMAT ", caps %"
|
|
GST_PTR_FORMAT, buf, gst_buffer_get_size (buf), from_caps);
|
|
g_signal_emit_by_name (src, "push-buffer", buf, &ret);
|
|
|
|
/* now see what happens. We either got an error somewhere or the pipeline
|
|
* prerolled */
|
|
bus = gst_element_get_bus (pipeline);
|
|
msg = gst_bus_timed_pop_filtered (bus,
|
|
timeout, GST_MESSAGE_ERROR | GST_MESSAGE_ASYNC_DONE);
|
|
|
|
if (msg) {
|
|
switch (GST_MESSAGE_TYPE (msg)) {
|
|
case GST_MESSAGE_ASYNC_DONE:
|
|
{
|
|
/* we're prerolled, get the frame from appsink */
|
|
g_signal_emit_by_name (sink, "pull-preroll", &result);
|
|
|
|
if (result) {
|
|
GST_DEBUG ("conversion successful: result = %p", result);
|
|
} else {
|
|
GST_ERROR ("prerolled but no result frame?!");
|
|
}
|
|
break;
|
|
}
|
|
case GST_MESSAGE_ERROR:{
|
|
gchar *dbg = NULL;
|
|
|
|
gst_message_parse_error (msg, &err, &dbg);
|
|
if (err) {
|
|
GST_ERROR ("Could not convert video frame: %s", err->message);
|
|
GST_DEBUG ("%s [debug: %s]", err->message, GST_STR_NULL (dbg));
|
|
if (error)
|
|
*error = err;
|
|
else
|
|
g_error_free (err);
|
|
}
|
|
g_free (dbg);
|
|
break;
|
|
}
|
|
default:{
|
|
g_return_val_if_reached (NULL);
|
|
}
|
|
}
|
|
gst_message_unref (msg);
|
|
} else {
|
|
GST_ERROR ("Could not convert video frame: timeout during conversion");
|
|
if (error)
|
|
*error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
|
|
"Could not convert video frame: timeout during conversion");
|
|
}
|
|
|
|
gst_element_set_state (pipeline, GST_STATE_NULL);
|
|
gst_object_unref (bus);
|
|
gst_object_unref (pipeline);
|
|
gst_caps_unref (to_caps_copy);
|
|
|
|
return result;
|
|
|
|
/* ERRORS */
|
|
no_pipeline:
|
|
state_change_failed:
|
|
{
|
|
gst_caps_unref (to_caps_copy);
|
|
|
|
if (error)
|
|
*error = err;
|
|
else
|
|
g_error_free (err);
|
|
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
typedef struct
|
|
{
|
|
gint ref_count;
|
|
GMutex mutex;
|
|
GstElement *pipeline;
|
|
GstVideoConvertSampleCallback callback;
|
|
gpointer user_data;
|
|
GDestroyNotify destroy_notify;
|
|
GMainContext *context;
|
|
GstSample *sample;
|
|
GSource *timeout_source;
|
|
gboolean finished;
|
|
|
|
/* Results */
|
|
GstSample *converted_sample;
|
|
GError *error;
|
|
} GstVideoConvertSampleContext;
|
|
|
|
static GstVideoConvertSampleContext *
|
|
gst_video_convert_frame_context_ref (GstVideoConvertSampleContext * ctx)
|
|
{
|
|
g_atomic_int_inc (&ctx->ref_count);
|
|
|
|
return ctx;
|
|
}
|
|
|
|
static void
|
|
gst_video_convert_frame_context_unref (GstVideoConvertSampleContext * ctx)
|
|
{
|
|
if (!g_atomic_int_dec_and_test (&ctx->ref_count))
|
|
return;
|
|
|
|
g_mutex_clear (&ctx->mutex);
|
|
if (ctx->timeout_source)
|
|
g_source_destroy (ctx->timeout_source);
|
|
if (ctx->sample)
|
|
gst_sample_unref (ctx->sample);
|
|
if (ctx->converted_sample)
|
|
gst_sample_unref (ctx->converted_sample);
|
|
g_clear_error (&ctx->error);
|
|
g_main_context_unref (ctx->context);
|
|
|
|
/* The pipeline was already destroyed in finish() earlier and we
|
|
* must not end up here without finish() being called */
|
|
g_warn_if_fail (ctx->pipeline == NULL);
|
|
|
|
g_slice_free (GstVideoConvertSampleContext, ctx);
|
|
}
|
|
|
|
static gboolean
|
|
convert_frame_dispatch_callback (GstVideoConvertSampleContext * ctx)
|
|
{
|
|
GstSample *sample;
|
|
GError *error;
|
|
|
|
g_return_val_if_fail (ctx->converted_sample != NULL
|
|
|| ctx->error != NULL, FALSE);
|
|
|
|
sample = ctx->converted_sample;
|
|
error = ctx->error;
|
|
ctx->converted_sample = NULL;
|
|
ctx->error = NULL;
|
|
|
|
ctx->callback (sample, error, ctx->user_data);
|
|
|
|
if (ctx->destroy_notify)
|
|
ctx->destroy_notify (ctx->user_data);
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static void
|
|
convert_frame_stop_pipeline (GstElement * element, gpointer user_data)
|
|
{
|
|
gst_element_set_state (element, GST_STATE_NULL);
|
|
}
|
|
|
|
static void
|
|
convert_frame_finish (GstVideoConvertSampleContext * context,
|
|
GstSample * sample, GError * error)
|
|
{
|
|
GSource *source;
|
|
|
|
g_return_if_fail (!context->finished);
|
|
g_return_if_fail (sample != NULL || error != NULL);
|
|
|
|
context->finished = TRUE;
|
|
context->converted_sample = sample;
|
|
context->error = error;
|
|
|
|
if (context->timeout_source)
|
|
g_source_destroy (context->timeout_source);
|
|
context->timeout_source = NULL;
|
|
|
|
source = g_timeout_source_new (0);
|
|
g_source_set_callback (source,
|
|
(GSourceFunc) convert_frame_dispatch_callback,
|
|
gst_video_convert_frame_context_ref (context),
|
|
(GDestroyNotify) gst_video_convert_frame_context_unref);
|
|
g_source_attach (source, context->context);
|
|
g_source_unref (source);
|
|
|
|
/* Asynchronously stop the pipeline here: this will set its
|
|
* state to NULL and get rid of its last reference, which in turn
|
|
* will get rid of all remaining references to our context and free
|
|
* it too. We can't do this directly here as we might be called from
|
|
* a streaming thread.
|
|
*
|
|
* We don't use the main loop here because the user might shut down it
|
|
* immediately after getting the result of the conversion above.
|
|
*/
|
|
if (context->pipeline) {
|
|
gst_element_call_async (context->pipeline, convert_frame_stop_pipeline,
|
|
NULL, NULL);
|
|
gst_object_unref (context->pipeline);
|
|
context->pipeline = NULL;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
convert_frame_timeout_callback (GstVideoConvertSampleContext * context)
|
|
{
|
|
GError *error;
|
|
|
|
g_mutex_lock (&context->mutex);
|
|
|
|
if (context->finished)
|
|
goto done;
|
|
|
|
GST_ERROR ("Could not convert video frame: timeout");
|
|
|
|
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
|
|
"Could not convert video frame: timeout");
|
|
|
|
convert_frame_finish (context, NULL, error);
|
|
|
|
done:
|
|
g_mutex_unlock (&context->mutex);
|
|
return FALSE;
|
|
}
|
|
|
|
static gboolean
|
|
convert_frame_bus_callback (GstBus * bus, GstMessage * message,
|
|
GstVideoConvertSampleContext * context)
|
|
{
|
|
g_mutex_lock (&context->mutex);
|
|
|
|
if (context->finished)
|
|
goto done;
|
|
|
|
switch (GST_MESSAGE_TYPE (message)) {
|
|
case GST_MESSAGE_ERROR:{
|
|
GError *error;
|
|
gchar *dbg = NULL;
|
|
|
|
gst_message_parse_error (message, &error, &dbg);
|
|
|
|
GST_ERROR ("Could not convert video frame: %s", error->message);
|
|
GST_DEBUG ("%s [debug: %s]", error->message, GST_STR_NULL (dbg));
|
|
|
|
convert_frame_finish (context, NULL, error);
|
|
|
|
g_free (dbg);
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
|
|
done:
|
|
g_mutex_unlock (&context->mutex);
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static void
|
|
convert_frame_need_data_callback (GstElement * src, guint size,
|
|
GstVideoConvertSampleContext * context)
|
|
{
|
|
GstFlowReturn ret = GST_FLOW_ERROR;
|
|
GError *error;
|
|
GstBuffer *buffer;
|
|
|
|
g_mutex_lock (&context->mutex);
|
|
|
|
if (context->finished)
|
|
goto done;
|
|
|
|
buffer = gst_sample_get_buffer (context->sample);
|
|
g_signal_emit_by_name (src, "push-buffer", buffer, &ret);
|
|
gst_sample_unref (context->sample);
|
|
context->sample = NULL;
|
|
|
|
if (ret != GST_FLOW_OK) {
|
|
GST_ERROR ("Could not push video frame: %s", gst_flow_get_name (ret));
|
|
|
|
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
|
|
"Could not push video frame: %s", gst_flow_get_name (ret));
|
|
|
|
convert_frame_finish (context, NULL, error);
|
|
}
|
|
|
|
done:
|
|
g_mutex_unlock (&context->mutex);
|
|
|
|
g_signal_handlers_disconnect_by_func (src, convert_frame_need_data_callback,
|
|
context);
|
|
}
|
|
|
|
static GstFlowReturn
|
|
convert_frame_new_preroll_callback (GstElement * sink,
|
|
GstVideoConvertSampleContext * context)
|
|
{
|
|
GstSample *sample = NULL;
|
|
GError *error = NULL;
|
|
|
|
g_mutex_lock (&context->mutex);
|
|
|
|
if (context->finished)
|
|
goto done;
|
|
|
|
g_signal_emit_by_name (sink, "pull-preroll", &sample);
|
|
|
|
if (!sample) {
|
|
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_FAILED,
|
|
"Could not get converted video sample");
|
|
}
|
|
convert_frame_finish (context, sample, error);
|
|
|
|
done:
|
|
g_mutex_unlock (&context->mutex);
|
|
|
|
g_signal_handlers_disconnect_by_func (sink, convert_frame_need_data_callback,
|
|
context);
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
/**
|
|
* gst_video_convert_sample_async:
|
|
* @sample: a #GstSample
|
|
* @to_caps: the #GstCaps to convert to
|
|
* @timeout: the maximum amount of time allowed for the processing.
|
|
* @callback: %GstVideoConvertSampleCallback that will be called after conversion.
|
|
* @user_data: extra data that will be passed to the @callback
|
|
* @destroy_notify: %GDestroyNotify to be called after @user_data is not needed anymore
|
|
*
|
|
* Converts a raw video buffer into the specified output caps.
|
|
*
|
|
* The output caps can be any raw video formats or any image formats (jpeg, png, ...).
|
|
*
|
|
* The width, height and pixel-aspect-ratio can also be specified in the output caps.
|
|
*
|
|
* @callback will be called after conversion, when an error occured or if conversion didn't
|
|
* finish after @timeout. @callback will always be called from the thread default
|
|
* %GMainContext, see g_main_context_get_thread_default(). If GLib before 2.22 is used,
|
|
* this will always be the global default main context.
|
|
*
|
|
* @destroy_notify will be called after the callback was called and @user_data is not needed
|
|
* anymore.
|
|
*/
|
|
void
|
|
gst_video_convert_sample_async (GstSample * sample,
|
|
const GstCaps * to_caps, GstClockTime timeout,
|
|
GstVideoConvertSampleCallback callback, gpointer user_data,
|
|
GDestroyNotify destroy_notify)
|
|
{
|
|
GMainContext *context = NULL;
|
|
GError *error = NULL;
|
|
GstBus *bus;
|
|
GstBuffer *buf;
|
|
GstCaps *from_caps, *to_caps_copy = NULL;
|
|
GstElement *pipeline, *src, *sink;
|
|
guint i, n;
|
|
GSource *source;
|
|
GstVideoConvertSampleContext *ctx;
|
|
|
|
g_return_if_fail (sample != NULL);
|
|
buf = gst_sample_get_buffer (sample);
|
|
g_return_if_fail (buf != NULL);
|
|
|
|
g_return_if_fail (to_caps != NULL);
|
|
|
|
from_caps = gst_sample_get_caps (sample);
|
|
g_return_if_fail (from_caps != NULL);
|
|
g_return_if_fail (callback != NULL);
|
|
|
|
context = g_main_context_get_thread_default ();
|
|
|
|
if (!context)
|
|
context = g_main_context_default ();
|
|
|
|
to_caps_copy = gst_caps_new_empty ();
|
|
n = gst_caps_get_size (to_caps);
|
|
for (i = 0; i < n; i++) {
|
|
GstStructure *s = gst_caps_get_structure (to_caps, i);
|
|
|
|
s = gst_structure_copy (s);
|
|
gst_structure_remove_field (s, "framerate");
|
|
gst_caps_append_structure (to_caps_copy, s);
|
|
}
|
|
|
|
/* There's a reference cycle between the context and the pipeline, which is
|
|
* broken up once the finish() is called on the context. At latest when the
|
|
* timeout triggers the context will be freed */
|
|
ctx = g_slice_new0 (GstVideoConvertSampleContext);
|
|
ctx->ref_count = 1;
|
|
g_mutex_init (&ctx->mutex);
|
|
ctx->sample = gst_sample_ref (sample);
|
|
ctx->callback = callback;
|
|
ctx->user_data = user_data;
|
|
ctx->destroy_notify = destroy_notify;
|
|
ctx->context = g_main_context_ref (context);
|
|
ctx->finished = FALSE;
|
|
|
|
pipeline =
|
|
build_convert_frame_pipeline (&src, &sink, from_caps,
|
|
gst_buffer_get_video_crop_meta (buf), to_caps_copy, &error);
|
|
if (!pipeline)
|
|
goto no_pipeline;
|
|
ctx->pipeline = pipeline;
|
|
|
|
bus = gst_element_get_bus (pipeline);
|
|
|
|
if (timeout != GST_CLOCK_TIME_NONE) {
|
|
ctx->timeout_source = g_timeout_source_new (timeout / GST_MSECOND);
|
|
g_source_set_callback (ctx->timeout_source,
|
|
(GSourceFunc) convert_frame_timeout_callback,
|
|
gst_video_convert_frame_context_ref (ctx),
|
|
(GDestroyNotify) gst_video_convert_frame_context_unref);
|
|
g_source_attach (ctx->timeout_source, context);
|
|
}
|
|
|
|
g_signal_connect_data (src, "need-data",
|
|
G_CALLBACK (convert_frame_need_data_callback),
|
|
gst_video_convert_frame_context_ref (ctx),
|
|
(GClosureNotify) gst_video_convert_frame_context_unref, 0);
|
|
g_signal_connect_data (sink, "new-preroll",
|
|
G_CALLBACK (convert_frame_new_preroll_callback),
|
|
gst_video_convert_frame_context_ref (ctx),
|
|
(GClosureNotify) gst_video_convert_frame_context_unref, 0);
|
|
|
|
source = gst_bus_create_watch (bus);
|
|
g_source_set_callback (source, (GSourceFunc) convert_frame_bus_callback,
|
|
gst_video_convert_frame_context_ref (ctx),
|
|
(GDestroyNotify) gst_video_convert_frame_context_unref);
|
|
g_source_attach (source, context);
|
|
g_source_unref (source);
|
|
gst_object_unref (bus);
|
|
|
|
if (gst_element_set_state (pipeline,
|
|
GST_STATE_PAUSED) == GST_STATE_CHANGE_FAILURE)
|
|
goto state_change_failed;
|
|
|
|
gst_caps_unref (to_caps_copy);
|
|
|
|
gst_video_convert_frame_context_unref (ctx);
|
|
|
|
return;
|
|
/* ERRORS */
|
|
no_pipeline:
|
|
{
|
|
gst_caps_unref (to_caps_copy);
|
|
|
|
g_mutex_lock (&ctx->mutex);
|
|
convert_frame_finish (ctx, NULL, error);
|
|
g_mutex_unlock (&ctx->mutex);
|
|
gst_video_convert_frame_context_unref (ctx);
|
|
|
|
return;
|
|
}
|
|
state_change_failed:
|
|
{
|
|
gst_caps_unref (to_caps_copy);
|
|
|
|
error = g_error_new (GST_CORE_ERROR, GST_CORE_ERROR_STATE_CHANGE,
|
|
"failed to change state to PLAYING");
|
|
|
|
g_mutex_lock (&ctx->mutex);
|
|
convert_frame_finish (ctx, NULL, error);
|
|
g_mutex_unlock (&ctx->mutex);
|
|
gst_video_convert_frame_context_unref (ctx);
|
|
|
|
return;
|
|
}
|
|
}
|