/* GStreamer Editing Services * Copyright (C) 2009 Edward Hervey * 2009 Nokia Corporation * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:gespipeline * @title: GESPipeline * @short_description: Convenience GstPipeline for editing. * * #GESPipeline allows developers to view and render #GESTimeline * in a simple fashion. * Its usage is inspired by the 'playbin' element from gst-plugins-base. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include "ges-internal.h" #include "ges-pipeline.h" #include "ges-screenshot.h" #include "ges-audio-track.h" #include "ges-video-track.h" GST_DEBUG_CATEGORY_STATIC (ges_pipeline_debug); #undef GST_CAT_DEFAULT #define GST_CAT_DEFAULT ges_pipeline_debug #define DEFAULT_TIMELINE_MODE GES_PIPELINE_MODE_PREVIEW #define IN_RENDERING_MODE(timeline) ((timeline->priv->mode) & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER)) #define CHECK_THREAD(pipeline) g_assert(pipeline->priv->valid_thread == g_thread_self()) /* Structure corresponding to a timeline - sink link */ typedef struct { GESTrack *track; GstElement *tee; GstPad *srcpad; /* Timeline source pad */ GstPad *playsinkpad; GstPad *encodebinpad; guint query_position_id; } OutputChain; struct _GESPipelinePrivate { GESTimeline *timeline; GstElement *playsink; GstElement *encodebin; /* Note : urisink is only created when a URI has been provided */ GstElement *urisink; GESPipelineFlags mode; GMutex dyn_mutex; GList *chains; GList *not_rendered_tracks; GstEncodingProfile *profile; GThread *valid_thread; }; enum { PROP_0, PROP_AUDIO_SINK, PROP_VIDEO_SINK, PROP_TIMELINE, PROP_MODE, PROP_AUDIO_FILTER, PROP_VIDEO_FILTER, PROP_LAST }; static GParamSpec *properties[PROP_LAST]; static GstStateChangeReturn ges_pipeline_change_state (GstElement * element, GstStateChange transition); static OutputChain *get_output_chain_for_track (GESPipeline * self, GESTrack * track); static OutputChain *new_output_chain_for_track (GESPipeline * self, GESTrack * track); static void _link_track (GESPipeline * self, GESTrack * track); static void _unlink_track (GESPipeline * self, GESTrack * track); /**************************************************** * Video Overlay vmethods implementation * ****************************************************/ static void _overlay_expose (GstVideoOverlay * overlay) { GESPipeline *pipeline = GES_PIPELINE (overlay); gst_video_overlay_expose (GST_VIDEO_OVERLAY (pipeline->priv->playsink)); } static void _overlay_handle_events (GstVideoOverlay * overlay, gboolean handle_events) { GESPipeline *pipeline = GES_PIPELINE (overlay); gst_video_overlay_handle_events (GST_VIDEO_OVERLAY (pipeline->priv->playsink), handle_events); } static void _overlay_set_render_rectangle (GstVideoOverlay * overlay, gint x, gint y, gint width, gint height) { GESPipeline *pipeline = GES_PIPELINE (overlay); gst_video_overlay_set_render_rectangle (GST_VIDEO_OVERLAY (pipeline->priv-> playsink), x, y, width, height); } static void _overlay_set_window_handle (GstVideoOverlay * overlay, guintptr handle) { GESPipeline *pipeline = GES_PIPELINE (overlay); gst_video_overlay_set_window_handle (GST_VIDEO_OVERLAY (pipeline-> priv->playsink), handle); } static void video_overlay_init (gpointer g_iface, gpointer g_iface_data) { GstVideoOverlayInterface *iface = (GstVideoOverlayInterface *) g_iface; iface->expose = _overlay_expose; iface->handle_events = _overlay_handle_events; iface->set_render_rectangle = _overlay_set_render_rectangle; iface->set_window_handle = _overlay_set_window_handle; } G_DEFINE_TYPE_WITH_CODE (GESPipeline, ges_pipeline, GST_TYPE_PIPELINE, G_ADD_PRIVATE (GESPipeline) G_IMPLEMENT_INTERFACE (GST_TYPE_VIDEO_OVERLAY, video_overlay_init)); static void ges_pipeline_get_property (GObject * object, guint property_id, GValue * value, GParamSpec * pspec) { GESPipeline *self = GES_PIPELINE (object); switch (property_id) { case PROP_AUDIO_SINK: g_object_get_property (G_OBJECT (self->priv->playsink), "audio-sink", value); break; case PROP_VIDEO_SINK: g_object_get_property (G_OBJECT (self->priv->playsink), "video-sink", value); break; case PROP_TIMELINE: g_value_set_object (value, self->priv->timeline); break; case PROP_MODE: g_value_set_flags (value, self->priv->mode); break; case PROP_AUDIO_FILTER: g_object_get_property (G_OBJECT (self->priv->playsink), "audio-filter", value); break; case PROP_VIDEO_FILTER: g_object_get_property (G_OBJECT (self->priv->playsink), "video-filter", value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); } } static void ges_pipeline_set_property (GObject * object, guint property_id, const GValue * value, GParamSpec * pspec) { GESPipeline *self = GES_PIPELINE (object); switch (property_id) { case PROP_AUDIO_SINK: g_object_set_property (G_OBJECT (self->priv->playsink), "audio-sink", value); break; case PROP_VIDEO_SINK: g_object_set_property (G_OBJECT (self->priv->playsink), "video-sink", value); break; case PROP_TIMELINE: ges_pipeline_set_timeline (GES_PIPELINE (object), g_value_get_object (value)); break; case PROP_MODE: ges_pipeline_set_mode (GES_PIPELINE (object), g_value_get_flags (value)); break; case PROP_AUDIO_FILTER: g_object_set (self->priv->playsink, "audio-filter", GST_ELEMENT (g_value_get_object (value)), NULL); break; case PROP_VIDEO_FILTER: g_object_set (self->priv->playsink, "video-filter", GST_ELEMENT (g_value_get_object (value)), NULL); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); } } static void _timeline_track_added_cb (GESTimeline * timeline, GESTrack * track, GESPipeline * pipeline) { track_disable_last_gap (track, ! !(pipeline->priv->mode & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER))); _link_track (pipeline, track); } static void _timeline_track_removed_cb (GESTimeline * timeline, GESTrack * track, GESPipeline * pipeline) { _unlink_track (pipeline, track); } static void ges_pipeline_dispose (GObject * object) { GESPipeline *self = GES_PIPELINE (object); if (self->priv->playsink) { if (self->priv->mode & (GES_PIPELINE_MODE_PREVIEW)) gst_bin_remove (GST_BIN (object), self->priv->playsink); else gst_object_unref (self->priv->playsink); self->priv->playsink = NULL; } if (self->priv->encodebin) { if (self->priv->mode & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER)) gst_bin_remove (GST_BIN (object), self->priv->encodebin); else gst_object_unref (self->priv->encodebin); self->priv->encodebin = NULL; } if (self->priv->profile) { gst_encoding_profile_unref (self->priv->profile); self->priv->profile = NULL; } if (self->priv->timeline) { g_signal_handlers_disconnect_by_func (self->priv->timeline, _timeline_track_added_cb, self); g_signal_handlers_disconnect_by_func (self->priv->timeline, _timeline_track_removed_cb, self); gst_element_set_state (GST_ELEMENT (self->priv->timeline), GST_STATE_NULL); } G_OBJECT_CLASS (ges_pipeline_parent_class)->dispose (object); } static void ges_pipeline_class_init (GESPipelineClass * klass) { GObjectClass *object_class = G_OBJECT_CLASS (klass); GstElementClass *element_class = GST_ELEMENT_CLASS (klass); GST_DEBUG_CATEGORY_INIT (ges_pipeline_debug, "gespipeline", GST_DEBUG_FG_YELLOW, "ges pipeline"); object_class->dispose = ges_pipeline_dispose; object_class->get_property = ges_pipeline_get_property; object_class->set_property = ges_pipeline_set_property; /** * GESPipeline:audio-sink: * * Audio sink for the preview. */ properties[PROP_AUDIO_SINK] = g_param_spec_object ("audio-sink", "Audio Sink", "Audio sink for the preview.", GST_TYPE_ELEMENT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS); /** * GESPipeline:video-sink: * * Video sink for the preview. */ properties[PROP_VIDEO_SINK] = g_param_spec_object ("video-sink", "Video Sink", "Video sink for the preview.", GST_TYPE_ELEMENT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS); /** * GESPipeline:timeline: * * Timeline to use in this pipeline. See also * ges_pipeline_set_timeline() for more info. */ properties[PROP_TIMELINE] = g_param_spec_object ("timeline", "Timeline", "Timeline to use in this pipeline. See also " "ges_pipeline_set_timeline() for more info.", GES_TYPE_TIMELINE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS); /** * GESPipeline:mode: * * Pipeline mode. See ges_pipeline_set_mode() for more * info. */ properties[PROP_MODE] = g_param_spec_flags ("mode", "Mode", "Pipeline mode. See ges_pipeline_set_mode() for more info.", GES_TYPE_PIPELINE_FLAGS, DEFAULT_TIMELINE_MODE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS); /** * GESPipeline::audio-filter * * The audio filter(s) to apply during playback right before the audio sink * * Since: 1.6.0 */ properties[PROP_AUDIO_FILTER] = g_param_spec_object ("audio-filter", "Audio filter", "the audio filter(s) to apply, if possible", GST_TYPE_ELEMENT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS); /** * GESPipeline::video-filter * * The video filter(s) to apply during playback right before the video sink * * Since: 1.6.0 */ properties[PROP_VIDEO_FILTER] = g_param_spec_object ("video-filter", "Video filter", "the Video filter(s) to apply, if possible", GST_TYPE_ELEMENT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS); g_object_class_install_properties (object_class, PROP_LAST, properties); element_class->change_state = GST_DEBUG_FUNCPTR (ges_pipeline_change_state); /* TODO : Add state_change handlers * Don't change state if we don't have a timeline */ } static void ges_pipeline_init (GESPipeline * self) { GST_INFO_OBJECT (self, "Creating new 'playsink'"); self->priv = ges_pipeline_get_instance_private (self); self->priv->valid_thread = g_thread_self (); self->priv->playsink = gst_element_factory_make ("playsink", "internal-sinks"); self->priv->encodebin = gst_element_factory_make ("encodebin", "internal-encodebin"); g_object_set (self->priv->encodebin, "avoid-reencoding", TRUE, NULL); if (G_UNLIKELY (self->priv->playsink == NULL)) goto no_playsink; if (G_UNLIKELY (self->priv->encodebin == NULL)) goto no_encodebin; ges_pipeline_set_mode (self, DEFAULT_TIMELINE_MODE); return; no_playsink: { GST_ERROR_OBJECT (self, "Can't create playsink instance !"); return; } no_encodebin: { GST_ERROR_OBJECT (self, "Can't create encodebin instance !"); return; } } /** * ges_pipeline_new: * * Creates a new conveninence #GESPipeline. * * Returns: (transfer floating): the new #GESPipeline. */ GESPipeline * ges_pipeline_new (void) { return GES_PIPELINE (gst_element_factory_make ("gespipeline", NULL)); } #define TRACK_COMPATIBLE_PROFILE(tracktype, profile) \ ( (GST_IS_ENCODING_AUDIO_PROFILE (profile) && (tracktype) == GES_TRACK_TYPE_AUDIO) || \ (GST_IS_ENCODING_VIDEO_PROFILE (profile) && (tracktype) == GES_TRACK_TYPE_VIDEO)) static gboolean _track_is_compatible_with_profile (GESPipeline * self, GESTrack * track, GstEncodingProfile * prof) { if (TRACK_COMPATIBLE_PROFILE (track->type, prof)) { if (self->priv->mode == GES_PIPELINE_MODE_SMART_RENDER) { GstCaps *ocaps, *rcaps; GST_DEBUG ("Smart Render mode, setting input caps"); ocaps = gst_encoding_profile_get_input_caps (prof); ocaps = gst_caps_make_writable (ocaps); if (track->type == GES_TRACK_TYPE_AUDIO) rcaps = gst_caps_new_empty_simple ("audio/x-raw"); else rcaps = gst_caps_new_empty_simple ("video/x-raw"); gst_caps_append (ocaps, rcaps); ges_track_set_caps (track, ocaps); gst_caps_unref (ocaps); } else { GstCaps *caps = NULL; /* Raw preview or rendering mode */ if (track->type == GES_TRACK_TYPE_VIDEO) caps = gst_caps_new_empty_simple ("video/x-raw"); else if (track->type == GES_TRACK_TYPE_AUDIO) caps = gst_caps_new_empty_simple ("audio/x-raw"); if (caps) { ges_track_set_caps (track, caps); gst_caps_unref (caps); } } return TRUE; } return FALSE; } static gboolean ges_pipeline_update_caps (GESPipeline * self) { GList *ltrack, *tracks, *lstream; if (!self->priv->profile) return TRUE; GST_DEBUG ("Updating track caps"); tracks = ges_timeline_get_tracks (self->priv->timeline); /* Take each stream of the encoding profile and find a matching * track to set the caps on */ for (ltrack = tracks; ltrack; ltrack = ltrack->next) { GESTrack *track = (GESTrack *) ltrack->data; GList *allstreams; if (!GST_IS_ENCODING_CONTAINER_PROFILE (self->priv->profile)) { if (_track_is_compatible_with_profile (self, track, self->priv->profile)) { gst_object_unref (track); goto done; } else { gst_object_unref (track); continue; } } allstreams = (GList *) gst_encoding_container_profile_get_profiles ( (GstEncodingContainerProfile *) self->priv->profile); /* Find a matching stream setting */ for (lstream = allstreams; lstream; lstream = lstream->next) { GstEncodingProfile *prof = (GstEncodingProfile *) lstream->data; if (_track_is_compatible_with_profile (self, track, prof)) break; } gst_object_unref (track); } done: if (tracks) g_list_free (tracks); GST_DEBUG ("Done updating caps"); return TRUE; } static void _link_tracks (GESPipeline * pipeline) { GList *tmp; GST_DEBUG_OBJECT (pipeline, "Linking tracks"); if (!pipeline->priv->timeline) { GST_INFO_OBJECT (pipeline, "Not timeline set yet, doing nothing"); return; } for (tmp = pipeline->priv->timeline->tracks; tmp; tmp = tmp->next) _link_track (pipeline, tmp->data); if (IN_RENDERING_MODE (pipeline)) { GString *unlinked_issues = NULL; GstIterator *pads; gboolean done = FALSE; GValue paditem = { 0, }; pads = gst_element_iterate_sink_pads (pipeline->priv->encodebin); while (!done) { switch (gst_iterator_next (pads, &paditem)) { case GST_ITERATOR_OK: { GstPad *testpad = g_value_get_object (&paditem); if (!gst_pad_is_linked (testpad)) { GstCaps *sinkcaps = gst_pad_query_caps (testpad, NULL); gchar *caps_string = gst_caps_to_string (sinkcaps); gchar *path_string = gst_object_get_path_string (GST_OBJECT (testpad)); gst_caps_unref (sinkcaps); if (!unlinked_issues) unlinked_issues = g_string_new ("Following encodebin pads are not linked:\n"); g_string_append_printf (unlinked_issues, " - %s: %s", path_string, caps_string); g_free (caps_string); g_free (path_string); } g_value_reset (&paditem); } break; case GST_ITERATOR_DONE: case GST_ITERATOR_ERROR: done = TRUE; break; case GST_ITERATOR_RESYNC: gst_iterator_resync (pads); break; } } g_value_reset (&paditem); gst_iterator_free (pads); if (unlinked_issues) { GST_ELEMENT_ERROR (pipeline, STREAM, FAILED, (NULL), ("%s", unlinked_issues->str)); g_string_free (unlinked_issues, TRUE); } } } static void _unlink_tracks (GESPipeline * pipeline) { GList *tmp; GST_DEBUG_OBJECT (pipeline, "Disconnecting all tracks"); if (!pipeline->priv->timeline) { GST_INFO_OBJECT (pipeline, "Not timeline set yet, doing nothing"); return; } for (tmp = pipeline->priv->timeline->tracks; tmp; tmp = tmp->next) _unlink_track (pipeline, tmp->data); } static GstStateChangeReturn ges_pipeline_change_state (GstElement * element, GstStateChange transition) { GESPipeline *self; GstStateChangeReturn ret; self = GES_PIPELINE (element); switch (transition) { case GST_STATE_CHANGE_READY_TO_PAUSED: if (G_UNLIKELY (self->priv->timeline == NULL)) { GST_ERROR_OBJECT (element, "No GESTimeline set on the pipeline, cannot play !"); ret = GST_STATE_CHANGE_FAILURE; goto done; } if (IN_RENDERING_MODE (self)) { GST_DEBUG ("rendering => Updating pipeline caps"); /* Set caps on all tracks according to profile if present */ if (!ges_pipeline_update_caps (self)) { GST_ERROR_OBJECT (element, "Error setting the caps for rendering"); ret = GST_STATE_CHANGE_FAILURE; goto done; } } _link_tracks (self); break; case GST_STATE_CHANGE_PAUSED_TO_READY: { GList *tmp; for (tmp = self->priv->not_rendered_tracks; tmp; tmp = tmp->next) gst_element_set_locked_state (tmp->data, FALSE); } break; case GST_STATE_CHANGE_PAUSED_TO_PLAYING: { GstElement *queue = gst_bin_get_by_name (GST_BIN (self->priv->playsink), "vqueue"); if (queue) { GST_INFO_OBJECT (self, "Setting playsink video queue max-size-time to" " 2 seconds."); g_object_set (G_OBJECT (queue), "max-size-buffers", 0, "max-size-bytes", 0, "max-size-time", (gint64) 2 * GST_SECOND, NULL); gst_object_unref (queue); } break; } default: break; } ret = GST_ELEMENT_CLASS (ges_pipeline_parent_class)->change_state (element, transition); switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY: case GST_STATE_CHANGE_READY_TO_NULL: case GST_STATE_CHANGE_NULL_TO_NULL: _unlink_tracks (self); break; default: break; } done: return ret; } static OutputChain * new_output_chain_for_track (GESPipeline * self, GESTrack * track) { OutputChain *chain; chain = g_new0 (OutputChain, 1); chain->track = track; return chain; } static OutputChain * get_output_chain_for_track (GESPipeline * self, GESTrack * track) { GList *tmp; for (tmp = self->priv->chains; tmp; tmp = tmp->next) { OutputChain *chain = (OutputChain *) tmp->data; if (chain->track == track) return chain; } return NULL; } /* Fetches a compatible pad on the target element which isn't already * linked */ static GstPad * get_compatible_unlinked_pad (GstElement * element, GESTrack * track) { GstPad *res = NULL; GstIterator *pads; gboolean done = FALSE; const GstCaps *srccaps; GValue paditem = { 0, }; if (G_UNLIKELY (track == NULL)) goto no_track; GST_DEBUG_OBJECT (element, " track %" GST_PTR_FORMAT, track); pads = gst_element_iterate_sink_pads (element); srccaps = ges_track_get_caps (track); GST_DEBUG_OBJECT (track, "srccaps %" GST_PTR_FORMAT, srccaps); while (!done) { switch (gst_iterator_next (pads, &paditem)) { case GST_ITERATOR_OK: { GstPad *testpad = g_value_get_object (&paditem); if (!gst_pad_is_linked (testpad)) { GstCaps *sinkcaps = gst_pad_query_caps (testpad, NULL); GST_DEBUG_OBJECT (track, "sinkccaps %" GST_PTR_FORMAT, sinkcaps); if (gst_caps_can_intersect (srccaps, sinkcaps)) { res = gst_object_ref (testpad); done = TRUE; } gst_caps_unref (sinkcaps); } g_value_reset (&paditem); } break; case GST_ITERATOR_DONE: case GST_ITERATOR_ERROR: done = TRUE; break; case GST_ITERATOR_RESYNC: gst_iterator_resync (pads); break; } } g_value_reset (&paditem); gst_iterator_free (pads); return res; no_track: { GST_ERROR ("No track to check against"); return NULL; } } static GstClockTime _query_position_cb (GstElement * composition, GESPipeline * self) { gint64 position; if (gst_element_query_position (GST_ELEMENT (self), GST_FORMAT_TIME, &position)) return position; return GST_CLOCK_TIME_NONE; } static void _link_track (GESPipeline * self, GESTrack * track) { GstPad *pad; OutputChain *chain; GstPad *sinkpad; GstCaps *caps; GstPadLinkReturn lret; gboolean reconfigured = FALSE; pad = ges_timeline_get_pad_for_track (self->priv->timeline, track); caps = gst_pad_query_caps (pad, NULL); GST_DEBUG_OBJECT (self, "new pad %s:%s , caps:%" GST_PTR_FORMAT, GST_DEBUG_PAD_NAME (pad), caps); gst_caps_unref (caps); if (G_UNLIKELY (!pad)) { GST_ELEMENT_ERROR (self, STREAM, FAILED, (NULL), ("Trying to link %" GST_PTR_FORMAT " but no pad is exposed for it.", track)); return; } /* Don't connect track if it's not going to be used */ if (track->type == GES_TRACK_TYPE_VIDEO && !(self->priv->mode & GES_PIPELINE_MODE_PREVIEW_VIDEO) && !(self->priv->mode & GES_PIPELINE_MODE_RENDER) && !(self->priv->mode & GES_PIPELINE_MODE_SMART_RENDER)) { GST_DEBUG_OBJECT (self, "Video track... but we don't need it. Not linking"); } if (track->type == GES_TRACK_TYPE_AUDIO && !(self->priv->mode & GES_PIPELINE_MODE_PREVIEW_AUDIO) && !(self->priv->mode & GES_PIPELINE_MODE_RENDER) && !(self->priv->mode & GES_PIPELINE_MODE_SMART_RENDER)) { GST_DEBUG_OBJECT (self, "Audio track... but we don't need it. Not linking"); } /* Get an existing chain or create it */ if (!(chain = get_output_chain_for_track (self, track))) chain = new_output_chain_for_track (self, track); if (chain->tee) { GST_INFO_OBJECT (self, "Chain is already built (%" GST_PTR_FORMAT ")", chain->encodebinpad ? chain->encodebinpad : chain->playsinkpad); return; } chain->query_position_id = g_signal_connect (ges_track_get_composition (track), "query-position", G_CALLBACK (_query_position_cb), self); chain->srcpad = pad; gst_object_unref (pad); /* Adding tee */ chain->tee = gst_element_factory_make ("tee", NULL); gst_bin_add (GST_BIN_CAST (self), chain->tee); gst_element_sync_state_with_parent (chain->tee); /* Linking pad to tee */ sinkpad = gst_element_get_static_pad (chain->tee, "sink"); lret = gst_pad_link (pad, sinkpad); if (lret != GST_PAD_LINK_OK) { GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, (NULL), ("Could not link the tee (%s)", gst_pad_link_get_name (lret))); goto error; } gst_object_unref (sinkpad); /* Connect playsink */ if (self->priv->mode & GES_PIPELINE_MODE_PREVIEW) { const gchar *sinkpad_name; GstPad *tmppad; GST_DEBUG_OBJECT (self, "Connecting to playsink"); switch (track->type) { case GES_TRACK_TYPE_VIDEO: sinkpad_name = "video_sink"; break; case GES_TRACK_TYPE_AUDIO: sinkpad_name = "audio_sink"; break; case GES_TRACK_TYPE_TEXT: sinkpad_name = "text_sink"; break; default: GST_WARNING_OBJECT (self, "Can't handle tracks of type %d yet", track->type); goto error; } /* Request a sinkpad from playsink */ if (G_UNLIKELY (!(sinkpad = gst_element_get_request_pad (self->priv->playsink, sinkpad_name)))) { GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, (NULL), ("Could not get a pad from playsink for %s", sinkpad_name)); goto error; } tmppad = gst_element_get_request_pad (chain->tee, "src_%u"); lret = gst_pad_link_full (tmppad, sinkpad, GST_PAD_LINK_CHECK_NOTHING); if (G_UNLIKELY (lret != GST_PAD_LINK_OK)) { gst_object_unref (tmppad); GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, (NULL), ("Could not link %" GST_PTR_FORMAT " and %" GST_PTR_FORMAT " (%s)", tmppad, sinkpad, gst_pad_link_get_name (lret))); goto error; } gst_object_unref (tmppad); GST_DEBUG ("Reconfiguring playsink"); /* reconfigure playsink */ g_signal_emit_by_name (self->priv->playsink, "reconfigure", &reconfigured); GST_DEBUG ("'reconfigure' returned %d", reconfigured); /* We still hold a reference on the sinkpad */ chain->playsinkpad = sinkpad; } /* Connect to encodebin */ if (IN_RENDERING_MODE (self)) { GstPad *tmppad; GST_DEBUG_OBJECT (self, "Connecting to encodebin"); if (!chain->encodebinpad) { /* Check for unused static pads */ sinkpad = get_compatible_unlinked_pad (self->priv->encodebin, track); if (sinkpad == NULL) { GstCaps *caps = gst_pad_query_caps (pad, NULL); /* If no compatible static pad is available, request a pad */ g_signal_emit_by_name (self->priv->encodebin, "request-pad", caps, &sinkpad); if (G_UNLIKELY (sinkpad == NULL)) { gst_element_set_locked_state (GST_ELEMENT (track), TRUE); self->priv->not_rendered_tracks = g_list_append (self->priv->not_rendered_tracks, track); GST_INFO_OBJECT (self, "Couldn't get a pad from encodebin for: %" GST_PTR_FORMAT, caps); gst_caps_unref (caps); goto error; } gst_caps_unref (caps); } chain->encodebinpad = sinkpad; GST_INFO_OBJECT (track, "Linked to %" GST_PTR_FORMAT, sinkpad); } tmppad = gst_element_get_request_pad (chain->tee, "src_%u"); if (G_UNLIKELY (gst_pad_link_full (tmppad, chain->encodebinpad, GST_PAD_LINK_CHECK_NOTHING) != GST_PAD_LINK_OK)) { GST_ERROR_OBJECT (self, "Couldn't link track pad to encodebin"); goto error; } gst_object_unref (tmppad); } /* If chain wasn't already present, insert it in list */ if (!get_output_chain_for_track (self, track)) self->priv->chains = g_list_append (self->priv->chains, chain); GST_DEBUG ("done"); return; error: { if (chain->tee) { gst_element_set_state (chain->tee, GST_STATE_NULL); gst_bin_remove (GST_BIN_CAST (self), chain->tee); } if (sinkpad) gst_object_unref (sinkpad); g_free (chain); } } static void _unlink_track (GESPipeline * self, GESTrack * track) { OutputChain *chain; GST_DEBUG_OBJECT (self, "Unlinking removed %" GST_PTR_FORMAT, track); if (G_UNLIKELY (!(chain = get_output_chain_for_track (self, track)))) { GST_DEBUG_OBJECT (self, "Track wasn't used"); return; } /* Unlink encodebin */ if (chain->encodebinpad) { GstPad *peer = gst_pad_get_peer (chain->encodebinpad); gst_pad_unlink (peer, chain->encodebinpad); gst_object_unref (peer); gst_element_release_request_pad (self->priv->encodebin, chain->encodebinpad); gst_object_unref (chain->encodebinpad); } /* Unlink playsink */ if (chain->playsinkpad) { GstPad *peer = gst_pad_get_peer (chain->playsinkpad); gst_pad_unlink (peer, chain->playsinkpad); gst_object_unref (peer); gst_element_release_request_pad (self->priv->playsink, chain->playsinkpad); gst_object_unref (chain->playsinkpad); } gst_element_set_state (chain->tee, GST_STATE_NULL); gst_bin_remove (GST_BIN (self), chain->tee); if (chain->query_position_id) { g_signal_handler_disconnect (ges_track_get_composition (track), chain->query_position_id); chain->query_position_id = 0; } self->priv->chains = g_list_remove (self->priv->chains, chain); g_free (chain); GST_DEBUG ("done"); } /** * ges_pipeline_set_timeline: * @pipeline: a #GESPipeline * @timeline: (transfer full): the #GESTimeline to set on the @pipeline. * * Sets the timeline to use in this pipeline. * * The reference to the @timeline will be stolen by the @pipeline. * * Returns: %TRUE if the @timeline could be successfully set on the @pipeline, * else %FALSE. */ gboolean ges_pipeline_set_timeline (GESPipeline * pipeline, GESTimeline * timeline) { g_return_val_if_fail (GES_IS_PIPELINE (pipeline), FALSE); g_return_val_if_fail (GES_IS_TIMELINE (timeline), FALSE); g_return_val_if_fail (pipeline->priv->timeline == NULL, FALSE); CHECK_THREAD (pipeline); GST_DEBUG ("pipeline:%p, timeline:%p", timeline, pipeline); if (G_UNLIKELY (!gst_bin_add (GST_BIN_CAST (pipeline), GST_ELEMENT (timeline)))) { return FALSE; } pipeline->priv->timeline = timeline; g_signal_connect (timeline, "track-added", G_CALLBACK (_timeline_track_added_cb), pipeline); g_signal_connect (timeline, "track-removed", G_CALLBACK (_timeline_track_removed_cb), pipeline); /* FIXME Check if we should rollback if we can't sync state */ gst_element_sync_state_with_parent (GST_ELEMENT (timeline)); return TRUE; } /** * ges_pipeline_set_render_settings: * @pipeline: a #GESPipeline * @output_uri: the URI to which the timeline will be rendered * @profile: the #GstEncodingProfile to use to render the timeline. * * Specify where the pipeline shall be rendered and with what settings. * * A copy of @profile and @output_uri will be done internally, the caller can * safely free those values afterwards. * * This method must be called before setting the pipeline mode to * #GES_PIPELINE_MODE_RENDER * * Returns: %TRUE if the settings were aknowledged properly, else %FALSE */ gboolean ges_pipeline_set_render_settings (GESPipeline * pipeline, const gchar * output_uri, GstEncodingProfile * profile) { GError *err = NULL; GstEncodingProfile *set_profile; g_return_val_if_fail (GES_IS_PIPELINE (pipeline), FALSE); CHECK_THREAD (pipeline); /* FIXME Properly handle multi track, for now GESPipeline * only hanles single track per type, so we should just set the * presence to 1. */ if (GST_IS_ENCODING_CONTAINER_PROFILE (profile)) { const GList *tmpprofiles = gst_encoding_container_profile_get_profiles (GST_ENCODING_CONTAINER_PROFILE (profile)); GList *tmptrack, *tracks = ges_timeline_get_tracks (pipeline->priv->timeline); for (; tmpprofiles; tmpprofiles = tmpprofiles->next) { for (tmptrack = tracks; tmptrack; tmptrack = tmptrack->next) { if ((GST_IS_ENCODING_AUDIO_PROFILE (tmpprofiles->data) && GES_IS_AUDIO_TRACK (tmptrack->data)) || (GST_IS_ENCODING_VIDEO_PROFILE (tmpprofiles->data) && GES_IS_VIDEO_TRACK (tmptrack->data))) { GST_DEBUG_OBJECT (pipeline, "Setting presence to 1!"); gst_encoding_profile_set_presence (tmpprofiles->data, 1); gst_encoding_profile_set_allow_dynamic_output (tmpprofiles->data, FALSE); } } } g_list_free_full (tracks, gst_object_unref); } /* Clear previous URI sink if it existed */ /* FIXME : We should figure out if it was added to the pipeline, * and if so, remove it. */ if (pipeline->priv->urisink) { gst_object_unref (pipeline->priv->urisink); pipeline->priv->urisink = NULL; } pipeline->priv->urisink = gst_element_make_from_uri (GST_URI_SINK, output_uri, "urisink", &err); if (G_UNLIKELY (pipeline->priv->urisink == NULL)) { GST_ERROR_OBJECT (pipeline, "Couldn't not create sink for URI %s: '%s'", output_uri, ((err && err->message) ? err->message : "failed to create element")); g_clear_error (&err); return FALSE; } if (pipeline->priv->profile) gst_encoding_profile_unref (pipeline->priv->profile); g_object_set (pipeline->priv->encodebin, "avoid-reencoding", !(!(pipeline->priv->mode & GES_PIPELINE_MODE_SMART_RENDER)), NULL); g_object_set (pipeline->priv->encodebin, "profile", profile, NULL); g_object_get (pipeline->priv->encodebin, "profile", &set_profile, NULL); if (set_profile == NULL) { GST_ERROR_OBJECT (pipeline, "Profile %" GST_PTR_FORMAT " could no be set", profile); return FALSE; } /* We got a referencer when getting back the profile */ pipeline->priv->profile = profile; return TRUE; } /** * ges_pipeline_get_mode: * @pipeline: a #GESPipeline * * Returns: the #GESPipelineFlags currently in use. **/ GESPipelineFlags ges_pipeline_get_mode (GESPipeline * pipeline) { return pipeline->priv->mode; } /** * ges_pipeline_set_mode: * @pipeline: a #GESPipeline * @mode: the #GESPipelineFlags to use * * switches the @pipeline to the specified @mode. The default mode when * creating a #GESPipeline is #GES_PIPELINE_MODE_PREVIEW. * * Note: The @pipeline will be set to #GST_STATE_NULL during this call due to * the internal changes that happen. The caller will therefore have to * set the @pipeline to the requested state after calling this method. * * Returns: %TRUE if the mode was properly set, else %FALSE. **/ gboolean ges_pipeline_set_mode (GESPipeline * pipeline, GESPipelineFlags mode) { GList *tmp; g_return_val_if_fail (GES_IS_PIPELINE (pipeline), FALSE); CHECK_THREAD (pipeline); GST_DEBUG_OBJECT (pipeline, "current mode : %d, mode : %d", pipeline->priv->mode, mode); /* fast-path, nothing to change */ if (mode == pipeline->priv->mode) return TRUE; /* FIXME: It would be nice if we are only (de)activating preview * modes to not set the whole pipeline to NULL, but instead just * do the proper (un)linking to playsink. */ /* Switch pipeline to NULL since we're changing the configuration */ gst_element_set_state (GST_ELEMENT_CAST (pipeline), GST_STATE_NULL); if (pipeline->priv->timeline) { gboolean disabled = ! !(mode & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER)); for (tmp = pipeline->priv->timeline->tracks; tmp; tmp = tmp->next) track_disable_last_gap (GES_TRACK (tmp->data), disabled); } /* remove no-longer needed components */ if (pipeline->priv->mode & GES_PIPELINE_MODE_PREVIEW && !(mode & GES_PIPELINE_MODE_PREVIEW)) { /* Disable playsink */ GST_DEBUG ("Disabling playsink"); gst_object_ref (pipeline->priv->playsink); gst_bin_remove (GST_BIN_CAST (pipeline), pipeline->priv->playsink); } if ((pipeline->priv->mode & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER)) && !(mode & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER))) { GList *tmp; GstCaps *caps; for (tmp = pipeline->priv->timeline->tracks; tmp; tmp = tmp->next) { GESTrackType type = GES_TRACK (tmp->data)->type; if (type == GES_TRACK_TYPE_AUDIO) caps = gst_caps_new_empty_simple ("audio/x-raw"); else if (type == GES_TRACK_TYPE_VIDEO) caps = gst_caps_new_empty_simple ("video/x-raw"); else continue; ges_track_set_caps (GES_TRACK (tmp->data), caps); gst_caps_unref (caps); } /* Disable render bin */ GST_DEBUG ("Disabling rendering bin"); gst_object_ref (pipeline->priv->encodebin); gst_object_ref (pipeline->priv->urisink); gst_bin_remove_many (GST_BIN_CAST (pipeline), pipeline->priv->encodebin, pipeline->priv->urisink, NULL); } /* Add new elements */ if (!(pipeline->priv->mode & GES_PIPELINE_MODE_PREVIEW) && (mode & GES_PIPELINE_MODE_PREVIEW)) { /* Add playsink */ GST_DEBUG ("Adding playsink"); if (!gst_bin_add (GST_BIN_CAST (pipeline), pipeline->priv->playsink)) { GST_ERROR_OBJECT (pipeline, "Couldn't add playsink"); return FALSE; } } if (!(pipeline->priv->mode & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER)) && (mode & (GES_PIPELINE_MODE_RENDER | GES_PIPELINE_MODE_SMART_RENDER))) { /* Adding render bin */ GST_DEBUG ("Adding render bin"); if (G_UNLIKELY (pipeline->priv->urisink == NULL)) { GST_ERROR_OBJECT (pipeline, "Output URI not set !"); return FALSE; } if (!gst_bin_add (GST_BIN_CAST (pipeline), pipeline->priv->encodebin)) { GST_ERROR_OBJECT (pipeline, "Couldn't add encodebin"); return FALSE; } if (!gst_bin_add (GST_BIN_CAST (pipeline), pipeline->priv->urisink)) { GST_ERROR_OBJECT (pipeline, "Couldn't add URI sink"); return FALSE; } g_object_set (pipeline->priv->encodebin, "avoid-reencoding", !(!(mode & GES_PIPELINE_MODE_SMART_RENDER)), NULL); gst_element_link_pads_full (pipeline->priv->encodebin, "src", pipeline->priv->urisink, "sink", GST_PAD_LINK_CHECK_NOTHING); } /* FIXUPS */ /* FIXME * If we are rendering, set playsink to sync=False, * If we are NOT rendering, set playsink to sync=TRUE */ pipeline->priv->mode = mode; return TRUE; } /** * ges_pipeline_get_thumbnail: * @self: a #GESPipeline in %GST_STATE_PLAYING or %GST_STATE_PAUSED * @caps: (transfer none): caps specifying current format. Use %GST_CAPS_ANY * for native size. * * Returns a #GstSample with the currently playing image in the format specified by * caps. The caller should free the sample with #gst_sample_unref when finished. If ANY * caps are specified, the information will be returned in the whatever format * is currently used by the sink. This information can be retrieve from caps * associated with the buffer. * * Returns: (transfer full) (nullable): a #GstSample or %NULL */ GstSample * ges_pipeline_get_thumbnail (GESPipeline * self, GstCaps * caps) { GstElement *sink; g_return_val_if_fail (GES_IS_PIPELINE (self), FALSE); CHECK_THREAD (self); sink = self->priv->playsink; if (!sink) { GST_WARNING ("thumbnailing can only be done if we have a playsink"); return NULL; } return ges_play_sink_convert_frame (sink, caps); } /** * ges_pipeline_save_thumbnail: * @self: a #GESPipeline in %GST_STATE_PLAYING or %GST_STATE_PAUSED * @width: the requested width or -1 for native size * @height: the requested height or -1 for native size * @format: a string specifying the desired mime type (for example, * image/jpeg) * @location: the path to save the thumbnail * @error: (out) (allow-none) (transfer full): An error to be set in case * something wrong happens or %NULL * * Saves the current frame to the specified @location. * * Returns: %TRUE if the thumbnail was properly save, else %FALSE. */ gboolean ges_pipeline_save_thumbnail (GESPipeline * self, int width, int height, const gchar * format, const gchar * location, GError ** error) { GstMapInfo map_info; GstBuffer *b; GstSample *sample; GstCaps *caps; gboolean res = TRUE; g_return_val_if_fail (GES_IS_PIPELINE (self), FALSE); CHECK_THREAD (self); caps = gst_caps_from_string (format); if (width > 1) gst_caps_set_simple (caps, "width", G_TYPE_INT, width, NULL); if (height > 1) gst_caps_set_simple (caps, "height", G_TYPE_INT, height, NULL); if (!(sample = ges_pipeline_get_thumbnail (self, caps))) { gst_caps_unref (caps); return FALSE; } b = gst_sample_get_buffer (sample); if (gst_buffer_map (b, &map_info, GST_MAP_READ)) { if (!g_file_set_contents (location, (const char *) map_info.data, map_info.size, error)) { GST_WARNING ("Could not save thumbnail: %s", error ? (*error)->message : ""); res = FALSE; } } gst_caps_unref (caps); gst_buffer_unmap (b, &map_info); gst_sample_unref (sample); return res; } /** * ges_pipeline_get_thumbnail_rgb24: * @self: a #GESPipeline in %GST_STATE_PLAYING or %GST_STATE_PAUSED * @width: the requested width or -1 for native size * @height: the requested height or -1 for native size * * A convenience method for @ges_pipeline_get_thumbnail which * returns a buffer in 24-bit RGB, optionally scaled to the specified width * and height. If -1 is specified for either dimension, it will be left at * native size. You can retreive this information from the caps associated * with the buffer. * * The caller is responsible for unreffing the returned sample with * #gst_sample_unref. * * Returns: (transfer full) (nullable): a #GstSample or %NULL */ GstSample * ges_pipeline_get_thumbnail_rgb24 (GESPipeline * self, gint width, gint height) { GstSample *ret; GstCaps *caps; g_return_val_if_fail (GES_IS_PIPELINE (self), FALSE); CHECK_THREAD (self); caps = gst_caps_new_simple ("video/x-raw", "format", G_TYPE_STRING, "RGB", NULL); if (width != -1) gst_caps_set_simple (caps, "width", G_TYPE_INT, (gint) width, NULL); if (height != -1) gst_caps_set_simple (caps, "height", G_TYPE_INT, (gint) height, NULL); ret = ges_pipeline_get_thumbnail (self, caps); gst_caps_unref (caps); return ret; } /** * ges_pipeline_preview_get_video_sink: * @self: a #GESPipeline * * Obtains a pointer to playsink's video sink element that is used for * displaying video when the #GESPipeline is in %GES_PIPELINE_MODE_PREVIEW * * The caller is responsible for unreffing the returned element with * #gst_object_unref. * * Returns: (transfer full): a pointer to the playsink video sink #GstElement */ GstElement * ges_pipeline_preview_get_video_sink (GESPipeline * self) { GstElement *sink = NULL; g_return_val_if_fail (GES_IS_PIPELINE (self), FALSE); CHECK_THREAD (self); g_object_get (self->priv->playsink, "video-sink", &sink, NULL); return sink; }; /** * ges_pipeline_preview_set_video_sink: * @self: a #GESPipeline in %GST_STATE_NULL * @sink: (transfer none): a video sink #GstElement * * Sets playsink's video sink element that is used for displaying video when * the #GESPipeline is in %GES_PIPELINE_MODE_PREVIEW */ void ges_pipeline_preview_set_video_sink (GESPipeline * self, GstElement * sink) { g_return_if_fail (GES_IS_PIPELINE (self)); CHECK_THREAD (self); g_object_set (self->priv->playsink, "video-sink", sink, NULL); }; /** * ges_pipeline_preview_get_audio_sink: * @self: a #GESPipeline * * Obtains a pointer to playsink's audio sink element that is used for * displaying audio when the #GESPipeline is in %GES_PIPELINE_MODE_PREVIEW * * The caller is responsible for unreffing the returned element with * #gst_object_unref. * * Returns: (transfer full): a pointer to the playsink audio sink #GstElement */ GstElement * ges_pipeline_preview_get_audio_sink (GESPipeline * self) { GstElement *sink = NULL; g_return_val_if_fail (GES_IS_PIPELINE (self), FALSE); CHECK_THREAD (self); g_object_get (self->priv->playsink, "audio-sink", &sink, NULL); return sink; }; /** * ges_pipeline_preview_set_audio_sink: * @self: a #GESPipeline in %GST_STATE_NULL * @sink: (transfer none): a audio sink #GstElement * * Sets playsink's audio sink element that is used for displaying audio when * the #GESPipeline is in %GES_PIPELINE_MODE_PREVIEW */ void ges_pipeline_preview_set_audio_sink (GESPipeline * self, GstElement * sink) { g_return_if_fail (GES_IS_PIPELINE (self)); CHECK_THREAD (self); g_object_set (self->priv->playsink, "audio-sink", sink, NULL); };