gstreamer/subprojects/gst-plugins-bad/gst-libs/gst/play/gstplay.c

4810 lines
130 KiB
C
Raw Normal View History

/* GStreamer
*
* Copyright (C) 2014-2015 Sebastian Dröge <sebastian@centricular.com>
* Copyright (C) 2015 Brijesh Singh <brijesh.ksingh@gmail.com>
* Copyright (C) 2019-2020 Stephan Hesse <stephan@emliri.com>
* Copyright (C) 2020 Philippe Normand <philn@igalia.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
/**
* SECTION:gstplay
* @title: GstPlay
* @short_description: Player
* @symbols:
* - GstPlay
*
* The goal of the GstPlay library is to ease the integration of multimedia
* playback features in applications. Thus, if you need to build a media player
* from the ground-up, GstPlay provides the features you will most likely need.
*
* An example player is available in gst-examples/playback/player/gst-play/.
*
* Internally the GstPlay makes use of the `playbin` element. `playbin3` can be
* selected if the `GST_PLAY_USE_PLAYBIN3=1` environment variable has been set.
*
* **Important note**: If your application relies on the GstBus to get
* notifications from GstPlay, you need to add some explicit clean-up code in
* order to prevent the GstPlay object from leaking. See below for the details.
* If you use the GstPlaySignalAdapter, no special clean-up is required.
*
* When the GstPlaySignalAdapter is not used, the GstBus owned by GstPlay should
* be set to flushing state before any attempt to drop the last reference of the
* GstPlay object. An example in C:
*
* ```c
* ...
* GstBus *bus = gst_play_get_message_bus (player);
* gst_bus_set_flushing (bus, TRUE);
* gst_object_unref (bus);
* gst_object_unref (player);
* ```
*
* The messages managed by the player contain a reference to itself, and if the
* bus watch is just removed together with dropping the player then the bus will
* simply keep them around forever (and the bus never goes away because the
* player has a strong reference to it, so there's a reference cycle as long as
* there are messages). Setting the bus to flushing state forces it to get rid
* of its queued messages, thus breaking any possible reference cycle.
*
* Since: 1.20
*/
/* TODO:
*
* - Equalizer
* - Gapless playback
* - Frame stepping
* - Subtitle font, connection speed
* - Deinterlacing
* - Buffering control (-> progressive downloading)
* - Playlist/queue object
* - Custom video sink (e.g. embed in GL scene)
*
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include "gstplay.h"
#include "gstplay-video-renderer-private.h"
#include "gstplay-media-info-private.h"
#include "gstplay-message-private.h"
#include <gst/gst.h>
#include <gst/video/video.h>
#include <gst/video/colorbalance.h>
#include <gst/tag/tag.h>
#include <gst/pbutils/descriptions.h>
#include <string.h>
GST_DEBUG_CATEGORY_STATIC (gst_play_debug);
#define GST_CAT_DEFAULT gst_play_debug
#define DEFAULT_URI NULL
#define DEFAULT_POSITION GST_CLOCK_TIME_NONE
#define DEFAULT_DURATION GST_CLOCK_TIME_NONE
#define DEFAULT_VOLUME 1.0
#define DEFAULT_MUTE FALSE
#define DEFAULT_RATE 1.0
#define DEFAULT_POSITION_UPDATE_INTERVAL_MS 100
#define DEFAULT_AUDIO_VIDEO_OFFSET 0
#define DEFAULT_SUBTITLE_VIDEO_OFFSET 0
/**
* gst_play_error_quark:
* Since: 1.20
*/
GQuark
gst_play_error_quark (void)
{
return g_quark_from_static_string ("gst-play-error-quark");
}
static GQuark QUARK_CONFIG;
/* Keep ConfigQuarkId and _config_quark_strings ordered and synced */
typedef enum
{
CONFIG_QUARK_USER_AGENT = 0,
CONFIG_QUARK_POSITION_INTERVAL_UPDATE,
CONFIG_QUARK_ACCURATE_SEEK,
CONFIG_QUARK_MAX
} ConfigQuarkId;
static const gchar *_config_quark_strings[] = {
"user-agent",
"position-interval-update",
"accurate-seek",
};
static GQuark _config_quark_table[CONFIG_QUARK_MAX];
#define CONFIG_QUARK(q) _config_quark_table[CONFIG_QUARK_##q]
enum
{
PROP_0,
PROP_VIDEO_RENDERER,
PROP_URI,
PROP_SUBURI,
PROP_POSITION,
PROP_DURATION,
PROP_MEDIA_INFO,
PROP_CURRENT_AUDIO_TRACK,
PROP_CURRENT_VIDEO_TRACK,
PROP_CURRENT_SUBTITLE_TRACK,
PROP_VOLUME,
PROP_MUTE,
PROP_RATE,
PROP_PIPELINE,
PROP_VIDEO_MULTIVIEW_MODE,
PROP_VIDEO_MULTIVIEW_FLAGS,
PROP_AUDIO_VIDEO_OFFSET,
PROP_SUBTITLE_VIDEO_OFFSET,
PROP_LAST
};
enum
{
GST_PLAY_FLAG_VIDEO = (1 << 0),
GST_PLAY_FLAG_AUDIO = (1 << 1),
GST_PLAY_FLAG_SUBTITLE = (1 << 2),
GST_PLAY_FLAG_VIS = (1 << 3)
};
struct _GstPlay
{
GstObject parent;
GstPlayVideoRenderer *video_renderer;
gchar *uri;
gchar *redirect_uri;
gchar *suburi;
GThread *thread;
GMutex lock;
GCond cond;
GMainContext *context;
GMainLoop *loop;
GstBus *api_bus;
GstElement *playbin;
GstBus *bus;
GstState target_state, current_state;
gboolean is_live, is_eos;
GSource *tick_source, *ready_timeout_source;
GstClockTime cached_duration;
gint64 cached_position;
gdouble rate;
GstPlayState app_state;
gint buffering_percent;
GstTagList *global_tags;
GstPlayMediaInfo *media_info;
GstElement *current_vis_element;
GstStructure *config;
/* Protected by lock */
gboolean seek_pending; /* Only set from main context */
GstClockTime last_seek_time; /* Only set from main context */
GSource *seek_source;
GstClockTime seek_position;
/* For playbin3 */
gboolean use_playbin3;
GstStreamCollection *collection;
gchar *video_sid;
gchar *audio_sid;
gchar *subtitle_sid;
gulong stream_notify_id;
};
struct _GstPlayClass
{
GstObjectClass parent_class;
};
#define parent_class gst_play_parent_class
G_DEFINE_TYPE (GstPlay, gst_play, GST_TYPE_OBJECT);
static GParamSpec *param_specs[PROP_LAST] = { NULL, };
static void gst_play_dispose (GObject * object);
static void gst_play_finalize (GObject * object);
static void gst_play_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec);
static void gst_play_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec);
static void gst_play_constructed (GObject * object);
static gpointer gst_play_main (gpointer data);
static void gst_play_set_playbin_video_sink (GstPlay * self);
static void gst_play_seek_internal_locked (GstPlay * self);
static void gst_play_stop_internal (GstPlay * self, gboolean transient);
static gboolean gst_play_pause_internal (gpointer user_data);
static gboolean gst_play_play_internal (gpointer user_data);
static gboolean gst_play_seek_internal (gpointer user_data);
static void gst_play_set_rate_internal (GstPlay * self);
static void change_state (GstPlay * self, GstPlayState state);
static GstPlayMediaInfo *gst_play_media_info_create (GstPlay * self);
static void gst_play_streams_info_create (GstPlay * self,
GstPlayMediaInfo * media_info, const gchar * prop, GType type);
static void gst_play_stream_info_update (GstPlay * self, GstPlayStreamInfo * s);
static void gst_play_stream_info_update_tags_and_caps (GstPlay * self,
GstPlayStreamInfo * s);
static GstPlayStreamInfo *gst_play_stream_info_find (GstPlayMediaInfo *
media_info, GType type, gint stream_index);
static GstPlayStreamInfo *gst_play_stream_info_get_current (GstPlay *
self, const gchar * prop, GType type);
static void gst_play_video_info_update (GstPlay * self,
GstPlayStreamInfo * stream_info);
static void gst_play_audio_info_update (GstPlay * self,
GstPlayStreamInfo * stream_info);
static void gst_play_subtitle_info_update (GstPlay * self,
GstPlayStreamInfo * stream_info);
/* For playbin3 */
static void gst_play_streams_info_create_from_collection (GstPlay * self,
GstPlayMediaInfo * media_info, GstStreamCollection * collection);
static void gst_play_stream_info_update_from_stream (GstPlay * self,
GstPlayStreamInfo * s, GstStream * stream);
static GstPlayStreamInfo *gst_play_stream_info_find_from_stream_id
(GstPlayMediaInfo * media_info, const gchar * stream_id);
static GstPlayStreamInfo *gst_play_stream_info_get_current_from_stream_id
(GstPlay * self, const gchar * stream_id, GType type);
static void stream_notify_cb (GstStreamCollection * collection,
GstStream * stream, GParamSpec * pspec, GstPlay * self);
static void on_media_info_updated (GstPlay * self);
static void *get_title (GstTagList * tags);
static void *get_container_format (GstTagList * tags);
static void *get_from_tags (GstPlay * self, GstPlayMediaInfo * media_info,
void *(*func) (GstTagList *));
static void *get_cover_sample (GstTagList * tags);
static void remove_seek_source (GstPlay * self);
static gboolean query_position (GstPlay * self, GstClockTime * position);
static void
gst_play_init (GstPlay * self)
{
GST_TRACE_OBJECT (self, "Initializing");
self = gst_play_get_instance_private (self);
g_mutex_init (&self->lock);
g_cond_init (&self->cond);
self->context = g_main_context_new ();
self->loop = g_main_loop_new (self->context, FALSE);
self->api_bus = gst_bus_new ();
/* *INDENT-OFF* */
self->config = gst_structure_new_id (QUARK_CONFIG,
CONFIG_QUARK (POSITION_INTERVAL_UPDATE), G_TYPE_UINT, DEFAULT_POSITION_UPDATE_INTERVAL_MS,
CONFIG_QUARK (ACCURATE_SEEK), G_TYPE_BOOLEAN, FALSE,
NULL);
/* *INDENT-ON* */
self->seek_pending = FALSE;
self->seek_position = GST_CLOCK_TIME_NONE;
self->last_seek_time = GST_CLOCK_TIME_NONE;
self->cached_position = 0;
self->cached_duration = GST_CLOCK_TIME_NONE;
GST_TRACE_OBJECT (self, "Initialized");
}
/*
* Works same as gst_structure_set to set field/type/value triplets on message data
*/
static void
api_bus_post_message (GstPlay * self, GstPlayMessage message_type,
const gchar * firstfield, ...)
{
GstStructure *message_data = NULL;
GstMessage *msg = NULL;
va_list varargs;
GST_INFO ("Posting API-bus message-type: %s",
gst_play_message_get_name (message_type));
message_data = gst_structure_new (GST_PLAY_MESSAGE_DATA,
GST_PLAY_MESSAGE_DATA_TYPE, GST_TYPE_PLAY_MESSAGE, message_type, NULL);
va_start (varargs, firstfield);
gst_structure_set_valist (message_data, firstfield, varargs);
va_end (varargs);
msg = gst_message_new_custom (GST_MESSAGE_APPLICATION,
GST_OBJECT (self), message_data);
GST_DEBUG ("Created message with payload: [ %" GST_PTR_FORMAT " ]",
message_data);
gst_bus_post (self->api_bus, msg);
}
static void
config_quark_initialize (void)
{
gint i;
QUARK_CONFIG = g_quark_from_static_string ("play-config");
if (G_N_ELEMENTS (_config_quark_strings) != CONFIG_QUARK_MAX)
g_warning ("the quark table is not consistent! %d != %d",
(int) G_N_ELEMENTS (_config_quark_strings), CONFIG_QUARK_MAX);
for (i = 0; i < CONFIG_QUARK_MAX; i++) {
_config_quark_table[i] =
g_quark_from_static_string (_config_quark_strings[i]);
}
}
static void
gst_play_class_init (GstPlayClass * klass)
{
GObjectClass *gobject_class = (GObjectClass *) klass;
gobject_class->set_property = gst_play_set_property;
gobject_class->get_property = gst_play_get_property;
gobject_class->dispose = gst_play_dispose;
gobject_class->finalize = gst_play_finalize;
gobject_class->constructed = gst_play_constructed;
param_specs[PROP_VIDEO_RENDERER] =
g_param_spec_object ("video-renderer",
"Video Renderer", "Video renderer to use for rendering videos",
GST_TYPE_PLAY_VIDEO_RENDERER, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_URI] = g_param_spec_string ("uri", "URI", "Current URI",
DEFAULT_URI, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_SUBURI] = g_param_spec_string ("suburi", "Subtitle URI",
"Current Subtitle URI", NULL, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_POSITION] =
g_param_spec_uint64 ("position", "Position", "Current Position",
0, G_MAXUINT64, DEFAULT_POSITION,
G_PARAM_READABLE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_MEDIA_INFO] =
g_param_spec_object ("media-info", "Media Info",
"Current media information", GST_TYPE_PLAY_MEDIA_INFO,
G_PARAM_READABLE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_CURRENT_AUDIO_TRACK] =
g_param_spec_object ("current-audio-track", "Current Audio Track",
"Current audio track information", GST_TYPE_PLAY_AUDIO_INFO,
G_PARAM_READABLE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_CURRENT_VIDEO_TRACK] =
g_param_spec_object ("current-video-track", "Current Video Track",
"Current video track information", GST_TYPE_PLAY_VIDEO_INFO,
G_PARAM_READABLE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_CURRENT_SUBTITLE_TRACK] =
g_param_spec_object ("current-subtitle-track", "Current Subtitle Track",
"Current audio subtitle information", GST_TYPE_PLAY_SUBTITLE_INFO,
G_PARAM_READABLE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_DURATION] =
g_param_spec_uint64 ("duration", "Duration", "Duration",
0, G_MAXUINT64, DEFAULT_DURATION,
G_PARAM_READABLE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_VOLUME] =
g_param_spec_double ("volume", "Volume", "Volume",
0, 10.0, DEFAULT_VOLUME, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_MUTE] =
g_param_spec_boolean ("mute", "Mute", "Mute",
DEFAULT_MUTE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_PIPELINE] =
g_param_spec_object ("pipeline", "Pipeline",
"GStreamer pipeline that is used",
GST_TYPE_ELEMENT, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_RATE] =
g_param_spec_double ("rate", "rate", "Playback rate",
-64.0, 64.0, DEFAULT_RATE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_VIDEO_MULTIVIEW_MODE] =
g_param_spec_enum ("video-multiview-mode",
"Multiview Mode Override",
"Re-interpret a video stream as one of several frame-packed stereoscopic modes.",
GST_TYPE_VIDEO_MULTIVIEW_FRAME_PACKING,
GST_VIDEO_MULTIVIEW_FRAME_PACKING_NONE,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_VIDEO_MULTIVIEW_FLAGS] =
g_param_spec_flags ("video-multiview-flags",
"Multiview Flags Override",
"Override details of the multiview frame layout",
GST_TYPE_VIDEO_MULTIVIEW_FLAGS, GST_VIDEO_MULTIVIEW_FLAGS_NONE,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_AUDIO_VIDEO_OFFSET] =
g_param_spec_int64 ("audio-video-offset", "Audio Video Offset",
"The synchronisation offset between audio and video in nanoseconds",
G_MININT64, G_MAXINT64, 0, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
param_specs[PROP_SUBTITLE_VIDEO_OFFSET] =
g_param_spec_int64 ("subtitle-video-offset", "Text Video Offset",
"The synchronisation offset between text and video in nanoseconds",
G_MININT64, G_MAXINT64, 0, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS);
g_object_class_install_properties (gobject_class, PROP_LAST, param_specs);
config_quark_initialize ();
}
static void
gst_play_dispose (GObject * object)
{
GstPlay *self = GST_PLAY (object);
GST_TRACE_OBJECT (self, "Stopping main thread");
gst_bus_set_flushing (self->api_bus, TRUE);
if (self->loop) {
g_main_loop_quit (self->loop);
if (self->thread != g_thread_self ())
g_thread_join (self->thread);
else
g_thread_unref (self->thread);
self->thread = NULL;
g_main_loop_unref (self->loop);
self->loop = NULL;
g_main_context_unref (self->context);
self->context = NULL;
}
gst_clear_object (&self->api_bus);
G_OBJECT_CLASS (parent_class)->dispose (object);
}
static void
gst_play_finalize (GObject * object)
{
GstPlay *self = GST_PLAY (object);
GST_TRACE_OBJECT (self, "Finalizing");
g_free (self->uri);
g_free (self->redirect_uri);
g_free (self->suburi);
g_free (self->video_sid);
g_free (self->audio_sid);
g_free (self->subtitle_sid);
if (self->global_tags)
gst_tag_list_unref (self->global_tags);
if (self->video_renderer)
g_object_unref (self->video_renderer);
if (self->current_vis_element)
gst_object_unref (self->current_vis_element);
if (self->config)
gst_structure_free (self->config);
if (self->collection)
gst_object_unref (self->collection);
if (self->media_info)
g_object_unref (self->media_info);
g_mutex_clear (&self->lock);
g_cond_clear (&self->cond);
G_OBJECT_CLASS (parent_class)->finalize (object);
}
static void
gst_play_constructed (GObject * object)
{
GstPlay *self = GST_PLAY (object);
GST_TRACE_OBJECT (self, "Constructed");
g_mutex_lock (&self->lock);
self->thread = g_thread_new ("GstPlay", gst_play_main, self);
while (!self->loop || !g_main_loop_is_running (self->loop))
g_cond_wait (&self->cond, &self->lock);
gst_play_set_playbin_video_sink (self);
g_mutex_unlock (&self->lock);
G_OBJECT_CLASS (parent_class)->constructed (object);
}
static gboolean
gst_play_set_uri_internal (gpointer user_data)
{
GstPlay *self = user_data;
gst_play_stop_internal (self, FALSE);
g_mutex_lock (&self->lock);
GST_DEBUG_OBJECT (self, "Changing URI to '%s'", GST_STR_NULL (self->uri));
g_object_set (self->playbin, "uri", self->uri, NULL);
api_bus_post_message (self, GST_PLAY_MESSAGE_URI_LOADED,
GST_PLAY_MESSAGE_DATA_URI, G_TYPE_STRING, self->uri, NULL);
g_object_set (self->playbin, "suburi", NULL, NULL);
g_mutex_unlock (&self->lock);
return G_SOURCE_REMOVE;
}
static gboolean
gst_play_set_suburi_internal (gpointer user_data)
{
GstPlay *self = user_data;
GstClockTime position;
GstState target_state;
/* save the state and position */
target_state = self->target_state;
position = gst_play_get_position (self);
gst_play_stop_internal (self, TRUE);
g_mutex_lock (&self->lock);
GST_DEBUG_OBJECT (self, "Changing SUBURI to '%s'",
GST_STR_NULL (self->suburi));
g_object_set (self->playbin, "suburi", self->suburi, NULL);
g_mutex_unlock (&self->lock);
/* restore state and position */
if (position != GST_CLOCK_TIME_NONE)
gst_play_seek (self, position);
if (target_state == GST_STATE_PAUSED)
gst_play_pause_internal (self);
else if (target_state == GST_STATE_PLAYING)
gst_play_play_internal (self);
return G_SOURCE_REMOVE;
}
static void
gst_play_set_rate_internal (GstPlay * self)
{
self->seek_position = gst_play_get_position (self);
/* If there is no seek being dispatch to the main context currently do that,
* otherwise we just updated the rate so that it will be taken by
* the seek handler from the main context instead of the old one.
*/
if (!self->seek_source) {
/* If no seek is pending then create new seek source */
if (!self->seek_pending) {
self->seek_source = g_idle_source_new ();
g_source_set_callback (self->seek_source,
(GSourceFunc) gst_play_seek_internal, self, NULL);
g_source_attach (self->seek_source, self->context);
}
}
}
static void
gst_play_set_playbin_video_sink (GstPlay * self)
{
GstElement *video_sink = NULL;
if (self->video_renderer != NULL) {
video_sink =
gst_play_video_renderer_create_video_sink (self->video_renderer, self);
}
if (video_sink) {
gst_object_ref_sink (video_sink);
g_object_set (self->playbin, "video-sink", video_sink, NULL);
gst_object_unref (video_sink);
}
}
static void
gst_play_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec)
{
GstPlay *self = GST_PLAY (object);
switch (prop_id) {
case PROP_VIDEO_RENDERER:
g_mutex_lock (&self->lock);
g_clear_object (&self->video_renderer);
self->video_renderer = g_value_dup_object (value);
// When the video_renderer is a GstPlayerWrappedVideoRenderer it cannot be set
// at construction time because it requires a valid pipeline which is created
// only after GstPlay has been constructed. That is why the video renderer is
// set *after* GstPlay has been constructed.
if (self->thread) {
gst_play_set_playbin_video_sink (self);
}
g_mutex_unlock (&self->lock);
break;
case PROP_URI:{
g_mutex_lock (&self->lock);
g_free (self->uri);
g_free (self->redirect_uri);
self->redirect_uri = NULL;
g_free (self->suburi);
self->suburi = NULL;
self->uri = g_value_dup_string (value);
GST_DEBUG_OBJECT (self, "Set uri=%s", GST_STR_NULL (self->uri));
g_mutex_unlock (&self->lock);
g_main_context_invoke_full (self->context, G_PRIORITY_DEFAULT,
gst_play_set_uri_internal, self, NULL);
break;
}
case PROP_SUBURI:{
g_mutex_lock (&self->lock);
g_free (self->suburi);
self->suburi = g_value_dup_string (value);
GST_DEBUG_OBJECT (self, "Set suburi=%s", self->suburi);
g_mutex_unlock (&self->lock);
g_main_context_invoke_full (self->context, G_PRIORITY_DEFAULT,
gst_play_set_suburi_internal, self, NULL);
break;
}
case PROP_VOLUME:
GST_DEBUG_OBJECT (self, "Set volume=%lf", g_value_get_double (value));
g_object_set_property (G_OBJECT (self->playbin), "volume", value);
break;
case PROP_RATE:
g_mutex_lock (&self->lock);
self->rate = g_value_get_double (value);
GST_DEBUG_OBJECT (self, "Set rate=%lf", g_value_get_double (value));
gst_play_set_rate_internal (self);
g_mutex_unlock (&self->lock);
break;
case PROP_MUTE:
GST_DEBUG_OBJECT (self, "Set mute=%d", g_value_get_boolean (value));
g_object_set_property (G_OBJECT (self->playbin), "mute", value);
break;
case PROP_VIDEO_MULTIVIEW_MODE:
GST_DEBUG_OBJECT (self, "Set multiview mode=%u",
g_value_get_enum (value));
g_object_set_property (G_OBJECT (self->playbin), "video-multiview-mode",
value);
break;
case PROP_VIDEO_MULTIVIEW_FLAGS:
GST_DEBUG_OBJECT (self, "Set multiview flags=%x",
g_value_get_flags (value));
g_object_set_property (G_OBJECT (self->playbin), "video-multiview-flags",
value);
break;
case PROP_AUDIO_VIDEO_OFFSET:
g_object_set_property (G_OBJECT (self->playbin), "av-offset", value);
break;
case PROP_SUBTITLE_VIDEO_OFFSET:
g_object_set_property (G_OBJECT (self->playbin), "text-offset", value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static void
gst_play_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec)
{
GstPlay *self = GST_PLAY (object);
switch (prop_id) {
case PROP_URI:
g_mutex_lock (&self->lock);
g_value_set_string (value, self->uri);
g_mutex_unlock (&self->lock);
break;
case PROP_SUBURI:
g_mutex_lock (&self->lock);
g_value_set_string (value, self->suburi);
g_mutex_unlock (&self->lock);
GST_DEBUG_OBJECT (self, "Returning suburi=%s",
g_value_get_string (value));
break;
case PROP_POSITION:{
GstClockTime position = GST_CLOCK_TIME_NONE;
query_position (self, &position);
g_value_set_uint64 (value, position);
GST_TRACE_OBJECT (self, "Returning position=%" GST_TIME_FORMAT,
GST_TIME_ARGS (g_value_get_uint64 (value)));
break;
}
case PROP_DURATION:{
g_value_set_uint64 (value, self->cached_duration);
GST_TRACE_OBJECT (self, "Returning duration=%" GST_TIME_FORMAT,
GST_TIME_ARGS (g_value_get_uint64 (value)));
break;
}
case PROP_MEDIA_INFO:{
GstPlayMediaInfo *media_info = gst_play_get_media_info (self);
g_value_take_object (value, media_info);
break;
}
case PROP_CURRENT_AUDIO_TRACK:{
GstPlayAudioInfo *audio_info = gst_play_get_current_audio_track (self);
g_value_take_object (value, audio_info);
break;
}
case PROP_CURRENT_VIDEO_TRACK:{
GstPlayVideoInfo *video_info = gst_play_get_current_video_track (self);
g_value_take_object (value, video_info);
break;
}
case PROP_CURRENT_SUBTITLE_TRACK:{
GstPlaySubtitleInfo *subtitle_info =
gst_play_get_current_subtitle_track (self);
g_value_take_object (value, subtitle_info);
break;
}
case PROP_VOLUME:
g_object_get_property (G_OBJECT (self->playbin), "volume", value);
GST_TRACE_OBJECT (self, "Returning volume=%lf",
g_value_get_double (value));
break;
case PROP_RATE:
g_mutex_lock (&self->lock);
g_value_set_double (value, self->rate);
g_mutex_unlock (&self->lock);
break;
case PROP_MUTE:
g_object_get_property (G_OBJECT (self->playbin), "mute", value);
GST_TRACE_OBJECT (self, "Returning mute=%d", g_value_get_boolean (value));
break;
case PROP_PIPELINE:
g_value_set_object (value, self->playbin);
break;
case PROP_VIDEO_MULTIVIEW_MODE:{
g_object_get_property (G_OBJECT (self->playbin), "video-multiview-mode",
value);
GST_TRACE_OBJECT (self, "Return multiview mode=%d",
g_value_get_enum (value));
break;
}
case PROP_VIDEO_MULTIVIEW_FLAGS:{
g_object_get_property (G_OBJECT (self->playbin), "video-multiview-flags",
value);
GST_TRACE_OBJECT (self, "Return multiview flags=%x",
g_value_get_flags (value));
break;
}
case PROP_AUDIO_VIDEO_OFFSET:
g_object_get_property (G_OBJECT (self->playbin), "av-offset", value);
break;
case PROP_SUBTITLE_VIDEO_OFFSET:
g_object_get_property (G_OBJECT (self->playbin), "text-offset", value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static gboolean
main_loop_running_cb (gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GST_TRACE_OBJECT (self, "Main loop running now");
g_mutex_lock (&self->lock);
g_cond_signal (&self->cond);
g_mutex_unlock (&self->lock);
return G_SOURCE_REMOVE;
}
static void
change_state (GstPlay * self, GstPlayState state)
{
if (state == self->app_state)
return;
GST_DEBUG_OBJECT (self, "Changing app state from %s to %s",
gst_play_state_get_name (self->app_state),
gst_play_state_get_name (state));
self->app_state = state;
api_bus_post_message (self, GST_PLAY_MESSAGE_STATE_CHANGED,
GST_PLAY_MESSAGE_DATA_PLAY_STATE, GST_TYPE_PLAY_STATE,
self->app_state, NULL);
}
static gboolean
tick_cb (gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstClockTime position;
if (query_position (self, &position)) {
api_bus_post_message (self, GST_PLAY_MESSAGE_POSITION_UPDATED,
GST_PLAY_MESSAGE_DATA_POSITION, GST_TYPE_CLOCK_TIME, position, NULL);
}
return G_SOURCE_CONTINUE;
}
/*
* Returns true when position is queried and differed from cached position.
* Sets position to cached value, and to queried value if position can be queried
* and different.
*/
static gboolean
query_position (GstPlay * self, GstClockTime * position)
{
gint64 current_position;
*position = self->cached_position;
if (self->target_state >= GST_STATE_PAUSED
&& gst_element_query_position (self->playbin, GST_FORMAT_TIME,
&current_position)) {
GST_LOG_OBJECT (self, "Queried position %" GST_TIME_FORMAT,
GST_TIME_ARGS (current_position));
if (self->cached_position != current_position) {
self->cached_position = current_position;
*position = (GstClockTime) current_position;
return TRUE;
}
}
return FALSE;
}
static void
add_tick_source (GstPlay * self)
{
guint position_update_interval_ms;
if (self->tick_source)
return;
position_update_interval_ms =
gst_play_config_get_position_update_interval (self->config);
if (!position_update_interval_ms)
return;
self->tick_source = g_timeout_source_new (position_update_interval_ms);
g_source_set_callback (self->tick_source, (GSourceFunc) tick_cb, self, NULL);
g_source_attach (self->tick_source, self->context);
}
static void
remove_tick_source (GstPlay * self)
{
if (!self->tick_source)
return;
g_source_destroy (self->tick_source);
g_source_unref (self->tick_source);
self->tick_source = NULL;
}
static gboolean
ready_timeout_cb (gpointer user_data)
{
GstPlay *self = user_data;
if (self->target_state <= GST_STATE_READY) {
GST_DEBUG_OBJECT (self, "Setting pipeline to NULL state");
self->target_state = GST_STATE_NULL;
self->current_state = GST_STATE_NULL;
gst_element_set_state (self->playbin, GST_STATE_NULL);
}
return G_SOURCE_REMOVE;
}
static void
add_ready_timeout_source (GstPlay * self)
{
if (self->ready_timeout_source)
return;
self->ready_timeout_source = g_timeout_source_new_seconds (60);
g_source_set_callback (self->ready_timeout_source,
(GSourceFunc) ready_timeout_cb, self, NULL);
g_source_attach (self->ready_timeout_source, self->context);
}
static void
remove_ready_timeout_source (GstPlay * self)
{
if (!self->ready_timeout_source)
return;
g_source_destroy (self->ready_timeout_source);
g_source_unref (self->ready_timeout_source);
self->ready_timeout_source = NULL;
}
static void
on_error (GstPlay * self, GError * err, const GstStructure * details)
{
GST_ERROR_OBJECT (self, "Error: %s (%s, %d)", err->message,
g_quark_to_string (err->domain), err->code);
api_bus_post_message (self, GST_PLAY_MESSAGE_ERROR,
GST_PLAY_MESSAGE_DATA_ERROR, G_TYPE_ERROR, err,
GST_PLAY_MESSAGE_DATA_ERROR_DETAILS, GST_TYPE_STRUCTURE, details, NULL);
g_error_free (err);
remove_tick_source (self);
remove_ready_timeout_source (self);
self->target_state = GST_STATE_NULL;
self->current_state = GST_STATE_NULL;
self->is_live = FALSE;
self->is_eos = FALSE;
gst_element_set_state (self->playbin, GST_STATE_NULL);
change_state (self, GST_PLAY_STATE_STOPPED);
self->buffering_percent = 100;
g_mutex_lock (&self->lock);
if (self->media_info) {
g_object_unref (self->media_info);
self->media_info = NULL;
}
if (self->global_tags) {
gst_tag_list_unref (self->global_tags);
self->global_tags = NULL;
}
self->seek_pending = FALSE;
remove_seek_source (self);
self->seek_position = GST_CLOCK_TIME_NONE;
self->last_seek_time = GST_CLOCK_TIME_NONE;
g_mutex_unlock (&self->lock);
}
static void
dump_dot_file (GstPlay * self, const gchar * name)
{
gchar *full_name;
full_name = g_strdup_printf ("gst-play.%p.%s", self, name);
GST_DEBUG_BIN_TO_DOT_FILE_WITH_TS (GST_BIN (self->playbin),
GST_DEBUG_GRAPH_SHOW_ALL, full_name);
g_free (full_name);
}
static void
error_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GError *err, *play_err;
gchar *name, *debug, *message, *full_message;
const GstStructure *details = NULL;
dump_dot_file (self, "error");
gst_message_parse_error (msg, &err, &debug);
gst_message_parse_error_details (msg, &details);
name = gst_object_get_path_string (msg->src);
message = gst_error_get_message (err->domain, err->code);
if (debug)
full_message =
g_strdup_printf ("Error from element %s: %s\n%s\n%s", name, message,
err->message, debug);
else
full_message =
g_strdup_printf ("Error from element %s: %s\n%s", name, message,
err->message);
GST_ERROR_OBJECT (self, "ERROR: from element %s: %s", name, err->message);
if (debug != NULL)
GST_ERROR_OBJECT (self, "Additional debug info: %s", debug);
play_err =
g_error_new_literal (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED, full_message);
on_error (self, play_err, details);
g_clear_error (&err);
g_free (debug);
g_free (name);
g_free (full_message);
g_free (message);
}
static void
warning_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GError *err, *play_err;
gchar *name, *debug, *message, *full_message;
const GstStructure *details = NULL;
dump_dot_file (self, "warning");
gst_message_parse_warning (msg, &err, &debug);
gst_message_parse_warning_details (msg, &details);
name = gst_object_get_path_string (msg->src);
message = gst_error_get_message (err->domain, err->code);
if (debug)
full_message =
g_strdup_printf ("Warning from element %s: %s\n%s\n%s", name, message,
err->message, debug);
else
full_message =
g_strdup_printf ("Warning from element %s: %s\n%s", name, message,
err->message);
GST_WARNING_OBJECT (self, "WARNING: from element %s: %s", name, err->message);
if (debug != NULL)
GST_WARNING_OBJECT (self, "Additional debug info: %s", debug);
play_err =
g_error_new_literal (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED, full_message);
GST_WARNING_OBJECT (self, "Warning: %s (%s, %d)", err->message,
g_quark_to_string (err->domain), err->code);
if (details != NULL) {
api_bus_post_message (self, GST_PLAY_MESSAGE_WARNING,
GST_PLAY_MESSAGE_DATA_WARNING, G_TYPE_ERROR, play_err,
GST_PLAY_MESSAGE_DATA_WARNING_DETAILS, GST_TYPE_STRUCTURE, details,
NULL);
} else {
api_bus_post_message (self, GST_PLAY_MESSAGE_WARNING,
GST_PLAY_MESSAGE_DATA_WARNING, G_TYPE_ERROR, play_err, NULL);
}
g_clear_error (&play_err);
g_clear_error (&err);
g_free (debug);
g_free (name);
g_free (full_message);
g_free (message);
}
static void
eos_cb (G_GNUC_UNUSED GstBus * bus, G_GNUC_UNUSED GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GST_DEBUG_OBJECT (self, "End of stream");
tick_cb (self);
remove_tick_source (self);
api_bus_post_message (self, GST_PLAY_MESSAGE_END_OF_STREAM, NULL);
change_state (self, GST_PLAY_STATE_STOPPED);
self->buffering_percent = 100;
self->is_eos = TRUE;
}
static void
buffering_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
gint percent;
if (self->target_state < GST_STATE_PAUSED)
return;
if (self->is_live)
return;
gst_message_parse_buffering (msg, &percent);
GST_LOG_OBJECT (self, "Buffering %d%%", percent);
if (percent < 100 && self->target_state >= GST_STATE_PAUSED) {
GstStateChangeReturn state_ret;
GST_DEBUG_OBJECT (self, "Waiting for buffering to finish");
state_ret = gst_element_set_state (self->playbin, GST_STATE_PAUSED);
if (state_ret == GST_STATE_CHANGE_FAILURE) {
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to handle buffering"), NULL);
return;
}
change_state (self, GST_PLAY_STATE_BUFFERING);
}
if (self->buffering_percent != percent) {
self->buffering_percent = percent;
api_bus_post_message (self, GST_PLAY_MESSAGE_BUFFERING,
GST_PLAY_MESSAGE_DATA_BUFFERING_PERCENT, G_TYPE_UINT, percent, NULL);
}
g_mutex_lock (&self->lock);
if (percent == 100 && (self->seek_position != GST_CLOCK_TIME_NONE ||
self->seek_pending)) {
g_mutex_unlock (&self->lock);
GST_DEBUG_OBJECT (self, "Buffering finished - seek pending");
} else if (percent == 100 && self->target_state >= GST_STATE_PLAYING
&& self->current_state >= GST_STATE_PAUSED) {
GstStateChangeReturn state_ret;
g_mutex_unlock (&self->lock);
GST_DEBUG_OBJECT (self, "Buffering finished - going to PLAYING");
state_ret = gst_element_set_state (self->playbin, GST_STATE_PLAYING);
/* Application state change is happening when the state change happened */
if (state_ret == GST_STATE_CHANGE_FAILURE)
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to handle buffering"), NULL);
} else if (percent == 100 && self->target_state >= GST_STATE_PAUSED) {
g_mutex_unlock (&self->lock);
GST_DEBUG_OBJECT (self, "Buffering finished - staying PAUSED");
change_state (self, GST_PLAY_STATE_PAUSED);
} else {
g_mutex_unlock (&self->lock);
}
}
static void
clock_lost_cb (G_GNUC_UNUSED GstBus * bus, G_GNUC_UNUSED GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstStateChangeReturn state_ret;
GST_DEBUG_OBJECT (self, "Clock lost");
if (self->target_state >= GST_STATE_PLAYING) {
state_ret = gst_element_set_state (self->playbin, GST_STATE_PAUSED);
if (state_ret != GST_STATE_CHANGE_FAILURE)
state_ret = gst_element_set_state (self->playbin, GST_STATE_PLAYING);
if (state_ret == GST_STATE_CHANGE_FAILURE)
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to handle clock loss"), NULL);
}
}
static void
check_video_dimensions_changed (GstPlay * self)
{
GstElement *video_sink;
GstPad *video_sink_pad;
GstCaps *caps;
GstVideoInfo info;
guint width = 0, height = 0;
g_object_get (self->playbin, "video-sink", &video_sink, NULL);
if (!video_sink)
goto out;
video_sink_pad = gst_element_get_static_pad (video_sink, "sink");
if (!video_sink_pad) {
gst_object_unref (video_sink);
goto out;
}
caps = gst_pad_get_current_caps (video_sink_pad);
if (caps) {
if (gst_video_info_from_caps (&info, caps)) {
info.width = info.width * info.par_n / info.par_d;
GST_DEBUG_OBJECT (self, "Video dimensions changed: %dx%d", info.width,
info.height);
width = info.width;
height = info.height;
}
gst_caps_unref (caps);
}
gst_object_unref (video_sink_pad);
gst_object_unref (video_sink);
out:
api_bus_post_message (self, GST_PLAY_MESSAGE_VIDEO_DIMENSIONS_CHANGED,
GST_PLAY_MESSAGE_DATA_VIDEO_WIDTH, G_TYPE_UINT, width,
GST_PLAY_MESSAGE_DATA_VIDEO_HEIGHT, G_TYPE_UINT, height, NULL);
}
static void
notify_caps_cb (G_GNUC_UNUSED GObject * object,
G_GNUC_UNUSED GParamSpec * pspec, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
check_video_dimensions_changed (self);
}
static void
on_duration_changed (GstPlay * self, GstClockTime duration)
{
gboolean updated = FALSE;
if (self->cached_duration == duration)
return;
GST_DEBUG_OBJECT (self, "Duration changed %" GST_TIME_FORMAT,
GST_TIME_ARGS (duration));
g_mutex_lock (&self->lock);
self->cached_duration = duration;
if (self->media_info) {
self->media_info->duration = duration;
updated = TRUE;
}
g_mutex_unlock (&self->lock);
api_bus_post_message (self, GST_PLAY_MESSAGE_DURATION_CHANGED,
GST_PLAY_MESSAGE_DATA_DURATION, GST_TYPE_CLOCK_TIME,
gst_play_get_duration (self), NULL);
if (updated) {
on_media_info_updated (self);
}
}
static void
on_seek_done (GstPlay * self)
{
api_bus_post_message (self, GST_PLAY_MESSAGE_SEEK_DONE,
GST_PLAY_MESSAGE_DATA_POSITION, GST_TYPE_CLOCK_TIME,
gst_play_get_position (self), NULL);
}
static void
state_changed_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstState old_state, new_state, pending_state;
gst_message_parse_state_changed (msg, &old_state, &new_state, &pending_state);
if (GST_MESSAGE_SRC (msg) == GST_OBJECT (self->playbin)) {
gchar *transition_name;
GST_DEBUG_OBJECT (self, "Changed state old: %s new: %s pending: %s",
gst_element_state_get_name (old_state),
gst_element_state_get_name (new_state),
gst_element_state_get_name (pending_state));
transition_name = g_strdup_printf ("%s_%s",
gst_element_state_get_name (old_state),
gst_element_state_get_name (new_state));
dump_dot_file (self, transition_name);
g_free (transition_name);
self->current_state = new_state;
if (old_state == GST_STATE_READY && new_state == GST_STATE_PAUSED
&& pending_state == GST_STATE_VOID_PENDING) {
GstElement *video_sink;
GstPad *video_sink_pad;
gint64 duration = -1;
GST_DEBUG_OBJECT (self, "Initial PAUSED - pre-rolled");
g_mutex_lock (&self->lock);
if (self->media_info)
g_object_unref (self->media_info);
self->media_info = gst_play_media_info_create (self);
g_mutex_unlock (&self->lock);
on_media_info_updated (self);
g_object_get (self->playbin, "video-sink", &video_sink, NULL);
if (video_sink) {
video_sink_pad = gst_element_get_static_pad (video_sink, "sink");
if (video_sink_pad) {
g_signal_connect (video_sink_pad, "notify::caps",
(GCallback) notify_caps_cb, self);
gst_object_unref (video_sink_pad);
}
gst_object_unref (video_sink);
}
check_video_dimensions_changed (self);
if (gst_element_query_duration (self->playbin, GST_FORMAT_TIME,
&duration)) {
on_duration_changed (self, duration);
} else {
self->cached_duration = GST_CLOCK_TIME_NONE;
}
}
if (new_state == GST_STATE_PAUSED
&& pending_state == GST_STATE_VOID_PENDING) {
remove_tick_source (self);
g_mutex_lock (&self->lock);
if (self->seek_pending) {
self->seek_pending = FALSE;
if (!self->media_info->seekable) {
GST_DEBUG_OBJECT (self, "Media is not seekable");
remove_seek_source (self);
self->seek_position = GST_CLOCK_TIME_NONE;
self->last_seek_time = GST_CLOCK_TIME_NONE;
} else if (self->seek_source) {
GST_DEBUG_OBJECT (self, "Seek finished but new seek is pending");
gst_play_seek_internal_locked (self);
} else {
GST_DEBUG_OBJECT (self, "Seek finished");
on_seek_done (self);
}
}
if (self->seek_position != GST_CLOCK_TIME_NONE) {
GST_DEBUG_OBJECT (self, "Seeking now that we reached PAUSED state");
gst_play_seek_internal_locked (self);
g_mutex_unlock (&self->lock);
} else if (!self->seek_pending) {
g_mutex_unlock (&self->lock);
tick_cb (self);
if (self->target_state >= GST_STATE_PLAYING
&& self->buffering_percent == 100) {
GstStateChangeReturn state_ret;
state_ret = gst_element_set_state (self->playbin, GST_STATE_PLAYING);
if (state_ret == GST_STATE_CHANGE_FAILURE)
on_error (self, g_error_new (GST_PLAY_ERROR,
GST_PLAY_ERROR_FAILED, "Failed to play"), NULL);
} else if (self->buffering_percent == 100) {
change_state (self, GST_PLAY_STATE_PAUSED);
}
} else {
g_mutex_unlock (&self->lock);
}
} else if (new_state == GST_STATE_PLAYING
&& pending_state == GST_STATE_VOID_PENDING) {
/* api_bus_post_message (self, GST_PLAY_MESSAGE_POSITION_UPDATED, */
/* GST_PLAY_MESSAGE_DATA_POSITION, GST_TYPE_CLOCK_TIME, 0, NULL); */
/* If no seek is currently pending, add the tick source. This can happen
* if we seeked already but the state-change message was still queued up */
if (!self->seek_pending) {
add_tick_source (self);
change_state (self, GST_PLAY_STATE_PLAYING);
}
} else if (new_state == GST_STATE_READY && old_state > GST_STATE_READY) {
change_state (self, GST_PLAY_STATE_STOPPED);
} else {
/* Otherwise we neither reached PLAYING nor PAUSED, so must
* wait for something to happen... i.e. are BUFFERING now */
change_state (self, GST_PLAY_STATE_BUFFERING);
}
}
}
static void
duration_changed_cb (G_GNUC_UNUSED GstBus * bus, G_GNUC_UNUSED GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
gint64 duration = GST_CLOCK_TIME_NONE;
if (gst_element_query_duration (self->playbin, GST_FORMAT_TIME, &duration)) {
on_duration_changed (self, duration);
}
}
static void
latency_cb (G_GNUC_UNUSED GstBus * bus, G_GNUC_UNUSED GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GST_DEBUG_OBJECT (self, "Latency changed");
gst_bin_recalculate_latency (GST_BIN (self->playbin));
}
static void
request_state_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstState state;
GstStateChangeReturn state_ret;
gst_message_parse_request_state (msg, &state);
GST_DEBUG_OBJECT (self, "State %s requested",
gst_element_state_get_name (state));
self->target_state = state;
state_ret = gst_element_set_state (self->playbin, state);
if (state_ret == GST_STATE_CHANGE_FAILURE)
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to change to requested state %s",
gst_element_state_get_name (state)), NULL);
}
static void
media_info_update (GstPlay * self, GstPlayMediaInfo * info)
{
g_free (info->title);
info->title = get_from_tags (self, info, get_title);
g_free (info->container);
info->container = get_from_tags (self, info, get_container_format);
if (info->image_sample)
gst_sample_unref (info->image_sample);
info->image_sample = get_from_tags (self, info, get_cover_sample);
GST_DEBUG_OBJECT (self, "title: %s, container: %s "
"image_sample: %p", info->title, info->container, info->image_sample);
}
static void
tags_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstTagList *tags = NULL;
gst_message_parse_tag (msg, &tags);
GST_DEBUG_OBJECT (self, "received %s tags",
gst_tag_list_get_scope (tags) ==
GST_TAG_SCOPE_GLOBAL ? "global" : "stream");
if (gst_tag_list_get_scope (tags) == GST_TAG_SCOPE_GLOBAL) {
g_mutex_lock (&self->lock);
if (self->media_info) {
if (self->media_info->tags)
gst_tag_list_unref (self->media_info->tags);
self->media_info->tags = gst_tag_list_ref (tags);
media_info_update (self, self->media_info);
g_mutex_unlock (&self->lock);
on_media_info_updated (self);
} else {
if (self->global_tags)
gst_tag_list_unref (self->global_tags);
self->global_tags = gst_tag_list_ref (tags);
g_mutex_unlock (&self->lock);
}
}
gst_tag_list_unref (tags);
}
static void
element_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
const GstStructure *s;
s = gst_message_get_structure (msg);
if (gst_structure_has_name (s, "redirect")) {
const gchar *new_location;
new_location = gst_structure_get_string (s, "new-location");
if (!new_location) {
const GValue *locations_list, *location_val;
guint i, size;
locations_list = gst_structure_get_value (s, "locations");
size = gst_value_list_get_size (locations_list);
for (i = 0; i < size; ++i) {
const GstStructure *location_s;
location_val = gst_value_list_get_value (locations_list, i);
if (!GST_VALUE_HOLDS_STRUCTURE (location_val))
continue;
location_s = (const GstStructure *) g_value_get_boxed (location_val);
if (!gst_structure_has_name (location_s, "redirect"))
continue;
new_location = gst_structure_get_string (location_s, "new-location");
if (new_location)
break;
}
}
if (new_location) {
GstState target_state;
GST_DEBUG_OBJECT (self, "Redirect to '%s'", new_location);
/* Remember target state and restore after setting the URI */
target_state = self->target_state;
gst_play_stop_internal (self, TRUE);
g_mutex_lock (&self->lock);
g_free (self->redirect_uri);
self->redirect_uri = g_strdup (new_location);
g_object_set (self->playbin, "uri", self->redirect_uri, NULL);
g_mutex_unlock (&self->lock);
if (target_state == GST_STATE_PAUSED)
gst_play_pause_internal (self);
else if (target_state == GST_STATE_PLAYING)
gst_play_play_internal (self);
}
}
}
/* Must be called with lock */
static gboolean
update_stream_collection (GstPlay * self, GstStreamCollection * collection)
{
if (self->collection && self->collection == collection)
return FALSE;
if (self->collection && self->stream_notify_id)
g_signal_handler_disconnect (self->collection, self->stream_notify_id);
gst_object_replace ((GstObject **) & self->collection,
(GstObject *) collection);
if (self->media_info) {
gst_object_unref (self->media_info);
self->media_info = gst_play_media_info_create (self);
}
self->stream_notify_id =
g_signal_connect (self->collection, "stream-notify",
G_CALLBACK (stream_notify_cb), self);
return TRUE;
}
static void
stream_collection_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstStreamCollection *collection = NULL;
gboolean updated = FALSE;
gst_message_parse_stream_collection (msg, &collection);
if (!collection)
return;
g_mutex_lock (&self->lock);
updated = update_stream_collection (self, collection);
gst_object_unref (collection);
g_mutex_unlock (&self->lock);
if (self->media_info && updated)
on_media_info_updated (self);
}
static void
streams_selected_cb (G_GNUC_UNUSED GstBus * bus, GstMessage * msg,
gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstStreamCollection *collection = NULL;
gboolean updated = FALSE;
guint i, len;
gst_message_parse_streams_selected (msg, &collection);
if (!collection)
return;
g_mutex_lock (&self->lock);
updated = update_stream_collection (self, collection);
gst_object_unref (collection);
g_free (self->video_sid);
g_free (self->audio_sid);
g_free (self->subtitle_sid);
self->video_sid = NULL;
self->audio_sid = NULL;
self->subtitle_sid = NULL;
len = gst_message_streams_selected_get_size (msg);
for (i = 0; i < len; i++) {
GstStream *stream;
GstStreamType stream_type;
const gchar *stream_id;
gchar **current_sid;
stream = gst_message_streams_selected_get_stream (msg, i);
stream_type = gst_stream_get_stream_type (stream);
stream_id = gst_stream_get_stream_id (stream);
if (stream_type & GST_STREAM_TYPE_AUDIO)
current_sid = &self->audio_sid;
else if (stream_type & GST_STREAM_TYPE_VIDEO)
current_sid = &self->video_sid;
else if (stream_type & GST_STREAM_TYPE_TEXT)
current_sid = &self->subtitle_sid;
else {
GST_WARNING_OBJECT (self,
"Unknown stream-id %s with type 0x%x", stream_id, stream_type);
continue;
}
if (G_UNLIKELY (*current_sid)) {
GST_FIXME_OBJECT (self,
"Multiple streams are selected for type %s, choose the first one",
gst_stream_type_get_name (stream_type));
continue;
}
*current_sid = g_strdup (stream_id);
}
g_mutex_unlock (&self->lock);
if (self->media_info && updated)
on_media_info_updated (self);
}
static void
play_set_flag (GstPlay * self, gint pos)
{
gint flags;
g_object_get (self->playbin, "flags", &flags, NULL);
flags |= pos;
g_object_set (self->playbin, "flags", flags, NULL);
GST_DEBUG_OBJECT (self, "setting flags=%#x", flags);
}
static void
play_clear_flag (GstPlay * self, gint pos)
{
gint flags;
g_object_get (self->playbin, "flags", &flags, NULL);
flags &= ~pos;
g_object_set (self->playbin, "flags", flags, NULL);
GST_DEBUG_OBJECT (self, "setting flags=%#x", flags);
}
/*
* on_media_info_updated:
*
* create a new copy of self->media_info object and post it to the user
* application.
*/
static void
on_media_info_updated (GstPlay * self)
{
GstPlayMediaInfo *media_info_copy;
g_mutex_lock (&self->lock);
media_info_copy = gst_play_media_info_copy (self->media_info);
g_mutex_unlock (&self->lock);
api_bus_post_message (self, GST_PLAY_MESSAGE_MEDIA_INFO_UPDATED,
GST_PLAY_MESSAGE_DATA_MEDIA_INFO, GST_TYPE_PLAY_MEDIA_INFO,
media_info_copy, NULL);
g_object_unref (media_info_copy);
}
static GstCaps *
get_caps (GstPlay * self, gint stream_index, GType type)
{
GstPad *pad = NULL;
GstCaps *caps = NULL;
if (type == GST_TYPE_PLAY_VIDEO_INFO)
g_signal_emit_by_name (G_OBJECT (self->playbin),
"get-video-pad", stream_index, &pad);
else if (type == GST_TYPE_PLAY_AUDIO_INFO)
g_signal_emit_by_name (G_OBJECT (self->playbin),
"get-audio-pad", stream_index, &pad);
else
g_signal_emit_by_name (G_OBJECT (self->playbin),
"get-text-pad", stream_index, &pad);
if (pad) {
caps = gst_pad_get_current_caps (pad);
gst_object_unref (pad);
}
return caps;
}
static void
gst_play_subtitle_info_update (GstPlay * self, GstPlayStreamInfo * stream_info)
{
GstPlaySubtitleInfo *info = (GstPlaySubtitleInfo *) stream_info;
if (stream_info->tags) {
/* free the old language info */
g_free (info->language);
info->language = NULL;
/* First try to get the language full name from tag, if name is not
* available then try language code. If we find the language code
* then use gstreamer api to translate code to full name.
*/
gst_tag_list_get_string (stream_info->tags, GST_TAG_LANGUAGE_NAME,
&info->language);
if (!info->language) {
gchar *lang_code = NULL;
gst_tag_list_get_string (stream_info->tags, GST_TAG_LANGUAGE_CODE,
&lang_code);
if (lang_code) {
info->language = g_strdup (gst_tag_get_language_name (lang_code));
g_free (lang_code);
}
}
/* If we are still failed to find language name then check if external
* subtitle is loaded and compare the stream index between current sub
* stream index with our stream index and if matches then declare it as
* external subtitle and use the filename.
*/
if (!info->language) {
gint text_index = -1;
gchar *suburi = NULL;
g_object_get (G_OBJECT (self->playbin), "current-suburi", &suburi, NULL);
if (suburi) {
if (self->use_playbin3) {
if (self->subtitle_sid &&
g_str_equal (self->subtitle_sid, stream_info->stream_id)) {
info->language = g_path_get_basename (suburi);
}
} else {
g_object_get (G_OBJECT (self->playbin), "current-text", &text_index,
NULL);
if (text_index == gst_play_stream_info_get_index (stream_info))
info->language = g_path_get_basename (suburi);
}
g_free (suburi);
}
}
} else {
g_free (info->language);
info->language = NULL;
}
GST_DEBUG_OBJECT (self, "language=%s", info->language);
}
static void
gst_play_video_info_update (GstPlay * self, GstPlayStreamInfo * stream_info)
{
GstPlayVideoInfo *info = (GstPlayVideoInfo *) stream_info;
if (stream_info->caps) {
GstStructure *s;
s = gst_caps_get_structure (stream_info->caps, 0);
if (s) {
gint width, height;
gint fps_n, fps_d;
gint par_n, par_d;
if (gst_structure_get_int (s, "width", &width))
info->width = width;
else
info->width = -1;
if (gst_structure_get_int (s, "height", &height))
info->height = height;
else
info->height = -1;
if (gst_structure_get_fraction (s, "framerate", &fps_n, &fps_d)) {
info->framerate_num = fps_n;
info->framerate_denom = fps_d;
} else {
info->framerate_num = 0;
info->framerate_denom = 1;
}
if (gst_structure_get_fraction (s, "pixel-aspect-ratio", &par_n, &par_d)) {
info->par_num = par_n;
info->par_denom = par_d;
} else {
info->par_num = 1;
info->par_denom = 1;
}
}
} else {
info->width = info->height = -1;
info->par_num = info->par_denom = 1;
info->framerate_num = 0;
info->framerate_denom = 1;
}
if (stream_info->tags) {
guint bitrate, max_bitrate;
if (gst_tag_list_get_uint (stream_info->tags, GST_TAG_BITRATE, &bitrate))
info->bitrate = bitrate;
else
info->bitrate = -1;
if (gst_tag_list_get_uint (stream_info->tags, GST_TAG_MAXIMUM_BITRATE,
&max_bitrate) || gst_tag_list_get_uint (stream_info->tags,
GST_TAG_NOMINAL_BITRATE, &max_bitrate))
info->max_bitrate = max_bitrate;
else
info->max_bitrate = -1;
} else {
info->bitrate = info->max_bitrate = -1;
}
GST_DEBUG_OBJECT (self, "width=%d height=%d fps=%.2f par=%d:%d "
"bitrate=%d max_bitrate=%d", info->width, info->height,
(gdouble) info->framerate_num / info->framerate_denom,
info->par_num, info->par_denom, info->bitrate, info->max_bitrate);
}
static void
gst_play_audio_info_update (GstPlay * self, GstPlayStreamInfo * stream_info)
{
GstPlayAudioInfo *info = (GstPlayAudioInfo *) stream_info;
if (stream_info->caps) {
GstStructure *s;
s = gst_caps_get_structure (stream_info->caps, 0);
if (s) {
gint rate, channels;
if (gst_structure_get_int (s, "rate", &rate))
info->sample_rate = rate;
else
info->sample_rate = -1;
if (gst_structure_get_int (s, "channels", &channels))
info->channels = channels;
else
info->channels = 0;
}
} else {
info->sample_rate = -1;
info->channels = 0;
}
if (stream_info->tags) {
guint bitrate, max_bitrate;
if (gst_tag_list_get_uint (stream_info->tags, GST_TAG_BITRATE, &bitrate))
info->bitrate = bitrate;
else
info->bitrate = -1;
if (gst_tag_list_get_uint (stream_info->tags, GST_TAG_MAXIMUM_BITRATE,
&max_bitrate) || gst_tag_list_get_uint (stream_info->tags,
GST_TAG_NOMINAL_BITRATE, &max_bitrate))
info->max_bitrate = max_bitrate;
else
info->max_bitrate = -1;
/* if we have old language the free it */
g_free (info->language);
info->language = NULL;
/* First try to get the language full name from tag, if name is not
* available then try language code. If we find the language code
* then use gstreamer api to translate code to full name.
*/
gst_tag_list_get_string (stream_info->tags, GST_TAG_LANGUAGE_NAME,
&info->language);
if (!info->language) {
gchar *lang_code = NULL;
gst_tag_list_get_string (stream_info->tags, GST_TAG_LANGUAGE_CODE,
&lang_code);
if (lang_code) {
info->language = g_strdup (gst_tag_get_language_name (lang_code));
g_free (lang_code);
}
}
} else {
g_free (info->language);
info->language = NULL;
info->max_bitrate = info->bitrate = -1;
}
GST_DEBUG_OBJECT (self, "language=%s rate=%d channels=%d bitrate=%d "
"max_bitrate=%d", info->language, info->sample_rate, info->channels,
info->bitrate, info->max_bitrate);
}
static GstPlayStreamInfo *
gst_play_stream_info_find (GstPlayMediaInfo * media_info,
GType type, gint stream_index)
{
GList *list, *l;
GstPlayStreamInfo *info = NULL;
if (!media_info)
return NULL;
list = gst_play_media_info_get_stream_list (media_info);
for (l = list; l != NULL; l = l->next) {
info = (GstPlayStreamInfo *) l->data;
if ((G_OBJECT_TYPE (info) == type) && (info->stream_index == stream_index)) {
return info;
}
}
return NULL;
}
static GstPlayStreamInfo *
gst_play_stream_info_find_from_stream_id (GstPlayMediaInfo * media_info,
const gchar * stream_id)
{
GList *list, *l;
GstPlayStreamInfo *info = NULL;
if (!media_info)
return NULL;
list = gst_play_media_info_get_stream_list (media_info);
for (l = list; l != NULL; l = l->next) {
info = (GstPlayStreamInfo *) l->data;
if (g_str_equal (info->stream_id, stream_id)) {
return info;
}
}
return NULL;
}
static gboolean
is_track_enabled (GstPlay * self, gint pos)
{
gint flags;
g_object_get (G_OBJECT (self->playbin), "flags", &flags, NULL);
if ((flags & pos))
return TRUE;
return FALSE;
}
static GstPlayStreamInfo *
gst_play_stream_info_get_current (GstPlay * self, const gchar * prop,
GType type)
{
gint current;
GstPlayStreamInfo *info;
if (!self->media_info)
return NULL;
g_object_get (G_OBJECT (self->playbin), prop, &current, NULL);
g_mutex_lock (&self->lock);
info = gst_play_stream_info_find (self->media_info, type, current);
if (info)
info = gst_play_stream_info_copy (info);
g_mutex_unlock (&self->lock);
return info;
}
static GstPlayStreamInfo *
gst_play_stream_info_get_current_from_stream_id (GstPlay * self,
const gchar * stream_id, GType type)
{
GstPlayStreamInfo *info;
if (!self->media_info || !stream_id)
return NULL;
g_mutex_lock (&self->lock);
info = gst_play_stream_info_find_from_stream_id (self->media_info, stream_id);
if (info && G_OBJECT_TYPE (info) == type)
info = gst_play_stream_info_copy (info);
else
info = NULL;
g_mutex_unlock (&self->lock);
return info;
}
static void
stream_notify_cb (GstStreamCollection * collection, GstStream * stream,
GParamSpec * pspec, GstPlay * self)
{
GstPlayStreamInfo *info;
const gchar *stream_id;
gboolean emit_signal = FALSE;
if (!self->media_info)
return;
if (G_PARAM_SPEC_VALUE_TYPE (pspec) != GST_TYPE_CAPS &&
G_PARAM_SPEC_VALUE_TYPE (pspec) != GST_TYPE_TAG_LIST)
return;
stream_id = gst_stream_get_stream_id (stream);
g_mutex_lock (&self->lock);
info = gst_play_stream_info_find_from_stream_id (self->media_info, stream_id);
if (info) {
gst_play_stream_info_update_from_stream (self, info, stream);
emit_signal = TRUE;
}
g_mutex_unlock (&self->lock);
if (emit_signal)
on_media_info_updated (self);
}
static void
gst_play_stream_info_update (GstPlay * self, GstPlayStreamInfo * s)
{
if (GST_IS_PLAY_VIDEO_INFO (s))
gst_play_video_info_update (self, s);
else if (GST_IS_PLAY_AUDIO_INFO (s))
gst_play_audio_info_update (self, s);
else
gst_play_subtitle_info_update (self, s);
}
static gchar *
stream_info_get_codec (GstPlayStreamInfo * s)
{
const gchar *type;
GstTagList *tags;
gchar *codec = NULL;
if (GST_IS_PLAY_VIDEO_INFO (s))
type = GST_TAG_VIDEO_CODEC;
else if (GST_IS_PLAY_AUDIO_INFO (s))
type = GST_TAG_AUDIO_CODEC;
else
type = GST_TAG_SUBTITLE_CODEC;
tags = gst_play_stream_info_get_tags (s);
if (tags) {
gst_tag_list_get_string (tags, type, &codec);
if (!codec)
gst_tag_list_get_string (tags, GST_TAG_CODEC, &codec);
}
if (!codec) {
GstCaps *caps;
caps = gst_play_stream_info_get_caps (s);
if (caps) {
codec = gst_pb_utils_get_codec_description (caps);
}
}
return codec;
}
static void
gst_play_stream_info_update_tags_and_caps (GstPlay * self,
GstPlayStreamInfo * s)
{
GstTagList *tags;
gint stream_index;
stream_index = gst_play_stream_info_get_index (s);
if (GST_IS_PLAY_VIDEO_INFO (s))
g_signal_emit_by_name (self->playbin, "get-video-tags",
stream_index, &tags);
else if (GST_IS_PLAY_AUDIO_INFO (s))
g_signal_emit_by_name (self->playbin, "get-audio-tags",
stream_index, &tags);
else
g_signal_emit_by_name (self->playbin, "get-text-tags", stream_index, &tags);
if (s->tags)
gst_tag_list_unref (s->tags);
s->tags = tags;
if (s->caps)
gst_caps_unref (s->caps);
s->caps = get_caps (self, stream_index, G_OBJECT_TYPE (s));
g_free (s->codec);
s->codec = stream_info_get_codec (s);
GST_DEBUG_OBJECT (self, "%s index: %d tags: %p caps: %p",
gst_play_stream_info_get_stream_type (s), stream_index, s->tags, s->caps);
gst_play_stream_info_update (self, s);
}
static void
gst_play_streams_info_create (GstPlay * self,
GstPlayMediaInfo * media_info, const gchar * prop, GType type)
{
gint i;
gint total = -1;
GstPlayStreamInfo *s;
if (!media_info)
return;
g_object_get (G_OBJECT (self->playbin), prop, &total, NULL);
GST_DEBUG_OBJECT (self, "%s: %d", prop, total);
for (i = 0; i < total; i++) {
/* check if stream already exist in the list */
s = gst_play_stream_info_find (media_info, type, i);
if (!s) {
/* create a new stream info instance */
s = gst_play_stream_info_new (i, type);
/* add the object in stream list */
media_info->stream_list = g_list_append (media_info->stream_list, s);
/* based on type, add the object in its corresponding stream_ list */
if (GST_IS_PLAY_AUDIO_INFO (s))
media_info->audio_stream_list = g_list_append
(media_info->audio_stream_list, s);
else if (GST_IS_PLAY_VIDEO_INFO (s))
media_info->video_stream_list = g_list_append
(media_info->video_stream_list, s);
else
media_info->subtitle_stream_list = g_list_append
(media_info->subtitle_stream_list, s);
GST_DEBUG_OBJECT (self, "create %s stream stream_index: %d",
gst_play_stream_info_get_stream_type (s), i);
}
gst_play_stream_info_update_tags_and_caps (self, s);
}
}
static void
gst_play_stream_info_update_from_stream (GstPlay * self,
GstPlayStreamInfo * s, GstStream * stream)
{
if (s->tags)
gst_tag_list_unref (s->tags);
s->tags = gst_stream_get_tags (stream);
if (s->caps)
gst_caps_unref (s->caps);
s->caps = gst_stream_get_caps (stream);
g_free (s->codec);
s->codec = stream_info_get_codec (s);
GST_DEBUG_OBJECT (self, "%s index: %d tags: %p caps: %p",
gst_play_stream_info_get_stream_type (s), s->stream_index,
s->tags, s->caps);
gst_play_stream_info_update (self, s);
}
static void
gst_play_streams_info_create_from_collection (GstPlay * self,
GstPlayMediaInfo * media_info, GstStreamCollection * collection)
{
guint i;
guint total;
GstPlayStreamInfo *s;
guint n_audio = 0;
guint n_video = 0;
guint n_text = 0;
if (!media_info || !collection)
return;
total = gst_stream_collection_get_size (collection);
for (i = 0; i < total; i++) {
GstStream *stream = gst_stream_collection_get_stream (collection, i);
GstStreamType stream_type = gst_stream_get_stream_type (stream);
const gchar *stream_id = gst_stream_get_stream_id (stream);
if (stream_type & GST_STREAM_TYPE_AUDIO) {
s = gst_play_stream_info_new (n_audio, GST_TYPE_PLAY_AUDIO_INFO);
n_audio++;
} else if (stream_type & GST_STREAM_TYPE_VIDEO) {
s = gst_play_stream_info_new (n_video, GST_TYPE_PLAY_VIDEO_INFO);
n_video++;
} else if (stream_type & GST_STREAM_TYPE_TEXT) {
s = gst_play_stream_info_new (n_text, GST_TYPE_PLAY_SUBTITLE_INFO);
n_text++;
} else {
GST_DEBUG_OBJECT (self, "Unknown type stream %d", i);
continue;
}
s->stream_id = g_strdup (stream_id);
/* add the object in stream list */
media_info->stream_list = g_list_append (media_info->stream_list, s);
/* based on type, add the object in its corresponding stream_ list */
if (GST_IS_PLAY_AUDIO_INFO (s))
media_info->audio_stream_list = g_list_append
(media_info->audio_stream_list, s);
else if (GST_IS_PLAY_VIDEO_INFO (s))
media_info->video_stream_list = g_list_append
(media_info->video_stream_list, s);
else
media_info->subtitle_stream_list = g_list_append
(media_info->subtitle_stream_list, s);
GST_DEBUG_OBJECT (self, "create %s stream stream_index: %d",
gst_play_stream_info_get_stream_type (s), s->stream_index);
gst_play_stream_info_update_from_stream (self, s, stream);
}
}
static void
video_changed_cb (G_GNUC_UNUSED GObject * object, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
g_mutex_lock (&self->lock);
gst_play_streams_info_create (self, self->media_info,
"n-video", GST_TYPE_PLAY_VIDEO_INFO);
g_mutex_unlock (&self->lock);
}
static void
audio_changed_cb (G_GNUC_UNUSED GObject * object, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
g_mutex_lock (&self->lock);
gst_play_streams_info_create (self, self->media_info,
"n-audio", GST_TYPE_PLAY_AUDIO_INFO);
g_mutex_unlock (&self->lock);
}
static void
subtitle_changed_cb (G_GNUC_UNUSED GObject * object, gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
g_mutex_lock (&self->lock);
gst_play_streams_info_create (self, self->media_info,
"n-text", GST_TYPE_PLAY_SUBTITLE_INFO);
g_mutex_unlock (&self->lock);
}
static void *
get_title (GstTagList * tags)
{
gchar *title = NULL;
gst_tag_list_get_string (tags, GST_TAG_TITLE, &title);
if (!title)
gst_tag_list_get_string (tags, GST_TAG_TITLE_SORTNAME, &title);
return title;
}
static void *
get_container_format (GstTagList * tags)
{
gchar *container = NULL;
gst_tag_list_get_string (tags, GST_TAG_CONTAINER_FORMAT, &container);
/* TODO: If container is not available then maybe consider
* parsing caps or file extension to guess the container format.
*/
return container;
}
static void *
get_from_tags (GstPlay * self, GstPlayMediaInfo * media_info,
void *(*func) (GstTagList *))
{
GList *l;
void *ret = NULL;
if (media_info->tags) {
ret = func (media_info->tags);
if (ret)
return ret;
}
/* if global tag does not exit then try video and audio streams */
GST_DEBUG_OBJECT (self, "trying video tags");
for (l = gst_play_media_info_get_video_streams (media_info); l != NULL;
l = l->next) {
GstTagList *tags;
tags = gst_play_stream_info_get_tags ((GstPlayStreamInfo *) l->data);
if (tags)
ret = func (tags);
if (ret)
return ret;
}
GST_DEBUG_OBJECT (self, "trying audio tags");
for (l = gst_play_media_info_get_audio_streams (media_info); l != NULL;
l = l->next) {
GstTagList *tags;
tags = gst_play_stream_info_get_tags ((GstPlayStreamInfo *) l->data);
if (tags)
ret = func (tags);
if (ret)
return ret;
}
GST_DEBUG_OBJECT (self, "failed to get the information from tags");
return NULL;
}
static void *
get_cover_sample (GstTagList * tags)
{
GstSample *cover_sample = NULL;
gst_tag_list_get_sample (tags, GST_TAG_IMAGE, &cover_sample);
if (!cover_sample)
gst_tag_list_get_sample (tags, GST_TAG_PREVIEW_IMAGE, &cover_sample);
return cover_sample;
}
static GstPlayMediaInfo *
gst_play_media_info_create (GstPlay * self)
{
GstPlayMediaInfo *media_info;
GstQuery *query;
GST_DEBUG_OBJECT (self, "begin");
media_info = gst_play_media_info_new (self->uri);
media_info->duration = gst_play_get_duration (self);
media_info->tags = self->global_tags;
media_info->is_live = self->is_live;
self->global_tags = NULL;
query = gst_query_new_seeking (GST_FORMAT_TIME);
if (gst_element_query (self->playbin, query))
gst_query_parse_seeking (query, NULL, &media_info->seekable, NULL, NULL);
gst_query_unref (query);
if (self->use_playbin3) {
gst_play_streams_info_create_from_collection (self, media_info,
self->collection);
} else {
/* create audio/video/sub streams */
gst_play_streams_info_create (self, media_info, "n-video",
GST_TYPE_PLAY_VIDEO_INFO);
gst_play_streams_info_create (self, media_info, "n-audio",
GST_TYPE_PLAY_AUDIO_INFO);
gst_play_streams_info_create (self, media_info, "n-text",
GST_TYPE_PLAY_SUBTITLE_INFO);
}
media_info->title = get_from_tags (self, media_info, get_title);
media_info->container =
get_from_tags (self, media_info, get_container_format);
media_info->image_sample = get_from_tags (self, media_info, get_cover_sample);
GST_DEBUG_OBJECT (self, "uri: %s title: %s duration: %" GST_TIME_FORMAT
" seekable: %s live: %s container: %s image_sample %p",
media_info->uri, media_info->title, GST_TIME_ARGS (media_info->duration),
media_info->seekable ? "yes" : "no", media_info->is_live ? "yes" : "no",
media_info->container, media_info->image_sample);
GST_DEBUG_OBJECT (self, "end");
return media_info;
}
static void
tags_changed_cb (GstPlay * self, gint stream_index, GType type)
{
GstPlayStreamInfo *s;
if (!self->media_info)
return;
/* update the stream information */
g_mutex_lock (&self->lock);
s = gst_play_stream_info_find (self->media_info, type, stream_index);
gst_play_stream_info_update_tags_and_caps (self, s);
g_mutex_unlock (&self->lock);
on_media_info_updated (self);
}
static void
video_tags_changed_cb (G_GNUC_UNUSED GstElement * playbin, gint stream_index,
gpointer user_data)
{
tags_changed_cb (GST_PLAY (user_data), stream_index,
GST_TYPE_PLAY_VIDEO_INFO);
}
static void
audio_tags_changed_cb (G_GNUC_UNUSED GstElement * playbin, gint stream_index,
gpointer user_data)
{
tags_changed_cb (GST_PLAY (user_data), stream_index,
GST_TYPE_PLAY_AUDIO_INFO);
}
static void
subtitle_tags_changed_cb (G_GNUC_UNUSED GstElement * playbin, gint stream_index,
gpointer user_data)
{
tags_changed_cb (GST_PLAY (user_data), stream_index,
GST_TYPE_PLAY_SUBTITLE_INFO);
}
static void
volume_notify_cb (G_GNUC_UNUSED GObject * obj, G_GNUC_UNUSED GParamSpec * pspec,
GstPlay * self)
{
api_bus_post_message (self, GST_PLAY_MESSAGE_VOLUME_CHANGED,
GST_PLAY_MESSAGE_DATA_VOLUME, G_TYPE_DOUBLE,
gst_play_get_volume (self), NULL);
}
static void
mute_notify_cb (G_GNUC_UNUSED GObject * obj, G_GNUC_UNUSED GParamSpec * pspec,
GstPlay * self)
{
api_bus_post_message (self, GST_PLAY_MESSAGE_MUTE_CHANGED,
GST_PLAY_MESSAGE_DATA_IS_MUTED, G_TYPE_BOOLEAN,
gst_play_get_mute (self), NULL);
}
static void
source_setup_cb (GstElement * playbin, GstElement * source, GstPlay * self)
{
gchar *user_agent;
user_agent = gst_play_config_get_user_agent (self->config);
if (user_agent) {
GParamSpec *prop;
prop = g_object_class_find_property (G_OBJECT_GET_CLASS (source),
"user-agent");
if (prop && prop->value_type == G_TYPE_STRING) {
GST_INFO_OBJECT (self, "Setting source user-agent: %s", user_agent);
g_object_set (source, "user-agent", user_agent, NULL);
}
g_free (user_agent);
}
}
static gpointer
gst_play_main (gpointer data)
{
GstPlay *self = GST_PLAY (data);
GstBus *bus;
GSource *source;
GstElement *scaletempo;
const gchar *env;
GST_TRACE_OBJECT (self, "Starting main thread");
g_main_context_push_thread_default (self->context);
source = g_idle_source_new ();
g_source_set_callback (source, (GSourceFunc) main_loop_running_cb, self,
NULL);
g_source_attach (source, self->context);
g_source_unref (source);
env = g_getenv ("GST_PLAY_USE_PLAYBIN3");
if (env && g_str_has_prefix (env, "1"))
self->use_playbin3 = TRUE;
if (self->use_playbin3) {
GST_DEBUG_OBJECT (self, "playbin3 enabled");
self->playbin = gst_element_factory_make ("playbin3", "playbin3");
} else {
self->playbin = gst_element_factory_make ("playbin", "playbin");
}
if (!self->playbin) {
g_error ("GstPlay: 'playbin' element not found, please check your setup");
g_assert_not_reached ();
}
gst_object_ref_sink (self->playbin);
if (self->video_renderer) {
gst_play_set_playbin_video_sink (self);
}
scaletempo = gst_element_factory_make ("scaletempo", NULL);
if (scaletempo) {
g_object_set (self->playbin, "audio-filter", scaletempo, NULL);
} else {
g_warning ("GstPlay: scaletempo element not available. Audio pitch "
"will not be preserved during trick modes");
}
self->bus = bus = gst_element_get_bus (self->playbin);
gst_bus_add_signal_watch (bus);
g_signal_connect (G_OBJECT (bus), "message::error", G_CALLBACK (error_cb),
self);
g_signal_connect (G_OBJECT (bus), "message::warning", G_CALLBACK (warning_cb),
self);
g_signal_connect (G_OBJECT (bus), "message::eos", G_CALLBACK (eos_cb), self);
g_signal_connect (G_OBJECT (bus), "message::state-changed",
G_CALLBACK (state_changed_cb), self);
g_signal_connect (G_OBJECT (bus), "message::buffering",
G_CALLBACK (buffering_cb), self);
g_signal_connect (G_OBJECT (bus), "message::clock-lost",
G_CALLBACK (clock_lost_cb), self);
g_signal_connect (G_OBJECT (bus), "message::duration-changed",
G_CALLBACK (duration_changed_cb), self);
g_signal_connect (G_OBJECT (bus), "message::latency",
G_CALLBACK (latency_cb), self);
g_signal_connect (G_OBJECT (bus), "message::request-state",
G_CALLBACK (request_state_cb), self);
g_signal_connect (G_OBJECT (bus), "message::element",
G_CALLBACK (element_cb), self);
g_signal_connect (G_OBJECT (bus), "message::tag", G_CALLBACK (tags_cb), self);
if (self->use_playbin3) {
g_signal_connect (G_OBJECT (bus), "message::stream-collection",
G_CALLBACK (stream_collection_cb), self);
g_signal_connect (G_OBJECT (bus), "message::streams-selected",
G_CALLBACK (streams_selected_cb), self);
} else {
g_signal_connect (self->playbin, "video-changed",
G_CALLBACK (video_changed_cb), self);
g_signal_connect (self->playbin, "audio-changed",
G_CALLBACK (audio_changed_cb), self);
g_signal_connect (self->playbin, "text-changed",
G_CALLBACK (subtitle_changed_cb), self);
g_signal_connect (self->playbin, "video-tags-changed",
G_CALLBACK (video_tags_changed_cb), self);
g_signal_connect (self->playbin, "audio-tags-changed",
G_CALLBACK (audio_tags_changed_cb), self);
g_signal_connect (self->playbin, "text-tags-changed",
G_CALLBACK (subtitle_tags_changed_cb), self);
}
g_signal_connect (self->playbin, "notify::volume",
G_CALLBACK (volume_notify_cb), self);
g_signal_connect (self->playbin, "notify::mute",
G_CALLBACK (mute_notify_cb), self);
g_signal_connect (self->playbin, "source-setup",
G_CALLBACK (source_setup_cb), self);
self->target_state = GST_STATE_NULL;
self->current_state = GST_STATE_NULL;
change_state (self, GST_PLAY_STATE_STOPPED);
self->buffering_percent = 100;
self->is_eos = FALSE;
self->is_live = FALSE;
self->rate = 1.0;
GST_TRACE_OBJECT (self, "Starting main loop");
g_main_loop_run (self->loop);
GST_TRACE_OBJECT (self, "Stopped main loop");
gst_bus_remove_signal_watch (bus);
gst_object_unref (bus);
remove_tick_source (self);
remove_ready_timeout_source (self);
g_mutex_lock (&self->lock);
if (self->media_info) {
g_object_unref (self->media_info);
self->media_info = NULL;
}
remove_seek_source (self);
g_mutex_unlock (&self->lock);
g_main_context_pop_thread_default (self->context);
self->target_state = GST_STATE_NULL;
self->current_state = GST_STATE_NULL;
if (self->playbin) {
gst_element_set_state (self->playbin, GST_STATE_NULL);
gst_object_unref (self->playbin);
self->playbin = NULL;
}
GST_TRACE_OBJECT (self, "Stopped main thread");
return NULL;
}
static gpointer
gst_play_init_once (G_GNUC_UNUSED gpointer user_data)
{
gst_init (NULL, NULL);
GST_DEBUG_CATEGORY_INIT (gst_play_debug, "gst-play", 0, "GstPlay");
gst_play_error_quark ();
return NULL;
}
/**
* gst_play_new:
* @video_renderer: (transfer full) (allow-none): GstPlayVideoRenderer to use
*
* Creates a new #GstPlay instance.
*
* Video is going to be rendered by @video_renderer, or if %NULL is provided
* no special video set up will be done and some default handling will be
* performed.
*
* Returns: (transfer full): a new #GstPlay instance
* Since: 1.20
*/
GstPlay *
gst_play_new (GstPlayVideoRenderer * video_renderer)
{
static GOnce once = G_ONCE_INIT;
GstPlay *self;
g_once (&once, gst_play_init_once, NULL);
self = g_object_new (GST_TYPE_PLAY, "video-renderer", video_renderer, NULL);
gst_object_ref_sink (self);
if (video_renderer)
g_object_unref (video_renderer);
return self;
}
/**
* gst_play_get_message_bus:
* @play: #GstPlay instance
*
* GstPlay API exposes a #GstBus instance which purpose is to provide data
* structures representing play-internal events in form of #GstMessage<!-- -->s of
* type GST_MESSAGE_APPLICATION.
*
* Each message carries a "play-message" field of type #GstPlayMessage.
* Further fields of the message data are specific to each possible value of
* that enumeration.
*
* Applications can consume the messages asynchronously within their own
* event-loop / UI-thread etc. Note that in case the application does not
* consume the messages, the bus will accumulate these internally and eventually
* fill memory. To avoid that, the bus has to be set "flushing".
*
* Returns: (transfer full): The play message bus instance
*
* Since: 1.20
*/
GstBus *
gst_play_get_message_bus (GstPlay * self)
{
return g_object_ref (self->api_bus);
}
static gboolean
gst_play_play_internal (gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstStateChangeReturn state_ret;
GST_DEBUG_OBJECT (self, "Play");
g_mutex_lock (&self->lock);
if (!self->uri) {
g_mutex_unlock (&self->lock);
return G_SOURCE_REMOVE;
}
g_mutex_unlock (&self->lock);
remove_ready_timeout_source (self);
self->target_state = GST_STATE_PLAYING;
if (self->current_state < GST_STATE_PAUSED)
change_state (self, GST_PLAY_STATE_BUFFERING);
if (self->current_state >= GST_STATE_PAUSED && !self->is_eos
&& self->buffering_percent >= 100
&& !(self->seek_position != GST_CLOCK_TIME_NONE || self->seek_pending)) {
state_ret = gst_element_set_state (self->playbin, GST_STATE_PLAYING);
} else {
state_ret = gst_element_set_state (self->playbin, GST_STATE_PAUSED);
}
if (state_ret == GST_STATE_CHANGE_FAILURE) {
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to play"), NULL);
return G_SOURCE_REMOVE;
} else if (state_ret == GST_STATE_CHANGE_NO_PREROLL) {
self->is_live = TRUE;
GST_DEBUG_OBJECT (self, "Pipeline is live");
}
if (self->is_eos) {
gboolean ret;
GST_DEBUG_OBJECT (self, "Was EOS, seeking to beginning");
self->is_eos = FALSE;
ret =
gst_element_seek_simple (self->playbin, GST_FORMAT_TIME,
GST_SEEK_FLAG_FLUSH, 0);
if (!ret) {
GST_ERROR_OBJECT (self, "Seek to beginning failed");
gst_play_stop_internal (self, TRUE);
gst_play_play_internal (self);
}
}
return G_SOURCE_REMOVE;
}
/**
* gst_play_play:
* @play: #GstPlay instance
*
* Request to play the loaded stream.
* Since: 1.20
*/
void
gst_play_play (GstPlay * self)
{
g_return_if_fail (GST_IS_PLAY (self));
g_main_context_invoke_full (self->context, G_PRIORITY_DEFAULT,
gst_play_play_internal, self, NULL);
}
static gboolean
gst_play_pause_internal (gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
GstStateChangeReturn state_ret;
GST_DEBUG_OBJECT (self, "Pause");
g_mutex_lock (&self->lock);
if (!self->uri) {
g_mutex_unlock (&self->lock);
return G_SOURCE_REMOVE;
}
g_mutex_unlock (&self->lock);
tick_cb (self);
remove_tick_source (self);
remove_ready_timeout_source (self);
self->target_state = GST_STATE_PAUSED;
if (self->current_state < GST_STATE_PAUSED)
change_state (self, GST_PLAY_STATE_BUFFERING);
state_ret = gst_element_set_state (self->playbin, GST_STATE_PAUSED);
if (state_ret == GST_STATE_CHANGE_FAILURE) {
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to pause"), NULL);
return G_SOURCE_REMOVE;
} else if (state_ret == GST_STATE_CHANGE_NO_PREROLL) {
self->is_live = TRUE;
GST_DEBUG_OBJECT (self, "Pipeline is live");
}
if (self->is_eos) {
gboolean ret;
GST_DEBUG_OBJECT (self, "Was EOS, seeking to beginning");
self->is_eos = FALSE;
ret =
gst_element_seek_simple (self->playbin, GST_FORMAT_TIME,
GST_SEEK_FLAG_FLUSH, 0);
if (!ret) {
GST_ERROR_OBJECT (self, "Seek to beginning failed");
gst_play_stop_internal (self, TRUE);
gst_play_pause_internal (self);
}
}
return G_SOURCE_REMOVE;
}
/**
* gst_play_pause:
* @play: #GstPlay instance
*
* Pauses the current stream.
* Since: 1.20
*/
void
gst_play_pause (GstPlay * self)
{
g_return_if_fail (GST_IS_PLAY (self));
g_main_context_invoke_full (self->context, G_PRIORITY_DEFAULT,
gst_play_pause_internal, self, NULL);
}
static void
gst_play_stop_internal (GstPlay * self, gboolean transient)
{
/* directly return if we're already stopped */
if (self->current_state <= GST_STATE_READY &&
self->target_state <= GST_STATE_READY)
return;
GST_DEBUG_OBJECT (self, "Stop (transient %d)", transient);
tick_cb (self);
remove_tick_source (self);
add_ready_timeout_source (self);
self->target_state = GST_STATE_NULL;
self->current_state = GST_STATE_READY;
self->is_live = FALSE;
self->is_eos = FALSE;
gst_bus_set_flushing (self->bus, TRUE);
gst_element_set_state (self->playbin, GST_STATE_READY);
gst_bus_set_flushing (self->bus, FALSE);
change_state (self, transient
&& self->app_state !=
GST_PLAY_STATE_STOPPED ? GST_PLAY_STATE_BUFFERING :
GST_PLAY_STATE_STOPPED);
self->buffering_percent = 100;
self->cached_duration = GST_CLOCK_TIME_NONE;
g_mutex_lock (&self->lock);
if (self->media_info) {
g_object_unref (self->media_info);
self->media_info = NULL;
}
if (self->global_tags) {
gst_tag_list_unref (self->global_tags);
self->global_tags = NULL;
}
self->seek_pending = FALSE;
remove_seek_source (self);
self->seek_position = GST_CLOCK_TIME_NONE;
self->last_seek_time = GST_CLOCK_TIME_NONE;
self->rate = 1.0;
if (self->collection) {
if (self->stream_notify_id)
g_signal_handler_disconnect (self->collection, self->stream_notify_id);
self->stream_notify_id = 0;
gst_object_unref (self->collection);
self->collection = NULL;
}
g_free (self->video_sid);
g_free (self->audio_sid);
g_free (self->subtitle_sid);
self->video_sid = NULL;
self->audio_sid = NULL;
self->subtitle_sid = NULL;
g_mutex_unlock (&self->lock);
}
static gboolean
gst_play_stop_internal_dispatch (gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
gst_play_stop_internal (self, FALSE);
return G_SOURCE_REMOVE;
}
/**
* gst_play_stop:
* @play: #GstPlay instance
*
* Stops playing the current stream and resets to the first position
* in the stream.
* Since: 1.20
*/
void
gst_play_stop (GstPlay * self)
{
g_return_if_fail (GST_IS_PLAY (self));
g_main_context_invoke_full (self->context, G_PRIORITY_DEFAULT,
gst_play_stop_internal_dispatch, self, NULL);
}
/* Must be called with lock from main context, releases lock! */
static void
gst_play_seek_internal_locked (GstPlay * self)
{
gboolean ret;
GstClockTime position;
gdouble rate;
GstStateChangeReturn state_ret;
GstEvent *s_event;
GstSeekFlags flags = 0;
gboolean accurate = FALSE;
remove_seek_source (self);
/* Only seek in PAUSED */
if (self->current_state < GST_STATE_PAUSED) {
return;
} else if (self->current_state != GST_STATE_PAUSED) {
g_mutex_unlock (&self->lock);
state_ret = gst_element_set_state (self->playbin, GST_STATE_PAUSED);
if (state_ret == GST_STATE_CHANGE_FAILURE) {
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to seek"), NULL);
g_mutex_lock (&self->lock);
return;
}
g_mutex_lock (&self->lock);
return;
}
self->last_seek_time = gst_util_get_timestamp ();
position = self->seek_position;
self->seek_position = GST_CLOCK_TIME_NONE;
self->seek_pending = TRUE;
rate = self->rate;
g_mutex_unlock (&self->lock);
remove_tick_source (self);
self->is_eos = FALSE;
flags |= GST_SEEK_FLAG_FLUSH;
accurate = gst_play_config_get_seek_accurate (self->config);
if (accurate) {
flags |= GST_SEEK_FLAG_ACCURATE;
} else {
flags &= ~GST_SEEK_FLAG_ACCURATE;
}
if (rate != 1.0) {
flags |= GST_SEEK_FLAG_TRICKMODE;
}
if (rate >= 0.0) {
s_event = gst_event_new_seek (rate, GST_FORMAT_TIME, flags,
GST_SEEK_TYPE_SET, position, GST_SEEK_TYPE_SET, GST_CLOCK_TIME_NONE);
} else {
s_event = gst_event_new_seek (rate, GST_FORMAT_TIME, flags,
GST_SEEK_TYPE_SET, G_GINT64_CONSTANT (0), GST_SEEK_TYPE_SET, position);
}
GST_DEBUG_OBJECT (self, "Seek with rate %.2lf to %" GST_TIME_FORMAT,
rate, GST_TIME_ARGS (position));
ret = gst_element_send_event (self->playbin, s_event);
if (!ret)
on_error (self, g_error_new (GST_PLAY_ERROR, GST_PLAY_ERROR_FAILED,
"Failed to seek to %" GST_TIME_FORMAT, GST_TIME_ARGS (position)),
NULL);
g_mutex_lock (&self->lock);
}
static gboolean
gst_play_seek_internal (gpointer user_data)
{
GstPlay *self = GST_PLAY (user_data);
g_mutex_lock (&self->lock);
gst_play_seek_internal_locked (self);
g_mutex_unlock (&self->lock);
return G_SOURCE_REMOVE;
}
/**
* gst_play_set_rate:
* @play: #GstPlay instance
* @rate: playback rate
*
* Playback at specified rate
* Since: 1.20
*/
void
gst_play_set_rate (GstPlay * self, gdouble rate)
{
g_return_if_fail (GST_IS_PLAY (self));
g_return_if_fail (rate != 0.0);
g_object_set (self, "rate", rate, NULL);
}
/**
* gst_play_get_rate:
* @play: #GstPlay instance
*
* Returns: current playback rate
* Since: 1.20
*/
gdouble
gst_play_get_rate (GstPlay * self)
{
gdouble val;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_RATE);
g_object_get (self, "rate", &val, NULL);
return val;
}
/**
* gst_play_seek:
* @play: #GstPlay instance
* @position: position to seek in nanoseconds
*
* Seeks the currently-playing stream to the absolute @position time
* in nanoseconds.
* Since: 1.20
*/
void
gst_play_seek (GstPlay * self, GstClockTime position)
{
g_return_if_fail (GST_IS_PLAY (self));
g_return_if_fail (GST_CLOCK_TIME_IS_VALID (position));
g_mutex_lock (&self->lock);
if (self->media_info && !self->media_info->seekable) {
GST_DEBUG_OBJECT (self, "Media is not seekable");
g_mutex_unlock (&self->lock);
return;
}
self->seek_position = position;
/* If there is no seek being dispatch to the main context currently do that,
* otherwise we just updated the seek position so that it will be taken by
* the seek handler from the main context instead of the old one.
*/
if (!self->seek_source) {
GstClockTime now = gst_util_get_timestamp ();
/* If no seek is pending or it was started more than 250 mseconds ago seek
* immediately, otherwise wait until the 250 mseconds have passed */
if (!self->seek_pending || (now - self->last_seek_time > 250 * GST_MSECOND)) {
self->seek_source = g_idle_source_new ();
g_source_set_callback (self->seek_source,
(GSourceFunc) gst_play_seek_internal, self, NULL);
GST_TRACE_OBJECT (self, "Dispatching seek to position %" GST_TIME_FORMAT,
GST_TIME_ARGS (position));
g_source_attach (self->seek_source, self->context);
} else {
guint delay = 250000 - (now - self->last_seek_time) / 1000;
/* Note that last_seek_time must be set to something at this point and
* it must be smaller than 250 mseconds */
self->seek_source = g_timeout_source_new (delay);
g_source_set_callback (self->seek_source,
(GSourceFunc) gst_play_seek_internal, self, NULL);
GST_TRACE_OBJECT (self,
"Delaying seek to position %" GST_TIME_FORMAT " by %u us",
GST_TIME_ARGS (position), delay);
g_source_attach (self->seek_source, self->context);
}
}
g_mutex_unlock (&self->lock);
}
static void
remove_seek_source (GstPlay * self)
{
if (!self->seek_source)
return;
g_source_destroy (self->seek_source);
g_source_unref (self->seek_source);
self->seek_source = NULL;
}
/**
* gst_play_get_uri:
* @play: #GstPlay instance
*
* Gets the URI of the currently-playing stream.
*
* Returns: (transfer full) (nullable): a string containing the URI of the
* currently-playing stream. g_free() after usage.
* Since: 1.20
*/
gchar *
gst_play_get_uri (GstPlay * self)
{
gchar *val;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_URI);
g_object_get (self, "uri", &val, NULL);
return val;
}
/**
* gst_play_set_uri:
* @play: #GstPlay instance
* @uri: (nullable): next URI to play.
*
* Sets the next URI to play.
* Since: 1.20
*/
void
gst_play_set_uri (GstPlay * self, const gchar * val)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "uri", val, NULL);
}
/**
* gst_play_set_subtitle_uri:
* @play: #GstPlay instance
* @uri: (nullable): subtitle URI
*
* Sets the external subtitle URI. This should be combined with a call to
* gst_play_set_subtitle_track_enabled(@play, TRUE) so the subtitles are actually
* rendered.
* Since: 1.20
*/
void
gst_play_set_subtitle_uri (GstPlay * self, const gchar * suburi)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "suburi", suburi, NULL);
}
/**
* gst_play_get_subtitle_uri:
* @play: #GstPlay instance
*
* current subtitle URI
*
* Returns: (transfer full) (nullable): URI of the current external subtitle.
* g_free() after usage.
* Since: 1.20
*/
gchar *
gst_play_get_subtitle_uri (GstPlay * self)
{
gchar *val = NULL;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
g_object_get (self, "suburi", &val, NULL);
return val;
}
/**
* gst_play_get_position:
* @play: #GstPlay instance
*
* Returns: the absolute position time, in nanoseconds, of the
* currently-playing stream.
* Since: 1.20
*/
GstClockTime
gst_play_get_position (GstPlay * self)
{
GstClockTime val;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_POSITION);
g_object_get (self, "position", &val, NULL);
return val;
}
/**
* gst_play_get_duration:
* @play: #GstPlay instance
*
* Retrieves the duration of the media stream that self represents.
*
* Returns: the duration of the currently-playing media stream, in
* nanoseconds.
* Since: 1.20
*/
GstClockTime
gst_play_get_duration (GstPlay * self)
{
GstClockTime val;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_DURATION);
g_object_get (self, "duration", &val, NULL);
return val;
}
/**
* gst_play_get_volume:
* @play: #GstPlay instance
*
* Returns the current volume level, as a percentage between 0 and 1.
*
* Returns: the volume as percentage between 0 and 1.
* Since: 1.20
*/
gdouble
gst_play_get_volume (GstPlay * self)
{
gdouble val;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_VOLUME);
g_object_get (self, "volume", &val, NULL);
return val;
}
/**
* gst_play_set_volume:
* @play: #GstPlay instance
* @val: the new volume level, as a percentage between 0 and 1
*
* Sets the volume level of the stream as a percentage between 0 and 1.
* Since: 1.20
*/
void
gst_play_set_volume (GstPlay * self, gdouble val)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "volume", val, NULL);
}
/**
* gst_play_get_mute:
* @play: #GstPlay instance
*
* Returns: %TRUE if the currently-playing stream is muted.
* Since: 1.20
*/
gboolean
gst_play_get_mute (GstPlay * self)
{
gboolean val;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_MUTE);
g_object_get (self, "mute", &val, NULL);
return val;
}
/**
* gst_play_set_mute:
* @play: #GstPlay instance
* @val: Mute state the should be set
*
* %TRUE if the currently-playing stream should be muted.
* Since: 1.20
*/
void
gst_play_set_mute (GstPlay * self, gboolean val)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "mute", val, NULL);
}
/**
* gst_play_get_pipeline:
* @play: #GstPlay instance
*
* Returns: (transfer full): The internal playbin instance.
*
* The caller should free it with g_object_unref()
* Since: 1.20
*/
GstElement *
gst_play_get_pipeline (GstPlay * self)
{
GstElement *val;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
g_object_get (self, "pipeline", &val, NULL);
return val;
}
/**
* gst_play_get_media_info:
* @play: #GstPlay instance
*
* A Function to get the current media info #GstPlayMediaInfo instance.
*
* Returns: (transfer full) (nullable): media info instance.
*
* The caller should free it with g_object_unref()
* Since: 1.20
*/
GstPlayMediaInfo *
gst_play_get_media_info (GstPlay * self)
{
GstPlayMediaInfo *info;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
if (!self->media_info)
return NULL;
g_mutex_lock (&self->lock);
info = gst_play_media_info_copy (self->media_info);
g_mutex_unlock (&self->lock);
return info;
}
/**
* gst_play_get_current_audio_track:
* @play: #GstPlay instance
*
* A Function to get current audio #GstPlayAudioInfo instance.
*
* Returns: (transfer full) (nullable): current audio track.
*
* The caller should free it with g_object_unref()
* Since: 1.20
*/
GstPlayAudioInfo *
gst_play_get_current_audio_track (GstPlay * self)
{
GstPlayAudioInfo *info;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
if (!is_track_enabled (self, GST_PLAY_FLAG_AUDIO))
return NULL;
if (self->use_playbin3) {
info = (GstPlayAudioInfo *)
gst_play_stream_info_get_current_from_stream_id (self,
self->audio_sid, GST_TYPE_PLAY_AUDIO_INFO);
} else {
info = (GstPlayAudioInfo *) gst_play_stream_info_get_current (self,
"current-audio", GST_TYPE_PLAY_AUDIO_INFO);
}
return info;
}
/**
* gst_play_get_current_video_track:
* @play: #GstPlay instance
*
* A Function to get current video #GstPlayVideoInfo instance.
*
* Returns: (transfer full) (nullable): current video track.
*
* The caller should free it with g_object_unref()
* Since: 1.20
*/
GstPlayVideoInfo *
gst_play_get_current_video_track (GstPlay * self)
{
GstPlayVideoInfo *info;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
if (!is_track_enabled (self, GST_PLAY_FLAG_VIDEO))
return NULL;
if (self->use_playbin3) {
info = (GstPlayVideoInfo *)
gst_play_stream_info_get_current_from_stream_id (self,
self->video_sid, GST_TYPE_PLAY_VIDEO_INFO);
} else {
info = (GstPlayVideoInfo *) gst_play_stream_info_get_current (self,
"current-video", GST_TYPE_PLAY_VIDEO_INFO);
}
return info;
}
/**
* gst_play_get_current_subtitle_track:
* @play: #GstPlay instance
*
* A Function to get current subtitle #GstPlaySubtitleInfo instance.
*
* Returns: (transfer full) (nullable): current subtitle track.
*
* The caller should free it with g_object_unref()
* Since: 1.20
*/
GstPlaySubtitleInfo *
gst_play_get_current_subtitle_track (GstPlay * self)
{
GstPlaySubtitleInfo *info;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
if (!is_track_enabled (self, GST_PLAY_FLAG_SUBTITLE))
return NULL;
if (self->use_playbin3) {
info = (GstPlaySubtitleInfo *)
gst_play_stream_info_get_current_from_stream_id (self,
self->subtitle_sid, GST_TYPE_PLAY_SUBTITLE_INFO);
} else {
info = (GstPlaySubtitleInfo *) gst_play_stream_info_get_current (self,
"current-text", GST_TYPE_PLAY_SUBTITLE_INFO);
}
return info;
}
/* Must be called with lock */
static gboolean
gst_play_select_streams (GstPlay * self)
{
GList *stream_list = NULL;
gboolean ret = FALSE;
if (self->audio_sid)
stream_list = g_list_append (stream_list, g_strdup (self->audio_sid));
if (self->video_sid)
stream_list = g_list_append (stream_list, g_strdup (self->video_sid));
if (self->subtitle_sid)
stream_list = g_list_append (stream_list, g_strdup (self->subtitle_sid));
g_mutex_unlock (&self->lock);
if (stream_list) {
ret = gst_element_send_event (self->playbin,
gst_event_new_select_streams (stream_list));
g_list_free_full (stream_list, g_free);
} else {
GST_ERROR_OBJECT (self, "No available streams for select-streams");
}
g_mutex_lock (&self->lock);
return ret;
}
/**
* gst_play_set_audio_track:
* @play: #GstPlay instance
* @stream_index: stream index
*
* Returns: %TRUE or %FALSE
*
* Sets the audio track @stream_index.
* Since: 1.20
*/
gboolean
gst_play_set_audio_track (GstPlay * self, gint stream_index)
{
GstPlayStreamInfo *info;
gboolean ret = TRUE;
g_return_val_if_fail (GST_IS_PLAY (self), 0);
g_mutex_lock (&self->lock);
info = gst_play_stream_info_find (self->media_info,
GST_TYPE_PLAY_AUDIO_INFO, stream_index);
g_mutex_unlock (&self->lock);
if (!info) {
GST_ERROR_OBJECT (self, "invalid audio stream index %d", stream_index);
return FALSE;
}
if (self->use_playbin3) {
g_mutex_lock (&self->lock);
g_free (self->audio_sid);
self->audio_sid = g_strdup (info->stream_id);
ret = gst_play_select_streams (self);
g_mutex_unlock (&self->lock);
} else {
g_object_set (G_OBJECT (self->playbin), "current-audio", stream_index,
NULL);
}
GST_DEBUG_OBJECT (self, "set stream index '%d'", stream_index);
return ret;
}
/**
* gst_play_set_video_track:
* @play: #GstPlay instance
* @stream_index: stream index
*
* Returns: %TRUE or %FALSE
*
* Sets the video track @stream_index.
* Since: 1.20
*/
gboolean
gst_play_set_video_track (GstPlay * self, gint stream_index)
{
GstPlayStreamInfo *info;
gboolean ret = TRUE;
g_return_val_if_fail (GST_IS_PLAY (self), 0);
/* check if stream_index exist in our internal media_info list */
g_mutex_lock (&self->lock);
info = gst_play_stream_info_find (self->media_info,
GST_TYPE_PLAY_VIDEO_INFO, stream_index);
g_mutex_unlock (&self->lock);
if (!info) {
GST_ERROR_OBJECT (self, "invalid video stream index %d", stream_index);
return FALSE;
}
if (self->use_playbin3) {
g_mutex_lock (&self->lock);
g_free (self->video_sid);
self->video_sid = g_strdup (info->stream_id);
ret = gst_play_select_streams (self);
g_mutex_unlock (&self->lock);
} else {
g_object_set (G_OBJECT (self->playbin), "current-video", stream_index,
NULL);
}
GST_DEBUG_OBJECT (self, "set stream index '%d'", stream_index);
return ret;
}
/**
* gst_play_set_subtitle_track:
* @play: #GstPlay instance
* @stream_index: stream index
*
* Returns: %TRUE or %FALSE
*
* Sets the subtitle stack @stream_index.
* Since: 1.20
*/
gboolean
gst_play_set_subtitle_track (GstPlay * self, gint stream_index)
{
GstPlayStreamInfo *info;
gboolean ret = TRUE;
g_return_val_if_fail (GST_IS_PLAY (self), 0);
g_mutex_lock (&self->lock);
info = gst_play_stream_info_find (self->media_info,
GST_TYPE_PLAY_SUBTITLE_INFO, stream_index);
g_mutex_unlock (&self->lock);
if (!info) {
GST_ERROR_OBJECT (self, "invalid subtitle stream index %d", stream_index);
return FALSE;
}
if (self->use_playbin3) {
g_mutex_lock (&self->lock);
g_free (self->subtitle_sid);
self->subtitle_sid = g_strdup (info->stream_id);
ret = gst_play_select_streams (self);
g_mutex_unlock (&self->lock);
} else {
g_object_set (G_OBJECT (self->playbin), "current-text", stream_index, NULL);
}
GST_DEBUG_OBJECT (self, "set stream index '%d'", stream_index);
return ret;
}
/**
* gst_play_set_audio_track_enabled:
* @play: #GstPlay instance
* @enabled: TRUE or FALSE
*
* Enable or disable the current audio track.
* Since: 1.20
*/
void
gst_play_set_audio_track_enabled (GstPlay * self, gboolean enabled)
{
g_return_if_fail (GST_IS_PLAY (self));
if (enabled)
play_set_flag (self, GST_PLAY_FLAG_AUDIO);
else
play_clear_flag (self, GST_PLAY_FLAG_AUDIO);
GST_DEBUG_OBJECT (self, "track is '%s'", enabled ? "Enabled" : "Disabled");
}
/**
* gst_play_set_video_track_enabled:
* @play: #GstPlay instance
* @enabled: TRUE or FALSE
*
* Enable or disable the current video track.
* Since: 1.20
*/
void
gst_play_set_video_track_enabled (GstPlay * self, gboolean enabled)
{
g_return_if_fail (GST_IS_PLAY (self));
if (enabled)
play_set_flag (self, GST_PLAY_FLAG_VIDEO);
else
play_clear_flag (self, GST_PLAY_FLAG_VIDEO);
GST_DEBUG_OBJECT (self, "track is '%s'", enabled ? "Enabled" : "Disabled");
}
/**
* gst_play_set_subtitle_track_enabled:
* @play: #GstPlay instance
* @enabled: TRUE or FALSE
*
* Enable or disable the current subtitle track.
* Since: 1.20
*/
void
gst_play_set_subtitle_track_enabled (GstPlay * self, gboolean enabled)
{
g_return_if_fail (GST_IS_PLAY (self));
if (enabled)
play_set_flag (self, GST_PLAY_FLAG_SUBTITLE);
else
play_clear_flag (self, GST_PLAY_FLAG_SUBTITLE);
GST_DEBUG_OBJECT (self, "track is '%s'", enabled ? "Enabled" : "Disabled");
}
/**
* gst_play_set_visualization:
* @play: #GstPlay instance
* @name: (nullable): visualization element obtained from
* #gst_play_visualizations_get()
*
* Returns: %TRUE if the visualizations was set correctly. Otherwise,
* %FALSE.
* Since: 1.20
*/
gboolean
gst_play_set_visualization (GstPlay * self, const gchar * name)
{
g_return_val_if_fail (GST_IS_PLAY (self), FALSE);
g_mutex_lock (&self->lock);
if (self->current_vis_element) {
gst_object_unref (self->current_vis_element);
self->current_vis_element = NULL;
}
if (name) {
self->current_vis_element = gst_element_factory_make (name, NULL);
if (!self->current_vis_element)
goto error_no_element;
gst_object_ref_sink (self->current_vis_element);
}
g_object_set (self->playbin, "vis-plugin", self->current_vis_element, NULL);
g_mutex_unlock (&self->lock);
GST_DEBUG_OBJECT (self, "set vis-plugin to '%s'", name);
return TRUE;
error_no_element:
g_mutex_unlock (&self->lock);
GST_WARNING_OBJECT (self, "could not find visualization '%s'", name);
return FALSE;
}
/**
* gst_play_get_current_visualization:
* @play: #GstPlay instance
*
* Returns: (transfer full) (nullable): Name of the currently enabled
* visualization.
* g_free() after usage.
* Since: 1.20
*/
gchar *
gst_play_get_current_visualization (GstPlay * self)
{
gchar *name = NULL;
GstElement *vis_plugin = NULL;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
if (!is_track_enabled (self, GST_PLAY_FLAG_VIS))
return NULL;
g_object_get (self->playbin, "vis-plugin", &vis_plugin, NULL);
if (vis_plugin) {
GstElementFactory *factory = gst_element_get_factory (vis_plugin);
if (factory)
name = g_strdup (gst_plugin_feature_get_name (factory));
gst_object_unref (vis_plugin);
}
GST_DEBUG_OBJECT (self, "vis-plugin '%s' %p", name, vis_plugin);
return name;
}
/**
* gst_play_set_visualization_enabled:
* @play: #GstPlay instance
* @enabled: TRUE or FALSE
*
* Enable or disable the visualization.
* Since: 1.20
*/
void
gst_play_set_visualization_enabled (GstPlay * self, gboolean enabled)
{
g_return_if_fail (GST_IS_PLAY (self));
if (enabled)
play_set_flag (self, GST_PLAY_FLAG_VIS);
else
play_clear_flag (self, GST_PLAY_FLAG_VIS);
GST_DEBUG_OBJECT (self, "visualization is '%s'",
enabled ? "Enabled" : "Disabled");
}
struct CBChannelMap
{
const gchar *label; /* channel label name */
const gchar *name; /* get_name () */
};
static const struct CBChannelMap cb_channel_map[] = {
/* GST_PLAY_COLOR_BALANCE_BRIGHTNESS */ {"BRIGHTNESS", "brightness"},
/* GST_PLAY_COLOR_BALANCE_CONTRAST */ {"CONTRAST", "contrast"},
/* GST_PLAY_COLOR_BALANCE_SATURATION */ {"SATURATION", "saturation"},
/* GST_PLAY_COLOR_BALANCE_HUE */ {"HUE", "hue"},
};
static GstColorBalanceChannel *
gst_play_color_balance_find_channel (GstPlay * self,
GstPlayColorBalanceType type)
{
GstColorBalanceChannel *channel;
const GList *l, *channels;
if (type < GST_PLAY_COLOR_BALANCE_BRIGHTNESS ||
type > GST_PLAY_COLOR_BALANCE_HUE)
return NULL;
channels =
gst_color_balance_list_channels (GST_COLOR_BALANCE (self->playbin));
for (l = channels; l; l = l->next) {
channel = l->data;
if (g_strrstr (channel->label, cb_channel_map[type].label))
return channel;
}
return NULL;
}
/**
* gst_play_has_color_balance:
* @play:#GstPlay instance
*
* Checks whether the @play has color balance support available.
*
* Returns: %TRUE if @play has color balance support. Otherwise,
* %FALSE.
* Since: 1.20
*/
gboolean
gst_play_has_color_balance (GstPlay * self)
{
const GList *channels;
g_return_val_if_fail (GST_IS_PLAY (self), FALSE);
if (!GST_IS_COLOR_BALANCE (self->playbin))
return FALSE;
channels =
gst_color_balance_list_channels (GST_COLOR_BALANCE (self->playbin));
return (channels != NULL);
}
/**
* gst_play_set_color_balance:
* @play: #GstPlay instance
* @type: #GstPlayColorBalanceType
* @value: The new value for the @type, ranged [0,1]
*
* Sets the current value of the indicated channel @type to the passed
* value.
* Since: 1.20
*/
void
gst_play_set_color_balance (GstPlay * self, GstPlayColorBalanceType type,
gdouble value)
{
GstColorBalanceChannel *channel;
gdouble new_val;
g_return_if_fail (GST_IS_PLAY (self));
g_return_if_fail (value >= 0.0 && value <= 1.0);
if (!GST_IS_COLOR_BALANCE (self->playbin))
return;
channel = gst_play_color_balance_find_channel (self, type);
if (!channel)
return;
value = CLAMP (value, 0.0, 1.0);
/* Convert to channel range */
new_val = channel->min_value + value * ((gdouble) channel->max_value -
(gdouble) channel->min_value);
gst_color_balance_set_value (GST_COLOR_BALANCE (self->playbin), channel,
new_val);
}
/**
* gst_play_get_color_balance:
* @play: #GstPlay instance
* @type: #GstPlayColorBalanceType
*
* Retrieve the current value of the indicated @type.
*
* Returns: The current value of @type, between [0,1]. In case of
* error -1 is returned.
* Since: 1.20
*/
gdouble
gst_play_get_color_balance (GstPlay * self, GstPlayColorBalanceType type)
{
GstColorBalanceChannel *channel;
gint value;
g_return_val_if_fail (GST_IS_PLAY (self), -1);
if (!GST_IS_COLOR_BALANCE (self->playbin))
return -1;
channel = gst_play_color_balance_find_channel (self, type);
if (!channel)
return -1;
value = gst_color_balance_get_value (GST_COLOR_BALANCE (self->playbin),
channel);
return ((gdouble) value -
(gdouble) channel->min_value) / ((gdouble) channel->max_value -
(gdouble) channel->min_value);
}
/**
* gst_play_get_multiview_mode:
* @play: #GstPlay instance
*
* Retrieve the current value of the indicated @type.
*
* Returns: The current value of @type, Default: -1 "none"
*
* Since: 1.20
*/
GstVideoMultiviewFramePacking
gst_play_get_multiview_mode (GstPlay * self)
{
GstVideoMultiviewFramePacking val = GST_VIDEO_MULTIVIEW_FRAME_PACKING_NONE;
g_return_val_if_fail (GST_IS_PLAY (self),
GST_VIDEO_MULTIVIEW_FRAME_PACKING_NONE);
g_object_get (self, "video-multiview-mode", &val, NULL);
return val;
}
/**
* gst_play_set_multiview_mode:
* @play: #GstPlay instance
* @mode: The new value for the @type
*
* Sets the current value of the indicated mode @type to the passed
* value.
*
* Since: 1.20
*/
void
gst_play_set_multiview_mode (GstPlay * self, GstVideoMultiviewFramePacking mode)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "video-multiview-mode", mode, NULL);
}
/**
* gst_play_get_multiview_flags:
* @play: #GstPlay instance
*
* Retrieve the current value of the indicated @type.
*
* Returns: The current value of @type, Default: 0x00000000 "none
*
* Since: 1.20
*/
GstVideoMultiviewFlags
gst_play_get_multiview_flags (GstPlay * self)
{
GstVideoMultiviewFlags val = GST_VIDEO_MULTIVIEW_FLAGS_NONE;
g_return_val_if_fail (GST_IS_PLAY (self), val);
g_object_get (self, "video-multiview-flags", &val, NULL);
return val;
}
/**
* gst_play_set_multiview_flags:
* @play: #GstPlay instance
* @flags: The new value for the @type
*
* Sets the current value of the indicated mode @type to the passed
* value.
*
* Since: 1.20
*/
void
gst_play_set_multiview_flags (GstPlay * self, GstVideoMultiviewFlags flags)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "video-multiview-flags", flags, NULL);
}
/**
* gst_play_get_audio_video_offset:
* @play: #GstPlay instance
*
* Retrieve the current value of audio-video-offset property
*
* Returns: The current value of audio-video-offset in nanoseconds
*
* Since: 1.20
*/
gint64
gst_play_get_audio_video_offset (GstPlay * self)
{
gint64 val = 0;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_AUDIO_VIDEO_OFFSET);
g_object_get (self, "audio-video-offset", &val, NULL);
return val;
}
/**
* gst_play_set_audio_video_offset:
* @play: #GstPlay instance
* @offset: #gint64 in nanoseconds
*
* Sets audio-video-offset property by value of @offset
*
* Since: 1.20
*/
void
gst_play_set_audio_video_offset (GstPlay * self, gint64 offset)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "audio-video-offset", offset, NULL);
}
/**
* gst_play_get_subtitle_video_offset:
* @play: #GstPlay instance
*
* Retrieve the current value of subtitle-video-offset property
*
* Returns: The current value of subtitle-video-offset in nanoseconds
*
* Since: 1.20
*/
gint64
gst_play_get_subtitle_video_offset (GstPlay * self)
{
gint64 val = 0;
g_return_val_if_fail (GST_IS_PLAY (self), DEFAULT_SUBTITLE_VIDEO_OFFSET);
g_object_get (self, "subtitle-video-offset", &val, NULL);
return val;
}
/**
* gst_play_set_subtitle_video_offset:
* @play: #GstPlay instance
* @offset: #gint64 in nanoseconds
*
* Sets subtitle-video-offset property by value of @offset
*
* Since: 1.20
*/
void
gst_play_set_subtitle_video_offset (GstPlay * self, gint64 offset)
{
g_return_if_fail (GST_IS_PLAY (self));
g_object_set (self, "subtitle-video-offset", offset, NULL);
}
#define C_ENUM(v) ((gint) v)
#define C_FLAGS(v) ((guint) v)
GType
gst_play_color_balance_type_get_type (void)
{
static gsize id = 0;
static const GEnumValue values[] = {
{C_ENUM (GST_PLAY_COLOR_BALANCE_HUE), "GST_PLAY_COLOR_BALANCE_HUE",
"hue"},
{C_ENUM (GST_PLAY_COLOR_BALANCE_BRIGHTNESS),
"GST_PLAY_COLOR_BALANCE_BRIGHTNESS", "brightness"},
{C_ENUM (GST_PLAY_COLOR_BALANCE_SATURATION),
"GST_PLAY_COLOR_BALANCE_SATURATION", "saturation"},
{C_ENUM (GST_PLAY_COLOR_BALANCE_CONTRAST),
"GST_PLAY_COLOR_BALANCE_CONTRAST", "contrast"},
{0, NULL, NULL}
};
if (g_once_init_enter (&id)) {
GType tmp = g_enum_register_static ("GstPlayColorBalanceType", values);
g_once_init_leave (&id, tmp);
}
return (GType) id;
}
/**
* gst_play_color_balance_type_get_name:
* @type: a #GstPlayColorBalanceType
*
* Gets a string representing the given color balance type.
*
* Returns: (transfer none): a string with the name of the color
* balance type.
* Since: 1.20
*/
const gchar *
gst_play_color_balance_type_get_name (GstPlayColorBalanceType type)
{
g_return_val_if_fail (type >= GST_PLAY_COLOR_BALANCE_BRIGHTNESS &&
type <= GST_PLAY_COLOR_BALANCE_HUE, NULL);
return cb_channel_map[type].name;
}
GType
gst_play_state_get_type (void)
{
static gsize id = 0;
static const GEnumValue values[] = {
{C_ENUM (GST_PLAY_STATE_STOPPED), "GST_PLAY_STATE_STOPPED", "stopped"},
{C_ENUM (GST_PLAY_STATE_BUFFERING), "GST_PLAY_STATE_BUFFERING",
"buffering"},
{C_ENUM (GST_PLAY_STATE_PAUSED), "GST_PLAY_STATE_PAUSED", "paused"},
{C_ENUM (GST_PLAY_STATE_PLAYING), "GST_PLAY_STATE_PLAYING", "playing"},
{0, NULL, NULL}
};
if (g_once_init_enter (&id)) {
GType tmp = g_enum_register_static ("GstPlayState", values);
g_once_init_leave (&id, tmp);
}
return (GType) id;
}
GType
gst_play_message_get_type (void)
{
static gsize id = 0;
static const GEnumValue values[] = {
{C_ENUM (GST_PLAY_MESSAGE_URI_LOADED), "GST_PLAY_MESSAGE_URI_LOADED",
"uri-loaded"},
{C_ENUM (GST_PLAY_MESSAGE_POSITION_UPDATED),
"GST_PLAY_MESSAGE_POSITION_UPDATED", "position-updated"},
{C_ENUM (GST_PLAY_MESSAGE_DURATION_CHANGED),
"GST_PLAY_MESSAGE_DURATION_CHANGED", "duration-changed"},
{C_ENUM (GST_PLAY_MESSAGE_STATE_CHANGED),
"GST_PLAY_MESSAGE_STATE_CHANGED", "state-changed"},
{C_ENUM (GST_PLAY_MESSAGE_BUFFERING), "GST_PLAY_MESSAGE_BUFFERING",
"buffering"},
{C_ENUM (GST_PLAY_MESSAGE_END_OF_STREAM),
"GST_PLAY_MESSAGE_END_OF_STREAM", "end-of-stream"},
{C_ENUM (GST_PLAY_MESSAGE_ERROR), "GST_PLAY_MESSAGE_ERROR", "error"},
{C_ENUM (GST_PLAY_MESSAGE_WARNING), "GST_PLAY_MESSAGE_WARNING",
"warning"},
{C_ENUM (GST_PLAY_MESSAGE_VIDEO_DIMENSIONS_CHANGED),
"GST_PLAY_MESSAGE_VIDEO_DIMENSIONS_CHANGED",
"video-dimensions-changed"},
{C_ENUM (GST_PLAY_MESSAGE_MEDIA_INFO_UPDATED),
"GST_PLAY_MESSAGE_MEDIA_INFO_UPDATED", "media-info-updated"},
{C_ENUM (GST_PLAY_MESSAGE_VOLUME_CHANGED),
"GST_PLAY_MESSAGE_VOLUME_CHANGED", "volume-changed"},
{C_ENUM (GST_PLAY_MESSAGE_MUTE_CHANGED),
"GST_PLAY_MESSAGE_MUTE_CHANGED", "mute-changed"},
{C_ENUM (GST_PLAY_MESSAGE_SEEK_DONE), "GST_PLAY_MESSAGE_SEEK_DONE",
"seek-done"},
{0, NULL, NULL}
};
if (g_once_init_enter (&id)) {
GType tmp = g_enum_register_static ("GstPlayMessage", values);
g_once_init_leave (&id, tmp);
}
return (GType) id;
}
/**
* gst_play_state_get_name:
* @state: a #GstPlayState
*
* Gets a string representing the given state.
*
* Returns: (transfer none): a string with the name of the state.
* Since: 1.20
*/
const gchar *
gst_play_state_get_name (GstPlayState state)
{
switch (state) {
case GST_PLAY_STATE_STOPPED:
return "stopped";
case GST_PLAY_STATE_BUFFERING:
return "buffering";
case GST_PLAY_STATE_PAUSED:
return "paused";
case GST_PLAY_STATE_PLAYING:
return "playing";
}
g_assert_not_reached ();
return NULL;
}
/**
* gst_play_message_get_name:
* @message_type: a #GstPlayMessage
*
* Returns: (transfer none): a string with the name of the message.
* Since: 1.20
*/
const gchar *
gst_play_message_get_name (GstPlayMessage message_type)
{
GEnumClass *enum_class;
GEnumValue *enum_value;
enum_class = g_type_class_ref (GST_TYPE_PLAY_MESSAGE);
enum_value = g_enum_get_value (enum_class, message_type);
g_assert (enum_value != NULL);
g_type_class_unref (enum_class);
return enum_value->value_name;
}
GType
gst_play_error_get_type (void)
{
static gsize id = 0;
static const GEnumValue values[] = {
{C_ENUM (GST_PLAY_ERROR_FAILED), "GST_PLAY_ERROR_FAILED", "failed"},
{0, NULL, NULL}
};
if (g_once_init_enter (&id)) {
GType tmp = g_enum_register_static ("GstPlayError", values);
g_once_init_leave (&id, tmp);
}
return (GType) id;
}
/**
* gst_play_error_get_name:
* @error: a #GstPlayError
*
* Gets a string representing the given error.
*
* Returns: (transfer none): a string with the given error.
* Since: 1.20
*/
const gchar *
gst_play_error_get_name (GstPlayError error)
{
switch (error) {
case GST_PLAY_ERROR_FAILED:
return "failed";
}
g_assert_not_reached ();
return NULL;
}
/**
* gst_play_set_config:
* @play: #GstPlay instance
* @config: (transfer full): a #GstStructure
*
* Set the configuration of the play. If the play is already configured, and
* the configuration haven't change, this function will return %TRUE. If the
* play is not in the GST_PLAY_STATE_STOPPED, this method will return %FALSE
* and active configuration will remain.
*
* @config is a #GstStructure that contains the configuration parameters for
* the play.
*
* This function takes ownership of @config.
*
* Returns: %TRUE when the configuration could be set.
* Since: 1.20
*/
gboolean
gst_play_set_config (GstPlay * self, GstStructure * config)
{
g_return_val_if_fail (GST_IS_PLAY (self), FALSE);
g_return_val_if_fail (config != NULL, FALSE);
g_mutex_lock (&self->lock);
if (self->app_state != GST_PLAY_STATE_STOPPED) {
GST_INFO_OBJECT (self, "can't change config while play is %s",
gst_play_state_get_name (self->app_state));
g_mutex_unlock (&self->lock);
return FALSE;
}
if (self->config)
gst_structure_free (self->config);
self->config = config;
g_mutex_unlock (&self->lock);
return TRUE;
}
/**
* gst_play_get_config:
* @play: #GstPlay instance
*
* Get a copy of the current configuration of the play. This configuration
* can either be modified and used for the gst_play_set_config() call
* or it must be freed after usage.
*
* Returns: (transfer full): a copy of the current configuration of @play. Use
* gst_structure_free() after usage or gst_play_set_config().
*
* Since: 1.20
*/
GstStructure *
gst_play_get_config (GstPlay * self)
{
GstStructure *ret;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
g_mutex_lock (&self->lock);
ret = gst_structure_copy (self->config);
g_mutex_unlock (&self->lock);
return ret;
}
/**
* gst_play_config_set_user_agent:
* @config: a #GstPlay configuration
* @agent: (nullable): the string to use as user agent
*
* Set the user agent to pass to the server if @play needs to connect
* to a server during playback. This is typically used when playing HTTP
* or RTSP streams.
*
* Since: 1.20
*/
void
gst_play_config_set_user_agent (GstStructure * config, const gchar * agent)
{
g_return_if_fail (config != NULL);
g_return_if_fail (agent != NULL);
gst_structure_id_set (config,
CONFIG_QUARK (USER_AGENT), G_TYPE_STRING, agent, NULL);
}
/**
* gst_play_config_get_user_agent:
* @config: a #GstPlay configuration
*
* Return the user agent which has been configured using
* gst_play_config_set_user_agent() if any.
*
* Returns: (transfer full) (nullable): the configured agent, or %NULL
* Since: 1.20
*/
gchar *
gst_play_config_get_user_agent (const GstStructure * config)
{
gchar *agent = NULL;
g_return_val_if_fail (config != NULL, NULL);
gst_structure_id_get (config,
CONFIG_QUARK (USER_AGENT), G_TYPE_STRING, &agent, NULL);
return agent;
}
/**
* gst_play_config_set_position_update_interval:
* @config: a #GstPlay configuration
* @interval: interval in ms
*
* set desired interval in milliseconds between two position-updated messages.
* pass 0 to stop updating the position.
* Since: 1.20
*/
void
gst_play_config_set_position_update_interval (GstStructure * config,
guint interval)
{
g_return_if_fail (config != NULL);
g_return_if_fail (interval <= 10000);
gst_structure_id_set (config,
CONFIG_QUARK (POSITION_INTERVAL_UPDATE), G_TYPE_UINT, interval, NULL);
}
/**
* gst_play_config_get_position_update_interval:
* @config: a #GstPlay configuration
*
* Returns: current position update interval in milliseconds
*
* Since: 1.20
*/
guint
gst_play_config_get_position_update_interval (const GstStructure * config)
{
guint interval = DEFAULT_POSITION_UPDATE_INTERVAL_MS;
g_return_val_if_fail (config != NULL, DEFAULT_POSITION_UPDATE_INTERVAL_MS);
gst_structure_id_get (config,
CONFIG_QUARK (POSITION_INTERVAL_UPDATE), G_TYPE_UINT, &interval, NULL);
return interval;
}
/**
* gst_play_config_set_seek_accurate:
* @config: a #GstPlay configuration
* @accurate: accurate seek or not
*
* Enable or disable accurate seeking. When enabled, elements will try harder
* to seek as accurately as possible to the requested seek position. Generally
* it will be slower especially for formats that don't have any indexes or
* timestamp markers in the stream.
*
* If accurate seeking is disabled, elements will seek as close as the request
* position without slowing down seeking too much.
*
* Accurate seeking is disabled by default.
*
* Since: 1.20
*/
void
gst_play_config_set_seek_accurate (GstStructure * config, gboolean accurate)
{
g_return_if_fail (config != NULL);
gst_structure_id_set (config,
CONFIG_QUARK (ACCURATE_SEEK), G_TYPE_BOOLEAN, accurate, NULL);
}
/**
* gst_play_config_get_seek_accurate:
* @config: a #GstPlay configuration
*
* Returns: %TRUE if accurate seeking is enabled
*
* Since: 1.20
*/
gboolean
gst_play_config_get_seek_accurate (const GstStructure * config)
{
gboolean accurate = FALSE;
g_return_val_if_fail (config != NULL, FALSE);
gst_structure_id_get (config,
CONFIG_QUARK (ACCURATE_SEEK), G_TYPE_BOOLEAN, &accurate, NULL);
return accurate;
}
/**
* gst_play_get_video_snapshot:
* @play: #GstPlay instance
* @format: output format of the video snapshot
* @config: (allow-none): Additional configuration
*
* Get a snapshot of the currently selected video stream, if any. The format can be
* selected with @format and optional configuration is possible with @config
* Currently supported settings are:
* - width, height of type G_TYPE_INT
* - pixel-aspect-ratio of type GST_TYPE_FRACTION
* Except for GST_PLAY_THUMBNAIL_RAW_NATIVE format, if no config is set, pixel-aspect-ratio would be 1/1
*
* Returns: (transfer full) (nullable): Current video snapshot sample or %NULL on failure
*
* Since: 1.20
*/
GstSample *
gst_play_get_video_snapshot (GstPlay * self,
GstPlaySnapshotFormat format, const GstStructure * config)
{
gint video_tracks = 0;
GstSample *sample = NULL;
GstCaps *caps = NULL;
gint width = -1;
gint height = -1;
gint par_n = 1;
gint par_d = 1;
g_return_val_if_fail (GST_IS_PLAY (self), NULL);
g_object_get (self->playbin, "n-video", &video_tracks, NULL);
if (video_tracks == 0) {
GST_DEBUG_OBJECT (self, "total video track num is 0");
return NULL;
}
switch (format) {
case GST_PLAY_THUMBNAIL_RAW_xRGB:
caps = gst_caps_new_simple ("video/x-raw",
"format", G_TYPE_STRING, "xRGB", NULL);
break;
case GST_PLAY_THUMBNAIL_RAW_BGRx:
caps = gst_caps_new_simple ("video/x-raw",
"format", G_TYPE_STRING, "BGRx", NULL);
break;
case GST_PLAY_THUMBNAIL_JPG:
caps = gst_caps_new_empty_simple ("image/jpeg");
break;
case GST_PLAY_THUMBNAIL_PNG:
caps = gst_caps_new_empty_simple ("image/png");
break;
case GST_PLAY_THUMBNAIL_RAW_NATIVE:
default:
caps = gst_caps_new_empty_simple ("video/x-raw");
break;
}
if (NULL != config) {
if (!gst_structure_get_int (config, "width", &width))
width = -1;
if (!gst_structure_get_int (config, "height", &height))
height = -1;
if (!gst_structure_get_fraction (config, "pixel-aspect-ratio", &par_n,
&par_d)) {
if (format != GST_PLAY_THUMBNAIL_RAW_NATIVE) {
par_n = 1;
par_d = 1;
} else {
par_n = 0;
par_d = 0;
}
}
}
if (width > 0 && height > 0) {
gst_caps_set_simple (caps, "width", G_TYPE_INT, width,
"height", G_TYPE_INT, height, NULL);
}
if (format != GST_PLAY_THUMBNAIL_RAW_NATIVE) {
gst_caps_set_simple (caps, "pixel-aspect-ratio", GST_TYPE_FRACTION,
par_n, par_d, NULL);
} else if (NULL != config && par_n != 0 && par_d != 0) {
gst_caps_set_simple (caps, "pixel-aspect-ratio", GST_TYPE_FRACTION,
par_n, par_d, NULL);
}
g_signal_emit_by_name (self->playbin, "convert-sample", caps, &sample);
gst_caps_unref (caps);
if (!sample) {
GST_WARNING_OBJECT (self, "Failed to retrieve or convert video frame");
return NULL;
}
return sample;
}
/**
* gst_play_is_play_message:
* @msg: A #GstMessage
*
* Returns: A #gboolean indicating wheter the passes message represents a #GstPlay message or not.
*
* Since: 1.20
*/
gboolean
gst_play_is_play_message (GstMessage * msg)
{
const GstStructure *data = NULL;
g_return_val_if_fail (GST_IS_MESSAGE (msg), FALSE);
data = gst_message_get_structure (msg);
g_return_val_if_fail (data, FALSE);
return g_str_equal (gst_structure_get_name (data), GST_PLAY_MESSAGE_DATA);
}
#define PARSE_MESSAGE_FIELD(msg, field, value_type, value) G_STMT_START { \
const GstStructure *data = NULL; \
g_return_if_fail (gst_play_is_play_message (msg)); \
data = gst_message_get_structure (msg); \
gst_structure_get (data, field, value_type, value, NULL); \
} G_STMT_END
/**
* gst_play_message_parse_type:
* @msg: A #GstMessage
* @type: (out) (optional): the resulting message type
*
* Parse the given @msg and extract its #GstPlayMessage type.
*
* Since: 1.20
*/
void
gst_play_message_parse_type (GstMessage * msg, GstPlayMessage * type)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_TYPE,
GST_TYPE_PLAY_MESSAGE, type);
}
/**
* gst_play_message_parse_duration_updated:
* @msg: A #GstMessage
* @duration: (out) (optional): the resulting duration
*
* Parse the given duration @msg and extract the corresponding #GstClockTime
*
* Since: 1.20
*/
void
gst_play_message_parse_duration_updated (GstMessage * msg,
GstClockTime * duration)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_DURATION,
GST_TYPE_CLOCK_TIME, duration);
}
/**
* gst_play_message_parse_position_updated:
* @msg: A #GstMessage
* @position: (out) (optional): the resulting position
*
* Parse the given position @msg and extract the corresponding #GstClockTime
*
* Since: 1.20
*/
void
gst_play_message_parse_position_updated (GstMessage * msg,
GstClockTime * position)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_POSITION,
GST_TYPE_CLOCK_TIME, position);
}
/**
* gst_play_message_parse_state_changed:
* @msg: A #GstMessage
* @state: (out) (optional): the resulting play state
*
* Parse the given state @msg and extract the corresponding #GstPlayState
*
* Since: 1.20
*/
void
gst_play_message_parse_state_changed (GstMessage * msg, GstPlayState * state)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_PLAY_STATE,
GST_TYPE_PLAY_STATE, state);
}
/**
* gst_play_message_parse_buffering_percent:
* @msg: A #GstMessage
* @percent: (out) (optional): the resulting buffering percent
*
* Parse the given buffering-percent @msg and extract the corresponding value
*
* Since: 1.20
*/
void
gst_play_message_parse_buffering_percent (GstMessage * msg, guint * percent)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_BUFFERING_PERCENT,
G_TYPE_UINT, percent);
}
/**
* gst_play_message_parse_error:
* @msg: A #GstMessage
* @error: (out) (optional) (transfer full): the resulting error
* @details: (out) (optional) (nullable) (transfer full): A #GstStructure containing additional details about the error
*
* Parse the given error @msg and extract the corresponding #GError
*
* Since: 1.20
*/
void
gst_play_message_parse_error (GstMessage * msg, GError ** error,
GstStructure ** details)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_ERROR, G_TYPE_ERROR, error);
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_ERROR, GST_TYPE_STRUCTURE,
details);
}
/**
* gst_play_message_parse_warning:
* @msg: A #GstMessage
* @error: (out) (optional) (transfer full): the resulting warning
* @details: (out) (optional) (nullable) (transfer full): A #GstStructure containing additional details about the warning
*
* Parse the given error @msg and extract the corresponding #GError warning
*
* Since: 1.20
*/
void
gst_play_message_parse_warning (GstMessage * msg, GError ** error,
GstStructure ** details)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_WARNING, G_TYPE_ERROR, error);
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_WARNING, GST_TYPE_STRUCTURE,
details);
}
/**
* gst_play_message_parse_video_dimensions_changed:
* @msg: A #GstMessage
* @width: (out) (optional): the resulting video width
* @height: (out) (optional): the resulting video height
*
* Parse the given @msg and extract the corresponding video dimensions
*
* Since: 1.20
*/
void
gst_play_message_parse_video_dimensions_changed (GstMessage * msg,
guint * width, guint * height)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_VIDEO_WIDTH,
G_TYPE_UINT, width);
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_VIDEO_HEIGHT,
G_TYPE_UINT, height);
}
/**
* gst_play_message_parse_media_info_updated:
* @msg: A #GstMessage
* @info: (out) (optional) (transfer full): the resulting media info
*
* Parse the given @msg and extract the corresponding media information
*
* Since: 1.20
*/
void
gst_play_message_parse_media_info_updated (GstMessage * msg,
GstPlayMediaInfo ** info)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_MEDIA_INFO,
GST_TYPE_PLAY_MEDIA_INFO, info);
}
/**
* gst_play_message_parse_volume_changed:
* @msg: A #GstMessage
* @volume: (out) (optional): the resulting audio volume
*
* Parse the given @msg and extract the corresponding audio volume
*
* Since: 1.20
*/
void
gst_play_message_parse_volume_changed (GstMessage * msg, gdouble * volume)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_VOLUME, G_TYPE_DOUBLE,
volume);
}
/**
* gst_play_message_parse_muted_changed:
* @msg: A #GstMessage
* @muted: (out) (optional): the resulting audio muted state
*
* Parse the given @msg and extract the corresponding audio muted state
*
* Since: 1.20
*/
void
gst_play_message_parse_muted_changed (GstMessage * msg, gboolean * muted)
{
PARSE_MESSAGE_FIELD (msg, GST_PLAY_MESSAGE_DATA_IS_MUTED, G_TYPE_BOOLEAN,
muted);
}