mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-19 00:01:23 +00:00
1152 lines
35 KiB
C
1152 lines
35 KiB
C
/* GStreamer DVB subtitles overlay
|
|
* Copyright (c) 2010 Mart Raudsepp <mart.raudsepp@collabora.co.uk>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
|
|
* Boston, MA 02111-1307, USA.
|
|
*/
|
|
|
|
/**
|
|
* SECTION:element-dvbsuboverlay
|
|
*
|
|
* Renders DVB subtitles on top of a video stream.
|
|
*
|
|
* <refsect2>
|
|
* <title>Example launch line</title>
|
|
* |[ FIXME
|
|
* gst-launch -v filesrc location=/path/to/ts ! mpegtsdemux name=d ! queue ! mp3parse ! mad ! audioconvert ! autoaudiosink \
|
|
* d. ! queue ! mpeg2dec ! ffmpegcolorspace ! r. \
|
|
* d. ! queue ! "private/x-dvbsub" ! dvbsuboverlay name=r ! ffmpegcolorspace ! autovideosink
|
|
* ]| This pipeline demuxes a MPEG-TS file with MPEG2 video, MP3 audio and embedded DVB subtitles and renders the subtitles on top of the video.
|
|
* </refsect2>
|
|
*/
|
|
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
# include <config.h>
|
|
#endif
|
|
|
|
#include "gstdvbsuboverlay.h"
|
|
|
|
#include <string.h>
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (gst_dvbsub_overlay_debug);
|
|
GST_DEBUG_CATEGORY_STATIC (gst_dvbsub_overlay_lib_debug);
|
|
#define GST_CAT_DEFAULT gst_dvbsub_overlay_debug
|
|
|
|
/* Filter signals and props */
|
|
enum
|
|
{
|
|
LAST_SIGNAL
|
|
};
|
|
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_ENABLE
|
|
};
|
|
|
|
static GstStaticPadTemplate src_factory = GST_STATIC_PAD_TEMPLATE ("src",
|
|
GST_PAD_SRC,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("I420"))
|
|
);
|
|
|
|
static GstStaticPadTemplate video_sink_factory =
|
|
GST_STATIC_PAD_TEMPLATE ("video_sink",
|
|
GST_PAD_SINK,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("I420"))
|
|
);
|
|
|
|
static GstStaticPadTemplate text_sink_factory =
|
|
GST_STATIC_PAD_TEMPLATE ("text_sink",
|
|
GST_PAD_SINK,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS ("private/x-dvbsub")
|
|
);
|
|
|
|
static void gst_dvbsub_overlay_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec);
|
|
static void gst_dvbsub_overlay_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec);
|
|
|
|
static void gst_dvbsub_overlay_finalize (GObject * object);
|
|
|
|
static GstStateChangeReturn gst_dvbsub_overlay_change_state (GstElement *
|
|
element, GstStateChange transition);
|
|
|
|
GST_BOILERPLATE (GstDVBSubOverlay, gst_dvbsub_overlay, GstElement,
|
|
GST_TYPE_ELEMENT);
|
|
|
|
static GstCaps *gst_dvbsub_overlay_getcaps (GstPad * pad);
|
|
|
|
static gboolean gst_dvbsub_overlay_setcaps_video (GstPad * pad, GstCaps * caps);
|
|
|
|
static GstFlowReturn gst_dvbsub_overlay_chain_video (GstPad * pad,
|
|
GstBuffer * buf);
|
|
static GstFlowReturn gst_dvbsub_overlay_chain_text (GstPad * pad,
|
|
GstBuffer * buf);
|
|
|
|
static gboolean gst_dvbsub_overlay_event_video (GstPad * pad, GstEvent * event);
|
|
static gboolean gst_dvbsub_overlay_event_text (GstPad * pad, GstEvent * event);
|
|
static gboolean gst_dvbsub_overlay_event_src (GstPad * pad, GstEvent * event);
|
|
|
|
static void new_dvb_subtitles_cb (DvbSub * dvb_sub, DVBSubtitles * subs,
|
|
gpointer user_data);
|
|
|
|
static GstFlowReturn gst_dvbsub_overlay_bufferalloc_video (GstPad * pad,
|
|
guint64 offset, guint size, GstCaps * caps, GstBuffer ** buffer);
|
|
|
|
static gboolean gst_dvbsub_overlay_query_src (GstPad * pad, GstQuery * query);
|
|
|
|
static void
|
|
gst_dvbsub_overlay_base_init (gpointer gclass)
|
|
{
|
|
GstElementClass *element_class = (GstElementClass *) gclass;
|
|
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&src_factory));
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&video_sink_factory));
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&text_sink_factory));
|
|
|
|
gst_element_class_set_details_simple (element_class,
|
|
"DVB Subtitles Overlay",
|
|
"Mixer/Video/Overlay/Subtitle",
|
|
"Renders DVB subtitles", "Mart Raudsepp <mart.raudsepp@collabora.co.uk>");
|
|
}
|
|
|
|
/* initialize the plugin's class */
|
|
static void
|
|
gst_dvbsub_overlay_class_init (GstDVBSubOverlayClass * klass)
|
|
{
|
|
GObjectClass *gobject_class = (GObjectClass *) klass;
|
|
GstElementClass *gstelement_class = (GstElementClass *) klass;
|
|
|
|
gobject_class->set_property = gst_dvbsub_overlay_set_property;
|
|
gobject_class->get_property = gst_dvbsub_overlay_get_property;
|
|
gobject_class->finalize = gst_dvbsub_overlay_finalize;
|
|
|
|
g_object_class_install_property (gobject_class, PROP_ENABLE, g_param_spec_boolean ("enable", "Enable", /* FIXME: "enable" vs "silent"? */
|
|
"Enable rendering of subtitles", TRUE,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
|
|
gstelement_class->change_state =
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_change_state);
|
|
}
|
|
|
|
static void
|
|
_dvbsub_log_cb (GLogLevelFlags level, const gchar * fmt, va_list args,
|
|
gpointer render)
|
|
{
|
|
gchar *message = g_strdup_vprintf (fmt, args);
|
|
|
|
if (level & G_LOG_LEVEL_ERROR)
|
|
GST_CAT_ERROR (gst_dvbsub_overlay_lib_debug, "%s", message);
|
|
else if (level & G_LOG_LEVEL_WARNING)
|
|
GST_CAT_WARNING (gst_dvbsub_overlay_lib_debug, "%s", message);
|
|
else if (level & G_LOG_LEVEL_INFO)
|
|
GST_CAT_INFO (gst_dvbsub_overlay_lib_debug, "%s", message);
|
|
else if (level & G_LOG_LEVEL_DEBUG)
|
|
GST_CAT_DEBUG (gst_dvbsub_overlay_lib_debug, "%s", message);
|
|
else
|
|
GST_CAT_LOG (gst_dvbsub_overlay_lib_debug,
|
|
"log level %d: %s", level, message);
|
|
|
|
g_free (message);
|
|
}
|
|
|
|
static void
|
|
gst_dvbsub_overlay_flush_subtitles (GstDVBSubOverlay * render)
|
|
{
|
|
DVBSubtitles *subs;
|
|
|
|
g_mutex_lock (render->dvbsub_mutex);
|
|
while ((subs = g_queue_pop_head (render->pending_subtitles))) {
|
|
dvb_subtitles_free (subs);
|
|
}
|
|
|
|
if (render->dvb_sub)
|
|
g_object_unref (render->dvb_sub);
|
|
render->dvb_sub = dvb_sub_new ();
|
|
if (!render->dvb_sub) {
|
|
GST_WARNING_OBJECT (render, "cannot create dvbsub instance");
|
|
g_assert_not_reached ();
|
|
}
|
|
|
|
{
|
|
DvbSubCallbacks dvbsub_callbacks = { &new_dvb_subtitles_cb, };
|
|
dvb_sub_set_callbacks (render->dvb_sub, &dvbsub_callbacks, render);
|
|
}
|
|
|
|
g_mutex_unlock (render->dvbsub_mutex);
|
|
}
|
|
|
|
static void
|
|
gst_dvbsub_overlay_init (GstDVBSubOverlay * render,
|
|
GstDVBSubOverlayClass * gclass)
|
|
{
|
|
GST_DEBUG_OBJECT (render, "init");
|
|
|
|
render->srcpad = gst_pad_new_from_static_template (&src_factory, "src");
|
|
render->video_sinkpad =
|
|
gst_pad_new_from_static_template (&video_sink_factory, "video_sink");
|
|
render->text_sinkpad =
|
|
gst_pad_new_from_static_template (&text_sink_factory, "text_sink");
|
|
|
|
gst_pad_set_setcaps_function (render->video_sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_setcaps_video));
|
|
|
|
gst_pad_set_getcaps_function (render->srcpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_getcaps));
|
|
gst_pad_set_getcaps_function (render->video_sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_getcaps));
|
|
|
|
gst_pad_set_chain_function (render->video_sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_chain_video));
|
|
gst_pad_set_chain_function (render->text_sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_chain_text));
|
|
|
|
gst_pad_set_event_function (render->video_sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_event_video));
|
|
gst_pad_set_event_function (render->text_sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_event_text));
|
|
gst_pad_set_event_function (render->srcpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_event_src));
|
|
|
|
gst_pad_set_bufferalloc_function (render->video_sinkpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_bufferalloc_video));
|
|
|
|
gst_pad_set_query_function (render->srcpad,
|
|
GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_query_src));
|
|
|
|
gst_element_add_pad (GST_ELEMENT (render), render->srcpad);
|
|
gst_element_add_pad (GST_ELEMENT (render), render->video_sinkpad);
|
|
gst_element_add_pad (GST_ELEMENT (render), render->text_sinkpad);
|
|
|
|
render->width = 0;
|
|
render->height = 0;
|
|
|
|
render->current_subtitle = NULL;
|
|
render->pending_subtitles = g_queue_new ();
|
|
|
|
render->enable = TRUE;
|
|
|
|
render->dvbsub_mutex = g_mutex_new ();
|
|
gst_dvbsub_overlay_flush_subtitles (render);
|
|
|
|
gst_segment_init (&render->video_segment, GST_FORMAT_TIME);
|
|
gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME);
|
|
|
|
GST_DEBUG_OBJECT (render, "init complete");
|
|
}
|
|
|
|
static void
|
|
gst_dvbsub_overlay_finalize (GObject * object)
|
|
{
|
|
GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (object);
|
|
DVBSubtitles *subs;
|
|
|
|
while ((subs = g_queue_pop_head (overlay->pending_subtitles))) {
|
|
dvb_subtitles_free (subs);
|
|
}
|
|
g_queue_free (overlay->pending_subtitles);
|
|
|
|
if (overlay->dvb_sub) {
|
|
g_object_unref (overlay->dvb_sub);
|
|
}
|
|
|
|
if (overlay->dvbsub_mutex)
|
|
g_mutex_free (overlay->dvbsub_mutex);
|
|
|
|
G_OBJECT_CLASS (parent_class)->finalize (object);
|
|
}
|
|
|
|
static void
|
|
gst_dvbsub_overlay_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_ENABLE:
|
|
overlay->enable = g_value_get_boolean (value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_dvbsub_overlay_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_ENABLE:
|
|
g_value_set_boolean (value, overlay->enable);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static GstStateChangeReturn
|
|
gst_dvbsub_overlay_change_state (GstElement * element,
|
|
GstStateChange transition)
|
|
{
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (element);
|
|
GstStateChangeReturn ret;
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_READY_TO_PAUSED:
|
|
gst_segment_init (&render->video_segment, GST_FORMAT_TIME);
|
|
gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME);
|
|
break;
|
|
case GST_STATE_CHANGE_NULL_TO_READY:
|
|
case GST_STATE_CHANGE_PAUSED_TO_PLAYING:
|
|
default:
|
|
break;
|
|
}
|
|
|
|
ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_PAUSED_TO_READY:
|
|
gst_dvbsub_overlay_flush_subtitles (render);
|
|
gst_segment_init (&render->video_segment, GST_FORMAT_TIME);
|
|
gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME);
|
|
render->format = GST_VIDEO_FORMAT_UNKNOWN;
|
|
break;
|
|
case GST_STATE_CHANGE_PLAYING_TO_PAUSED:
|
|
case GST_STATE_CHANGE_READY_TO_NULL:
|
|
default:
|
|
break;
|
|
}
|
|
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_dvbsub_overlay_query_src (GstPad * pad, GstQuery * query)
|
|
{
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad));
|
|
gboolean ret;
|
|
|
|
ret = gst_pad_peer_query (render->video_sinkpad, query);
|
|
|
|
gst_object_unref (render);
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_dvbsub_overlay_event_src (GstPad * pad, GstEvent * event)
|
|
{
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad));
|
|
gboolean ret = FALSE;
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_SEEK:{
|
|
GstSeekFlags flags;
|
|
|
|
GST_DEBUG_OBJECT (render, "seek received, driving from here");
|
|
|
|
gst_event_parse_seek (event, NULL, NULL, &flags, NULL, NULL, NULL, NULL);
|
|
|
|
/* Flush downstream, only for flushing seek */
|
|
if (flags & GST_SEEK_FLAG_FLUSH)
|
|
gst_pad_push_event (render->srcpad, gst_event_new_flush_start ());
|
|
|
|
gst_dvbsub_overlay_flush_subtitles (render);
|
|
|
|
/* Seek on each sink pad */
|
|
gst_event_ref (event);
|
|
ret = gst_pad_push_event (render->video_sinkpad, event);
|
|
if (ret) {
|
|
ret = gst_pad_push_event (render->text_sinkpad, event);
|
|
} else {
|
|
gst_event_unref (event);
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
gst_event_ref (event);
|
|
ret = gst_pad_push_event (render->video_sinkpad, event);
|
|
gst_pad_push_event (render->text_sinkpad, event);
|
|
break;
|
|
}
|
|
|
|
gst_object_unref (render);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_dvbsub_overlay_getcaps (GstPad * pad)
|
|
{
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad));
|
|
GstPad *otherpad;
|
|
GstCaps *caps;
|
|
|
|
if (pad == render->srcpad)
|
|
otherpad = render->video_sinkpad;
|
|
else
|
|
otherpad = render->srcpad;
|
|
|
|
/* we can do what the peer can */
|
|
caps = gst_pad_peer_get_caps (otherpad);
|
|
if (caps) {
|
|
GstCaps *temp;
|
|
const GstCaps *templ;
|
|
|
|
/* filtered against our padtemplate */
|
|
templ = gst_pad_get_pad_template_caps (otherpad);
|
|
temp = gst_caps_intersect (caps, templ);
|
|
gst_caps_unref (caps);
|
|
/* this is what we can do */
|
|
caps = temp;
|
|
} else {
|
|
/* no peer, our padtemplate is enough then */
|
|
caps = gst_caps_copy (gst_pad_get_pad_template_caps (pad));
|
|
}
|
|
|
|
gst_object_unref (render);
|
|
|
|
return caps;
|
|
}
|
|
|
|
static inline gint
|
|
rgb_to_y (gint r, gint g, gint b)
|
|
{
|
|
gint ret;
|
|
|
|
ret = (gint) (((19595 * r) >> 16) + ((38470 * g) >> 16) + ((7471 * b) >> 16));
|
|
ret = CLAMP (ret, 0, 255);
|
|
return ret;
|
|
}
|
|
|
|
static inline gint
|
|
rgb_to_u (gint r, gint g, gint b)
|
|
{
|
|
gint ret;
|
|
|
|
ret =
|
|
(gint) (-((11059 * r) >> 16) - ((21709 * g) >> 16) + ((32768 * b) >> 16) +
|
|
128);
|
|
ret = CLAMP (ret, 0, 255);
|
|
return ret;
|
|
}
|
|
|
|
static inline gint
|
|
rgb_to_v (gint r, gint g, gint b)
|
|
{
|
|
gint ret;
|
|
|
|
ret =
|
|
(gint) (((32768 * r) >> 16) - ((27439 * g) >> 16) - ((5329 * b) >> 16) +
|
|
128);
|
|
ret = CLAMP (ret, 0, 255);
|
|
return ret;
|
|
}
|
|
|
|
/* FIXME: DVB-SUB actually provides us AYUV from CLUT, but libdvbsub used to convert it to ARGB */
|
|
static void
|
|
blit_i420 (GstDVBSubOverlay * overlay, DVBSubtitles * subs, GstBuffer * buffer)
|
|
{
|
|
guint counter;
|
|
DVBSubtitleRect *sub_region;
|
|
gint r, g, b;
|
|
gint a1, a2, a3, a4;
|
|
gint y1, y2, y3, y4;
|
|
gint u1, u2, u3, u4;
|
|
gint v1, v2, v3, v4;
|
|
guint32 color;
|
|
const guint8 *src;
|
|
guint8 *dst_y, *dst_y2, *dst_u, *dst_v;
|
|
gint x, y, w, h;
|
|
gint w2, h2;
|
|
gint width = overlay->width;
|
|
gint height = overlay->height;
|
|
gint src_stride;
|
|
gint y_offset, y_height, y_width, y_stride;
|
|
gint u_offset, u_height, u_width, u_stride;
|
|
gint v_offset, v_height, v_width, v_stride;
|
|
|
|
y_offset =
|
|
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 0, width,
|
|
height);
|
|
u_offset =
|
|
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 1, width,
|
|
height);
|
|
v_offset =
|
|
gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 2, width,
|
|
height);
|
|
|
|
y_height =
|
|
gst_video_format_get_component_height (GST_VIDEO_FORMAT_I420, 0, height);
|
|
u_height =
|
|
gst_video_format_get_component_height (GST_VIDEO_FORMAT_I420, 1, height);
|
|
v_height =
|
|
gst_video_format_get_component_height (GST_VIDEO_FORMAT_I420, 2, height);
|
|
|
|
y_width =
|
|
gst_video_format_get_component_width (GST_VIDEO_FORMAT_I420, 0, width);
|
|
u_width =
|
|
gst_video_format_get_component_width (GST_VIDEO_FORMAT_I420, 1, width);
|
|
v_width =
|
|
gst_video_format_get_component_width (GST_VIDEO_FORMAT_I420, 2, width);
|
|
|
|
y_stride = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 0, width);
|
|
u_stride = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 1, width);
|
|
v_stride = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 2, width);
|
|
|
|
for (counter = 0; counter < subs->num_rects; counter++) {
|
|
gint dw, dh;
|
|
gint32 sx = 0, sy; /* 16.16 fixed point */
|
|
gint32 xstep, ystep; /* 16.16 fixed point */
|
|
|
|
sub_region = subs->rects[counter];
|
|
if (sub_region->y > height || sub_region->x > width)
|
|
continue;
|
|
|
|
/* blend subtitles onto the video frame */
|
|
w = MIN (sub_region->w, width - sub_region->x);
|
|
h = MIN (sub_region->h, height - sub_region->y);
|
|
|
|
/* TODO
|
|
dw = MIN (sub_region->dw, width - sub_region->x);
|
|
dh = MIN (sub_region->dh, height - sub_region->y);
|
|
*/
|
|
dw = w;
|
|
dh = h;
|
|
|
|
xstep = (w << 16) / dw;
|
|
ystep = (h << 16) / dh;
|
|
|
|
w2 = (dw + 1) / 2;
|
|
h2 = (dh + 1) / 2;
|
|
|
|
src_stride = sub_region->pict.rowstride;
|
|
|
|
src = sub_region->pict.data;
|
|
dst_y = buffer->data + y_offset + sub_region->y * y_stride + sub_region->x;
|
|
dst_y2 =
|
|
buffer->data + y_offset + (sub_region->y + 1) * y_stride +
|
|
sub_region->x;
|
|
dst_u =
|
|
buffer->data + u_offset + ((sub_region->y + 1) / 2) * u_stride +
|
|
(sub_region->x + 1) / 2;
|
|
dst_v =
|
|
buffer->data + v_offset + ((sub_region->y + 1) / 2) * v_stride +
|
|
(sub_region->x + 1) / 2;
|
|
|
|
sy = 0;
|
|
for (y = 0; y < dh - 1; y += 2) {
|
|
sx = 0;
|
|
for (x = 0; x < dw - 1; x += 2) {
|
|
color =
|
|
sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]];
|
|
a1 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y1 = rgb_to_y (r, g, b);
|
|
u1 = rgb_to_u (r, g, b);
|
|
v1 = rgb_to_v (r, g, b);
|
|
|
|
color =
|
|
sub_region->pict.palette[src[(sy >> 16) * src_stride + ((sx +
|
|
xstep) >> 16)]];
|
|
a2 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y2 = rgb_to_y (r, g, b);
|
|
u2 = rgb_to_u (r, g, b);
|
|
v2 = rgb_to_v (r, g, b);
|
|
|
|
color =
|
|
sub_region->pict.palette[src[((sy + ystep) >> 16) * src_stride +
|
|
(sx >> 16)]];
|
|
a3 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y3 = rgb_to_y (r, g, b);
|
|
u3 = rgb_to_u (r, g, b);
|
|
v3 = rgb_to_v (r, g, b);
|
|
|
|
color =
|
|
sub_region->pict.palette[src[((sy + ystep) >> 16) * src_stride +
|
|
((sx + xstep) >> 16)]];
|
|
a4 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y4 = rgb_to_y (r, g, b);
|
|
u4 = rgb_to_u (r, g, b);
|
|
v4 = rgb_to_v (r, g, b);
|
|
|
|
dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255;
|
|
dst_y[1] = (a2 * y2 + (255 - a2) * dst_y[1]) / 255;
|
|
dst_y2[0] = (a3 * y3 + (255 - a3) * dst_y2[0]) / 255;
|
|
dst_y2[1] = (a4 * y4 + (255 - a4) * dst_y2[1]) / 255;
|
|
|
|
a1 = (a1 + a2 + a3 + a4) / 4;
|
|
dst_u[0] =
|
|
(a1 * ((u1 + u2 + u3 + u4) / 4) + (255 - a1) * dst_u[0]) / 255;
|
|
dst_v[0] =
|
|
(a1 * ((v1 + v2 + v3 + v4) / 4) + (255 - a1) * dst_v[0]) / 255;
|
|
|
|
dst_y += 2;
|
|
dst_y2 += 2;
|
|
dst_u += 1;
|
|
dst_v += 1;
|
|
sx += 2 * xstep;
|
|
}
|
|
|
|
/* Odd width */
|
|
if (x < dw) {
|
|
color =
|
|
sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]];
|
|
a1 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y1 = rgb_to_y (r, g, b);
|
|
u1 = rgb_to_u (r, g, b);
|
|
v1 = rgb_to_v (r, g, b);
|
|
|
|
color =
|
|
sub_region->pict.palette[src[((sy + ystep) >> 16) * src_stride +
|
|
(sx >> 16)]];
|
|
a3 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y3 = rgb_to_y (r, g, b);
|
|
u3 = rgb_to_u (r, g, b);
|
|
v3 = rgb_to_v (r, g, b);
|
|
|
|
dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255;
|
|
dst_y2[0] = (a3 * y3 + (255 - a3) * dst_y2[0]) / 255;
|
|
|
|
a1 = (a1 + a3) / 2;
|
|
dst_u[0] = (a1 * ((u1 + u3) / 2) + (255 - a1) * dst_u[0]) / 255;
|
|
dst_v[0] = (a1 * ((v1 + v3) / 2) + (255 - a1) * dst_v[0]) / 255;
|
|
|
|
dst_y += 1;
|
|
dst_y2 += 1;
|
|
dst_u += 1;
|
|
dst_v += 1;
|
|
sx += xstep;
|
|
}
|
|
|
|
sy += 2 * ystep;
|
|
|
|
dst_y += y_stride + (y_stride - w);
|
|
dst_y2 += y_stride + (y_stride - w);
|
|
dst_u += u_stride - w2;
|
|
dst_v += v_stride - w2;
|
|
}
|
|
|
|
/* Odd height */
|
|
if (y < dh) {
|
|
for (x = 0; x < dw - 1; x += 2) {
|
|
color =
|
|
sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]];
|
|
a1 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y1 = rgb_to_y (r, g, b);
|
|
u1 = rgb_to_u (r, g, b);
|
|
v1 = rgb_to_v (r, g, b);
|
|
|
|
color =
|
|
sub_region->pict.palette[src[(sy >> 16) * src_stride + ((sx +
|
|
xstep) >> 16)]];
|
|
a2 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y2 = rgb_to_y (r, g, b);
|
|
u2 = rgb_to_u (r, g, b);
|
|
v2 = rgb_to_v (r, g, b);
|
|
|
|
dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255;
|
|
dst_y[1] = (a2 * y2 + (255 - a2) * dst_y[1]) / 255;
|
|
|
|
a1 = (a1 + a2) / 2;
|
|
dst_u[0] = (a1 * ((u1 + u2) / 2) + (255 - a1) * dst_u[0]) / 255;
|
|
dst_v[0] = (a1 * ((v1 + v2) / 2) + (255 - a1) * dst_v[0]) / 255;
|
|
|
|
dst_y += 2;
|
|
dst_u += 1;
|
|
dst_v += 1;
|
|
sx += 2 * xstep;
|
|
}
|
|
|
|
/* Odd height and width */
|
|
if (x < dw) {
|
|
color =
|
|
sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]];
|
|
a1 = (color >> 24) & 0xff;
|
|
r = (color >> 16) & 0xff;
|
|
g = (color >> 8) & 0xff;
|
|
b = color & 0xff;
|
|
|
|
y1 = rgb_to_y (r, g, b);
|
|
u1 = rgb_to_u (r, g, b);
|
|
v1 = rgb_to_v (r, g, b);
|
|
|
|
dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255;
|
|
|
|
dst_u[0] = (a1 * u1 + (255 - a1) * dst_u[0]) / 255;
|
|
dst_v[0] = (a1 * v1 + (255 - a1) * dst_v[0]) / 255;
|
|
|
|
dst_y += 1;
|
|
dst_u += 1;
|
|
dst_v += 1;
|
|
sx += xstep;
|
|
}
|
|
}
|
|
}
|
|
|
|
GST_LOG_OBJECT (overlay, "amount of rendered DVBSubtitleRect: %u", counter);
|
|
}
|
|
|
|
static gboolean
|
|
gst_dvbsub_overlay_setcaps_video (GstPad * pad, GstCaps * caps)
|
|
{
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad));
|
|
gboolean ret = FALSE;
|
|
|
|
render->width = 0;
|
|
render->height = 0;
|
|
|
|
if (!gst_video_format_parse_caps (caps, &render->format, &render->width,
|
|
&render->height) ||
|
|
!gst_video_parse_caps_framerate (caps, &render->fps_n, &render->fps_d)) {
|
|
GST_ERROR_OBJECT (render, "Can't parse caps: %" GST_PTR_FORMAT, caps);
|
|
ret = FALSE;
|
|
goto out;
|
|
}
|
|
|
|
gst_video_parse_caps_pixel_aspect_ratio (caps, &render->par_n,
|
|
&render->par_d);
|
|
|
|
ret = gst_pad_set_caps (render->srcpad, caps);
|
|
if (!ret)
|
|
goto out;
|
|
|
|
GST_DEBUG_OBJECT (render, "ass renderer setup complete");
|
|
|
|
out:
|
|
gst_object_unref (render);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void
|
|
gst_dvbsub_overlay_process_text (GstDVBSubOverlay * overlay, GstBuffer * buffer,
|
|
guint64 pts)
|
|
{
|
|
guint8 *data = (guint8 *) GST_BUFFER_DATA (buffer);
|
|
guint size = GST_BUFFER_SIZE (buffer);
|
|
|
|
GST_DEBUG_OBJECT (overlay,
|
|
"Processing subtitles with fake PTS=%" G_GUINT64_FORMAT
|
|
" which is a running time of %" GST_TIME_FORMAT,
|
|
pts, GST_TIME_ARGS (pts));
|
|
GST_DEBUG_OBJECT (overlay, "Feeding %u bytes to libdvbsub", size);
|
|
g_mutex_lock (overlay->dvbsub_mutex);
|
|
dvb_sub_feed_with_pts (overlay->dvb_sub, pts, data, size);
|
|
g_mutex_unlock (overlay->dvbsub_mutex);
|
|
gst_buffer_unref (buffer);
|
|
}
|
|
|
|
static void
|
|
new_dvb_subtitles_cb (DvbSub * dvb_sub, DVBSubtitles * subs, gpointer user_data)
|
|
{
|
|
GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (user_data);
|
|
|
|
GST_INFO_OBJECT (overlay,
|
|
"New DVB subtitles arrived with a page_time_out of %d and %d regions for PTS=%"
|
|
G_GUINT64_FORMAT ", which should be at running time %" GST_TIME_FORMAT,
|
|
subs->page_time_out, subs->num_rects, subs->pts,
|
|
GST_TIME_ARGS (subs->pts));
|
|
|
|
g_queue_push_tail (overlay->pending_subtitles, subs);
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_dvbsub_overlay_bufferalloc_video (GstPad * pad, guint64 offset, guint size,
|
|
GstCaps * caps, GstBuffer ** buffer)
|
|
{
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad));
|
|
GstFlowReturn ret = GST_FLOW_WRONG_STATE;
|
|
GstPad *allocpad;
|
|
|
|
GST_OBJECT_LOCK (render);
|
|
allocpad = render->srcpad ? gst_object_ref (render->srcpad) : NULL;
|
|
GST_OBJECT_UNLOCK (render);
|
|
|
|
if (allocpad) {
|
|
ret = gst_pad_alloc_buffer (allocpad, offset, size, caps, buffer);
|
|
gst_object_unref (allocpad);
|
|
}
|
|
|
|
gst_object_unref (render);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_dvbsub_overlay_chain_text (GstPad * pad, GstBuffer * buffer)
|
|
{
|
|
GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (GST_PAD_PARENT (pad));
|
|
GstClockTime sub_running_time;
|
|
|
|
GST_INFO_OBJECT (overlay, "private/x-dvbsub buffer with size %u",
|
|
GST_BUFFER_SIZE (buffer));
|
|
|
|
GST_LOG_OBJECT (overlay,
|
|
"Video segment: %" GST_SEGMENT_FORMAT " --- Subtitle segment: %"
|
|
GST_SEGMENT_FORMAT " --- BUFFER: ts=%" GST_TIME_FORMAT,
|
|
&overlay->video_segment, &overlay->subtitle_segment,
|
|
GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (buffer)));
|
|
|
|
/* DVB subtitle packets are required to carry the PTS */
|
|
if (G_UNLIKELY (!GST_BUFFER_TIMESTAMP_IS_VALID (buffer))) {
|
|
GST_WARNING_OBJECT (overlay,
|
|
"Text buffer without valid timestamp, dropping");
|
|
gst_buffer_unref (buffer);
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
/* As the passed start and stop is equal, we shouldn't need to care about out of segment at all,
|
|
* the subtitle data for the PTS is completely out of interest to us. A given display set must
|
|
* carry the same PTS value. */
|
|
/* FIXME: Consider with larger than 64kB display sets, which would be cut into multiple packets,
|
|
* FIXME: does our waiting + render code work when there are more than one packets before
|
|
* FIXME: rendering callback will get called? */
|
|
|
|
gst_segment_set_last_stop (&overlay->subtitle_segment, GST_FORMAT_TIME,
|
|
GST_BUFFER_TIMESTAMP (buffer));
|
|
|
|
sub_running_time =
|
|
gst_segment_to_running_time (&overlay->subtitle_segment, GST_FORMAT_TIME,
|
|
GST_BUFFER_TIMESTAMP (buffer));
|
|
|
|
GST_DEBUG_OBJECT (overlay, "SUBTITLE real running time: %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (sub_running_time));
|
|
|
|
/* FIXME: We are abusing libdvbsub pts value for tracking our gstreamer running time instead of real PTS. Should be mostly fine though... */
|
|
gst_dvbsub_overlay_process_text (overlay, buffer, sub_running_time);
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_dvbsub_overlay_chain_video (GstPad * pad, GstBuffer * buffer)
|
|
{
|
|
GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (GST_PAD_PARENT (pad));
|
|
GstFlowReturn ret = GST_FLOW_OK;
|
|
gint64 start, stop;
|
|
gint64 cstart, cstop;
|
|
gboolean in_seg;
|
|
GstClockTime vid_running_time, vid_running_time_end;
|
|
|
|
if (overlay->format == GST_VIDEO_FORMAT_UNKNOWN)
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
|
|
if (!GST_BUFFER_TIMESTAMP_IS_VALID (buffer))
|
|
goto missing_timestamp;
|
|
|
|
start = GST_BUFFER_TIMESTAMP (buffer);
|
|
|
|
GST_LOG_OBJECT (overlay,
|
|
"Video segment: %" GST_SEGMENT_FORMAT " --- Subtitle last_stop: %"
|
|
GST_TIME_FORMAT " --- BUFFER: ts=%" GST_TIME_FORMAT,
|
|
&overlay->video_segment,
|
|
GST_TIME_ARGS (overlay->subtitle_segment.last_stop),
|
|
GST_TIME_ARGS (start));
|
|
|
|
/* ignore buffers that are outside of the current segment */
|
|
if (!GST_BUFFER_DURATION_IS_VALID (buffer)) {
|
|
stop = GST_CLOCK_TIME_NONE;
|
|
} else {
|
|
stop = start + GST_BUFFER_DURATION (buffer);
|
|
}
|
|
|
|
in_seg = gst_segment_clip (&overlay->video_segment, GST_FORMAT_TIME,
|
|
start, stop, &cstart, &cstop);
|
|
if (!in_seg) {
|
|
GST_DEBUG_OBJECT (overlay, "Buffer outside configured segment -- dropping");
|
|
gst_buffer_unref (buffer);
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
buffer = gst_buffer_make_metadata_writable (buffer);
|
|
GST_BUFFER_TIMESTAMP (buffer) = cstart;
|
|
if (GST_BUFFER_DURATION_IS_VALID (buffer))
|
|
GST_BUFFER_DURATION (buffer) = cstop - cstart;
|
|
|
|
vid_running_time =
|
|
gst_segment_to_running_time (&overlay->video_segment, GST_FORMAT_TIME,
|
|
cstart);
|
|
if (GST_BUFFER_DURATION_IS_VALID (buffer))
|
|
vid_running_time_end =
|
|
gst_segment_to_running_time (&overlay->video_segment, GST_FORMAT_TIME,
|
|
cstop);
|
|
else
|
|
vid_running_time_end = vid_running_time;
|
|
|
|
GST_DEBUG_OBJECT (overlay, "Video running time: %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (vid_running_time));
|
|
|
|
gst_segment_set_last_stop (&overlay->video_segment, GST_FORMAT_TIME,
|
|
GST_BUFFER_TIMESTAMP (buffer));
|
|
|
|
g_mutex_lock (overlay->dvbsub_mutex);
|
|
if (!g_queue_is_empty (overlay->pending_subtitles)) {
|
|
DVBSubtitles *tmp, *candidate = NULL;
|
|
|
|
while (!g_queue_is_empty (overlay->pending_subtitles)) {
|
|
tmp = g_queue_peek_head (overlay->pending_subtitles);
|
|
|
|
if (tmp->pts > vid_running_time_end) {
|
|
/* For a future video frame */
|
|
break;
|
|
} else if (tmp->num_rects == 0) {
|
|
/* Clear screen */
|
|
if (overlay->current_subtitle)
|
|
dvb_subtitles_free (overlay->current_subtitle);
|
|
overlay->current_subtitle = NULL;
|
|
if (candidate)
|
|
dvb_subtitles_free (candidate);
|
|
candidate = NULL;
|
|
g_queue_pop_head (overlay->pending_subtitles);
|
|
} else if (tmp->pts + tmp->page_time_out * GST_SECOND *
|
|
overlay->subtitle_segment.abs_rate >= vid_running_time) {
|
|
if (candidate)
|
|
dvb_subtitles_free (candidate);
|
|
candidate = tmp;
|
|
g_queue_pop_head (overlay->pending_subtitles);
|
|
} else {
|
|
/* Too late */
|
|
dvb_subtitles_free (tmp);
|
|
tmp = NULL;
|
|
g_queue_pop_head (overlay->pending_subtitles);
|
|
}
|
|
}
|
|
|
|
if (candidate) {
|
|
GST_DEBUG_OBJECT (overlay,
|
|
"Time to show the next subtitle page (%" GST_TIME_FORMAT " >= %"
|
|
GST_TIME_FORMAT ") - it has %u regions",
|
|
GST_TIME_ARGS (vid_running_time), GST_TIME_ARGS (candidate->pts),
|
|
candidate->num_rects);
|
|
dvb_subtitles_free (overlay->current_subtitle);
|
|
overlay->current_subtitle = candidate;
|
|
/* FIXME: Pre-convert current_subtitle to a quick-blend format, num_rects=0 means that there are no regions, e.g, a subtitle "clear" happened */
|
|
}
|
|
}
|
|
|
|
/* Check that we haven't hit the fallback timeout for current subtitle page */
|
|
if (overlay->current_subtitle
|
|
&& vid_running_time >
|
|
(overlay->current_subtitle->pts +
|
|
overlay->current_subtitle->page_time_out * GST_SECOND *
|
|
overlay->subtitle_segment.abs_rate)) {
|
|
GST_INFO_OBJECT (overlay,
|
|
"Subtitle page not redefined before fallback page_time_out of %u seconds (missed data?) - deleting current page",
|
|
overlay->current_subtitle->page_time_out);
|
|
dvb_subtitles_free (overlay->current_subtitle);
|
|
overlay->current_subtitle = NULL;
|
|
}
|
|
|
|
/* Now render it */
|
|
if (overlay->current_subtitle) {
|
|
buffer = gst_buffer_make_writable (buffer);
|
|
blit_i420 (overlay, overlay->current_subtitle, buffer);
|
|
}
|
|
g_mutex_unlock (overlay->dvbsub_mutex);
|
|
|
|
ret = gst_pad_push (overlay->srcpad, buffer);
|
|
|
|
return ret;
|
|
|
|
missing_timestamp:
|
|
{
|
|
GST_WARNING_OBJECT (overlay, "video buffer without timestamp, discarding");
|
|
gst_buffer_unref (buffer);
|
|
return GST_FLOW_OK;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_dvbsub_overlay_event_video (GstPad * pad, GstEvent * event)
|
|
{
|
|
gboolean ret = FALSE;
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad));
|
|
|
|
GST_DEBUG_OBJECT (pad, "received video event %s",
|
|
GST_EVENT_TYPE_NAME (event));
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_NEWSEGMENT:
|
|
{
|
|
GstFormat format;
|
|
gdouble rate;
|
|
gint64 start, stop, time;
|
|
gboolean update;
|
|
|
|
GST_DEBUG_OBJECT (render, "received new segment");
|
|
|
|
gst_event_parse_new_segment (event, &update, &rate, &format, &start,
|
|
&stop, &time);
|
|
|
|
if (format == GST_FORMAT_TIME) {
|
|
GST_DEBUG_OBJECT (render, "VIDEO SEGMENT now: %" GST_SEGMENT_FORMAT,
|
|
&render->video_segment);
|
|
|
|
gst_segment_set_newsegment (&render->video_segment, update, rate,
|
|
format, start, stop, time);
|
|
|
|
GST_DEBUG_OBJECT (render, "VIDEO SEGMENT after: %" GST_SEGMENT_FORMAT,
|
|
&render->video_segment);
|
|
ret = gst_pad_push_event (render->srcpad, event);
|
|
} else {
|
|
GST_ELEMENT_WARNING (render, STREAM, MUX, (NULL),
|
|
("received non-TIME newsegment event on video input"));
|
|
ret = FALSE;
|
|
gst_event_unref (event);
|
|
}
|
|
break;
|
|
}
|
|
case GST_EVENT_FLUSH_STOP:
|
|
gst_segment_init (&render->video_segment, GST_FORMAT_TIME);
|
|
default:
|
|
ret = gst_pad_push_event (render->srcpad, event);
|
|
break;
|
|
}
|
|
|
|
gst_object_unref (render);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_dvbsub_overlay_event_text (GstPad * pad, GstEvent * event)
|
|
{
|
|
gboolean ret = FALSE;
|
|
GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad));
|
|
|
|
GST_DEBUG_OBJECT (pad, "received text event %s", GST_EVENT_TYPE_NAME (event));
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_NEWSEGMENT:
|
|
{
|
|
GstFormat format;
|
|
gdouble rate;
|
|
gint64 start, stop, time;
|
|
gboolean update;
|
|
|
|
GST_DEBUG_OBJECT (render, "received new segment");
|
|
|
|
gst_event_parse_new_segment (event, &update, &rate, &format, &start,
|
|
&stop, &time);
|
|
|
|
if (format == GST_FORMAT_TIME) {
|
|
GST_DEBUG_OBJECT (render, "SUBTITLE SEGMENT now: %" GST_SEGMENT_FORMAT,
|
|
&render->subtitle_segment);
|
|
|
|
gst_segment_set_newsegment (&render->subtitle_segment, update, rate,
|
|
format, start, stop, time);
|
|
|
|
GST_DEBUG_OBJECT (render,
|
|
"SUBTITLE SEGMENT after: %" GST_SEGMENT_FORMAT,
|
|
&render->subtitle_segment);
|
|
ret = TRUE;
|
|
gst_event_unref (event);
|
|
} else {
|
|
GST_ELEMENT_WARNING (render, STREAM, MUX, (NULL),
|
|
("received non-TIME newsegment event on subtitle sinkpad"));
|
|
ret = FALSE;
|
|
gst_event_unref (event);
|
|
}
|
|
break;
|
|
}
|
|
case GST_EVENT_FLUSH_STOP:
|
|
GST_DEBUG_OBJECT (render, "stop flushing");
|
|
gst_dvbsub_overlay_flush_subtitles (render);
|
|
gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME);
|
|
gst_event_unref (event);
|
|
ret = TRUE;
|
|
break;
|
|
case GST_EVENT_FLUSH_START:
|
|
GST_DEBUG_OBJECT (render, "begin flushing");
|
|
gst_event_unref (event);
|
|
ret = TRUE;
|
|
break;
|
|
case GST_EVENT_EOS:
|
|
GST_INFO_OBJECT (render, "text EOS");
|
|
gst_event_unref (event);
|
|
ret = TRUE;
|
|
break;
|
|
default:
|
|
ret = gst_pad_push_event (render->srcpad, event);
|
|
break;
|
|
}
|
|
|
|
gst_object_unref (render);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
plugin_init (GstPlugin * plugin)
|
|
{
|
|
GST_DEBUG_CATEGORY_INIT (gst_dvbsub_overlay_debug, "dvbsuboverlay",
|
|
0, "DVB subtitle overlay");
|
|
GST_DEBUG_CATEGORY_INIT (gst_dvbsub_overlay_lib_debug, "dvbsub_library",
|
|
0, "libdvbsub library");
|
|
|
|
dvb_sub_set_global_log_cb (_dvbsub_log_cb, NULL);
|
|
|
|
return gst_element_register (plugin, "dvbsuboverlay",
|
|
GST_RANK_PRIMARY, GST_TYPE_DVBSUB_OVERLAY);
|
|
}
|
|
|
|
GST_PLUGIN_DEFINE (GST_VERSION_MAJOR,
|
|
GST_VERSION_MINOR,
|
|
"dvbsuboverlay",
|
|
"DVB subtitle renderer",
|
|
plugin_init, VERSION, "LGPL", GST_PACKAGE_NAME, GST_PACKAGE_ORIGIN)
|