/* GStreamer DVB subtitles overlay * Copyright (c) 2010 Mart Raudsepp * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ /** * SECTION:element-dvbsuboverlay * * Renders DVB subtitles on top of a video stream. * * * Example launch line * |[ FIXME * gst-launch -v filesrc location=/path/to/ts ! mpegtsdemux name=d ! queue ! mp3parse ! mad ! audioconvert ! autoaudiosink \ * d. ! queue ! mpeg2dec ! ffmpegcolorspace ! r. \ * d. ! queue ! "subpicture/x-dvb" ! dvbsuboverlay name=r ! ffmpegcolorspace ! autovideosink * ]| This pipeline demuxes a MPEG-TS file with MPEG2 video, MP3 audio and embedded DVB subtitles and renders the subtitles on top of the video. * */ #ifdef HAVE_CONFIG_H # include #endif #include "gstdvbsuboverlay.h" #include GST_DEBUG_CATEGORY_STATIC (gst_dvbsub_overlay_debug); #define GST_CAT_DEFAULT gst_dvbsub_overlay_debug /* Filter signals and props */ enum { LAST_SIGNAL }; enum { PROP_0, PROP_ENABLE, PROP_MAX_PAGE_TIMEOUT, }; #define DEFAULT_ENABLE (TRUE) #define DEFAULT_MAX_PAGE_TIMEOUT (0) static GstStaticPadTemplate src_factory = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("I420")) ); static GstStaticPadTemplate video_sink_factory = GST_STATIC_PAD_TEMPLATE ("video_sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("I420")) ); static GstStaticPadTemplate text_sink_factory = GST_STATIC_PAD_TEMPLATE ("text_sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS ("subpicture/x-dvb") ); static void gst_dvbsub_overlay_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_dvbsub_overlay_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static void gst_dvbsub_overlay_finalize (GObject * object); static GstStateChangeReturn gst_dvbsub_overlay_change_state (GstElement * element, GstStateChange transition); GST_BOILERPLATE (GstDVBSubOverlay, gst_dvbsub_overlay, GstElement, GST_TYPE_ELEMENT); static GstCaps *gst_dvbsub_overlay_getcaps (GstPad * pad); static gboolean gst_dvbsub_overlay_setcaps_video (GstPad * pad, GstCaps * caps); static GstFlowReturn gst_dvbsub_overlay_chain_video (GstPad * pad, GstBuffer * buf); static GstFlowReturn gst_dvbsub_overlay_chain_text (GstPad * pad, GstBuffer * buf); static gboolean gst_dvbsub_overlay_event_video (GstPad * pad, GstEvent * event); static gboolean gst_dvbsub_overlay_event_text (GstPad * pad, GstEvent * event); static gboolean gst_dvbsub_overlay_event_src (GstPad * pad, GstEvent * event); static void new_dvb_subtitles_cb (DvbSub * dvb_sub, DVBSubtitles * subs, gpointer user_data); static GstFlowReturn gst_dvbsub_overlay_bufferalloc_video (GstPad * pad, guint64 offset, guint size, GstCaps * caps, GstBuffer ** buffer); static gboolean gst_dvbsub_overlay_query_src (GstPad * pad, GstQuery * query); static void gst_dvbsub_overlay_base_init (gpointer gclass) { GstElementClass *element_class = (GstElementClass *) gclass; gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&src_factory)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&video_sink_factory)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&text_sink_factory)); gst_element_class_set_details_simple (element_class, "DVB Subtitles Overlay", "Mixer/Video/Overlay/Subtitle", "Renders DVB subtitles", "Mart Raudsepp "); } /* initialize the plugin's class */ static void gst_dvbsub_overlay_class_init (GstDVBSubOverlayClass * klass) { GObjectClass *gobject_class = (GObjectClass *) klass; GstElementClass *gstelement_class = (GstElementClass *) klass; gobject_class->set_property = gst_dvbsub_overlay_set_property; gobject_class->get_property = gst_dvbsub_overlay_get_property; gobject_class->finalize = gst_dvbsub_overlay_finalize; g_object_class_install_property (gobject_class, PROP_ENABLE, g_param_spec_boolean ("enable", "Enable", /* FIXME: "enable" vs "silent"? */ "Enable rendering of subtitles", DEFAULT_ENABLE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_MAX_PAGE_TIMEOUT, g_param_spec_int ("max-page-timeout", "max-page-timeout", "Limit maximum display time of a subtitle page (0 - disabled, value in seconds)", 0, G_MAXINT, DEFAULT_MAX_PAGE_TIMEOUT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); gstelement_class->change_state = GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_change_state); } static void gst_dvbsub_overlay_flush_subtitles (GstDVBSubOverlay * render) { DVBSubtitles *subs; g_mutex_lock (render->dvbsub_mutex); while ((subs = g_queue_pop_head (render->pending_subtitles))) { dvb_subtitles_free (subs); } if (render->current_subtitle) dvb_subtitles_free (render->current_subtitle); render->current_subtitle = NULL; if (render->dvb_sub) dvb_sub_free (render->dvb_sub); render->dvb_sub = dvb_sub_new (); { DvbSubCallbacks dvbsub_callbacks = { &new_dvb_subtitles_cb, }; dvb_sub_set_callbacks (render->dvb_sub, &dvbsub_callbacks, render); } g_mutex_unlock (render->dvbsub_mutex); } static void gst_dvbsub_overlay_init (GstDVBSubOverlay * render, GstDVBSubOverlayClass * gclass) { GST_DEBUG_OBJECT (render, "init"); render->srcpad = gst_pad_new_from_static_template (&src_factory, "src"); render->video_sinkpad = gst_pad_new_from_static_template (&video_sink_factory, "video_sink"); render->text_sinkpad = gst_pad_new_from_static_template (&text_sink_factory, "text_sink"); gst_pad_set_setcaps_function (render->video_sinkpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_setcaps_video)); gst_pad_set_getcaps_function (render->srcpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_getcaps)); gst_pad_set_getcaps_function (render->video_sinkpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_getcaps)); gst_pad_set_chain_function (render->video_sinkpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_chain_video)); gst_pad_set_chain_function (render->text_sinkpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_chain_text)); gst_pad_set_event_function (render->video_sinkpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_event_video)); gst_pad_set_event_function (render->text_sinkpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_event_text)); gst_pad_set_event_function (render->srcpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_event_src)); gst_pad_set_bufferalloc_function (render->video_sinkpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_bufferalloc_video)); gst_pad_set_query_function (render->srcpad, GST_DEBUG_FUNCPTR (gst_dvbsub_overlay_query_src)); gst_element_add_pad (GST_ELEMENT (render), render->srcpad); gst_element_add_pad (GST_ELEMENT (render), render->video_sinkpad); gst_element_add_pad (GST_ELEMENT (render), render->text_sinkpad); render->width = 0; render->height = 0; render->current_subtitle = NULL; render->pending_subtitles = g_queue_new (); render->enable = DEFAULT_ENABLE; render->max_page_timeout = DEFAULT_MAX_PAGE_TIMEOUT; render->dvbsub_mutex = g_mutex_new (); gst_dvbsub_overlay_flush_subtitles (render); gst_segment_init (&render->video_segment, GST_FORMAT_TIME); gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME); GST_DEBUG_OBJECT (render, "init complete"); } static void gst_dvbsub_overlay_finalize (GObject * object) { GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (object); DVBSubtitles *subs; while ((subs = g_queue_pop_head (overlay->pending_subtitles))) { dvb_subtitles_free (subs); } g_queue_free (overlay->pending_subtitles); if (overlay->current_subtitle) dvb_subtitles_free (overlay->current_subtitle); overlay->current_subtitle = NULL; if (overlay->dvb_sub) dvb_sub_free (overlay->dvb_sub); if (overlay->dvbsub_mutex) g_mutex_free (overlay->dvbsub_mutex); G_OBJECT_CLASS (parent_class)->finalize (object); } static void gst_dvbsub_overlay_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (object); switch (prop_id) { case PROP_ENABLE: g_atomic_int_set (&overlay->enable, g_value_get_boolean (value)); break; case PROP_MAX_PAGE_TIMEOUT: g_atomic_int_set (&overlay->max_page_timeout, g_value_get_int (value)); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static void gst_dvbsub_overlay_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (object); switch (prop_id) { case PROP_ENABLE: g_value_set_boolean (value, g_atomic_int_get (&overlay->enable)); break; case PROP_MAX_PAGE_TIMEOUT: g_value_set_int (value, g_atomic_int_get (&overlay->max_page_timeout)); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static GstStateChangeReturn gst_dvbsub_overlay_change_state (GstElement * element, GstStateChange transition) { GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (element); GstStateChangeReturn ret; switch (transition) { case GST_STATE_CHANGE_READY_TO_PAUSED: gst_segment_init (&render->video_segment, GST_FORMAT_TIME); gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME); break; case GST_STATE_CHANGE_NULL_TO_READY: case GST_STATE_CHANGE_PAUSED_TO_PLAYING: default: break; } ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY: gst_dvbsub_overlay_flush_subtitles (render); gst_segment_init (&render->video_segment, GST_FORMAT_TIME); gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME); render->format = GST_VIDEO_FORMAT_UNKNOWN; break; case GST_STATE_CHANGE_PLAYING_TO_PAUSED: case GST_STATE_CHANGE_READY_TO_NULL: default: break; } return ret; } static gboolean gst_dvbsub_overlay_query_src (GstPad * pad, GstQuery * query) { GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad)); gboolean ret; ret = gst_pad_peer_query (render->video_sinkpad, query); gst_object_unref (render); return ret; } static gboolean gst_dvbsub_overlay_event_src (GstPad * pad, GstEvent * event) { GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad)); gboolean ret = FALSE; switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEEK:{ GstSeekFlags flags; GST_DEBUG_OBJECT (render, "seek received, driving from here"); gst_event_parse_seek (event, NULL, NULL, &flags, NULL, NULL, NULL, NULL); /* Flush downstream, only for flushing seek */ if (flags & GST_SEEK_FLAG_FLUSH) gst_pad_push_event (render->srcpad, gst_event_new_flush_start ()); gst_dvbsub_overlay_flush_subtitles (render); /* Seek on each sink pad */ gst_event_ref (event); ret = gst_pad_push_event (render->video_sinkpad, event); if (ret) { ret = gst_pad_push_event (render->text_sinkpad, event); } else { gst_event_unref (event); } break; } default: gst_event_ref (event); ret = gst_pad_push_event (render->video_sinkpad, event); gst_pad_push_event (render->text_sinkpad, event); break; } gst_object_unref (render); return ret; } static GstCaps * gst_dvbsub_overlay_getcaps (GstPad * pad) { GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad)); GstPad *otherpad; GstCaps *caps; if (pad == render->srcpad) otherpad = render->video_sinkpad; else otherpad = render->srcpad; /* we can do what the peer can */ caps = gst_pad_peer_get_caps (otherpad); if (caps) { GstCaps *temp; const GstCaps *templ; /* filtered against our padtemplate */ templ = gst_pad_get_pad_template_caps (otherpad); temp = gst_caps_intersect (caps, templ); gst_caps_unref (caps); /* this is what we can do */ caps = temp; } else { /* no peer, our padtemplate is enough then */ caps = gst_caps_copy (gst_pad_get_pad_template_caps (pad)); } gst_object_unref (render); return caps; } static void blit_i420 (GstDVBSubOverlay * overlay, DVBSubtitles * subs, GstBuffer * buffer) { guint counter; DVBSubtitleRect *sub_region; gint a1, a2, a3, a4; gint y1, y2, y3, y4; gint u1, u2, u3, u4; gint v1, v2, v3, v4; guint32 color; const guint8 *src; guint8 *dst_y, *dst_y2, *dst_u, *dst_v; gint x, y; gint w2, h2; gint width = overlay->width; gint height = overlay->height; gint src_stride; gint y_offset, y_height, y_width, y_stride; gint u_offset, u_height, u_width, u_stride; gint v_offset, v_height, v_width, v_stride; gint scale = 0; gint scale_x = 0, scale_y = 0; /* 16.16 fixed point */ y_offset = gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 0, width, height); u_offset = gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 1, width, height); v_offset = gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 2, width, height); y_height = gst_video_format_get_component_height (GST_VIDEO_FORMAT_I420, 0, height); u_height = gst_video_format_get_component_height (GST_VIDEO_FORMAT_I420, 1, height); v_height = gst_video_format_get_component_height (GST_VIDEO_FORMAT_I420, 2, height); y_width = gst_video_format_get_component_width (GST_VIDEO_FORMAT_I420, 0, width); u_width = gst_video_format_get_component_width (GST_VIDEO_FORMAT_I420, 1, width); v_width = gst_video_format_get_component_width (GST_VIDEO_FORMAT_I420, 2, width); y_stride = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 0, width); u_stride = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 1, width); v_stride = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 2, width); if (width != subs->display_def.display_width && height != subs->display_def.display_height) { scale = 1; if (subs->display_def.window_flag) { scale_x = (width << 16) / subs->display_def.window_width; scale_y = (height << 16) / subs->display_def.window_height; } else { scale_x = (width << 16) / subs->display_def.display_width; scale_y = (height << 16) / subs->display_def.display_height; } } for (counter = 0; counter < subs->num_rects; counter++) { gint dw, dh, dx, dy; gint32 sx = 0, sy; /* 16.16 fixed point */ gint32 xstep, ystep; /* 16.16 fixed point */ sub_region = &subs->rects[counter]; if (sub_region->y > height || sub_region->x > width) continue; /* blend subtitles onto the video frame */ dx = sub_region->x; dy = sub_region->y; dw = sub_region->w; dh = sub_region->h; if (scale) { dx = (dx * scale_x) >> 16; dy = (dy * scale_y) >> 16; dw = (dw * scale_x) >> 16; dh = (dh * scale_y) >> 16; /* apply subtitle window offsets after scaling */ if (subs->display_def.window_flag) { dx += subs->display_def.window_x; dy += subs->display_def.window_y; } } dw = MIN (dw, width - dx); dh = MIN (dh, height - dx); xstep = (sub_region->w << 16) / dw; ystep = (sub_region->h << 16) / dh; w2 = (dw + 1) / 2; h2 = (dh + 1) / 2; src_stride = sub_region->pict.rowstride; src = sub_region->pict.data; dst_y = buffer->data + y_offset + dy * y_stride + dx; dst_y2 = buffer->data + y_offset + (dy + 1) * y_stride + dx; dst_u = buffer->data + u_offset + ((dy + 1) / 2) * u_stride + (dx + 1) / 2; dst_v = buffer->data + v_offset + ((dy + 1) / 2) * v_stride + (dx + 1) / 2; sy = 0; for (y = 0; y < dh - 1; y += 2) { sx = 0; for (x = 0; x < dw - 1; x += 2) { color = sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]]; a1 = (color >> 24) & 0xff; y1 = (color >> 16) & 0xff; u1 = (color >> 8) & 0xff; v1 = color & 0xff; color = sub_region->pict.palette[src[(sy >> 16) * src_stride + ((sx + xstep) >> 16)]]; a2 = (color >> 24) & 0xff; y2 = (color >> 16) & 0xff; u2 = (color >> 8) & 0xff; v2 = color & 0xff; color = sub_region->pict.palette[src[((sy + ystep) >> 16) * src_stride + (sx >> 16)]]; a3 = (color >> 24) & 0xff; y3 = (color >> 16) & 0xff; u3 = (color >> 8) & 0xff; v3 = color & 0xff; color = sub_region->pict.palette[src[((sy + ystep) >> 16) * src_stride + ((sx + xstep) >> 16)]]; a4 = (color >> 24) & 0xff; y4 = (color >> 16) & 0xff; u4 = (color >> 8) & 0xff; v4 = color & 0xff; dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255; dst_y[1] = (a2 * y2 + (255 - a2) * dst_y[1]) / 255; dst_y2[0] = (a3 * y3 + (255 - a3) * dst_y2[0]) / 255; dst_y2[1] = (a4 * y4 + (255 - a4) * dst_y2[1]) / 255; a1 = (a1 + a2 + a3 + a4) / 4; dst_u[0] = (a1 * ((u1 + u2 + u3 + u4) / 4) + (255 - a1) * dst_u[0]) / 255; dst_v[0] = (a1 * ((v1 + v2 + v3 + v4) / 4) + (255 - a1) * dst_v[0]) / 255; dst_y += 2; dst_y2 += 2; dst_u += 1; dst_v += 1; sx += 2 * xstep; } /* Odd width */ if (x < dw) { color = sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]]; a1 = (color >> 24) & 0xff; y1 = (color >> 16) & 0xff; u1 = (color >> 8) & 0xff; v1 = color & 0xff; color = sub_region->pict.palette[src[((sy + ystep) >> 16) * src_stride + (sx >> 16)]]; a3 = (color >> 24) & 0xff; y3 = (color >> 16) & 0xff; u3 = (color >> 8) & 0xff; v3 = color & 0xff; dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255; dst_y2[0] = (a3 * y3 + (255 - a3) * dst_y2[0]) / 255; a1 = (a1 + a3) / 2; dst_u[0] = (a1 * ((u1 + u3) / 2) + (255 - a1) * dst_u[0]) / 255; dst_v[0] = (a1 * ((v1 + v3) / 2) + (255 - a1) * dst_v[0]) / 255; dst_y += 1; dst_y2 += 1; dst_u += 1; dst_v += 1; sx += xstep; } sy += 2 * ystep; dst_y += y_stride + (y_stride - dw); dst_y2 += y_stride + (y_stride - dw); dst_u += u_stride - w2; dst_v += v_stride - w2; } /* Odd height */ if (y < dh) { sx = 0; for (x = 0; x < dw - 1; x += 2) { color = sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]]; a1 = (color >> 24) & 0xff; y1 = (color >> 16) & 0xff; u1 = (color >> 8) & 0xff; v1 = color & 0xff; color = sub_region->pict.palette[src[(sy >> 16) * src_stride + ((sx + xstep) >> 16)]]; a2 = (color >> 24) & 0xff; y2 = (color >> 16) & 0xff; u2 = (color >> 8) & 0xff; v2 = color & 0xff; dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255; dst_y[1] = (a2 * y2 + (255 - a2) * dst_y[1]) / 255; a1 = (a1 + a2) / 2; dst_u[0] = (a1 * ((u1 + u2) / 2) + (255 - a1) * dst_u[0]) / 255; dst_v[0] = (a1 * ((v1 + v2) / 2) + (255 - a1) * dst_v[0]) / 255; dst_y += 2; dst_u += 1; dst_v += 1; sx += 2 * xstep; } /* Odd height and width */ if (x < dw) { color = sub_region->pict.palette[src[(sy >> 16) * src_stride + (sx >> 16)]]; a1 = (color >> 24) & 0xff; y1 = (color >> 16) & 0xff; u1 = (color >> 8) & 0xff; v1 = color & 0xff; dst_y[0] = (a1 * y1 + (255 - a1) * dst_y[0]) / 255; dst_u[0] = (a1 * u1 + (255 - a1) * dst_u[0]) / 255; dst_v[0] = (a1 * v1 + (255 - a1) * dst_v[0]) / 255; dst_y += 1; dst_u += 1; dst_v += 1; sx += xstep; } } } GST_LOG_OBJECT (overlay, "amount of rendered DVBSubtitleRect: %u", counter); } static gboolean gst_dvbsub_overlay_setcaps_video (GstPad * pad, GstCaps * caps) { GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad)); gboolean ret = FALSE; render->width = 0; render->height = 0; if (!gst_video_format_parse_caps (caps, &render->format, &render->width, &render->height) || !gst_video_parse_caps_framerate (caps, &render->fps_n, &render->fps_d)) { GST_ERROR_OBJECT (render, "Can't parse caps: %" GST_PTR_FORMAT, caps); ret = FALSE; goto out; } gst_video_parse_caps_pixel_aspect_ratio (caps, &render->par_n, &render->par_d); ret = gst_pad_set_caps (render->srcpad, caps); if (!ret) goto out; GST_DEBUG_OBJECT (render, "ass renderer setup complete"); out: gst_object_unref (render); return ret; } static void gst_dvbsub_overlay_process_text (GstDVBSubOverlay * overlay, GstBuffer * buffer, guint64 pts) { guint8 *data = (guint8 *) GST_BUFFER_DATA (buffer); guint size = GST_BUFFER_SIZE (buffer); GST_DEBUG_OBJECT (overlay, "Processing subtitles with fake PTS=%" G_GUINT64_FORMAT " which is a running time of %" GST_TIME_FORMAT, pts, GST_TIME_ARGS (pts)); GST_DEBUG_OBJECT (overlay, "Feeding %u bytes to libdvbsub", size); g_mutex_lock (overlay->dvbsub_mutex); dvb_sub_feed_with_pts (overlay->dvb_sub, pts, data, size); g_mutex_unlock (overlay->dvbsub_mutex); gst_buffer_unref (buffer); } static void new_dvb_subtitles_cb (DvbSub * dvb_sub, DVBSubtitles * subs, gpointer user_data) { GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (user_data); int max_page_timeout; max_page_timeout = g_atomic_int_get (&overlay->max_page_timeout); if (max_page_timeout > 0) subs->page_time_out = MIN (subs->page_time_out, max_page_timeout); GST_INFO_OBJECT (overlay, "New DVB subtitles arrived with a page_time_out of %d and %d regions for PTS=%" G_GUINT64_FORMAT ", which should be at running time %" GST_TIME_FORMAT, subs->page_time_out, subs->num_rects, subs->pts, GST_TIME_ARGS (subs->pts)); g_queue_push_tail (overlay->pending_subtitles, subs); } static GstFlowReturn gst_dvbsub_overlay_bufferalloc_video (GstPad * pad, guint64 offset, guint size, GstCaps * caps, GstBuffer ** buffer) { GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad)); GstFlowReturn ret = GST_FLOW_WRONG_STATE; GstPad *allocpad; GST_OBJECT_LOCK (render); allocpad = render->srcpad ? gst_object_ref (render->srcpad) : NULL; GST_OBJECT_UNLOCK (render); if (allocpad) { ret = gst_pad_alloc_buffer (allocpad, offset, size, caps, buffer); gst_object_unref (allocpad); } gst_object_unref (render); return ret; } static GstFlowReturn gst_dvbsub_overlay_chain_text (GstPad * pad, GstBuffer * buffer) { GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (GST_PAD_PARENT (pad)); GstClockTime sub_running_time; GST_INFO_OBJECT (overlay, "subpicture/x-dvb buffer with size %u", GST_BUFFER_SIZE (buffer)); GST_LOG_OBJECT (overlay, "Video segment: %" GST_SEGMENT_FORMAT " --- Subtitle segment: %" GST_SEGMENT_FORMAT " --- BUFFER: ts=%" GST_TIME_FORMAT, &overlay->video_segment, &overlay->subtitle_segment, GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (buffer))); /* DVB subtitle packets are required to carry the PTS */ if (G_UNLIKELY (!GST_BUFFER_TIMESTAMP_IS_VALID (buffer))) { GST_WARNING_OBJECT (overlay, "Text buffer without valid timestamp, dropping"); gst_buffer_unref (buffer); return GST_FLOW_OK; } /* As the passed start and stop is equal, we shouldn't need to care about out of segment at all, * the subtitle data for the PTS is completely out of interest to us. A given display set must * carry the same PTS value. */ /* FIXME: Consider with larger than 64kB display sets, which would be cut into multiple packets, * FIXME: does our waiting + render code work when there are more than one packets before * FIXME: rendering callback will get called? */ gst_segment_set_last_stop (&overlay->subtitle_segment, GST_FORMAT_TIME, GST_BUFFER_TIMESTAMP (buffer)); sub_running_time = gst_segment_to_running_time (&overlay->subtitle_segment, GST_FORMAT_TIME, GST_BUFFER_TIMESTAMP (buffer)); GST_DEBUG_OBJECT (overlay, "SUBTITLE real running time: %" GST_TIME_FORMAT, GST_TIME_ARGS (sub_running_time)); /* FIXME: We are abusing libdvbsub pts value for tracking our gstreamer running time instead of real PTS. Should be mostly fine though... */ gst_dvbsub_overlay_process_text (overlay, buffer, sub_running_time); return GST_FLOW_OK; } static GstFlowReturn gst_dvbsub_overlay_chain_video (GstPad * pad, GstBuffer * buffer) { GstDVBSubOverlay *overlay = GST_DVBSUB_OVERLAY (GST_PAD_PARENT (pad)); GstFlowReturn ret = GST_FLOW_OK; gint64 start, stop; gint64 cstart, cstop; gboolean in_seg; GstClockTime vid_running_time, vid_running_time_end; if (overlay->format == GST_VIDEO_FORMAT_UNKNOWN) return GST_FLOW_NOT_NEGOTIATED; if (!GST_BUFFER_TIMESTAMP_IS_VALID (buffer)) goto missing_timestamp; start = GST_BUFFER_TIMESTAMP (buffer); GST_LOG_OBJECT (overlay, "Video segment: %" GST_SEGMENT_FORMAT " --- Subtitle last_stop: %" GST_TIME_FORMAT " --- BUFFER: ts=%" GST_TIME_FORMAT, &overlay->video_segment, GST_TIME_ARGS (overlay->subtitle_segment.last_stop), GST_TIME_ARGS (start)); /* ignore buffers that are outside of the current segment */ if (!GST_BUFFER_DURATION_IS_VALID (buffer)) { stop = GST_CLOCK_TIME_NONE; } else { stop = start + GST_BUFFER_DURATION (buffer); } in_seg = gst_segment_clip (&overlay->video_segment, GST_FORMAT_TIME, start, stop, &cstart, &cstop); if (!in_seg) { GST_DEBUG_OBJECT (overlay, "Buffer outside configured segment -- dropping"); gst_buffer_unref (buffer); return GST_FLOW_OK; } buffer = gst_buffer_make_metadata_writable (buffer); GST_BUFFER_TIMESTAMP (buffer) = cstart; if (GST_BUFFER_DURATION_IS_VALID (buffer)) GST_BUFFER_DURATION (buffer) = cstop - cstart; vid_running_time = gst_segment_to_running_time (&overlay->video_segment, GST_FORMAT_TIME, cstart); if (GST_BUFFER_DURATION_IS_VALID (buffer)) vid_running_time_end = gst_segment_to_running_time (&overlay->video_segment, GST_FORMAT_TIME, cstop); else vid_running_time_end = vid_running_time; GST_DEBUG_OBJECT (overlay, "Video running time: %" GST_TIME_FORMAT, GST_TIME_ARGS (vid_running_time)); gst_segment_set_last_stop (&overlay->video_segment, GST_FORMAT_TIME, GST_BUFFER_TIMESTAMP (buffer)); g_mutex_lock (overlay->dvbsub_mutex); if (!g_queue_is_empty (overlay->pending_subtitles)) { DVBSubtitles *tmp, *candidate = NULL; while (!g_queue_is_empty (overlay->pending_subtitles)) { tmp = g_queue_peek_head (overlay->pending_subtitles); if (tmp->pts > vid_running_time_end) { /* For a future video frame */ break; } else if (tmp->num_rects == 0) { /* Clear screen */ if (overlay->current_subtitle) dvb_subtitles_free (overlay->current_subtitle); overlay->current_subtitle = NULL; if (candidate) dvb_subtitles_free (candidate); candidate = NULL; g_queue_pop_head (overlay->pending_subtitles); dvb_subtitles_free (tmp); tmp = NULL; } else if (tmp->pts + tmp->page_time_out * GST_SECOND * overlay->subtitle_segment.abs_rate >= vid_running_time) { if (candidate) dvb_subtitles_free (candidate); candidate = tmp; g_queue_pop_head (overlay->pending_subtitles); } else { /* Too late */ dvb_subtitles_free (tmp); tmp = NULL; g_queue_pop_head (overlay->pending_subtitles); } } if (candidate) { GST_DEBUG_OBJECT (overlay, "Time to show the next subtitle page (%" GST_TIME_FORMAT " >= %" GST_TIME_FORMAT ") - it has %u regions", GST_TIME_ARGS (vid_running_time), GST_TIME_ARGS (candidate->pts), candidate->num_rects); dvb_subtitles_free (overlay->current_subtitle); overlay->current_subtitle = candidate; /* FIXME: Pre-convert current_subtitle to a quick-blend format, num_rects=0 means that there are no regions, e.g, a subtitle "clear" happened */ } } /* Check that we haven't hit the fallback timeout for current subtitle page */ if (overlay->current_subtitle && vid_running_time > (overlay->current_subtitle->pts + overlay->current_subtitle->page_time_out * GST_SECOND * overlay->subtitle_segment.abs_rate)) { GST_INFO_OBJECT (overlay, "Subtitle page not redefined before fallback page_time_out of %u seconds (missed data?) - deleting current page", overlay->current_subtitle->page_time_out); dvb_subtitles_free (overlay->current_subtitle); overlay->current_subtitle = NULL; } /* Now render it */ if (g_atomic_int_get (&overlay->enable) && overlay->current_subtitle) { buffer = gst_buffer_make_writable (buffer); blit_i420 (overlay, overlay->current_subtitle, buffer); } g_mutex_unlock (overlay->dvbsub_mutex); ret = gst_pad_push (overlay->srcpad, buffer); return ret; missing_timestamp: { GST_WARNING_OBJECT (overlay, "video buffer without timestamp, discarding"); gst_buffer_unref (buffer); return GST_FLOW_OK; } } static gboolean gst_dvbsub_overlay_event_video (GstPad * pad, GstEvent * event) { gboolean ret = FALSE; GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad)); GST_DEBUG_OBJECT (pad, "received video event %s", GST_EVENT_TYPE_NAME (event)); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_NEWSEGMENT: { GstFormat format; gdouble rate; gint64 start, stop, time; gboolean update; GST_DEBUG_OBJECT (render, "received new segment"); gst_event_parse_new_segment (event, &update, &rate, &format, &start, &stop, &time); if (format == GST_FORMAT_TIME) { GST_DEBUG_OBJECT (render, "VIDEO SEGMENT now: %" GST_SEGMENT_FORMAT, &render->video_segment); gst_segment_set_newsegment (&render->video_segment, update, rate, format, start, stop, time); GST_DEBUG_OBJECT (render, "VIDEO SEGMENT after: %" GST_SEGMENT_FORMAT, &render->video_segment); ret = gst_pad_push_event (render->srcpad, event); } else { GST_ELEMENT_WARNING (render, STREAM, MUX, (NULL), ("received non-TIME newsegment event on video input")); ret = FALSE; gst_event_unref (event); } break; } case GST_EVENT_FLUSH_STOP: gst_segment_init (&render->video_segment, GST_FORMAT_TIME); default: ret = gst_pad_push_event (render->srcpad, event); break; } gst_object_unref (render); return ret; } static gboolean gst_dvbsub_overlay_event_text (GstPad * pad, GstEvent * event) { gboolean ret = FALSE; GstDVBSubOverlay *render = GST_DVBSUB_OVERLAY (gst_pad_get_parent (pad)); GST_DEBUG_OBJECT (pad, "received text event %s", GST_EVENT_TYPE_NAME (event)); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_NEWSEGMENT: { GstFormat format; gdouble rate; gint64 start, stop, time; gboolean update; GST_DEBUG_OBJECT (render, "received new segment"); gst_event_parse_new_segment (event, &update, &rate, &format, &start, &stop, &time); if (format == GST_FORMAT_TIME) { GST_DEBUG_OBJECT (render, "SUBTITLE SEGMENT now: %" GST_SEGMENT_FORMAT, &render->subtitle_segment); gst_segment_set_newsegment (&render->subtitle_segment, update, rate, format, start, stop, time); GST_DEBUG_OBJECT (render, "SUBTITLE SEGMENT after: %" GST_SEGMENT_FORMAT, &render->subtitle_segment); ret = TRUE; gst_event_unref (event); } else { GST_ELEMENT_WARNING (render, STREAM, MUX, (NULL), ("received non-TIME newsegment event on subtitle sinkpad")); ret = FALSE; gst_event_unref (event); } break; } case GST_EVENT_FLUSH_STOP: GST_DEBUG_OBJECT (render, "stop flushing"); gst_dvbsub_overlay_flush_subtitles (render); gst_segment_init (&render->subtitle_segment, GST_FORMAT_TIME); gst_event_unref (event); ret = TRUE; break; case GST_EVENT_FLUSH_START: GST_DEBUG_OBJECT (render, "begin flushing"); gst_event_unref (event); ret = TRUE; break; case GST_EVENT_EOS: GST_INFO_OBJECT (render, "text EOS"); gst_event_unref (event); ret = TRUE; break; default: ret = gst_pad_push_event (render->srcpad, event); break; } gst_object_unref (render); return ret; } static gboolean plugin_init (GstPlugin * plugin) { GST_DEBUG_CATEGORY_INIT (gst_dvbsub_overlay_debug, "dvbsuboverlay", 0, "DVB subtitle overlay"); return gst_element_register (plugin, "dvbsuboverlay", GST_RANK_PRIMARY, GST_TYPE_DVBSUB_OVERLAY); } GST_PLUGIN_DEFINE (GST_VERSION_MAJOR, GST_VERSION_MINOR, "dvbsuboverlay", "DVB subtitle renderer", plugin_init, VERSION, "LGPL", GST_PACKAGE_NAME, GST_PACKAGE_ORIGIN)