/* * GStreamer * Copyright (C) 2018 Edward Hervey * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:element-line21decoder * @title: line21decoder * */ #ifdef HAVE_CONFIG_H # include #endif #include #include #include #include "gstline21dec.h" GST_DEBUG_CATEGORY_STATIC (gst_line_21_decoder_debug); #define GST_CAT_DEFAULT gst_line_21_decoder_debug enum { PROP_0, }; #define SUPPORTED_FORMATS "{ I420, YUY2, YVYU, UYVY, VYUY, v210 }" static GstStaticPadTemplate sinktemplate = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (SUPPORTED_FORMATS))); static GstStaticPadTemplate srctemplate = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (SUPPORTED_FORMATS))); G_DEFINE_TYPE (GstLine21Decoder, gst_line_21_decoder, GST_TYPE_VIDEO_FILTER); #define parent_class gst_line_21_decoder_parent_class static void gst_line_21_decoder_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_line_21_decoder_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static GstStateChangeReturn gst_line_21_decoder_change_state (GstElement * element, GstStateChange transition); static void gst_line_21_decoder_finalize (GObject * self); static gboolean gst_line_21_decoder_set_info (GstVideoFilter * filter, GstCaps * incaps, GstVideoInfo * in_info, GstCaps * outcaps, GstVideoInfo * out_info); static GstFlowReturn gst_line_21_decoder_transform_ip (GstVideoFilter * filter, GstVideoFrame * frame); static GstFlowReturn gst_line_21_decoder_prepare_output_buffer (GstBaseTransform * trans, GstBuffer * in, GstBuffer ** out); static void gst_line_21_decoder_class_init (GstLine21DecoderClass * klass) { GObjectClass *gobject_class; GstElementClass *gstelement_class; GstBaseTransformClass *transform_class; GstVideoFilterClass *filter_class; gobject_class = (GObjectClass *) klass; gstelement_class = (GstElementClass *) klass; transform_class = (GstBaseTransformClass *) klass; filter_class = (GstVideoFilterClass *) klass; gobject_class->set_property = gst_line_21_decoder_set_property; gobject_class->get_property = gst_line_21_decoder_get_property; gobject_class->finalize = gst_line_21_decoder_finalize; gstelement_class->change_state = GST_DEBUG_FUNCPTR (gst_line_21_decoder_change_state); gst_element_class_set_static_metadata (gstelement_class, "Line 21 CC Decoder", "Filter", "Extract line21 CC from SD video streams", "Edward Hervey "); gst_element_class_add_static_pad_template (gstelement_class, &sinktemplate); gst_element_class_add_static_pad_template (gstelement_class, &srctemplate); transform_class->prepare_output_buffer = gst_line_21_decoder_prepare_output_buffer; filter_class->set_info = gst_line_21_decoder_set_info; filter_class->transform_frame_ip = gst_line_21_decoder_transform_ip; GST_DEBUG_CATEGORY_INIT (gst_line_21_decoder_debug, "line21decoder", 0, "Line 21 CC Decoder"); vbi_initialize_gst_debug (); } static void gst_line_21_decoder_reset (GstLine21Decoder * self) { self->line21_offset = -1; self->max_line_probes = 40; if (self->info) { gst_video_info_free (self->info); self->info = NULL; } g_free (self->converted_lines); self->converted_lines = NULL; } static void gst_line_21_decoder_init (GstLine21Decoder * filter) { gst_line_21_decoder_reset (filter); } static void gst_line_21_decoder_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { /* GstLine21Decoder *filter = GST_LINE21DECODER (object); */ switch (prop_id) { default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static void gst_line_21_decoder_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { /* GstLine21Decoder *filter = GST_LINE21DECODER (object); */ switch (prop_id) { default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static vbi_pixfmt vbi_pixfmt_from_gst_video_format (GstVideoFormat format, gboolean * convert_v210) { *convert_v210 = FALSE; switch (format) { case GST_VIDEO_FORMAT_I420: return VBI_PIXFMT_YUV420; case GST_VIDEO_FORMAT_YUY2: return VBI_PIXFMT_YUYV; case GST_VIDEO_FORMAT_YVYU: return VBI_PIXFMT_YVYU; case GST_VIDEO_FORMAT_UYVY: return VBI_PIXFMT_UYVY; case GST_VIDEO_FORMAT_VYUY: return VBI_PIXFMT_VYUY; /* for v210 we'll convert it to I420 luma */ case GST_VIDEO_FORMAT_v210: *convert_v210 = TRUE; return VBI_PIXFMT_YUV420; /* All the other formats are not really bullet-proof. Force conversion */ default: return 0; } #undef NATIVE_VBI_FMT } static gboolean gst_line_21_decoder_set_info (GstVideoFilter * filter, GstCaps * incaps, GstVideoInfo * in_info, GstCaps * outcaps, GstVideoInfo * out_info) { GstLine21Decoder *self = (GstLine21Decoder *) filter; vbi_pixfmt fmt = vbi_pixfmt_from_gst_video_format (GST_VIDEO_INFO_FORMAT (in_info), &self->convert_v210); GST_DEBUG_OBJECT (filter, "caps %" GST_PTR_FORMAT, incaps); GST_DEBUG_OBJECT (filter, "plane_stride:%u , comp_stride:%u , pstride:%u", GST_VIDEO_INFO_PLANE_STRIDE (in_info, 0), GST_VIDEO_INFO_COMP_STRIDE (in_info, 0), GST_VIDEO_INFO_COMP_PSTRIDE (in_info, 0)); GST_DEBUG_OBJECT (filter, "#planes : %d #components : %d", GST_VIDEO_INFO_N_PLANES (in_info), GST_VIDEO_INFO_N_COMPONENTS (in_info)); if (GST_VIDEO_INFO_WIDTH (in_info) != 720) { GST_DEBUG_OBJECT (filter, "Only 720 pixel wide formats are supported"); self->compatible_format = FALSE; return TRUE; } if (fmt == 0) { if (GST_VIDEO_INFO_FORMAT (in_info) == GST_VIDEO_FORMAT_v210) { GST_DEBUG_OBJECT (filter, "Format not supported natively, Adding conversion to YUY2"); self->compatible_format = TRUE; self->convert_v210 = TRUE; } else { GST_DEBUG_OBJECT (filter, "Unsupported format"); self->compatible_format = FALSE; } return TRUE; } if (GST_VIDEO_INFO_WIDTH (in_info) == 720 && GST_VIDEO_INFO_HEIGHT (in_info) >= 200) { GST_DEBUG_OBJECT (filter, "Compatible size!"); GST_DEBUG_OBJECT (filter, "Compatible format plane_stride:%u , comp_stride:%u , pstride:%u", GST_VIDEO_INFO_PLANE_STRIDE (in_info, 0), GST_VIDEO_INFO_COMP_STRIDE (in_info, 0), GST_VIDEO_INFO_COMP_PSTRIDE (in_info, 0)); self->compatible_format = TRUE; if (self->convert_v210) { self->info = gst_video_info_new (); gst_video_info_set_format (self->info, GST_VIDEO_FORMAT_I420, GST_VIDEO_INFO_WIDTH (in_info), GST_VIDEO_INFO_HEIGHT (in_info)); /* Allocate space for two *I420* Y lines (with stride) */ self->converted_lines = g_malloc0 (2 * GST_VIDEO_INFO_COMP_STRIDE (self->info, 0)); } else self->info = gst_video_info_copy (in_info); /* initialize the decoder */ vbi_raw_decoder_init (&self->zvbi_decoder); /* We either deal with PAL (625 lines) or NTSC (525 lines) */ self->zvbi_decoder.scanning = 525; /* The pixel format. Quite a few formats are handled by zvbi, but * some are not and require conversion (or cheating) */ self->zvbi_decoder.sampling_format = fmt; /* Sampling rate. For BT.601 it's 13.5MHz */ self->zvbi_decoder.sampling_rate = 13.5e6; /* Hz (i.e. BT.601) */ /* Stride */ self->zvbi_decoder.bytes_per_line = GST_VIDEO_INFO_COMP_STRIDE (self->info, 0); /* Sampling starts 9.7 µs from the front edge of the hor. sync pulse. You may have to adjust this. NOTE : This is actually ignored in the code ... */ self->zvbi_decoder.offset = 9.7e-6 * 13.5e6; /* The following values indicate what we are feeding to zvbi. * By setting start[0] = 21, we are telling zvbi that the very * beginning of the data we are feeding to it corresponds to * line 21 (which is where CC1/CC3 is located). * * Then by specifying count[0] = 1, we are telling it to only * scan 1 line from the beginning of the data. * * It is more efficient and flexible to do it this way, since * we can then control what we are feeding it (i.e. *we* will * figure out the offset to line 21, which might or might not * be the beginning of the buffer data, and feed data from * there). This would also allows us to have a "scanning" mode * where we repeatedly provide it with pairs of lines until it * finds something. */ self->zvbi_decoder.start[0] = 21; self->zvbi_decoder.count[0] = 1; /* Second field. */ self->zvbi_decoder.start[1] = 284; self->zvbi_decoder.count[1] = 1; /* FIXME : Adjust according to the info.interlace_mode ! */ self->zvbi_decoder.interlaced = TRUE; /* synchronous is essentially top-field-first. * WARNING : zvbi doesn't support bottom-field-first. */ self->zvbi_decoder.synchronous = TRUE; /* Specify the services you want. Adjust based on whether we * have PAL or NTSC */ vbi_raw_decoder_add_services (&self->zvbi_decoder, VBI_SLICED_CAPTION_525, /* strict */ 0); } else self->compatible_format = FALSE; return TRUE; } static GstFlowReturn gst_line_21_decoder_prepare_output_buffer (GstBaseTransform * trans, GstBuffer * in, GstBuffer ** out) { GstLine21Decoder *self = (GstLine21Decoder *) trans; GST_DEBUG_OBJECT (trans, "compatible_format:%d", self->compatible_format); if (self->compatible_format) { /* Make the output buffer writable */ *out = gst_buffer_make_writable (in); return GST_FLOW_OK; } return GST_BASE_TRANSFORM_CLASS (gst_line_21_decoder_parent_class)->prepare_output_buffer (trans, in, out); } static void convert_line_v210_luma (const guint8 * orig, guint8 * dest, guint width) { guint i; guint32 a, b, c, d; guint8 *y = dest; for (i = 0; i < width - 5; i += 6) { a = GST_READ_UINT32_LE (orig + (i / 6) * 16 + 0); b = GST_READ_UINT32_LE (orig + (i / 6) * 16 + 4); c = GST_READ_UINT32_LE (orig + (i / 6) * 16 + 8); d = GST_READ_UINT32_LE (orig + (i / 6) * 16 + 12); *y++ = (a >> 12) & 0xff; *y++ = (b >> 2) & 0xff; *y++ = (b >> 22) & 0xff; *y++ = (c >> 12) & 0xff; *y++ = (d >> 2) & 0xff; *y++ = (d >> 22) & 0xff; } } static guint8 * get_video_data (GstLine21Decoder * self, GstVideoFrame * frame, gint line) { guint8 *data = self->converted_lines; guint8 *v210; if (!self->convert_v210) return (guint8 *) GST_VIDEO_FRAME_PLANE_DATA (frame, 0) + line * GST_VIDEO_INFO_COMP_STRIDE (self->info, 0); v210 = (guint8 *) GST_VIDEO_FRAME_PLANE_DATA (frame, 0) + line * GST_VIDEO_FRAME_COMP_STRIDE (frame, 0); /* Convert v210 to I420 */ convert_line_v210_luma (v210, data, GST_VIDEO_FRAME_WIDTH (frame)); v210 += GST_VIDEO_FRAME_COMP_STRIDE (frame, 0); convert_line_v210_luma (v210, data + GST_VIDEO_INFO_COMP_STRIDE (self->info, 0), GST_VIDEO_FRAME_WIDTH (frame)); GST_MEMDUMP ("converted", self->converted_lines, 64); return self->converted_lines; } /* Call this to scan for CC * Returns TRUE if it was found and set, else FALSE */ static gboolean gst_line_21_decoder_scan (GstLine21Decoder * self, GstVideoFrame * frame) { gint i; vbi_sliced sliced[52]; gboolean found = FALSE; guint8 *data; GST_DEBUG_OBJECT (self, "Starting probing. max_line_probes:%d", self->max_line_probes); i = self->line21_offset; if (i == -1) { GST_DEBUG_OBJECT (self, "Scanning from the beginning"); i = 0; } for (; i < self->max_line_probes; i++) { gint n_lines; data = get_video_data (self, frame, i); /* Scan until we get n_lines == 2 */ n_lines = vbi_raw_decode (&self->zvbi_decoder, data, sliced); GST_DEBUG_OBJECT (self, "i:%d n_lines:%d", i, n_lines); if (n_lines == 2) { GST_DEBUG_OBJECT (self, "Found 2 CC lines at offset %d", i); self->line21_offset = i; found = TRUE; break; } } if (!found) { GST_DEBUG_OBJECT (self, "No CC found"); self->line21_offset = -1; } else { guint8 ccdata[6] = { 0xfc, 0x80, 0x80, 0xfd, 0x80, 0x80 }; /* Initialize the ccdata */ ccdata[1] = sliced[0].data[0]; ccdata[2] = sliced[0].data[1]; ccdata[4] = sliced[1].data[0]; ccdata[5] = sliced[1].data[1]; gst_buffer_add_video_caption_meta (frame->buffer, GST_VIDEO_CAPTION_TYPE_CEA608_IN_CEA708_RAW, ccdata, 6); GST_TRACE_OBJECT (self, "Got CC 0x%02x 0x%02x / 0x%02x 0x%02x '%c%c / %c%c'", ccdata[1], ccdata[2], ccdata[4], ccdata[5], g_ascii_isprint (ccdata[1] & 0x7f) ? ccdata[1] & 0x7f : '.', g_ascii_isprint (ccdata[2] & 0x7f) ? ccdata[2] & 0x7f : '.', g_ascii_isprint (ccdata[4] & 0x7f) ? ccdata[4] & 0x7f : '.', g_ascii_isprint (ccdata[5] & 0x7f) ? ccdata[5] & 0x7f : '.'); } return found; } static GstFlowReturn gst_line_21_decoder_transform_ip (GstVideoFilter * filter, GstVideoFrame * frame) { GstLine21Decoder *self = (GstLine21Decoder *) filter; if (!self->compatible_format) return GST_FLOW_OK; gst_line_21_decoder_scan (self, frame); return GST_FLOW_OK; } static GstStateChangeReturn gst_line_21_decoder_change_state (GstElement * element, GstStateChange transition) { GstStateChangeReturn ret; GstLine21Decoder *filter = GST_LINE21DECODER (element); switch (transition) { case GST_STATE_CHANGE_NULL_TO_READY: break; case GST_STATE_CHANGE_READY_TO_PAUSED: break; case GST_STATE_CHANGE_PAUSED_TO_PLAYING: break; default: break; } ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); if (ret != GST_STATE_CHANGE_SUCCESS) return ret; switch (transition) { case GST_STATE_CHANGE_PLAYING_TO_PAUSED: break; case GST_STATE_CHANGE_PAUSED_TO_READY: gst_line_21_decoder_reset (filter); break; case GST_STATE_CHANGE_READY_TO_NULL: default: break; } return ret; } static void gst_line_21_decoder_finalize (GObject * object) { G_OBJECT_CLASS (parent_class)->finalize (object); }