/* * GStreamer * Copyright (C) 2018 Sebastian Dröge * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ #ifdef HAVE_CONFIG_H # include #endif #include #include #include #include #include "gstccconverter.h" GST_DEBUG_CATEGORY_STATIC (gst_cc_converter_debug); #define GST_CAT_DEFAULT gst_cc_converter_debug /* Ordered by the amount of information they can contain */ #define CC_CAPS \ "closedcaption/x-cea-708,format=(string) cdp; " \ "closedcaption/x-cea-708,format=(string) cc_data; " \ "closedcaption/x-cea-608,format=(string) s334-1a; " \ "closedcaption/x-cea-608,format=(string) raw" #define VAL_OR_0(v) ((v) ? (*(v)) : 0) static GstStaticPadTemplate sinktemplate = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS (CC_CAPS)); static GstStaticPadTemplate srctemplate = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (CC_CAPS)); G_DEFINE_TYPE (GstCCConverter, gst_cc_converter, GST_TYPE_BASE_TRANSFORM); #define parent_class gst_cc_converter_parent_class static gboolean gst_cc_converter_transform_size (GstBaseTransform * base, GstPadDirection direction, GstCaps * caps, gsize size, GstCaps * othercaps, gsize * othersize) { /* We can't really convert from an output size to an input size */ if (direction != GST_PAD_SINK) return FALSE; /* Assume worst-case here and over-allocate, and in ::transform() we then * downsize the buffer as needed. The worst-case is one CDP packet, which * can be up to MAX_CDP_PACKET_LEN bytes large */ *othersize = MAX_CDP_PACKET_LEN; return TRUE; } static GstCaps * gst_cc_converter_transform_caps (GstBaseTransform * base, GstPadDirection direction, GstCaps * caps, GstCaps * filter) { static GstStaticCaps non_cdp_caps = GST_STATIC_CAPS ("closedcaption/x-cea-708, format=(string)cc_data; " "closedcaption/x-cea-608,format=(string) s334-1a; " "closedcaption/x-cea-608,format=(string) raw"); static GstStaticCaps cdp_caps = GST_STATIC_CAPS ("closedcaption/x-cea-708, format=(string)cdp"); static GstStaticCaps cdp_caps_framerate = GST_STATIC_CAPS ("closedcaption/x-cea-708, format=(string)cdp, " "framerate=(fraction){60/1, 60000/1001, 50/1, 30/1, 30000/1001, 25/1, 24/1, 24000/1001}"); GstCCConverter *self = GST_CCCONVERTER (base); guint i, n; GstCaps *res, *templ; templ = gst_pad_get_pad_template_caps (base->srcpad); res = gst_caps_new_empty (); n = gst_caps_get_size (caps); for (i = 0; i < n; i++) { const GstStructure *s = gst_caps_get_structure (caps, i); const GValue *framerate = gst_structure_get_value (s, "framerate"); if (gst_structure_has_name (s, "closedcaption/x-cea-608")) { if (direction == GST_PAD_SRC) { /* SRC direction: We produce upstream caps * * Downstream wanted CEA608 caps. If it had a framerate, we * also need upstream to provide exactly that same framerate * and otherwise we don't care. * * We can convert everything to CEA608. */ res = gst_caps_merge (res, gst_static_caps_get (&cdp_caps_framerate)); res = gst_caps_merge (res, gst_static_caps_get (&non_cdp_caps)); } else { /* SINK: We produce downstream caps * * Upstream provided CEA608 caps. We can convert that to CDP if * also a CDP compatible framerate was provided, and we can convert * it to anything else regardless. * * If upstream provided a framerate we can pass that through, possibly * filtered for the CDP case. */ if (framerate) { GstCaps *tmp; GstStructure *t; /* Create caps that contain the intersection of all framerates with * the CDP allowed framerates */ tmp = gst_caps_make_writable (gst_static_caps_get (&cdp_caps_framerate)); t = gst_caps_get_structure (tmp, 0); gst_structure_set_name (t, "closedcaption/x-cea-608"); gst_structure_remove_field (t, "format"); if (gst_structure_can_intersect (s, t)) { gst_caps_unref (tmp); tmp = gst_caps_make_writable (gst_static_caps_get (&cdp_caps_framerate)); res = gst_caps_merge (res, tmp); } else { gst_caps_unref (tmp); } /* And we can convert to everything else with the given framerate */ tmp = gst_caps_make_writable (gst_static_caps_get (&non_cdp_caps)); gst_caps_set_value (tmp, "framerate", framerate); res = gst_caps_merge (res, tmp); } else { res = gst_caps_merge (res, gst_static_caps_get (&non_cdp_caps)); } } } else if (gst_structure_has_name (s, "closedcaption/x-cea-708")) { if (direction == GST_PAD_SRC) { /* SRC direction: We produce upstream caps * * Downstream wanted CEA708 caps. If downstream wants *only* CDP we * either need CDP from upstream, or anything else with a CDP * framerate. * If downstream also wants non-CDP we can accept anything. * * We pass through any framerate as-is, except for filtering * for CDP framerates if downstream wants only CDP. */ if (g_strcmp0 (gst_structure_get_string (s, "format"), "cdp") == 0) { /* Downstream wants only CDP */ /* We need CDP from upstream in that case */ res = gst_caps_merge (res, gst_static_caps_get (&cdp_caps_framerate)); /* Or anything else with a CDP framerate */ if (framerate) { GstCaps *tmp; GstStructure *t; const GValue *cdp_framerate; /* Create caps that contain the intersection of all framerates with * the CDP allowed framerates */ tmp = gst_caps_make_writable (gst_static_caps_get (&cdp_caps_framerate)); t = gst_caps_get_structure (tmp, 0); /* There's an intersection between the framerates so we can convert * into CDP with exactly those framerates from anything else */ cdp_framerate = gst_structure_get_value (t, "framerate"); tmp = gst_caps_make_writable (gst_static_caps_get (&non_cdp_caps)); gst_caps_set_value (tmp, "framerate", cdp_framerate); res = gst_caps_merge (res, tmp); } else { GstCaps *tmp, *cdp_caps; const GValue *cdp_framerate; /* Get all CDP framerates, we can accept anything that has those * framerates */ cdp_caps = gst_static_caps_get (&cdp_caps_framerate); cdp_framerate = gst_structure_get_value (gst_caps_get_structure (cdp_caps, 0), "framerate"); tmp = gst_caps_make_writable (gst_static_caps_get (&non_cdp_caps)); gst_caps_set_value (tmp, "framerate", cdp_framerate); gst_caps_unref (cdp_caps); res = gst_caps_merge (res, tmp); } } else { /* Downstream wants not only CDP, we can do everything */ res = gst_caps_merge (res, gst_static_caps_get (&cdp_caps_framerate)); res = gst_caps_merge (res, gst_static_caps_get (&non_cdp_caps)); } } else { GstCaps *tmp; /* SINK: We produce downstream caps * * Upstream provided CEA708 caps. If upstream provided CDP we can * output CDP, no matter what (-> passthrough). If upstream did not * provide CDP, we can output CDP only if the framerate fits. * We can always produce everything else apart from CDP. * * If upstream provided a framerate we pass that through for non-CDP * output, and pass it through filtered for CDP output. */ if (gst_structure_can_intersect (s, gst_caps_get_structure (gst_static_caps_get (&cdp_caps), 0))) { /* Upstream provided CDP caps, we can do everything independent of * framerate */ res = gst_caps_merge (res, gst_static_caps_get (&cdp_caps_framerate)); } else if (framerate) { const GValue *cdp_framerate; GstStructure *t; /* Upstream did not provide CDP. We can only do CDP if upstream * happened to have a CDP framerate */ /* Create caps that contain the intersection of all framerates with * the CDP allowed framerates */ tmp = gst_caps_make_writable (gst_static_caps_get (&cdp_caps_framerate)); t = gst_caps_get_structure (tmp, 0); /* There's an intersection between the framerates so we can convert * into CDP with exactly those framerates */ cdp_framerate = gst_structure_get_value (t, "framerate"); gst_caps_set_value (tmp, "framerate", cdp_framerate); res = gst_caps_merge (res, tmp); } /* We can always convert CEA708 to all non-CDP formats */ res = gst_caps_merge (res, gst_static_caps_get (&non_cdp_caps)); } } else { g_assert_not_reached (); } } GST_DEBUG_OBJECT (self, "pre filter caps %" GST_PTR_FORMAT, res); /* We can convert anything into anything but it might involve loss of * information so always filter according to the order in our template caps * in the end */ if (filter) { GstCaps *tmp; filter = gst_caps_intersect_full (templ, filter, GST_CAPS_INTERSECT_FIRST); tmp = gst_caps_intersect_full (filter, res, GST_CAPS_INTERSECT_FIRST); gst_caps_unref (res); gst_caps_unref (filter); res = tmp; } gst_caps_unref (templ); GST_DEBUG_OBJECT (self, "Transformed in direction %s caps %" GST_PTR_FORMAT, direction == GST_PAD_SRC ? "src" : "sink", caps); GST_DEBUG_OBJECT (self, "filter %" GST_PTR_FORMAT, filter); GST_DEBUG_OBJECT (self, "to %" GST_PTR_FORMAT, res); return res; } static GstCaps * gst_cc_converter_fixate_caps (GstBaseTransform * base, GstPadDirection direction, GstCaps * incaps, GstCaps * outcaps) { GstCCConverter *self = GST_CCCONVERTER (base); const GstStructure *s; GstStructure *t; const GValue *framerate; GstCaps *intersection, *templ; GST_DEBUG_OBJECT (self, "Fixating in direction %s incaps %" GST_PTR_FORMAT, direction == GST_PAD_SRC ? "src" : "sink", incaps); GST_DEBUG_OBJECT (self, "and outcaps %" GST_PTR_FORMAT, outcaps); /* Prefer passthrough if we can */ if (gst_caps_is_subset (incaps, outcaps)) { gst_caps_unref (outcaps); return GST_BASE_TRANSFORM_CLASS (parent_class)->fixate_caps (base, direction, incaps, gst_caps_ref (incaps)); } /* Otherwise prefer caps in the order of our template caps */ templ = gst_pad_get_pad_template_caps (base->srcpad); intersection = gst_caps_intersect_full (templ, outcaps, GST_CAPS_INTERSECT_FIRST); gst_caps_unref (outcaps); outcaps = intersection; outcaps = GST_BASE_TRANSFORM_CLASS (parent_class)->fixate_caps (base, direction, incaps, outcaps); /* remove any framerate that might've been added by basetransform due to * intersecting with downstream */ s = gst_caps_get_structure (incaps, 0); framerate = gst_structure_get_value (s, "framerate"); outcaps = gst_caps_make_writable (outcaps); t = gst_caps_get_structure (outcaps, 0); if (!framerate) { gst_structure_remove_field (t, "framerate"); } GST_DEBUG_OBJECT (self, "Fixated caps %" GST_PTR_FORMAT " to %" GST_PTR_FORMAT, incaps, outcaps); return outcaps; } static gboolean gst_cc_converter_set_caps (GstBaseTransform * base, GstCaps * incaps, GstCaps * outcaps) { GstCCConverter *self = GST_CCCONVERTER (base); const GstStructure *s; gboolean passthrough; self->input_caption_type = gst_video_caption_type_from_caps (incaps); self->output_caption_type = gst_video_caption_type_from_caps (outcaps); if (self->input_caption_type == GST_VIDEO_CAPTION_TYPE_UNKNOWN || self->output_caption_type == GST_VIDEO_CAPTION_TYPE_UNKNOWN) goto invalid_caps; s = gst_caps_get_structure (incaps, 0); if (!gst_structure_get_fraction (s, "framerate", &self->in_fps_n, &self->in_fps_d)) self->in_fps_n = self->in_fps_d = 0; s = gst_caps_get_structure (outcaps, 0); if (!gst_structure_get_fraction (s, "framerate", &self->out_fps_n, &self->out_fps_d)) self->out_fps_n = self->out_fps_d = 0; gst_video_time_code_clear (&self->current_output_timecode); /* Caps can be different but we can passthrough as long as they can * intersect, i.e. have same caps name and format */ passthrough = gst_caps_can_intersect (incaps, outcaps); gst_base_transform_set_passthrough (base, passthrough); GST_DEBUG_OBJECT (self, "Got caps %" GST_PTR_FORMAT " to %" GST_PTR_FORMAT " (passthrough %d)", incaps, outcaps, passthrough); return TRUE; invalid_caps: { GST_ERROR_OBJECT (self, "Invalid caps: in %" GST_PTR_FORMAT " out: %" GST_PTR_FORMAT, incaps, outcaps); return FALSE; } } struct cdp_fps_entry { guint8 fps_idx; guint fps_n, fps_d; guint max_cc_count; guint max_ccp_count; guint max_cea608_count; }; static const struct cdp_fps_entry cdp_fps_table[] = { {0x1f, 24000, 1001, 25, 22, 3 /* FIXME: alternating max cea608 count! */ }, {0x2f, 24, 1, 25, 22, 2}, {0x3f, 25, 1, 24, 22, 2}, {0x4f, 30000, 1001, 20, 18, 2}, {0x5f, 30, 1, 20, 18, 2}, {0x6f, 50, 1, 12, 11, 1}, {0x7f, 60000, 1001, 10, 9, 1}, {0x8f, 60, 1, 10, 9, 1}, }; static const struct cdp_fps_entry null_fps_entry = { 0, 0, 0, 0 }; static const struct cdp_fps_entry * cdp_fps_entry_from_id (guint8 id) { int i; for (i = 0; i < G_N_ELEMENTS (cdp_fps_table); i++) { if (cdp_fps_table[i].fps_idx == id) return &cdp_fps_table[i]; } return &null_fps_entry; } static const struct cdp_fps_entry * cdp_fps_entry_from_fps (guint fps_n, guint fps_d) { int i; for (i = 0; i < G_N_ELEMENTS (cdp_fps_table); i++) { if (cdp_fps_table[i].fps_n == fps_n && cdp_fps_table[i].fps_d == fps_d) return &cdp_fps_table[i]; } return &null_fps_entry; } static void get_framerate_output_scale (GstCCConverter * self, const struct cdp_fps_entry *in_fps_entry, gint * scale_n, gint * scale_d) { if (self->in_fps_n == 0 || self->out_fps_d == 0) { *scale_n = 1; *scale_d = 1; return; } /* compute the relative rates of the two framerates */ if (!gst_util_fraction_multiply (in_fps_entry->fps_d, in_fps_entry->fps_n, self->out_fps_n, self->out_fps_d, scale_n, scale_d)) /* we should never overflow */ g_assert_not_reached (); } static gboolean interpolate_time_code_with_framerate (GstCCConverter * self, const GstVideoTimeCode * tc, gint out_fps_n, gint out_fps_d, gint scale_n, gint scale_d, GstVideoTimeCode * out) { gchar *tc_str; gint output_n, output_d; guint output_frame; GstVideoTimeCodeFlags flags; g_return_val_if_fail (tc != NULL, FALSE); g_return_val_if_fail (out != NULL, FALSE); /* out_n/d can only be 0 if scale_n/d are 1/1 */ g_return_val_if_fail ((scale_n == 1 && scale_d == 1) || (out_fps_n != 0 && out_fps_d != 0), FALSE); if (!tc || tc->config.fps_n == 0) return FALSE; if (!gst_util_fraction_multiply (tc->frames, 1, scale_n, scale_d, &output_n, &output_d)) /* we should never overflow */ g_assert_not_reached (); tc_str = gst_video_time_code_to_string (tc); GST_TRACE_OBJECT (self, "interpolating time code %s with scale %d/%d " "to frame %d/%d", tc_str, scale_n, scale_d, output_n, output_d); g_free (tc_str); if (out_fps_n == 0 || out_fps_d == 0) { out_fps_n = tc->config.fps_n; out_fps_d = tc->config.fps_d; } flags = tc->config.flags; if ((flags & GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME) != 0 && out_fps_d != 1001 && out_fps_n != 60000 && out_fps_n != 30000) { flags &= ~GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME; } else if ((flags & GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME) == 0 && out_fps_d == 1001 && (out_fps_n == 60000 || out_fps_n == 30000)) { /* XXX: theoretically, not quite correct however this is an assumption * we have elsewhere that these framerates are always drop-framed */ flags |= GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME; } output_frame = output_n / output_d; *out = (GstVideoTimeCode) GST_VIDEO_TIME_CODE_INIT; do { /* here we try to find the next available valid timecode. The dropped * (when they exist) frames in time codes are that the beginning of each * minute */ gst_video_time_code_clear (out); gst_video_time_code_init (out, out_fps_n, out_fps_d, tc->config.latest_daily_jam, flags, tc->hours, tc->minutes, tc->seconds, output_frame, tc->field_count); output_frame++; } while ((flags & GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME) != 0 && output_frame < 10 && !gst_video_time_code_is_valid (out)); tc_str = gst_video_time_code_to_string (out); GST_TRACE_OBJECT (self, "interpolated to %s", tc_str); g_free (tc_str); return TRUE; } /* remove padding bytes from a cc_data packet. Returns the length of the new * data in @cc_data */ static guint compact_cc_data (guint8 * cc_data, guint cc_data_len) { gboolean started_ccp = FALSE; guint out_len = 0; guint i; if (cc_data_len % 3 != 0) { GST_WARNING ("Invalid cc_data buffer size"); cc_data_len = cc_data_len - (cc_data_len % 3); } for (i = 0; i < cc_data_len / 3; i++) { gboolean cc_valid = (cc_data[i * 3] & 0x04) == 0x04; guint8 cc_type = cc_data[i * 3] & 0x03; if (!started_ccp && (cc_type == 0x00 || cc_type == 0x01)) { if (cc_valid) { /* copy over valid 608 data */ cc_data[out_len++] = cc_data[i * 3]; cc_data[out_len++] = cc_data[i * 3 + 1]; cc_data[out_len++] = cc_data[i * 3 + 2]; } continue; } if (cc_type & 0x10) started_ccp = TRUE; if (!cc_valid) continue; if (cc_type == 0x00 || cc_type == 0x01) { GST_WARNING ("Invalid cc_data. cea608 bytes after cea708"); return 0; } cc_data[out_len++] = cc_data[i * 3]; cc_data[out_len++] = cc_data[i * 3 + 1]; cc_data[out_len++] = cc_data[i * 3 + 2]; } GST_LOG ("compacted cc_data from %u to %u", cc_data_len, out_len); return out_len; } static gint cc_data_extract_cea608 (guint8 * cc_data, guint cc_data_len, guint8 * cea608_field1, guint * cea608_field1_len, guint8 * cea608_field2, guint * cea608_field2_len) { guint i, field_1_len = 0, field_2_len = 0; if (cea608_field1_len) { field_1_len = *cea608_field1_len; *cea608_field1_len = 0; } if (cea608_field2_len) { field_2_len = *cea608_field2_len; *cea608_field2_len = 0; } if (cc_data_len % 3 != 0) { GST_WARNING ("Invalid cc_data buffer size %u. Truncating to a multiple " "of 3", cc_data_len); cc_data_len = cc_data_len - (cc_data_len % 3); } for (i = 0; i < cc_data_len / 3; i++) { gboolean cc_valid = (cc_data[i * 3] & 0x04) == 0x04; guint8 cc_type = cc_data[i * 3] & 0x03; if (cc_type == 0x00) { if (!cc_valid) continue; if (cea608_field1 && cea608_field1_len) { if (*cea608_field1_len + 2 > field_1_len) { GST_WARNING ("Too many cea608 input bytes %u for field 1", *cea608_field1_len + 2); return -1; } cea608_field1[(*cea608_field1_len)++] = cc_data[i * 3 + 1]; cea608_field1[(*cea608_field1_len)++] = cc_data[i * 3 + 2]; } } else if (cc_type == 0x01) { if (!cc_valid) continue; if (cea608_field2 && cea608_field2_len) { if (*cea608_field2_len + 2 > field_2_len) { GST_WARNING ("Too many cea608 input bytes %u for field 2", *cea608_field2_len + 2); return -1; } cea608_field2[(*cea608_field2_len)++] = cc_data[i * 3 + 1]; cea608_field2[(*cea608_field2_len)++] = cc_data[i * 3 + 2]; } } else { /* all cea608 packets must be at the beginning of a cc_data */ break; } } g_assert_cmpint (i * 3, <=, cc_data_len); GST_LOG ("Extracted cea608-1 of length %u and cea608-2 of length %u", VAL_OR_0 (cea608_field1_len), VAL_OR_0 (cea608_field2_len)); return i * 3; } static void store_cc_data (GstCCConverter * self, const guint8 * ccp_data, guint ccp_data_len, const guint8 * cea608_1, guint cea608_1_len, const guint8 * cea608_2, guint cea608_2_len) { GST_DEBUG_OBJECT (self, "holding data of len ccp:%u, cea608 1:%u, " "cea608 2:%u until next input buffer", ccp_data_len, cea608_1_len, cea608_2_len); if (ccp_data && ccp_data_len > 0) { memcpy (self->scratch_ccp, ccp_data, ccp_data_len); self->scratch_ccp_len = ccp_data_len; } else { self->scratch_ccp_len = 0; } g_assert_cmpint (self->scratch_ccp_len, <, sizeof (self->scratch_ccp)); if (cea608_1 && cea608_1_len > 0) { memcpy (self->scratch_cea608_1, cea608_1, cea608_1_len); self->scratch_cea608_1_len = cea608_1_len; } else { self->scratch_cea608_1_len = 0; } g_assert_cmpint (self->scratch_cea608_1_len, <, sizeof (self->scratch_cea608_1)); if (cea608_2 && cea608_2_len > 0) { memcpy (self->scratch_cea608_2, cea608_2, cea608_2_len); self->scratch_cea608_2_len = cea608_2_len; } else { self->scratch_cea608_2_len = 0; } g_assert_cmpint (self->scratch_cea608_2_len, <, sizeof (self->scratch_cea608_2)); } static gboolean combine_cc_data (GstCCConverter * self, gboolean pad_cea608, const struct cdp_fps_entry *out_fps_entry, const guint8 * ccp_data, guint ccp_data_len, const guint8 * cea608_1, guint cea608_1_len, const guint8 * cea608_2, guint cea608_2_len, guint8 * out, guint * out_size) { guint i = 0, out_i = 0, max_size = 0, cea608_1_i = 0, cea608_2_i = 0; guint cea608_output_count; guint total_cea608_1_count, total_cea608_2_count; g_assert (out); g_assert (out_size); g_assert (!ccp_data || ccp_data_len % 3 == 0); g_assert (!cea608_1 || cea608_1_len % 2 == 0); g_assert (!cea608_2 || cea608_2_len % 2 == 0); cea608_1_len /= 2; cea608_2_len /= 2; #if 0 /* FIXME: if cea608 field 2 is generated, field 1 needs to be generated, * However that is not possible for 60fps (where only one cea608 field fits) * without adding previous output buffer tracking */ g_assert_cmpint (cea608_1_len >= cea608_2_len); #endif g_assert_cmpint (cea608_1_len + cea608_2_len, <=, out_fps_entry->max_cea608_count); total_cea608_1_count = cea608_1_len; total_cea608_2_count = cea608_2_len; #if 0 /* FIXME: if cea608 field 2 is generated, field 1 needs to be generated. */ if (cea608_1_len < cea608_2_len) total_cea608_1_count += cea608_2_len - cea608_1_len; #endif max_size = ccp_data_len + (total_cea608_1_count + total_cea608_2_count) * 3; if (*out_size < max_size) { GST_WARNING_OBJECT (self, "Output data too small (%u < %u)", *out_size, max_size); return FALSE; } /* FIXME: interlacing, tff, rff, ensuring cea608 field1 is generated if * field2 exists even across packets */ cea608_output_count = cea608_1_len + cea608_2_len; if (pad_cea608) { for (i = total_cea608_1_count + total_cea608_2_count; i < out_fps_entry->max_cea608_count; i++) { /* try to pad evenly */ if (i > cea608_1_len / 2) total_cea608_1_count++; else total_cea608_2_count++; cea608_output_count++; } } GST_LOG ("writing %u cea608-1 fields and %u cea608-2 fields", total_cea608_1_count, total_cea608_2_count); g_assert_cmpint (total_cea608_1_count + total_cea608_2_count, <=, out_fps_entry->max_cea608_count); while (cea608_1_i + cea608_2_i < cea608_output_count) { if (cea608_1_i < cea608_1_len) { out[out_i++] = 0xfc; out[out_i++] = cea608_1[cea608_1_i * 2]; out[out_i++] = cea608_1[cea608_1_i * 2 + 1]; cea608_1_i++; i++; } else if (cea608_1_i < total_cea608_1_count) { out[out_i++] = 0xfc; out[out_i++] = 0x80; out[out_i++] = 0x80; cea608_1_i++; i++; } if (cea608_2_i < cea608_2_len) { out[out_i++] = 0xfd; out[out_i++] = cea608_2[cea608_2_i * 2]; out[out_i++] = cea608_2[cea608_2_i * 2 + 1]; cea608_2_i++; i++; } else if (cea608_2_i < total_cea608_2_count) { out[out_i++] = 0xfd; out[out_i++] = 0x80; out[out_i++] = 0x80; cea608_2_i++; i++; } } g_assert_cmpint (out_i / 3, <=, out_fps_entry->max_cea608_count); *out_size = out_i; if (ccp_data) { memcpy (&out[out_i], ccp_data, ccp_data_len); *out_size += ccp_data_len; } g_assert_cmpint (*out_size, <, MAX_CDP_PACKET_LEN); return TRUE; } /* takes cc_data cea608_1, cea608_2 and attempts to fit it into a hypothetical * output packet. Any leftover data is stored for later addition. Returns * whether any output can be generated. @ccp_data_len, @cea608_1_len, * @cea608_2_len are also updated to reflect the size of that data to add to * the output packet */ static gboolean fit_and_scale_cc_data (GstCCConverter * self, const struct cdp_fps_entry *in_fps_entry, const struct cdp_fps_entry *out_fps_entry, const guint8 * ccp_data, guint * ccp_data_len, const guint8 * cea608_1, guint * cea608_1_len, const guint8 * cea608_2, guint * cea608_2_len, const GstVideoTimeCode * tc) { if (!in_fps_entry || in_fps_entry->fps_n == 0) { in_fps_entry = cdp_fps_entry_from_fps (self->in_fps_n, self->in_fps_d); if (!in_fps_entry || in_fps_entry->fps_n == 0) g_assert_not_reached (); } /* This is slightly looser than checking for the exact framerate as the cdp * spec allow for 0.1% difference between framerates to be considered equal */ if (in_fps_entry->max_cc_count == out_fps_entry->max_cc_count) { if (tc && tc->config.fps_n != 0) interpolate_time_code_with_framerate (self, tc, out_fps_entry->fps_n, out_fps_entry->fps_d, 1, 1, &self->current_output_timecode); } else { int input_frame_n, input_frame_d, output_frame_n, output_frame_d; int output_time_cmp, scale_n, scale_d, rate_cmp; /* TODO: handle input discont */ /* compute the relative frame count for each */ if (!gst_util_fraction_multiply (self->in_fps_d, self->in_fps_n, self->input_frames, 1, &input_frame_n, &input_frame_d)) /* we should never overflow */ g_assert_not_reached (); if (!gst_util_fraction_multiply (self->out_fps_d, self->out_fps_n, self->output_frames, 1, &output_frame_n, &output_frame_d)) /* we should never overflow */ g_assert_not_reached (); output_time_cmp = gst_util_fraction_compare (input_frame_n, input_frame_d, output_frame_n, output_frame_d); /* compute the relative rates of the two framerates */ get_framerate_output_scale (self, in_fps_entry, &scale_n, &scale_d); rate_cmp = gst_util_fraction_compare (scale_n, scale_d, 1, 1); GST_TRACE_OBJECT (self, "performing framerate conversion at scale %d/%d " "of cc data of with sizes, ccp:%u, cea608-1:%u, cea608-2:%u", scale_n, scale_d, VAL_OR_0 (ccp_data_len), VAL_OR_0 (cea608_1_len), VAL_OR_0 (cea608_2_len)); if (rate_cmp == 0) { /* we are not scaling. Should never happen with current conditions * above */ g_assert_not_reached (); } else if (output_time_cmp < 0) { /* we can't generate an output yet */ guint cd_len = ccp_data_len ? *ccp_data_len : 0; guint c1_len = cea608_1_len ? *cea608_1_len : 0; guint c2_len = cea608_2_len ? *cea608_2_len : 0; store_cc_data (self, ccp_data, cd_len, cea608_1, c1_len, cea608_2, c2_len); if (ccp_data_len) *ccp_data_len = 0; if (cea608_1_len) *cea608_1_len = 0; if (cea608_2_len) *cea608_2_len = 0; return FALSE; } else if (rate_cmp != 0) { /* we are changing the framerate and may overflow the max output packet * size. Split them where necessary. */ gint extra_ccp = 0, extra_cea608_1 = 0, extra_cea608_2 = 0; gint ccp_off = 0, cea608_1_off = 0, cea608_2_off = 0; if (output_time_cmp == 0) { /* we have completed a cycle and can reset our counters to avoid * overflow. Anything that fits into the output packet will be written */ GST_LOG_OBJECT (self, "cycle completed, resetting frame counters"); self->scratch_ccp_len = 0; self->scratch_cea608_1_len = 0; self->scratch_cea608_2_len = 0; self->input_frames = 0; self->output_frames = 0; } if (ccp_data_len) { extra_ccp = *ccp_data_len - 3 * out_fps_entry->max_ccp_count; extra_ccp = MAX (0, extra_ccp); ccp_off = *ccp_data_len - extra_ccp; } if (cea608_1_len) { extra_cea608_1 = *cea608_1_len - 2 * out_fps_entry->max_cea608_count; extra_cea608_1 = MAX (0, extra_cea608_1); cea608_1_off = *cea608_1_len - extra_cea608_1; } if (cea608_2_len) { /* this prefers using field1 data first. This may not be quite correct */ if (extra_cea608_1 > 0) { /* all the cea608 space is for field 1 */ extra_cea608_2 = *cea608_2_len; cea608_2_off = 0; } else if (cea608_1_len) { /* cea608 space is shared between field 1 and field 2 */ extra_cea608_2 = *cea608_1_len + *cea608_2_len - 2 * out_fps_entry->max_cea608_count; extra_cea608_2 = MAX (0, extra_cea608_2); cea608_2_off = *cea608_2_len - extra_cea608_2; } else { /* all of the cea608 space is for field 2 */ extra_cea608_2 = *cea608_2_len - 2 * out_fps_entry->max_cea608_count; extra_cea608_2 = MAX (0, extra_cea608_2); cea608_2_off = *cea608_2_len - extra_cea608_2; } } if (extra_ccp > 0 || extra_cea608_1 > 0 || extra_cea608_2 > 0) { /* packet would overflow, push extra bytes into the next packet */ GST_DEBUG_OBJECT (self, "buffer would overflow by %u ccp bytes, " "%u cea608 field 1 bytes, or %u cea608 field 2 bytes", extra_ccp, extra_cea608_1, extra_cea608_2); store_cc_data (self, &ccp_data[ccp_off], extra_ccp, &cea608_1[cea608_1_off], extra_cea608_1, &cea608_2[cea608_2_off], extra_cea608_2); if (ccp_data_len) *ccp_data_len = MIN (*ccp_data_len, ccp_off); if (cea608_1_len) *cea608_1_len = MIN (*cea608_1_len, cea608_1_off); if (cea608_2_len) *cea608_2_len = MIN (*cea608_2_len, cea608_2_off); } else { GST_DEBUG_OBJECT (self, "section sizes of %u ccp bytes, " "%u cea608 field 1 bytes, and %u cea608 field 2 bytes fit within " "output packet", VAL_OR_0 (ccp_data_len), VAL_OR_0 (cea608_1_len), VAL_OR_0 (cea608_2_len)); self->scratch_ccp_len = 0; self->scratch_cea608_1_len = 0; self->scratch_cea608_2_len = 0; } } else { g_assert_not_reached (); } if (tc && tc->config.fps_n != 0) interpolate_time_code_with_framerate (self, tc, out_fps_entry->fps_n, out_fps_entry->fps_d, scale_n, scale_d, &self->current_output_timecode); } g_assert_cmpint (VAL_OR_0 (ccp_data_len) + (VAL_OR_0 (cea608_1_len) + VAL_OR_0 (cea608_2_len)) / 2 * 3, <=, 3 * out_fps_entry->max_cc_count); GST_DEBUG_OBJECT (self, "write out packet with lengths ccp:%u, cea608-1:%u, " "cea608-2:%u", VAL_OR_0 (ccp_data_len), VAL_OR_0 (cea608_1_len), VAL_OR_0 (cea608_2_len)); return TRUE; } /* Converts raw CEA708 cc_data and an optional timecode into CDP */ static guint convert_cea708_cc_data_cea708_cdp_internal (GstCCConverter * self, const guint8 * cc_data, guint cc_data_len, guint8 * cdp, guint cdp_len, const GstVideoTimeCode * tc, const struct cdp_fps_entry *fps_entry) { GstByteWriter bw; guint8 flags, checksum; guint i, len; GST_DEBUG_OBJECT (self, "writing out cdp packet from cc_data with length %u", cc_data_len); gst_byte_writer_init_with_data (&bw, cdp, cdp_len, FALSE); gst_byte_writer_put_uint16_be_unchecked (&bw, 0x9669); /* Write a length of 0 for now */ gst_byte_writer_put_uint8_unchecked (&bw, 0); gst_byte_writer_put_uint8_unchecked (&bw, fps_entry->fps_idx); if (cc_data_len / 3 > fps_entry->max_cc_count) { GST_WARNING_OBJECT (self, "Too many cc_data triplet for framerate: %u > %u", cc_data_len / 3, fps_entry->max_cc_count); cc_data_len = 3 * fps_entry->max_cc_count; } /* ccdata_present | caption_service_active */ flags = 0x42; /* time_code_present */ if (tc && tc->config.fps_n > 0) flags |= 0x80; /* reserved */ flags |= 0x01; gst_byte_writer_put_uint8_unchecked (&bw, flags); gst_byte_writer_put_uint16_be_unchecked (&bw, self->cdp_hdr_sequence_cntr); if (tc && tc->config.fps_n > 0) { gst_byte_writer_put_uint8_unchecked (&bw, 0x71); gst_byte_writer_put_uint8_unchecked (&bw, 0xc0 | (((tc->hours % 10) & 0x3) << 4) | ((tc->hours - (tc->hours % 10)) & 0xf)); gst_byte_writer_put_uint8_unchecked (&bw, 0x80 | (((tc->minutes % 10) & 0x7) << 4) | ((tc->minutes - (tc->minutes % 10)) & 0xf)); gst_byte_writer_put_uint8_unchecked (&bw, (tc->field_count < 2 ? 0x00 : 0x80) | (((tc->seconds % 10) & 0x7) << 4) | ((tc->seconds - (tc->seconds % 10)) & 0xf)); gst_byte_writer_put_uint8_unchecked (&bw, ((tc->config.flags & GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME) ? 0x80 : 0x00) | (((tc->frames % 10) & 0x3) << 4) | ((tc->frames - (tc->frames % 10)) & 0xf)); } gst_byte_writer_put_uint8_unchecked (&bw, 0x72); gst_byte_writer_put_uint8_unchecked (&bw, 0xe0 | fps_entry->max_cc_count); gst_byte_writer_put_data_unchecked (&bw, cc_data, cc_data_len); while (fps_entry->max_cc_count > cc_data_len / 3) { gst_byte_writer_put_uint8_unchecked (&bw, 0xfa); gst_byte_writer_put_uint8_unchecked (&bw, 0x00); gst_byte_writer_put_uint8_unchecked (&bw, 0x00); cc_data_len += 3; } gst_byte_writer_put_uint8_unchecked (&bw, 0x74); gst_byte_writer_put_uint16_be_unchecked (&bw, self->cdp_hdr_sequence_cntr); self->cdp_hdr_sequence_cntr++; /* We calculate the checksum afterwards */ gst_byte_writer_put_uint8_unchecked (&bw, 0); len = gst_byte_writer_get_pos (&bw); gst_byte_writer_set_pos (&bw, 2); gst_byte_writer_put_uint8_unchecked (&bw, len); checksum = 0; for (i = 0; i < len; i++) { checksum += cdp[i]; } checksum &= 0xff; checksum = 256 - checksum; cdp[len - 1] = checksum; return len; } /* Converts CDP into raw CEA708 cc_data */ static guint convert_cea708_cdp_cea708_cc_data_internal (GstCCConverter * self, const guint8 * cdp, guint cdp_len, guint8 cc_data[MAX_CDP_PACKET_LEN], GstVideoTimeCode * tc, const struct cdp_fps_entry **out_fps_entry) { GstByteReader br; guint16 u16; guint8 u8; guint8 flags; guint len = 0; const struct cdp_fps_entry *fps_entry; *out_fps_entry = &null_fps_entry; memset (tc, 0, sizeof (*tc)); /* Header + footer length */ if (cdp_len < 11) return 0; gst_byte_reader_init (&br, cdp, cdp_len); u16 = gst_byte_reader_get_uint16_be_unchecked (&br); if (u16 != 0x9669) return 0; u8 = gst_byte_reader_get_uint8_unchecked (&br); if (u8 != cdp_len) return 0; u8 = gst_byte_reader_get_uint8_unchecked (&br); fps_entry = cdp_fps_entry_from_id (u8); if (!fps_entry || fps_entry->fps_n == 0) return 0; flags = gst_byte_reader_get_uint8_unchecked (&br); /* No cc_data? */ if ((flags & 0x40) == 0) return 0; /* cdp_hdr_sequence_cntr */ gst_byte_reader_skip_unchecked (&br, 2); /* time_code_present */ if (flags & 0x80) { guint8 hours, minutes, seconds, frames, fields; gboolean drop_frame; if (gst_byte_reader_get_remaining (&br) < 5) return 0; if (gst_byte_reader_get_uint8_unchecked (&br) != 0x71) return 0; u8 = gst_byte_reader_get_uint8_unchecked (&br); if ((u8 & 0xc) != 0xc) return 0; hours = ((u8 >> 4) & 0x3) * 10 + (u8 & 0xf); u8 = gst_byte_reader_get_uint8_unchecked (&br); if ((u8 & 0x80) != 0x80) return 0; minutes = ((u8 >> 4) & 0x7) * 10 + (u8 & 0xf); u8 = gst_byte_reader_get_uint8_unchecked (&br); if (u8 & 0x80) fields = 2; else fields = 1; seconds = ((u8 >> 4) & 0x7) * 10 + (u8 & 0xf); u8 = gst_byte_reader_get_uint8_unchecked (&br); if (u8 & 0x40) return 0; drop_frame = ! !(u8 & 0x80); frames = ((u8 >> 4) & 0x3) * 10 + (u8 & 0xf); gst_video_time_code_init (tc, fps_entry->fps_n, fps_entry->fps_d, NULL, drop_frame ? GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME : GST_VIDEO_TIME_CODE_FLAGS_NONE, hours, minutes, seconds, frames, fields); } /* ccdata_present */ if (flags & 0x40) { guint8 cc_count; if (gst_byte_reader_get_remaining (&br) < 2) return 0; if (gst_byte_reader_get_uint8_unchecked (&br) != 0x72) return 0; cc_count = gst_byte_reader_get_uint8_unchecked (&br); if ((cc_count & 0xe0) != 0xe0) return 0; cc_count &= 0x1f; len = 3 * cc_count; if (gst_byte_reader_get_remaining (&br) < len) return 0; memcpy (cc_data, gst_byte_reader_get_data_unchecked (&br, len), len); } *out_fps_entry = fps_entry; /* skip everything else we don't care about */ return len; } static gboolean copy_from_stored_data (GstCCConverter * self, guint8 * out_ccp, guint * ccp_size, guint8 * cea608_1, guint * cea608_1_len, guint8 * cea608_2, guint * cea608_2_len) { guint ccp_in_size = 0, cea608_1_in_size = 0, cea608_2_in_size = 0; g_assert ((out_ccp && ccp_size) || (!out_ccp && !ccp_size)); g_assert ((cea608_1 && cea608_1_len) || (!cea608_1 && !cea608_1_len)); g_assert ((cea608_2 && cea608_2_len) || (!cea608_2 && !cea608_2_len)); if (ccp_size) { ccp_in_size = *ccp_size; *ccp_size = 0; } if (cea608_1_len) { cea608_1_in_size = *cea608_1_len; *cea608_1_len = 0; } if (cea608_2_len) { cea608_2_in_size = *cea608_2_len; *cea608_2_len = 0; } if (out_ccp && self->scratch_ccp_len > 0) { GST_DEBUG_OBJECT (self, "copying from previous scratch ccp buffer of " "%u bytes", self->scratch_ccp_len); if (ccp_in_size < *ccp_size + self->scratch_ccp_len) { GST_WARNING_OBJECT (self, "output buffer too small %u < %u", ccp_in_size, *ccp_size + self->scratch_ccp_len); goto fail; } memcpy (&out_ccp[*ccp_size], self->scratch_ccp, self->scratch_ccp_len); *ccp_size += self->scratch_ccp_len; } if (cea608_1 && self->scratch_cea608_1_len > 0) { GST_DEBUG_OBJECT (self, "copying from previous scratch cea608 field 1 " "buffer of %u bytes", self->scratch_cea608_1_len); if (cea608_1_in_size < *cea608_1_len + self->scratch_cea608_1_len) { GST_WARNING_OBJECT (self, "output buffer too small %u < %u", cea608_1_in_size, *cea608_1_len + self->scratch_cea608_1_len); goto fail; } memcpy (&cea608_1[*cea608_1_len], self->scratch_cea608_1, self->scratch_cea608_1_len); *cea608_1_len += self->scratch_cea608_1_len; } if (cea608_2 && self->scratch_cea608_2_len > 0) { GST_DEBUG_OBJECT (self, "copying from previous scratch cea608 field 2 " "buffer of %u bytes", self->scratch_cea608_2_len); if (cea608_2_in_size < *cea608_2_len + self->scratch_cea608_2_len) { GST_WARNING_OBJECT (self, "output buffer too small %u < %u", cea608_2_in_size, *cea608_2_len + self->scratch_cea608_2_len); goto fail; } memcpy (&cea608_2[*cea608_2_len], self->scratch_cea608_2, self->scratch_cea608_2_len); *cea608_2_len += self->scratch_cea608_2_len; } return TRUE; fail: if (ccp_size) *ccp_size = 0; if (cea608_1_len) *cea608_1_len = 0; if (cea608_2_len) *cea608_2_len = 0; return FALSE; } static gboolean cc_data_to_cea608_ccp (GstCCConverter * self, guint8 * cc_data, guint cc_data_len, guint8 * out_ccp, guint * ccp_size, guint8 * cea608_1, guint * cea608_1_len, guint8 * cea608_2, guint * cea608_2_len, const struct cdp_fps_entry *in_fps_entry) { guint ccp_in_size = 0, cea608_1_in_size = 0, cea608_2_in_size = 0; g_assert (cc_data || cc_data_len == 0); if (ccp_size) ccp_in_size = *ccp_size; if (cea608_1_len) cea608_1_in_size = *cea608_1_len; if (cea608_2_len) cea608_2_in_size = *cea608_2_len; if (!copy_from_stored_data (self, out_ccp, ccp_size, cea608_1, cea608_1_len, cea608_2, cea608_2_len)) goto fail; if (cc_data) { gint ccp_offset = 0; guint new_cea608_1_len = 0, new_cea608_2_len = 0; guint8 *new_cea608_1 = cea608_1, *new_cea608_2 = cea608_2; if (cea608_1_len) { new_cea608_1_len = cea608_1_in_size - *cea608_1_len; new_cea608_1 = &cea608_1[*cea608_1_len]; } if (cea608_2_len) { new_cea608_2_len = cea608_2_in_size - *cea608_2_len; new_cea608_2 = &cea608_2[*cea608_2_len]; } cc_data_len = compact_cc_data (cc_data, cc_data_len); if (cc_data_len / 3 > in_fps_entry->max_cc_count) { GST_WARNING_OBJECT (self, "Too many cc_data triples in CDP packet %u", cc_data_len / 3); cc_data_len = 3 * in_fps_entry->max_cc_count; } ccp_offset = cc_data_extract_cea608 (cc_data, cc_data_len, new_cea608_1, &new_cea608_1_len, new_cea608_2, &new_cea608_2_len); if (ccp_offset < 0) { GST_WARNING_OBJECT (self, "Failed to extract cea608 from cc_data"); goto fail; } if ((new_cea608_1_len + new_cea608_2_len) / 2 > in_fps_entry->max_cea608_count) { GST_WARNING_OBJECT (self, "Too many cea608 triples in CDP packet %u", (new_cea608_1_len + new_cea608_2_len) / 2); if ((new_cea608_1_len + new_cea608_2_len) / 2 > in_fps_entry->max_cea608_count) { new_cea608_1_len = 2 * in_fps_entry->max_cea608_count; new_cea608_2_len = 0; } else { new_cea608_2_len = 2 * in_fps_entry->max_cea608_count - new_cea608_1_len; } } if (cea608_1_len) *cea608_1_len += new_cea608_1_len; if (cea608_2_len) *cea608_2_len += new_cea608_2_len; if (out_ccp) { if (ccp_in_size < *ccp_size + cc_data_len - ccp_offset) { GST_WARNING_OBJECT (self, "output buffer too small %u < %u", ccp_in_size, *ccp_size + cc_data_len - ccp_offset); goto fail; } memcpy (&out_ccp[*ccp_size], &cc_data[ccp_offset], cc_data_len - ccp_offset); *ccp_size += cc_data_len - ccp_offset; } } return TRUE; fail: if (ccp_size) *ccp_size = 0; if (cea608_1_len) *cea608_1_len = 0; if (cea608_2_len) *cea608_2_len = 0; return FALSE; } static gboolean cdp_to_cea608_cc_data (GstCCConverter * self, GstBuffer * inbuf, guint8 * out_ccp, guint * ccp_size, guint8 * cea608_1, guint * cea608_1_len, guint8 * cea608_2, guint * cea608_2_len, GstVideoTimeCode * out_tc, const struct cdp_fps_entry **in_fps_entry) { guint8 cc_data[MAX_CDP_PACKET_LEN]; guint cc_data_len = 0; GstMapInfo in; if (inbuf) { gst_buffer_map (inbuf, &in, GST_MAP_READ); cc_data_len = convert_cea708_cdp_cea708_cc_data_internal (self, in.data, in.size, cc_data, out_tc, in_fps_entry); gst_buffer_unmap (inbuf, &in); self->input_frames++; } return cc_data_to_cea608_ccp (self, inbuf ? cc_data : NULL, cc_data_len, out_ccp, ccp_size, cea608_1, cea608_1_len, cea608_2, cea608_2_len, inbuf ? *in_fps_entry : NULL); } static GstFlowReturn convert_cea608_raw_cea608_s334_1a (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; guint i, n; n = gst_buffer_get_size (inbuf); if (n & 1) { GST_WARNING_OBJECT (self, "Invalid raw CEA608 buffer size"); gst_buffer_set_size (outbuf, 0); return GST_FLOW_OK; } n /= 2; if (n > 3) { GST_WARNING_OBJECT (self, "Too many CEA608 pairs %u", n); n = 3; } gst_buffer_set_size (outbuf, 3 * n); gst_buffer_map (inbuf, &in, GST_MAP_READ); gst_buffer_map (outbuf, &out, GST_MAP_WRITE); /* We have to assume that each value is from the first field and * don't know from which line offset it originally is */ for (i = 0; i < n; i++) { out.data[i * 3] = 0x80; out.data[i * 3 + 1] = in.data[i * 2]; out.data[i * 3 + 2] = in.data[i * 2 + 1]; } gst_buffer_unmap (inbuf, &in); gst_buffer_unmap (outbuf, &out); return GST_FLOW_OK; } static GstFlowReturn convert_cea608_raw_cea708_cc_data (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; guint i, n; n = gst_buffer_get_size (inbuf); if (n & 1) { GST_WARNING_OBJECT (self, "Invalid raw CEA608 buffer size"); gst_buffer_set_size (outbuf, 0); return GST_FLOW_OK; } n /= 2; if (n > 3) { GST_WARNING_OBJECT (self, "Too many CEA608 pairs %u", n); n = 3; } gst_buffer_set_size (outbuf, 3 * n); gst_buffer_map (inbuf, &in, GST_MAP_READ); gst_buffer_map (outbuf, &out, GST_MAP_WRITE); /* We have to assume that each value is from the first field and * don't know from which line offset it originally is */ for (i = 0; i < n; i++) { out.data[i * 3] = 0xfc; out.data[i * 3 + 1] = in.data[i * 2]; out.data[i * 3 + 2] = in.data[i * 2 + 1]; } gst_buffer_unmap (inbuf, &in); gst_buffer_unmap (outbuf, &out); return GST_FLOW_OK; } static GstFlowReturn convert_cea608_raw_cea708_cdp (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; const GstVideoTimeCodeMeta *tc_meta; const struct cdp_fps_entry *in_fps_entry, *out_fps_entry; guint cc_data_len = MAX_CDP_PACKET_LEN; guint cea608_1_len = MAX_CDP_PACKET_LEN; guint8 cc_data[MAX_CDP_PACKET_LEN], cea608_1[MAX_CEA608_LEN]; in_fps_entry = cdp_fps_entry_from_fps (self->in_fps_n, self->in_fps_d); if (!in_fps_entry || in_fps_entry->fps_n == 0) g_assert_not_reached (); if (!copy_from_stored_data (self, NULL, 0, cea608_1, &cea608_1_len, NULL, 0)) goto drop; if (inbuf) { guint n = 0; n = gst_buffer_get_size (inbuf); if (n & 1) { GST_WARNING_OBJECT (self, "Invalid raw CEA608 buffer size"); gst_buffer_set_size (outbuf, 0); return GST_FLOW_OK; } n /= 2; if (n > in_fps_entry->max_cea608_count) { GST_WARNING_OBJECT (self, "Too many CEA608 pairs %u", n); n = 3; } gst_buffer_map (inbuf, &in, GST_MAP_READ); memcpy (&cea608_1[cea608_1_len], in.data, n * 2); gst_buffer_unmap (inbuf, &in); cea608_1_len += n * 2; self->input_frames++; tc_meta = gst_buffer_get_video_time_code_meta (inbuf); } else { tc_meta = NULL; } out_fps_entry = cdp_fps_entry_from_fps (self->out_fps_n, self->out_fps_d); if (!out_fps_entry || out_fps_entry->fps_n == 0) g_assert_not_reached (); if (!fit_and_scale_cc_data (self, in_fps_entry, out_fps_entry, NULL, 0, cea608_1, &cea608_1_len, NULL, 0, tc_meta ? &tc_meta->tc : NULL)) goto drop; if (!combine_cc_data (self, TRUE, out_fps_entry, NULL, 0, cea608_1, cea608_1_len, NULL, 0, cc_data, &cc_data_len)) goto drop; gst_buffer_map (outbuf, &out, GST_MAP_WRITE); cc_data_len = convert_cea708_cc_data_cea708_cdp_internal (self, cc_data, cc_data_len, out.data, out.size, &self->current_output_timecode, out_fps_entry); self->output_frames++; gst_buffer_unmap (outbuf, &out); out: gst_buffer_set_size (outbuf, cc_data_len); return GST_FLOW_OK; drop: cc_data_len = 0; goto out; } static GstFlowReturn convert_cea608_s334_1a_cea608_raw (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; guint i, n; guint cea608 = 0; n = gst_buffer_get_size (inbuf); if (n % 3 != 0) { GST_WARNING_OBJECT (self, "Invalid S334-1A CEA608 buffer size"); n = n - (n % 3); } n /= 3; if (n > 3) { GST_WARNING_OBJECT (self, "Too many S334-1A CEA608 triplets %u", n); n = 3; } gst_buffer_map (inbuf, &in, GST_MAP_READ); gst_buffer_map (outbuf, &out, GST_MAP_WRITE); for (i = 0; i < n; i++) { if (in.data[i * 3] & 0x80) { out.data[i * 2] = in.data[i * 3 + 1]; out.data[i * 2 + 1] = in.data[i * 3 + 2]; cea608++; } } gst_buffer_unmap (inbuf, &in); gst_buffer_unmap (outbuf, &out); gst_buffer_set_size (outbuf, 2 * cea608); return GST_FLOW_OK; } static GstFlowReturn convert_cea608_s334_1a_cea708_cc_data (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; guint i, n; n = gst_buffer_get_size (inbuf); if (n % 3 != 0) { GST_WARNING_OBJECT (self, "Invalid S334-1A CEA608 buffer size"); n = n - (n % 3); } n /= 3; if (n > 3) { GST_WARNING_OBJECT (self, "Too many S334-1A CEA608 triplets %u", n); n = 3; } gst_buffer_set_size (outbuf, 3 * n); gst_buffer_map (inbuf, &in, GST_MAP_READ); gst_buffer_map (outbuf, &out, GST_MAP_WRITE); for (i = 0; i < n; i++) { out.data[i * 3] = (in.data[i * 3] & 0x80) ? 0xfc : 0xfd; out.data[i * 3 + 1] = in.data[i * 3 + 1]; out.data[i * 3 + 2] = in.data[i * 3 + 2]; } gst_buffer_unmap (inbuf, &in); gst_buffer_unmap (outbuf, &out); return GST_FLOW_OK; } static GstFlowReturn convert_cea608_s334_1a_cea708_cdp (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; const GstVideoTimeCodeMeta *tc_meta; const struct cdp_fps_entry *in_fps_entry, *out_fps_entry; guint cc_data_len = MAX_CDP_PACKET_LEN; guint cea608_1_len = MAX_CDP_PACKET_LEN, cea608_2_len = MAX_CDP_PACKET_LEN; guint8 cc_data[MAX_CDP_PACKET_LEN]; guint8 cea608_1[MAX_CEA608_LEN], cea608_2[MAX_CEA608_LEN]; guint i, n; in_fps_entry = cdp_fps_entry_from_fps (self->in_fps_n, self->in_fps_d); if (!in_fps_entry || in_fps_entry->fps_n == 0) g_assert_not_reached (); if (!copy_from_stored_data (self, NULL, 0, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len)) goto drop; if (inbuf) { n = gst_buffer_get_size (inbuf); if (n % 3 != 0) { GST_WARNING_OBJECT (self, "Invalid S334-1A CEA608 buffer size"); n = n - (n % 3); } n /= 3; if (n > in_fps_entry->max_cea608_count) { GST_WARNING_OBJECT (self, "Too many S334-1A CEA608 triplets %u", n); n = 3; } gst_buffer_map (inbuf, &in, GST_MAP_READ); for (i = 0; i < n; i++) { if (in.data[i * 3] & 0x80) { cea608_1[cea608_1_len++] = in.data[i * 3 + 1]; cea608_1[cea608_1_len++] = in.data[i * 3 + 2]; } else { cea608_2[cea608_2_len++] = in.data[i * 3 + 1]; cea608_2[cea608_2_len++] = in.data[i * 3 + 2]; } } gst_buffer_unmap (inbuf, &in); self->input_frames++; tc_meta = gst_buffer_get_video_time_code_meta (inbuf); } else { tc_meta = NULL; } out_fps_entry = cdp_fps_entry_from_fps (self->out_fps_n, self->out_fps_d); if (!out_fps_entry || out_fps_entry->fps_n == 0) g_assert_not_reached (); if (!fit_and_scale_cc_data (self, in_fps_entry, out_fps_entry, NULL, 0, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, tc_meta ? &tc_meta->tc : NULL)) { goto drop; } if (!combine_cc_data (self, TRUE, out_fps_entry, NULL, 0, cea608_1, cea608_1_len, cea608_2, cea608_2_len, cc_data, &cc_data_len)) { goto drop; } gst_buffer_map (outbuf, &out, GST_MAP_WRITE); cc_data_len = convert_cea708_cc_data_cea708_cdp_internal (self, cc_data, cc_data_len, out.data, out.size, &self->current_output_timecode, out_fps_entry); self->output_frames++; gst_buffer_unmap (outbuf, &out); out: gst_buffer_set_size (outbuf, cc_data_len); return GST_FLOW_OK; drop: cc_data_len = 0; goto out; } static GstFlowReturn convert_cea708_cc_data_cea608_raw (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; guint i, n; guint cea608 = 0; n = gst_buffer_get_size (inbuf); if (n % 3 != 0) { GST_WARNING_OBJECT (self, "Invalid raw CEA708 buffer size"); n = n - (n % 3); } n /= 3; if (n > 25) { GST_WARNING_OBJECT (self, "Too many CEA708 triplets %u", n); n = 25; } gst_buffer_map (inbuf, &in, GST_MAP_READ); gst_buffer_map (outbuf, &out, GST_MAP_WRITE); for (i = 0; i < n; i++) { /* We can only really copy the first field here as there can't be any * signalling in raw CEA608 and we must not mix the streams of different * fields */ if (in.data[i * 3] == 0xfc) { out.data[cea608 * 2] = in.data[i * 3 + 1]; out.data[cea608 * 2 + 1] = in.data[i * 3 + 2]; cea608++; } } gst_buffer_unmap (inbuf, &in); gst_buffer_unmap (outbuf, &out); gst_buffer_set_size (outbuf, 2 * cea608); return GST_FLOW_OK; } static GstFlowReturn convert_cea708_cc_data_cea608_s334_1a (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; guint i, n; guint cea608 = 0; n = gst_buffer_get_size (inbuf); if (n % 3 != 0) { GST_WARNING_OBJECT (self, "Invalid raw CEA708 buffer size"); n = n - (n % 3); } n /= 3; if (n > 25) { GST_WARNING_OBJECT (self, "Too many CEA708 triplets %u", n); n = 25; } gst_buffer_map (inbuf, &in, GST_MAP_READ); gst_buffer_map (outbuf, &out, GST_MAP_WRITE); for (i = 0; i < n; i++) { if (in.data[i * 3] == 0xfc || in.data[i * 3] == 0xfd) { /* We have to assume a line offset of 0 */ out.data[cea608 * 3] = in.data[i * 3] == 0xfc ? 0x80 : 0x00; out.data[cea608 * 3 + 1] = in.data[i * 3 + 1]; out.data[cea608 * 3 + 2] = in.data[i * 3 + 2]; cea608++; } } gst_buffer_unmap (inbuf, &in); gst_buffer_unmap (outbuf, &out); gst_buffer_set_size (outbuf, 3 * cea608); return GST_FLOW_OK; } static GstFlowReturn convert_cea708_cc_data_cea708_cdp (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo in, out; const GstVideoTimeCodeMeta *tc_meta; const struct cdp_fps_entry *in_fps_entry, *out_fps_entry; guint in_cc_data_len; guint cc_data_len = MAX_CDP_PACKET_LEN, ccp_data_len = MAX_CDP_PACKET_LEN; guint cea608_1_len = MAX_CEA608_LEN, cea608_2_len = MAX_CEA608_LEN; guint8 cc_data[MAX_CDP_PACKET_LEN], ccp_data[MAX_CDP_PACKET_LEN]; guint8 cea608_1[MAX_CEA608_LEN], cea608_2[MAX_CEA608_LEN]; guint8 *in_cc_data; if (inbuf) { gst_buffer_map (inbuf, &in, GST_MAP_READ); in_cc_data = in.data; in_cc_data_len = in.size; tc_meta = gst_buffer_get_video_time_code_meta (inbuf); self->input_frames++; } else { in_cc_data = NULL; in_cc_data_len = 0; tc_meta = NULL; } in_fps_entry = cdp_fps_entry_from_fps (self->in_fps_n, self->in_fps_d); if (!in_fps_entry || in_fps_entry->fps_n == 0) g_assert_not_reached (); out_fps_entry = cdp_fps_entry_from_fps (self->out_fps_n, self->out_fps_d); if (!out_fps_entry || out_fps_entry->fps_n == 0) g_assert_not_reached (); if (!cc_data_to_cea608_ccp (self, in_cc_data, in_cc_data_len, ccp_data, &ccp_data_len, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, in_fps_entry)) { if (inbuf) gst_buffer_unmap (inbuf, &in); goto drop; } if (inbuf) gst_buffer_unmap (inbuf, &in); if (!fit_and_scale_cc_data (self, in_fps_entry, out_fps_entry, ccp_data, &ccp_data_len, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, tc_meta ? &tc_meta->tc : NULL)) goto drop; if (!combine_cc_data (self, TRUE, out_fps_entry, ccp_data, ccp_data_len, cea608_1, cea608_1_len, cea608_2, cea608_2_len, cc_data, &cc_data_len)) goto drop; gst_buffer_map (outbuf, &out, GST_MAP_WRITE); cc_data_len = convert_cea708_cc_data_cea708_cdp_internal (self, cc_data, cc_data_len, out.data, out.size, &self->current_output_timecode, out_fps_entry); self->output_frames++; gst_buffer_unmap (outbuf, &out); out: gst_buffer_set_size (outbuf, cc_data_len); return GST_FLOW_OK; drop: cc_data_len = 0; goto out; } static GstFlowReturn convert_cea708_cdp_cea608_raw (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo out; GstVideoTimeCode tc = GST_VIDEO_TIME_CODE_INIT; guint cea608_1_len; const struct cdp_fps_entry *in_fps_entry = NULL, *out_fps_entry; gst_buffer_map (outbuf, &out, GST_MAP_WRITE); cea608_1_len = (guint) out.size; if (!cdp_to_cea608_cc_data (self, inbuf, NULL, NULL, out.data, &cea608_1_len, NULL, NULL, &tc, &in_fps_entry)) { gst_buffer_set_size (outbuf, 0); return GST_FLOW_OK; } out_fps_entry = cdp_fps_entry_from_fps (self->out_fps_n, self->out_fps_d); if (!out_fps_entry || out_fps_entry->fps_n == 0) out_fps_entry = in_fps_entry; if (fit_and_scale_cc_data (self, in_fps_entry, out_fps_entry, NULL, 0, out.data, &cea608_1_len, NULL, NULL, &tc)) { self->output_frames++; } else { cea608_1_len = 0; } gst_buffer_unmap (outbuf, &out); gst_buffer_set_size (outbuf, cea608_1_len); if (self->current_output_timecode.config.fps_n != 0 && !gst_buffer_get_video_time_code_meta (inbuf)) { gst_buffer_add_video_time_code_meta (outbuf, &self->current_output_timecode); gst_video_time_code_increment_frame (&self->current_output_timecode); } return GST_FLOW_OK; } static GstFlowReturn convert_cea708_cdp_cea608_s334_1a (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo out; GstVideoTimeCode tc = GST_VIDEO_TIME_CODE_INIT; const struct cdp_fps_entry *in_fps_entry = NULL, *out_fps_entry; guint8 cea608_1[MAX_CEA608_LEN], cea608_2[MAX_CEA608_LEN]; guint cea608_1_len = MAX_CEA608_LEN, cea608_2_len = MAX_CEA608_LEN; guint i, cc_data_len; if (!cdp_to_cea608_cc_data (self, inbuf, NULL, NULL, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, &tc, &in_fps_entry)) goto drop; out_fps_entry = cdp_fps_entry_from_fps (self->out_fps_n, self->out_fps_d); if (!out_fps_entry || out_fps_entry->fps_n == 0) out_fps_entry = in_fps_entry; if (!fit_and_scale_cc_data (self, in_fps_entry, out_fps_entry, NULL, 0, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, &tc)) goto drop; cc_data_len = gst_buffer_get_sizes (outbuf, NULL, NULL); gst_buffer_map (outbuf, &out, GST_MAP_READWRITE); if (!combine_cc_data (self, FALSE, out_fps_entry, NULL, 0, cea608_1, cea608_1_len, cea608_2, cea608_2_len, out.data, &cc_data_len)) { gst_buffer_unmap (outbuf, &out); goto drop; } for (i = 0; i < cc_data_len / 3; i++) /* We have to assume a line offset of 0 */ out.data[i * 3] = out.data[i * 3] == 0xfc ? 0x80 : 0x00; gst_buffer_unmap (outbuf, &out); self->output_frames++; gst_buffer_set_size (outbuf, cc_data_len); if (self->current_output_timecode.config.fps_n != 0 && !gst_buffer_get_video_time_code_meta (inbuf)) { gst_buffer_add_video_time_code_meta (outbuf, &self->current_output_timecode); gst_video_time_code_increment_frame (&self->current_output_timecode); } return GST_FLOW_OK; drop: gst_buffer_set_size (outbuf, 0); return GST_FLOW_OK; } static GstFlowReturn convert_cea708_cdp_cea708_cc_data (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo out; GstVideoTimeCode tc = GST_VIDEO_TIME_CODE_INIT; const struct cdp_fps_entry *in_fps_entry = NULL, *out_fps_entry; guint8 cea608_1[MAX_CEA608_LEN], cea608_2[MAX_CEA608_LEN]; guint8 ccp_data[MAX_CDP_PACKET_LEN]; guint cea608_1_len = MAX_CEA608_LEN, cea608_2_len = MAX_CEA608_LEN; guint ccp_data_len = MAX_CDP_PACKET_LEN; guint out_len = 0; if (!cdp_to_cea608_cc_data (self, inbuf, ccp_data, &ccp_data_len, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, &tc, &in_fps_entry)) goto out; out_fps_entry = cdp_fps_entry_from_fps (self->out_fps_n, self->out_fps_d); if (!out_fps_entry || out_fps_entry->fps_n == 0) out_fps_entry = in_fps_entry; if (!fit_and_scale_cc_data (self, in_fps_entry, out_fps_entry, ccp_data, &ccp_data_len, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, &tc)) goto out; gst_buffer_map (outbuf, &out, GST_MAP_WRITE); out_len = (guint) out.size; if (!combine_cc_data (self, FALSE, out_fps_entry, ccp_data, ccp_data_len, cea608_1, cea608_1_len, cea608_2, cea608_2_len, out.data, &out_len)) { gst_buffer_unmap (outbuf, &out); out_len = 0; goto out; } gst_buffer_unmap (outbuf, &out); self->output_frames++; if (self->current_output_timecode.config.fps_n != 0 && !gst_buffer_get_video_time_code_meta (inbuf)) { gst_buffer_add_video_time_code_meta (outbuf, &self->current_output_timecode); gst_video_time_code_increment_frame (&self->current_output_timecode); } out: gst_buffer_set_size (outbuf, out_len); return GST_FLOW_OK; } static GstFlowReturn convert_cea708_cdp_cea708_cdp (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstMapInfo out; GstVideoTimeCode tc = GST_VIDEO_TIME_CODE_INIT; const struct cdp_fps_entry *in_fps_entry = NULL, *out_fps_entry; guint8 cea608_1[MAX_CEA608_LEN], cea608_2[MAX_CEA608_LEN]; guint8 ccp_data[MAX_CDP_PACKET_LEN], cc_data[MAX_CDP_PACKET_LEN]; guint cea608_1_len = MAX_CEA608_LEN, cea608_2_len = MAX_CEA608_LEN; guint ccp_data_len = MAX_CDP_PACKET_LEN, cc_data_len = MAX_CDP_PACKET_LEN; guint out_len = 0; if (!cdp_to_cea608_cc_data (self, inbuf, ccp_data, &ccp_data_len, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, &tc, &in_fps_entry)) goto out; out_fps_entry = cdp_fps_entry_from_fps (self->out_fps_n, self->out_fps_d); if (!out_fps_entry || out_fps_entry->fps_n == 0) out_fps_entry = in_fps_entry; if (!fit_and_scale_cc_data (self, in_fps_entry, out_fps_entry, ccp_data, &ccp_data_len, cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, &tc)) goto out; if (!combine_cc_data (self, TRUE, out_fps_entry, ccp_data, ccp_data_len, cea608_1, cea608_1_len, cea608_2, cea608_2_len, cc_data, &cc_data_len)) { goto out; } gst_buffer_map (outbuf, &out, GST_MAP_WRITE); out_len = convert_cea708_cc_data_cea708_cdp_internal (self, cc_data, cc_data_len, out.data, out.size, &self->current_output_timecode, out_fps_entry); gst_buffer_unmap (outbuf, &out); self->output_frames++; out: gst_buffer_set_size (outbuf, out_len); return GST_FLOW_OK; } static GstFlowReturn gst_cc_converter_transform (GstCCConverter * self, GstBuffer * inbuf, GstBuffer * outbuf) { GstVideoTimeCodeMeta *tc_meta = NULL; GstFlowReturn ret = GST_FLOW_OK; GST_DEBUG_OBJECT (self, "Converting %" GST_PTR_FORMAT " from %u to %u", inbuf, self->input_caption_type, self->output_caption_type); if (inbuf) tc_meta = gst_buffer_get_video_time_code_meta (inbuf); if (tc_meta) { if (self->current_output_timecode.config.fps_n <= 0) { /* XXX: this assumes the input time codes are well-formed and increase * at the rate of one frame for each input buffer */ const struct cdp_fps_entry *in_fps_entry; gint scale_n, scale_d; in_fps_entry = cdp_fps_entry_from_fps (self->in_fps_n, self->in_fps_d); if (!in_fps_entry || in_fps_entry->fps_n == 0) scale_n = scale_d = 1; else get_framerate_output_scale (self, in_fps_entry, &scale_n, &scale_d); if (tc_meta) interpolate_time_code_with_framerate (self, &tc_meta->tc, self->out_fps_n, self->out_fps_d, scale_n, scale_d, &self->current_output_timecode); } } switch (self->input_caption_type) { case GST_VIDEO_CAPTION_TYPE_CEA608_RAW: switch (self->output_caption_type) { case GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A: ret = convert_cea608_raw_cea608_s334_1a (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_RAW: ret = convert_cea608_raw_cea708_cc_data (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_CDP: ret = convert_cea608_raw_cea708_cdp (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA608_RAW: default: g_assert_not_reached (); break; } break; case GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A: switch (self->output_caption_type) { case GST_VIDEO_CAPTION_TYPE_CEA608_RAW: ret = convert_cea608_s334_1a_cea608_raw (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_RAW: ret = convert_cea608_s334_1a_cea708_cc_data (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_CDP: ret = convert_cea608_s334_1a_cea708_cdp (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A: default: g_assert_not_reached (); break; } break; case GST_VIDEO_CAPTION_TYPE_CEA708_RAW: switch (self->output_caption_type) { case GST_VIDEO_CAPTION_TYPE_CEA608_RAW: ret = convert_cea708_cc_data_cea608_raw (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A: ret = convert_cea708_cc_data_cea608_s334_1a (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_CDP: ret = convert_cea708_cc_data_cea708_cdp (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_RAW: default: g_assert_not_reached (); break; } break; case GST_VIDEO_CAPTION_TYPE_CEA708_CDP: switch (self->output_caption_type) { case GST_VIDEO_CAPTION_TYPE_CEA608_RAW: ret = convert_cea708_cdp_cea608_raw (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A: ret = convert_cea708_cdp_cea608_s334_1a (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_RAW: ret = convert_cea708_cdp_cea708_cc_data (self, inbuf, outbuf); break; case GST_VIDEO_CAPTION_TYPE_CEA708_CDP: ret = convert_cea708_cdp_cea708_cdp (self, inbuf, outbuf); break; default: g_assert_not_reached (); break; } break; default: g_assert_not_reached (); break; } if (ret != GST_FLOW_OK) { GST_DEBUG_OBJECT (self, "returning %s", gst_flow_get_name (ret)); return ret; } GST_DEBUG_OBJECT (self, "Converted to %" GST_PTR_FORMAT, outbuf); if (gst_buffer_get_size (outbuf) > 0) { if (self->current_output_timecode.config.fps_n > 0) { gst_buffer_add_video_time_code_meta (outbuf, &self->current_output_timecode); gst_video_time_code_increment_frame (&self->current_output_timecode); } return GST_FLOW_OK; } else { return GST_FLOW_OK; } } static gboolean gst_cc_converter_transform_meta (GstBaseTransform * base, GstBuffer * outbuf, GstMeta * meta, GstBuffer * inbuf) { const GstMetaInfo *info = meta->info; /* we do this manually for framerate scaling */ if (info->api == GST_VIDEO_TIME_CODE_META_API_TYPE) return FALSE; return GST_BASE_TRANSFORM_CLASS (parent_class)->transform_meta (base, outbuf, meta, inbuf); } static gboolean can_generate_output (GstCCConverter * self) { int input_frame_n, input_frame_d, output_frame_n, output_frame_d; int output_time_cmp; if (self->in_fps_n == 0 || self->out_fps_n == 0) return FALSE; /* compute the relative frame count for each */ if (!gst_util_fraction_multiply (self->in_fps_d, self->in_fps_n, self->input_frames, 1, &input_frame_n, &input_frame_d)) /* we should never overflow */ g_assert_not_reached (); if (!gst_util_fraction_multiply (self->out_fps_d, self->out_fps_n, self->output_frames, 1, &output_frame_n, &output_frame_d)) /* we should never overflow */ g_assert_not_reached (); output_time_cmp = gst_util_fraction_compare (input_frame_n, input_frame_d, output_frame_n, output_frame_d); /* if the next output frame is at or before the current input frame */ if (output_time_cmp >= 0) return TRUE; return FALSE; } static void reset_counters (GstCCConverter * self) { self->scratch_ccp_len = 0; self->scratch_cea608_1_len = 0; self->scratch_cea608_2_len = 0; self->input_frames = 0; self->output_frames = 1; gst_video_time_code_clear (&self->current_output_timecode); gst_clear_buffer (&self->previous_buffer); } static GstFlowReturn drain_input (GstCCConverter * self) { GstBaseTransform *trans = GST_BASE_TRANSFORM (self); GstFlowReturn ret = GST_FLOW_OK; while (self->scratch_ccp_len > 0 || self->scratch_cea608_1_len > 0 || self->scratch_cea608_2_len > 0 || can_generate_output (self)) { GstBuffer *outbuf; outbuf = gst_buffer_new_allocate (NULL, MAX_CDP_PACKET_LEN, NULL); ret = gst_cc_converter_transform (self, NULL, outbuf); if (gst_buffer_get_size (outbuf) <= 0) { /* try to move the output along */ self->input_frames++; gst_buffer_unref (outbuf); continue; } else if (ret != GST_FLOW_OK) { gst_buffer_unref (outbuf); return ret; } ret = gst_pad_push (GST_BASE_TRANSFORM_SRC_PAD (trans), outbuf); if (ret != GST_FLOW_OK) { return ret; } } return ret; } static GstFlowReturn gst_cc_converter_generate_output (GstBaseTransform * base, GstBuffer ** outbuf) { GstBaseTransformClass *bclass = GST_BASE_TRANSFORM_GET_CLASS (base); GstCCConverter *self = GST_CCCONVERTER (base); GstBuffer *inbuf = base->queued_buf; GstFlowReturn ret; *outbuf = NULL; base->queued_buf = NULL; if (!inbuf && !can_generate_output (self)) { return GST_FLOW_OK; } if (gst_base_transform_is_passthrough (base)) { *outbuf = inbuf; ret = GST_FLOW_OK; } else { if (inbuf && GST_BUFFER_IS_DISCONT (inbuf)) { ret = drain_input (self); reset_counters (self); if (ret != GST_FLOW_OK) return ret; } *outbuf = gst_buffer_new_allocate (NULL, MAX_CDP_PACKET_LEN, NULL); if (*outbuf == NULL) goto no_buffer; if (inbuf) gst_buffer_replace (&self->previous_buffer, inbuf); if (bclass->copy_metadata) { if (!bclass->copy_metadata (base, self->previous_buffer, *outbuf)) { /* something failed, post a warning */ GST_ELEMENT_WARNING (self, STREAM, NOT_IMPLEMENTED, ("could not copy metadata"), (NULL)); } } ret = gst_cc_converter_transform (self, inbuf, *outbuf); if (gst_buffer_get_size (*outbuf) <= 0) { gst_buffer_unref (*outbuf); *outbuf = NULL; ret = GST_FLOW_OK; } if (inbuf) gst_buffer_unref (inbuf); } return ret; no_buffer: { if (inbuf) gst_buffer_unref (inbuf); *outbuf = NULL; GST_WARNING_OBJECT (self, "could not allocate buffer"); return GST_FLOW_ERROR; } } static gboolean gst_cc_converter_sink_event (GstBaseTransform * trans, GstEvent * event) { GstCCConverter *self = GST_CCCONVERTER (trans); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_EOS: GST_DEBUG_OBJECT (self, "received EOS"); drain_input (self); /* fallthrough */ case GST_EVENT_FLUSH_START: reset_counters (self); break; default: break; } return GST_BASE_TRANSFORM_CLASS (parent_class)->sink_event (trans, event); } static gboolean gst_cc_converter_start (GstBaseTransform * base) { GstCCConverter *self = GST_CCCONVERTER (base); /* Resetting this is not really needed but makes debugging easier */ self->cdp_hdr_sequence_cntr = 0; self->current_output_timecode = (GstVideoTimeCode) GST_VIDEO_TIME_CODE_INIT; self->input_frames = 0; self->output_frames = 1; self->scratch_ccp_len = 0; self->scratch_cea608_1_len = 0; self->scratch_cea608_2_len = 0; return TRUE; } static gboolean gst_cc_converter_stop (GstBaseTransform * base) { GstCCConverter *self = GST_CCCONVERTER (base); gst_video_time_code_clear (&self->current_output_timecode); gst_clear_buffer (&self->previous_buffer); return TRUE; } static void gst_cc_converter_class_init (GstCCConverterClass * klass) { GstElementClass *gstelement_class; GstBaseTransformClass *basetransform_class; gstelement_class = (GstElementClass *) klass; basetransform_class = (GstBaseTransformClass *) klass; gst_element_class_set_static_metadata (gstelement_class, "Closed Caption Converter", "Filter/ClosedCaption", "Converts Closed Captions between different formats", "Sebastian Dröge "); gst_element_class_add_static_pad_template (gstelement_class, &sinktemplate); gst_element_class_add_static_pad_template (gstelement_class, &srctemplate); basetransform_class->start = GST_DEBUG_FUNCPTR (gst_cc_converter_start); basetransform_class->stop = GST_DEBUG_FUNCPTR (gst_cc_converter_stop); basetransform_class->sink_event = GST_DEBUG_FUNCPTR (gst_cc_converter_sink_event); basetransform_class->transform_size = GST_DEBUG_FUNCPTR (gst_cc_converter_transform_size); basetransform_class->transform_caps = GST_DEBUG_FUNCPTR (gst_cc_converter_transform_caps); basetransform_class->fixate_caps = GST_DEBUG_FUNCPTR (gst_cc_converter_fixate_caps); basetransform_class->set_caps = GST_DEBUG_FUNCPTR (gst_cc_converter_set_caps); basetransform_class->transform_meta = GST_DEBUG_FUNCPTR (gst_cc_converter_transform_meta); basetransform_class->generate_output = GST_DEBUG_FUNCPTR (gst_cc_converter_generate_output); basetransform_class->passthrough_on_same_caps = TRUE; GST_DEBUG_CATEGORY_INIT (gst_cc_converter_debug, "ccconverter", 0, "Closed Caption converter"); } static void gst_cc_converter_init (GstCCConverter * self) { }