gstreamer/subprojects/gst-plugins-bad/ext/closedcaption/gstcccombiner.c

1510 lines
49 KiB
C
Raw Normal View History

/*
* GStreamer
* Copyright (C) 2018 Sebastian Dröge <sebastian@centricular.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
#ifdef HAVE_CONFIG_H
# include <config.h>
#endif
#include <gst/gst.h>
#include <gst/base/base.h>
#include <gst/video/video.h>
#include <string.h>
#include "ccutils.h"
#include "gstcccombiner.h"
GST_DEBUG_CATEGORY_STATIC (gst_cc_combiner_debug);
#define GST_CAT_DEFAULT gst_cc_combiner_debug
static GstStaticPadTemplate sinktemplate = GST_STATIC_PAD_TEMPLATE ("sink",
GST_PAD_SINK,
GST_PAD_ALWAYS,
GST_STATIC_CAPS_ANY);
static GstStaticPadTemplate srctemplate = GST_STATIC_PAD_TEMPLATE ("src",
GST_PAD_SRC,
GST_PAD_ALWAYS,
GST_STATIC_CAPS_ANY);
static GstStaticPadTemplate captiontemplate =
GST_STATIC_PAD_TEMPLATE ("caption",
GST_PAD_SINK,
GST_PAD_REQUEST,
GST_STATIC_CAPS
("closedcaption/x-cea-608,format={ (string) raw, (string) s334-1a}; "
"closedcaption/x-cea-708,format={ (string) cc_data, (string) cdp }"));
#define parent_class gst_cc_combiner_parent_class
G_DEFINE_TYPE (GstCCCombiner, gst_cc_combiner, GST_TYPE_AGGREGATOR);
GST_ELEMENT_REGISTER_DEFINE (cccombiner, "cccombiner",
GST_RANK_NONE, GST_TYPE_CCCOMBINER);
enum
{
PROP_0,
PROP_SCHEDULE,
PROP_OUTPUT_PADDING,
PROP_MAX_SCHEDULED,
PROP_CEA608_PADDING_STRATEGY,
PROP_CEA608_VALID_PADDING_TIMEOUT,
PROP_SCHEDULE_TIMEOUT,
PROP_INPUT_META_PROCESSING,
};
#define DEFAULT_MAX_SCHEDULED 30
#define DEFAULT_SCHEDULE TRUE
#define DEFAULT_OUTPUT_PADDING TRUE
#define DEFAULT_CEA608_PADDING_STRATEGY CC_BUFFER_CEA608_PADDING_STRATEGY_VALID
#define DEFAULT_CEA608_VALID_PADDING_TIMEOUT GST_CLOCK_TIME_NONE
#define DEFAULT_SCHEDULE_TIMEOUT GST_CLOCK_TIME_NONE
#define DEFAULT_INPUT_META_PROCESSING CCCOMBINER_INPUT_PROCESSING_APPEND
#define GST_TYPE_CCCOMBINER_INPUT_META_PROCESSING (gst_cccombiner_input_meta_processing_get_type())
static GType
gst_cccombiner_input_meta_processing_get_type (void)
{
static GType cccombiner_input_meta_processing_type = 0;
static const GEnumValue cccombiner_input_meta_processing[] = {
{CCCOMBINER_INPUT_PROCESSING_APPEND,
"append aggregated CC to existing metas on video buffer", "append"},
{CCCOMBINER_INPUT_PROCESSING_DROP,
"drop existing CC metas on input video buffer", "drop"},
{CCCOMBINER_INPUT_PROCESSING_FAVOR,
"discard aggregated CC when input video buffers hold CC metas already",
"favor"},
{CCCOMBINER_INPUT_PROCESSING_FORCE,
"discard aggregated CC even when input video buffers do not hold CC",
"force"},
{0, NULL, NULL},
};
if (!cccombiner_input_meta_processing_type) {
cccombiner_input_meta_processing_type =
g_enum_register_static ("GstCCCombinerInputProcessing",
cccombiner_input_meta_processing);
}
return cccombiner_input_meta_processing_type;
}
typedef struct
{
GstVideoCaptionType caption_type;
GstBuffer *buffer;
} CaptionData;
typedef struct
{
GstBuffer *buffer;
GstClockTime running_time;
GstClockTime stream_time;
} CaptionQueueItem;
static void
caption_data_clear (CaptionData * data)
{
gst_buffer_unref (data->buffer);
}
static void
gst_cc_combiner_finalize (GObject * object)
{
GstCCCombiner *self = GST_CCCOMBINER (object);
gst_clear_object (&self->caption_pad);
g_array_unref (self->current_frame_captions);
self->current_frame_captions = NULL;
gst_clear_object (&self->cc_buffer);
G_OBJECT_CLASS (parent_class)->finalize (object);
}
#define GST_FLOW_NEED_DATA GST_FLOW_CUSTOM_SUCCESS
static guint
extract_cdp (GstCCCombiner * self, const guint8 * cdp, guint cdp_len,
guint8 * cc_data)
{
const struct cdp_fps_entry *out_fps_entry;
GstVideoTimeCode tc = GST_VIDEO_TIME_CODE_INIT;
return convert_cea708_cdp_to_cc_data (GST_OBJECT (self), cdp, cdp_len,
cc_data, &tc, &out_fps_entry);
}
#define MAX_CEA608_LEN 32
#define CDP_MODE (GST_CC_CDP_MODE_CC_DATA | GST_CC_CDP_MODE_TIME_CODE)
static GstBuffer *
make_cdp_buffer (GstCCCombiner * self, const guint8 * cc_data,
guint cc_data_len, const struct cdp_fps_entry *fps_entry,
const GstVideoTimeCode * tc)
{
guint len;
GstBuffer *ret = gst_buffer_new_allocate (NULL, MAX_CDP_PACKET_LEN, NULL);
GstMapInfo map;
gst_buffer_map (ret, &map, GST_MAP_WRITE);
len = convert_cea708_cc_data_to_cdp (GST_OBJECT (self), CDP_MODE,
self->cdp_hdr_sequence_cntr, cc_data, cc_data_len, map.data, map.size,
tc, fps_entry);
self->cdp_hdr_sequence_cntr++;
gst_buffer_unmap (ret, &map);
gst_buffer_set_size (ret, len);
return ret;
}
static GstBuffer *
make_buffer (GstCCCombiner * self, const guint8 * cc_data, guint cc_data_len)
{
GstBuffer *ret = gst_buffer_new_allocate (NULL, cc_data_len, NULL);
gst_buffer_fill (ret, 0, cc_data, cc_data_len);
return ret;
}
static void
write_cc_data_to (GstCCCombiner * self, GstBuffer * buffer)
{
GstMapInfo map;
guint len;
gst_buffer_map (buffer, &map, GST_MAP_WRITE);
len = map.size;
cc_buffer_take_cc_data (self->cc_buffer, self->cdp_fps_entry, map.data, &len);
gst_buffer_unmap (buffer, &map);
gst_buffer_set_size (buffer, len);
}
static void
prepend_s334_to_cea608 (guint field, guint8 * data, guint * len,
guint alloc_len)
{
int i;
g_assert (*len / 2 * 3 <= alloc_len);
for (i = *len / 2; i >= 0; i--) {
data[i * 3 + 2] = data[i * 2 + 1];
data[i * 3 + 1] = data[i * 2 + 0];
data[i * 3 + 0] = field == 0 ? 0x80 : 0x00;
}
*len = *len * 3 / 2;
}
static void
take_s334_both_fields (GstCCCombiner * self, GstBuffer * buffer)
{
GstMapInfo out = GST_MAP_INFO_INIT;
guint s334_len, cc_data_len, i;
gst_buffer_map (buffer, &out, GST_MAP_READWRITE);
cc_data_len = out.size;
cc_buffer_take_cc_data (self->cc_buffer, self->cdp_fps_entry,
out.data, &cc_data_len);
s334_len = drop_ccp_from_cc_data (out.data, cc_data_len);
if (s334_len < 0) {
s334_len = 0;
goto out;
}
for (i = 0; i < s334_len / 3; i++) {
guint byte = out.data[i * 3];
/* We have to assume a line offset of 0 */
out.data[i * 3] = (byte == 0xfc || byte == 0xf8) ? 0x80 : 0x00;
}
out:
gst_buffer_unmap (buffer, &out);
gst_buffer_set_size (buffer, s334_len);
}
static void
schedule_cdp (GstCCCombiner * self, const GstVideoTimeCode * tc,
const guint8 * data, guint len, GstClockTime pts, GstClockTime duration)
{
guint8 cc_data[MAX_CDP_PACKET_LEN];
guint cc_data_len;
cc_data_len = extract_cdp (self, data, len, cc_data);
if (cc_buffer_push_cc_data (self->cc_buffer, cc_data, cc_data_len))
self->current_scheduled++;
}
static void
schedule_cea608_s334_1a (GstCCCombiner * self, guint8 * data, guint len,
GstClockTime pts, GstClockTime duration)
{
guint8 field0_data[3], field1_data[3];
guint field0_len = 0, field1_len = 0;
guint i;
if (len % 3 != 0) {
GST_WARNING ("Invalid cc_data buffer size %u. Truncating to a multiple "
"of 3", len);
len = len - (len % 3);
}
for (i = 0; i < len / 3; i++) {
if (data[i * 3] & 0x80) {
if (data[i * 3 + 1] == 0x80 && data[i * 3 + 2] == 0x80)
continue;
field0_data[field0_len++] = data[i * 3 + 1];
field0_data[field0_len++] = data[i * 3 + 2];
} else {
if (data[i * 3 + 1] == 0x80 && data[i * 3 + 2] == 0x80)
continue;
field1_data[field1_len++] = data[i * 3 + 1];
field1_data[field1_len++] = data[i * 3 + 2];
}
}
if (cc_buffer_push_separated (self->cc_buffer, field0_data, field0_len,
field1_data, field1_len, NULL, 0))
self->current_scheduled++;
}
static void
schedule_cea708_raw (GstCCCombiner * self, guint8 * data, guint len,
GstClockTime pts, GstClockTime duration)
{
if (cc_buffer_push_cc_data (self->cc_buffer, data, len))
self->current_scheduled++;
}
static void
schedule_cea608_raw (GstCCCombiner * self, guint8 * data, guint len)
{
if (cc_buffer_push_separated (self->cc_buffer, data, len, NULL, 0, NULL, 0))
self->current_scheduled++;
}
static void
schedule_caption (GstCCCombiner * self, GstAggregatorPad * caption_pad,
GstBuffer * caption_buf, const GstVideoTimeCode * tc)
{
GstMapInfo map;
GstClockTime pts, duration, running_time;
pts = GST_BUFFER_PTS (caption_buf);
duration = GST_BUFFER_DURATION (caption_buf);
running_time =
gst_segment_to_running_time (&caption_pad->segment, GST_FORMAT_TIME, pts);
if (self->current_scheduled + 1 >= self->max_scheduled) {
GstClockTime stream_time;
GST_WARNING_OBJECT (self,
"scheduled queue runs too long, discarding stored");
stream_time =
gst_segment_to_stream_time (&caption_pad->segment, GST_FORMAT_TIME,
pts);
gst_element_post_message (GST_ELEMENT_CAST (self),
gst_message_new_qos (GST_OBJECT_CAST (self), FALSE,
running_time, stream_time, pts, duration));
cc_buffer_discard (self->cc_buffer);
self->current_scheduled = 0;
}
self->last_caption_ts = running_time;
gst_buffer_map (caption_buf, &map, GST_MAP_READ);
switch (self->caption_type) {
case GST_VIDEO_CAPTION_TYPE_CEA708_CDP:
schedule_cdp (self, tc, map.data, map.size, pts, duration);
break;
case GST_VIDEO_CAPTION_TYPE_CEA708_RAW:
schedule_cea708_raw (self, map.data, map.size, pts, duration);
break;
case GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A:
schedule_cea608_s334_1a (self, map.data, map.size, pts, duration);
break;
case GST_VIDEO_CAPTION_TYPE_CEA608_RAW:
schedule_cea608_raw (self, map.data, map.size);
break;
default:
break;
}
gst_buffer_unmap (caption_buf, &map);
}
static void
dequeue_caption (GstCCCombiner * self, GstVideoTimeCode * tc, gboolean drain)
{
guint8 cea608_1[MAX_CEA608_LEN], cea608_2[MAX_CEA608_LEN];
guint8 cc_data[MAX_CDP_PACKET_LEN];
guint cea608_1_len = MAX_CEA608_LEN, cea608_2_len = MAX_CEA608_LEN;
guint cc_data_len = MAX_CDP_PACKET_LEN;
CaptionData caption_data;
g_assert (self->current_frame_captions->len == 0);
if (drain && cc_buffer_is_empty (self->cc_buffer))
return;
if (self->prop_schedule_timeout != GST_CLOCK_TIME_NONE) {
if (self->last_caption_ts == GST_CLOCK_TIME_NONE) {
return;
}
if (self->current_video_running_time > self->last_caption_ts
&& self->current_video_running_time - self->last_caption_ts
> self->prop_schedule_timeout) {
GST_LOG_OBJECT (self, "Not outputting caption as last caption buffer ts %"
GST_TIME_FORMAT " is more than the schedule timeout %" GST_TIME_FORMAT
" from the current output time %" GST_TIME_FORMAT,
GST_TIME_ARGS (self->last_caption_ts),
GST_TIME_ARGS (self->prop_schedule_timeout),
GST_TIME_ARGS (self->current_video_running_time));
return;
}
}
caption_data.caption_type = self->caption_type;
switch (self->caption_type) {
case GST_VIDEO_CAPTION_TYPE_CEA708_CDP:
{
/* Only relevant in alternate and mixed mode, no need to look at the caps */
if (GST_BUFFER_FLAG_IS_SET (self->current_video_buffer,
GST_VIDEO_BUFFER_FLAG_INTERLACED)) {
if (!GST_VIDEO_BUFFER_IS_BOTTOM_FIELD (self->current_video_buffer)) {
cc_buffer_take_cc_data (self->cc_buffer, self->cdp_fps_entry,
cc_data, &cc_data_len);
caption_data.buffer =
make_cdp_buffer (self, cc_data, cc_data_len, self->cdp_fps_entry,
tc);
g_array_append_val (self->current_frame_captions, caption_data);
}
} else {
cc_buffer_take_cc_data (self->cc_buffer, self->cdp_fps_entry,
cc_data, &cc_data_len);
caption_data.buffer =
make_cdp_buffer (self, cc_data, cc_data_len, self->cdp_fps_entry,
tc);
g_array_append_val (self->current_frame_captions, caption_data);
}
break;
}
case GST_VIDEO_CAPTION_TYPE_CEA708_RAW:
{
/* Only relevant in alternate and mixed mode, no need to look at the caps */
if (GST_BUFFER_FLAG_IS_SET (self->current_video_buffer,
GST_VIDEO_BUFFER_FLAG_INTERLACED)) {
if (!GST_VIDEO_BUFFER_IS_BOTTOM_FIELD (self->current_video_buffer)) {
caption_data.buffer =
gst_buffer_new_allocate (NULL, MAX_CDP_PACKET_LEN, NULL);
write_cc_data_to (self, caption_data.buffer);
g_array_append_val (self->current_frame_captions, caption_data);
}
} else {
caption_data.buffer =
gst_buffer_new_allocate (NULL, MAX_CDP_PACKET_LEN, NULL);
write_cc_data_to (self, caption_data.buffer);
g_array_append_val (self->current_frame_captions, caption_data);
}
break;
}
case GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A:
{
if (self->progressive) {
cc_buffer_take_separated (self->cc_buffer, self->cdp_fps_entry,
cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, cc_data,
&cc_data_len);
prepend_s334_to_cea608 (0, cea608_1, &cea608_1_len, sizeof (cea608_1));
caption_data.buffer = make_buffer (self, cea608_1, cea608_1_len);
g_array_append_val (self->current_frame_captions, caption_data);
} else if (GST_BUFFER_FLAG_IS_SET (self->current_video_buffer,
GST_VIDEO_BUFFER_FLAG_INTERLACED) &&
GST_BUFFER_FLAG_IS_SET (self->current_video_buffer,
GST_VIDEO_BUFFER_FLAG_ONEFIELD)) {
cc_buffer_take_separated (self->cc_buffer, self->cdp_fps_entry,
cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, cc_data,
&cc_data_len);
if (GST_VIDEO_BUFFER_IS_TOP_FIELD (self->current_video_buffer)) {
prepend_s334_to_cea608 (0, cea608_1, &cea608_1_len,
sizeof (cea608_1));
caption_data.buffer = make_buffer (self, cea608_1, cea608_1_len);
} else {
prepend_s334_to_cea608 (1, cea608_2, &cea608_2_len,
sizeof (cea608_2));
caption_data.buffer = make_buffer (self, cea608_2, cea608_2_len);
}
g_array_append_val (self->current_frame_captions, caption_data);
} else {
caption_data.buffer =
gst_buffer_new_allocate (NULL, MAX_CDP_PACKET_LEN, NULL);
take_s334_both_fields (self, caption_data.buffer);
g_array_append_val (self->current_frame_captions, caption_data);
}
break;
}
case GST_VIDEO_CAPTION_TYPE_CEA608_RAW:
{
cc_buffer_take_separated (self->cc_buffer, self->cdp_fps_entry,
cea608_1, &cea608_1_len, cea608_2, &cea608_2_len, cc_data,
&cc_data_len);
if (self->progressive) {
caption_data.buffer = make_buffer (self, cea608_1, cea608_1_len);
g_array_append_val (self->current_frame_captions, caption_data);
} else if (GST_BUFFER_FLAG_IS_SET (self->current_video_buffer,
GST_VIDEO_BUFFER_FLAG_INTERLACED)) {
if (!GST_VIDEO_BUFFER_IS_BOTTOM_FIELD (self->current_video_buffer)) {
caption_data.buffer = make_buffer (self, cea608_1, cea608_1_len);
g_array_append_val (self->current_frame_captions, caption_data);
}
} else {
caption_data.buffer = make_buffer (self, cea608_1, cea608_1_len);
g_array_append_val (self->current_frame_captions, caption_data);
}
break;
}
default:
break;
}
}
static gboolean
remove_caption_meta (GstBuffer * buffer, GstMeta ** meta, gpointer user_data)
{
if ((*meta)->info->api == GST_VIDEO_CAPTION_META_API_TYPE)
*meta = NULL;
return TRUE;
}
static GstFlowReturn
gst_cc_combiner_collect_captions (GstCCCombiner * self, gboolean timeout)
{
GstAggregatorPad *src_pad =
GST_AGGREGATOR_PAD (GST_AGGREGATOR_SRC_PAD (self));
GstAggregatorPad *caption_pad;
GstBuffer *video_buf;
GstVideoTimeCodeMeta *tc_meta;
GstVideoTimeCode *tc = NULL;
gboolean caption_pad_is_eos = FALSE;
g_assert (self->current_video_buffer != NULL);
GST_OBJECT_LOCK (self);
caption_pad = self->caption_pad ? gst_object_ref (self->caption_pad) : NULL;
GST_OBJECT_UNLOCK (self);
/* No caption pad, forward buffer directly */
if (!caption_pad) {
GST_LOG_OBJECT (self, "No caption pad, passing through video");
video_buf = self->current_video_buffer;
gst_aggregator_selected_samples (GST_AGGREGATOR_CAST (self),
GST_BUFFER_PTS (video_buf), GST_BUFFER_DTS (video_buf),
GST_BUFFER_DURATION (video_buf), NULL);
self->current_video_buffer = NULL;
goto done;
}
tc_meta = gst_buffer_get_video_time_code_meta (self->current_video_buffer);
if (tc_meta) {
tc = &tc_meta->tc;
}
GST_LOG_OBJECT (self, "Trying to collect captions for queued video buffer");
do {
GstBuffer *caption_buf;
GstClockTime caption_time;
CaptionData caption_data;
caption_buf = gst_aggregator_pad_peek_buffer (caption_pad);
if (!caption_buf) {
if (gst_aggregator_pad_is_eos (caption_pad)) {
GST_DEBUG_OBJECT (self, "Caption pad is EOS, we're done");
caption_pad_is_eos = TRUE;
break;
} else if (!timeout) {
GST_DEBUG_OBJECT (self, "Need more caption data");
gst_object_unref (caption_pad);
return GST_FLOW_NEED_DATA;
} else {
GST_DEBUG_OBJECT (self, "No caption data on timeout");
break;
}
}
caption_time = GST_BUFFER_PTS (caption_buf);
if (!GST_CLOCK_TIME_IS_VALID (caption_time)) {
GST_ERROR_OBJECT (self, "Caption buffer without PTS");
gst_buffer_unref (caption_buf);
gst_object_unref (caption_pad);
return GST_FLOW_ERROR;
}
caption_time =
gst_segment_to_running_time (&caption_pad->segment, GST_FORMAT_TIME,
caption_time);
if (!GST_CLOCK_TIME_IS_VALID (caption_time)) {
GST_DEBUG_OBJECT (self, "Caption buffer outside segment, dropping");
gst_aggregator_pad_drop_buffer (caption_pad);
gst_buffer_unref (caption_buf);
continue;
}
if (gst_buffer_get_size (caption_buf) == 0 &&
GST_BUFFER_FLAG_IS_SET (caption_buf, GST_BUFFER_FLAG_GAP)) {
/* This is a gap, we can go ahead. We only consume it once its end point
* is behind the current video running time. Important to note that
* we can't deal with gaps with no duration (-1)
*/
if (!GST_CLOCK_TIME_IS_VALID (GST_BUFFER_DURATION (caption_buf))) {
GST_ERROR_OBJECT (self, "GAP buffer without a duration");
gst_buffer_unref (caption_buf);
gst_object_unref (caption_pad);
return GST_FLOW_ERROR;
}
gst_buffer_unref (caption_buf);
if (caption_time + GST_BUFFER_DURATION (caption_buf) <
self->current_video_running_time_end) {
gst_aggregator_pad_drop_buffer (caption_pad);
continue;
} else {
break;
}
}
/* Collected all caption buffers for this video buffer */
if (caption_time >= self->current_video_running_time_end) {
gst_buffer_unref (caption_buf);
break;
} else if (!self->schedule) {
if (GST_CLOCK_TIME_IS_VALID (self->previous_video_running_time_end)) {
if (caption_time < self->previous_video_running_time_end) {
GST_WARNING_OBJECT (self,
"Caption buffer before end of last video frame, dropping");
gst_aggregator_pad_drop_buffer (caption_pad);
gst_buffer_unref (caption_buf);
continue;
}
} else if (caption_time < self->current_video_running_time) {
GST_WARNING_OBJECT (self,
"Caption buffer before current video frame, dropping");
gst_aggregator_pad_drop_buffer (caption_pad);
gst_buffer_unref (caption_buf);
continue;
}
}
/* This caption buffer has to be collected */
GST_LOG_OBJECT (self,
"Collecting caption buffer %p %" GST_TIME_FORMAT " for video buffer %p",
caption_buf, GST_TIME_ARGS (caption_time), self->current_video_buffer);
caption_data.caption_type = self->caption_type;
gst_aggregator_pad_drop_buffer (caption_pad);
if (!self->schedule) {
caption_data.buffer = caption_buf;
g_array_append_val (self->current_frame_captions, caption_data);
} else {
schedule_caption (self, caption_pad, caption_buf, tc);
gst_buffer_unref (caption_buf);
}
} while (TRUE);
if (self->schedule) {
dequeue_caption (self, tc, caption_pad_is_eos);
}
gst_aggregator_selected_samples (GST_AGGREGATOR_CAST (self),
GST_BUFFER_PTS (self->current_video_buffer),
GST_BUFFER_DTS (self->current_video_buffer),
GST_BUFFER_DURATION (self->current_video_buffer), NULL);
GST_LOG_OBJECT (self, "Collected %u captions for buffer %p",
self->current_frame_captions->len, self->current_video_buffer);
switch (self->prop_input_meta_processing) {
case CCCOMBINER_INPUT_PROCESSING_APPEND:
break;
case CCCOMBINER_INPUT_PROCESSING_DROP:
self->current_video_buffer =
gst_buffer_make_writable (self->current_video_buffer);
gst_buffer_foreach_meta (self->current_video_buffer, remove_caption_meta,
NULL);
break;
case CCCOMBINER_INPUT_PROCESSING_FAVOR:
if (gst_buffer_get_meta (self->current_video_buffer,
GST_VIDEO_CAPTION_META_API_TYPE)) {
GST_LOG_OBJECT (self,
"Video buffer already has captions, dropping %d dequeued captions",
self->current_frame_captions->len);
g_array_set_size (self->current_frame_captions, 0);
}
break;
case CCCOMBINER_INPUT_PROCESSING_FORCE:
GST_LOG_OBJECT (self,
"Forced input captions, dropping %d dequeued captions",
self->current_frame_captions->len);
g_array_set_size (self->current_frame_captions, 0);
break;
}
if (self->current_frame_captions->len > 0) {
guint i;
if (self->schedule)
self->current_scheduled = MAX (1, self->current_scheduled) - 1;
video_buf = gst_buffer_make_writable (self->current_video_buffer);
self->current_video_buffer = NULL;
for (i = 0; i < self->current_frame_captions->len; i++) {
CaptionData *caption_data =
&g_array_index (self->current_frame_captions, CaptionData, i);
GstMapInfo map;
gst_buffer_map (caption_data->buffer, &map, GST_MAP_READ);
gst_buffer_add_video_caption_meta (video_buf, caption_data->caption_type,
map.data, map.size);
gst_buffer_unmap (caption_data->buffer, &map);
}
g_array_set_size (self->current_frame_captions, 0);
} else {
GST_LOG_OBJECT (self, "No captions for buffer %p",
self->current_video_buffer);
video_buf = self->current_video_buffer;
self->current_video_buffer = NULL;
}
gst_object_unref (caption_pad);
done:
src_pad->segment.position =
GST_BUFFER_PTS (video_buf) + GST_BUFFER_DURATION (video_buf);
return gst_aggregator_finish_buffer (GST_AGGREGATOR_CAST (self), video_buf);
}
static GstClockTime
gst_cc_combiner_get_next_time (GstAggregator * aggregator)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
GST_OBJECT_LOCK (self);
/* No point timing out if we can't combine captions */
if (!self->caption_pad)
goto wait_for_data;
/* We need a video buffer */
if (!self->current_video_buffer &&
!gst_aggregator_pad_has_buffer (self->video_pad))
goto wait_for_data;
GST_OBJECT_UNLOCK (self);
return gst_aggregator_simple_get_next_time (aggregator);
wait_for_data:
GST_OBJECT_UNLOCK (self);
return GST_CLOCK_TIME_NONE;
}
static GstFlowReturn
gst_cc_combiner_aggregate (GstAggregator * aggregator, gboolean timeout)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
GstFlowReturn flow_ret = GST_FLOW_OK;
/* If we have no current video buffer, queue one. If we have one but
* its end running time is not known yet, try to determine it from the
* next video buffer */
if (!self->current_video_buffer
|| !GST_CLOCK_TIME_IS_VALID (self->current_video_running_time_end)) {
GstAggregatorPad *video_pad;
GstClockTime video_start;
GstBuffer *video_buf;
video_pad = self->video_pad;
video_buf = gst_aggregator_pad_peek_buffer (video_pad);
if (!video_buf) {
if (gst_aggregator_pad_is_eos (video_pad)) {
GST_DEBUG_OBJECT (aggregator, "Video pad is EOS, we're done");
/* Assume that this buffer ends where it started +50ms (25fps) and handle it */
if (self->current_video_buffer) {
self->current_video_running_time_end =
self->current_video_running_time + 50 * GST_MSECOND;
flow_ret = gst_cc_combiner_collect_captions (self, timeout);
}
/* If we collected all captions for the remaining video frame we're
* done, otherwise get called another time and go directly into the
* outer branch for finishing the current video frame */
if (flow_ret == GST_FLOW_NEED_DATA)
flow_ret = GST_FLOW_OK;
else
flow_ret = GST_FLOW_EOS;
} else {
flow_ret = GST_FLOW_OK;
}
return flow_ret;
}
video_start = GST_BUFFER_PTS (video_buf);
if (!GST_CLOCK_TIME_IS_VALID (video_start)) {
gst_buffer_unref (video_buf);
GST_ERROR_OBJECT (aggregator, "Video buffer without PTS");
return GST_FLOW_ERROR;
}
video_start =
gst_segment_to_running_time (&video_pad->segment, GST_FORMAT_TIME,
video_start);
if (!GST_CLOCK_TIME_IS_VALID (video_start)) {
GST_DEBUG_OBJECT (aggregator, "Buffer outside segment, dropping");
gst_aggregator_pad_drop_buffer (video_pad);
gst_buffer_unref (video_buf);
return GST_FLOW_OK;
}
if (self->current_video_buffer) {
/* If we already have a video buffer just update the current end running
* time accordingly. That's what was missing and why we got here */
self->current_video_running_time_end = video_start;
gst_buffer_unref (video_buf);
GST_LOG_OBJECT (self,
"Determined end timestamp for video buffer: %p %" GST_TIME_FORMAT
" - %" GST_TIME_FORMAT, self->current_video_buffer,
GST_TIME_ARGS (self->current_video_running_time),
GST_TIME_ARGS (self->current_video_running_time_end));
} else {
/* Otherwise we had no buffer queued currently. Let's do that now
* so that we can collect captions for it */
gst_buffer_replace (&self->current_video_buffer, video_buf);
self->current_video_running_time = video_start;
gst_aggregator_pad_drop_buffer (video_pad);
gst_buffer_unref (video_buf);
if (GST_BUFFER_DURATION_IS_VALID (video_buf)) {
GstClockTime end_time =
GST_BUFFER_PTS (video_buf) + GST_BUFFER_DURATION (video_buf);
if (video_pad->segment.stop != -1 && end_time > video_pad->segment.stop)
end_time = video_pad->segment.stop;
self->current_video_running_time_end =
gst_segment_to_running_time (&video_pad->segment, GST_FORMAT_TIME,
end_time);
} else if (self->video_fps_n != 0 && self->video_fps_d != 0) {
GstClockTime end_time =
GST_BUFFER_PTS (video_buf) + gst_util_uint64_scale_int (GST_SECOND,
self->video_fps_d, self->video_fps_n);
if (video_pad->segment.stop != -1 && end_time > video_pad->segment.stop)
end_time = video_pad->segment.stop;
self->current_video_running_time_end =
gst_segment_to_running_time (&video_pad->segment, GST_FORMAT_TIME,
end_time);
} else {
self->current_video_running_time_end = GST_CLOCK_TIME_NONE;
}
GST_LOG_OBJECT (self,
"Queued new video buffer: %p %" GST_TIME_FORMAT " - %"
GST_TIME_FORMAT, self->current_video_buffer,
GST_TIME_ARGS (self->current_video_running_time),
GST_TIME_ARGS (self->current_video_running_time_end));
}
}
/* At this point we have a video buffer queued and can start collecting
* caption buffers for it */
g_assert (self->current_video_buffer != NULL);
g_assert (GST_CLOCK_TIME_IS_VALID (self->current_video_running_time));
g_assert (GST_CLOCK_TIME_IS_VALID (self->current_video_running_time_end));
flow_ret = gst_cc_combiner_collect_captions (self, timeout);
/* Only if we collected all captions we replace the current video buffer
* with NULL and continue with the next one on the next call */
if (flow_ret == GST_FLOW_NEED_DATA) {
flow_ret = GST_FLOW_OK;
} else {
gst_buffer_replace (&self->current_video_buffer, NULL);
self->previous_video_running_time_end =
self->current_video_running_time_end;
self->current_video_running_time = self->current_video_running_time_end =
GST_CLOCK_TIME_NONE;
}
return flow_ret;
}
static gboolean
gst_cc_combiner_sink_event (GstAggregator * aggregator,
GstAggregatorPad * agg_pad, GstEvent * event)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
switch (GST_EVENT_TYPE (event)) {
case GST_EVENT_CAPS:{
GstCaps *caps;
GstStructure *s;
gst_event_parse_caps (event, &caps);
s = gst_caps_get_structure (caps, 0);
if (strcmp (GST_OBJECT_NAME (agg_pad), "caption") == 0) {
GstVideoCaptionType caption_type =
gst_video_caption_type_from_caps (caps);
if (self->caption_type != GST_VIDEO_CAPTION_TYPE_UNKNOWN &&
caption_type != self->caption_type) {
GST_ERROR_OBJECT (self, "Changing caption type is not allowed");
GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, (NULL),
("Changing caption type is not allowed"));
return FALSE;
}
self->caption_type = caption_type;
} else {
gint fps_n, fps_d;
const gchar *interlace_mode;
fps_n = fps_d = 0;
gst_structure_get_fraction (s, "framerate", &fps_n, &fps_d);
interlace_mode = gst_structure_get_string (s, "interlace-mode");
self->progressive = !interlace_mode
|| !g_strcmp0 (interlace_mode, "progressive");
if (fps_n != self->video_fps_n || fps_d != self->video_fps_d) {
GstClockTime latency;
latency = gst_util_uint64_scale (GST_SECOND, fps_d, fps_n);
gst_aggregator_set_latency (aggregator, latency, latency);
}
self->video_fps_n = fps_n;
self->video_fps_d = fps_d;
self->cdp_fps_entry = cdp_fps_entry_from_fps (fps_n, fps_d);
if (!self->cdp_fps_entry || self->cdp_fps_entry->fps_n == 0) {
GST_WARNING_OBJECT (self, "Missing valid caption framerate in "
"video caps");
GST_ELEMENT_WARNING (self, CORE, NEGOTIATION, (NULL),
("Missing valid caption framerate in video caps"));
self->cdp_fps_entry = cdp_fps_entry_from_fps (60, 1);
}
gst_aggregator_set_src_caps (aggregator, caps);
}
break;
}
case GST_EVENT_SEGMENT:{
if (strcmp (GST_OBJECT_NAME (agg_pad), "sink") == 0) {
const GstSegment *segment;
gst_event_parse_segment (event, &segment);
gst_aggregator_update_segment (aggregator, segment);
}
break;
}
case GST_EVENT_STREAM_START:{
if (strcmp (GST_OBJECT_NAME (agg_pad), "caption") == 0) {
self->last_caption_ts = GST_CLOCK_TIME_NONE;
}
}
default:
break;
}
return GST_AGGREGATOR_CLASS (parent_class)->sink_event (aggregator, agg_pad,
event);
}
static gboolean
gst_cc_combiner_stop (GstAggregator * aggregator)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
self->video_fps_n = self->video_fps_d = 0;
self->current_video_running_time = self->current_video_running_time_end =
self->previous_video_running_time_end = GST_CLOCK_TIME_NONE;
gst_buffer_replace (&self->current_video_buffer, NULL);
g_array_set_size (self->current_frame_captions, 0);
self->caption_type = GST_VIDEO_CAPTION_TYPE_UNKNOWN;
cc_buffer_discard (self->cc_buffer);
self->current_scheduled = 0;
self->cdp_fps_entry = &null_fps_entry;
return TRUE;
}
static GstFlowReturn
gst_cc_combiner_flush (GstAggregator * aggregator)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
GstAggregatorPad *src_pad =
GST_AGGREGATOR_PAD (GST_AGGREGATOR_SRC_PAD (aggregator));
self->current_video_running_time = self->current_video_running_time_end =
self->previous_video_running_time_end = GST_CLOCK_TIME_NONE;
gst_buffer_replace (&self->current_video_buffer, NULL);
g_array_set_size (self->current_frame_captions, 0);
src_pad->segment.position = GST_CLOCK_TIME_NONE;
self->cdp_hdr_sequence_cntr = 0;
cc_buffer_discard (self->cc_buffer);
self->current_scheduled = 0;
return GST_FLOW_OK;
}
static GstAggregatorPad *
gst_cc_combiner_create_new_pad (GstAggregator * aggregator,
GstPadTemplate * templ, const gchar * req_name, const GstCaps * caps)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
GstAggregatorPad *agg_pad;
if (templ->direction != GST_PAD_SINK)
return NULL;
if (templ->presence != GST_PAD_REQUEST)
return NULL;
if (strcmp (templ->name_template, "caption") != 0)
return NULL;
GST_OBJECT_LOCK (self);
agg_pad = g_object_new (GST_TYPE_AGGREGATOR_PAD,
"name", "caption", "direction", GST_PAD_SINK, "template", templ, NULL);
self->caption_type = GST_VIDEO_CAPTION_TYPE_UNKNOWN;
self->caption_pad = gst_object_ref (agg_pad);
GST_OBJECT_UNLOCK (self);
return agg_pad;
}
static void
gst_cc_combiner_release_pad (GstElement * element, GstPad * pad)
{
GstCCCombiner *self = GST_CCCOMBINER (element);
GST_OBJECT_LOCK (self);
if (pad == GST_PAD_CAST (self->caption_pad)) {
gst_clear_object (&self->caption_pad);
}
GST_OBJECT_UNLOCK (self);
GST_ELEMENT_CLASS (parent_class)->release_pad (element, pad);
}
static gboolean
gst_cc_combiner_src_query (GstAggregator * aggregator, GstQuery * query)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
gboolean ret;
switch (GST_QUERY_TYPE (query)) {
case GST_QUERY_POSITION:
case GST_QUERY_DURATION:
case GST_QUERY_URI:
case GST_QUERY_CAPS:
case GST_QUERY_ALLOCATION:
ret = gst_pad_peer_query (GST_PAD_CAST (self->video_pad), query);
break;
case GST_QUERY_ACCEPT_CAPS:{
GstCaps *caps;
GstCaps *templ = gst_static_pad_template_get_caps (&srctemplate);
gst_query_parse_accept_caps (query, &caps);
gst_query_set_accept_caps_result (query, gst_caps_is_subset (caps,
templ));
gst_caps_unref (templ);
ret = TRUE;
break;
}
default:
ret = GST_AGGREGATOR_CLASS (parent_class)->src_query (aggregator, query);
break;
}
return ret;
}
static gboolean
gst_cc_combiner_sink_query (GstAggregator * aggregator,
GstAggregatorPad * aggpad, GstQuery * query)
{
GstCCCombiner *self = GST_CCCOMBINER (aggregator);
GstPad *srcpad = GST_AGGREGATOR_SRC_PAD (aggregator);
gboolean ret;
switch (GST_QUERY_TYPE (query)) {
case GST_QUERY_POSITION:
case GST_QUERY_DURATION:
case GST_QUERY_URI:
case GST_QUERY_ALLOCATION:
if (aggpad == self->video_pad) {
ret = gst_pad_peer_query (srcpad, query);
} else {
ret =
GST_AGGREGATOR_CLASS (parent_class)->sink_query (aggregator,
aggpad, query);
}
break;
case GST_QUERY_CAPS:
if (aggpad == self->video_pad) {
ret = gst_pad_peer_query (srcpad, query);
} else {
GstCaps *filter;
GstCaps *templ = gst_static_pad_template_get_caps (&captiontemplate);
gst_query_parse_caps (query, &filter);
if (filter) {
GstCaps *caps =
gst_caps_intersect_full (filter, templ, GST_CAPS_INTERSECT_FIRST);
gst_query_set_caps_result (query, caps);
gst_caps_unref (caps);
} else {
gst_query_set_caps_result (query, templ);
}
gst_caps_unref (templ);
ret = TRUE;
}
break;
case GST_QUERY_ACCEPT_CAPS:
if (aggpad == self->video_pad) {
ret = gst_pad_peer_query (srcpad, query);
} else {
GstCaps *caps;
GstCaps *templ = gst_static_pad_template_get_caps (&captiontemplate);
gst_query_parse_accept_caps (query, &caps);
gst_query_set_accept_caps_result (query, gst_caps_is_subset (caps,
templ));
gst_caps_unref (templ);
ret = TRUE;
}
break;
default:
ret = GST_AGGREGATOR_CLASS (parent_class)->sink_query (aggregator,
aggpad, query);
break;
}
return ret;
}
static GstSample *
gst_cc_combiner_peek_next_sample (GstAggregator * agg,
GstAggregatorPad * aggpad)
{
GstCCCombiner *self = GST_CCCOMBINER (agg);
GstSample *res = NULL;
if (aggpad != self->video_pad) {
/* Must be the caption pad */
if (self->current_frame_captions->len > 0) {
GstCaps *caps = gst_pad_get_current_caps (GST_PAD (aggpad));
GstBufferList *buflist = gst_buffer_list_new ();
guint i;
for (i = 0; i < self->current_frame_captions->len; i++) {
CaptionData *caption_data =
&g_array_index (self->current_frame_captions, CaptionData, i);
gst_buffer_list_add (buflist, gst_buffer_ref (caption_data->buffer));
}
res = gst_sample_new (NULL, caps, &aggpad->segment, NULL);
gst_caps_unref (caps);
gst_sample_set_buffer_list (res, buflist);
gst_buffer_list_unref (buflist);
}
} else {
if (self->current_video_buffer) {
GstCaps *caps = gst_pad_get_current_caps (GST_PAD (aggpad));
res = gst_sample_new (self->current_video_buffer,
caps, &aggpad->segment, NULL);
gst_caps_unref (caps);
}
}
return res;
}
static GstStateChangeReturn
gst_cc_combiner_change_state (GstElement * element, GstStateChange transition)
{
GstCCCombiner *self = GST_CCCOMBINER (element);
switch (transition) {
case GST_STATE_CHANGE_READY_TO_PAUSED:
self->schedule = self->prop_schedule;
self->max_scheduled = self->prop_max_scheduled;
self->output_padding = self->prop_output_padding;
cc_buffer_set_max_buffer_time (self->cc_buffer, GST_CLOCK_TIME_NONE);
cc_buffer_set_output_padding (self->cc_buffer, self->prop_output_padding,
self->prop_output_padding);
break;
default:
break;
}
return GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
}
static void
gst_cc_combiner_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec)
{
GstCCCombiner *self = GST_CCCOMBINER (object);
switch (prop_id) {
case PROP_SCHEDULE:
self->prop_schedule = g_value_get_boolean (value);
break;
case PROP_MAX_SCHEDULED:
self->prop_max_scheduled = g_value_get_uint (value);
break;
case PROP_OUTPUT_PADDING:
self->prop_output_padding = g_value_get_boolean (value);
break;
case PROP_CEA608_PADDING_STRATEGY:
self->prop_cea608_padding_strategy = g_value_get_flags (value);
cc_buffer_set_cea608_padding_strategy (self->cc_buffer,
self->prop_cea608_padding_strategy);
break;
case PROP_CEA608_VALID_PADDING_TIMEOUT:
self->prop_cea608_valid_padding_timeout = g_value_get_uint64 (value);
cc_buffer_set_cea608_valid_timeout (self->cc_buffer,
self->prop_cea608_valid_padding_timeout);
break;
case PROP_SCHEDULE_TIMEOUT:
self->prop_schedule_timeout = g_value_get_uint64 (value);
break;
case PROP_INPUT_META_PROCESSING:
self->prop_input_meta_processing = g_value_get_enum (value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static void
gst_cc_combiner_get_property (GObject * object, guint prop_id, GValue * value,
GParamSpec * pspec)
{
GstCCCombiner *self = GST_CCCOMBINER (object);
switch (prop_id) {
case PROP_SCHEDULE:
g_value_set_boolean (value, self->prop_schedule);
break;
case PROP_MAX_SCHEDULED:
g_value_set_uint (value, self->prop_max_scheduled);
break;
case PROP_OUTPUT_PADDING:
g_value_set_boolean (value, self->prop_output_padding);
break;
case PROP_CEA608_PADDING_STRATEGY:
g_value_set_flags (value, self->prop_cea608_padding_strategy);
break;
case PROP_CEA608_VALID_PADDING_TIMEOUT:
g_value_set_uint64 (value, self->prop_cea608_valid_padding_timeout);
break;
case PROP_SCHEDULE_TIMEOUT:
g_value_set_uint64 (value, self->prop_schedule_timeout);
break;
case PROP_INPUT_META_PROCESSING:
g_value_set_enum (value, self->prop_input_meta_processing);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
break;
}
}
static void
gst_cc_combiner_class_init (GstCCCombinerClass * klass)
{
GObjectClass *gobject_class;
GstElementClass *gstelement_class;
GstAggregatorClass *aggregator_class;
gobject_class = (GObjectClass *) klass;
gstelement_class = (GstElementClass *) klass;
aggregator_class = (GstAggregatorClass *) klass;
gobject_class->finalize = gst_cc_combiner_finalize;
gobject_class->set_property = gst_cc_combiner_set_property;
gobject_class->get_property = gst_cc_combiner_get_property;
gst_element_class_set_static_metadata (gstelement_class,
"Closed Caption Combiner",
"Filter",
"Combines GstVideoCaptionMeta with video input stream",
"Sebastian Dröge <sebastian@centricular.com>");
/**
* GstCCCombiner:schedule:
*
* Controls whether caption buffers should be smoothly scheduled
* in order to have exactly one per output video buffer.
*
* This can involve rewriting input captions, for example when the
* input is CDP sequence counters are rewritten, time codes are dropped
* and potentially re-injected if the input video frame had a time code
* meta.
*
* Caption buffers may also get split up in order to assign captions to
* the correct field when the input is interlaced.
*
* This can also imply that the input will drift from synchronization,
* when there isn't enough padding in the input stream to catch up. In
* that case the element will start dropping old caption buffers once
* the number of buffers in its internal queue reaches
* #GstCCCombiner:max-scheduled.
*
* When this is set to %FALSE, the behaviour of this element is essentially
* that of a funnel.
*
* Since: 1.20
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_SCHEDULE, g_param_spec_boolean ("schedule",
"Schedule",
"Schedule caption buffers so that exactly one is output per video frame",
DEFAULT_SCHEDULE,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
GST_PARAM_MUTABLE_READY));
/**
* GstCCCombiner:max-scheduled:
*
* Controls the number of scheduled buffers after which the element
* will start dropping old buffers from its internal queues. See
* #GstCCCombiner:schedule.
*
* Since: 1.20
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_MAX_SCHEDULED, g_param_spec_uint ("max-scheduled",
"Max Scheduled",
"Maximum number of buffers to queue for scheduling", 0, G_MAXUINT,
DEFAULT_MAX_SCHEDULED,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
GST_PARAM_MUTABLE_READY));
/**
* GstCCCombiner:output-padding:
*
* When #GstCCCombiner:schedule is %TRUE, this property controls
* whether the output closed caption meta stream will be padded.
*
* Since: 1.22
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_OUTPUT_PADDING, g_param_spec_boolean ("output-padding",
"Output padding",
"Whether to output padding packets when schedule=true",
DEFAULT_OUTPUT_PADDING,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
GST_PARAM_MUTABLE_READY));
/**
* GstCCCombiner:cea608-padding-strategy:
*
* Controls the transformations that may be done on padding CEA-608 data.
*
* Since: 1.26
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_CEA608_PADDING_STRATEGY,
g_param_spec_flags ("cea608-padding-strategy",
"CEA-608 Padding Strategy",
"What transformations to perform on CEA-608 padding data",
GST_TYPE_CC_BUFFER_CEA608_PADDING_STRATEGY,
DEFAULT_CEA608_PADDING_STRATEGY,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
GST_PARAM_MUTABLE_PLAYING));
/**
* GstCCCombiner:cea608-padding-valid-timeout:
*
* Timeout to apply when padding strategy contains "valid". After this time
* hase passed, CEA-608 padding will be signalled as invalid until new valid
* CEA-608 non-padding data is received.
*
* Since: 1.26
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_CEA608_VALID_PADDING_TIMEOUT,
g_param_spec_uint64 ("cea608-padding-valid-timeout",
"CEA-608 Padding Valid Timeout",
"How long after receiving valid non-padding CEA-608 data to keep writing valid CEA-608 padding bytes",
0, G_MAXUINT64, DEFAULT_CEA608_VALID_PADDING_TIMEOUT,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
GST_PARAM_MUTABLE_PLAYING));
/**
* GstCCCombiner:schedule-timeout:
*
* Timeout to apply when the caption pad is EOS and schedule=true. On
* reaching the timeout, no caption data will be placed on the outgoing
* buffers until receiving a new stream.
*
* Since: 1.26
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_SCHEDULE_TIMEOUT,
g_param_spec_uint64 ("schedule-timeout",
"Schedule Timeout",
"How long after not receiving caption data on the caption pad to continue adding (padding) caption data on output buffers",
0, G_MAXUINT64, DEFAULT_SCHEDULE_TIMEOUT,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
GST_PARAM_MUTABLE_PLAYING));
/**
* GstCCCombiner:input-meta-processing
*
* Controls how input closed caption meta is processed.
*
* Since: 1.26
*/
g_object_class_install_property (G_OBJECT_CLASS (klass),
PROP_INPUT_META_PROCESSING, g_param_spec_enum ("input-meta-processing",
"Input Meta Processing",
"Controls how input closed caption meta is processed",
GST_TYPE_CCCOMBINER_INPUT_META_PROCESSING,
DEFAULT_INPUT_META_PROCESSING,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
gst_element_class_add_static_pad_template_with_gtype (gstelement_class,
&sinktemplate, GST_TYPE_AGGREGATOR_PAD);
gst_element_class_add_static_pad_template_with_gtype (gstelement_class,
&srctemplate, GST_TYPE_AGGREGATOR_PAD);
gst_element_class_add_static_pad_template_with_gtype (gstelement_class,
&captiontemplate, GST_TYPE_AGGREGATOR_PAD);
gstelement_class->change_state =
GST_DEBUG_FUNCPTR (gst_cc_combiner_change_state);
gstelement_class->release_pad =
GST_DEBUG_FUNCPTR (gst_cc_combiner_release_pad);
aggregator_class->aggregate = gst_cc_combiner_aggregate;
aggregator_class->stop = gst_cc_combiner_stop;
aggregator_class->flush = gst_cc_combiner_flush;
aggregator_class->create_new_pad = gst_cc_combiner_create_new_pad;
aggregator_class->sink_event = gst_cc_combiner_sink_event;
aggregator_class->negotiate = NULL;
aggregator_class->get_next_time = gst_cc_combiner_get_next_time;
aggregator_class->src_query = gst_cc_combiner_src_query;
aggregator_class->sink_query = gst_cc_combiner_sink_query;
aggregator_class->peek_next_sample = gst_cc_combiner_peek_next_sample;
GST_DEBUG_CATEGORY_INIT (gst_cc_combiner_debug, "cccombiner",
0, "Closed Caption combiner");
gst_type_mark_as_plugin_api (GST_TYPE_CCCOMBINER_INPUT_META_PROCESSING, 0);
}
static void
gst_cc_combiner_init (GstCCCombiner * self)
{
GstPadTemplate *templ;
templ = gst_static_pad_template_get (&sinktemplate);
self->video_pad = g_object_new (GST_TYPE_AGGREGATOR_PAD,
"name", "sink", "direction", GST_PAD_SINK, "template", templ, NULL);
gst_object_unref (templ);
gst_element_add_pad (GST_ELEMENT_CAST (self), GST_PAD_CAST (self->video_pad));
self->current_frame_captions =
g_array_new (FALSE, FALSE, sizeof (CaptionData));
g_array_set_clear_func (self->current_frame_captions,
(GDestroyNotify) caption_data_clear);
self->current_video_running_time = self->current_video_running_time_end =
self->previous_video_running_time_end = GST_CLOCK_TIME_NONE;
self->caption_type = GST_VIDEO_CAPTION_TYPE_UNKNOWN;
self->prop_schedule = DEFAULT_SCHEDULE;
self->prop_max_scheduled = DEFAULT_MAX_SCHEDULED;
self->prop_output_padding = DEFAULT_OUTPUT_PADDING;
self->prop_cea608_padding_strategy = DEFAULT_CEA608_PADDING_STRATEGY;
self->prop_cea608_valid_padding_timeout =
DEFAULT_CEA608_VALID_PADDING_TIMEOUT;
self->prop_schedule_timeout = DEFAULT_SCHEDULE_TIMEOUT;
self->prop_input_meta_processing = DEFAULT_INPUT_META_PROCESSING;
self->cdp_hdr_sequence_cntr = 0;
self->cdp_fps_entry = &null_fps_entry;
self->last_caption_ts = GST_CLOCK_TIME_NONE;
self->cc_buffer = cc_buffer_new ();
cc_buffer_set_max_buffer_time (self->cc_buffer, GST_CLOCK_TIME_NONE);
cc_buffer_set_cea608_valid_timeout (self->cc_buffer,
self->prop_cea608_valid_padding_timeout);
cc_buffer_set_cea608_padding_strategy (self->cc_buffer,
self->prop_cea608_padding_strategy);
}