mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-19 00:01:23 +00:00
e3fe89aba8
Add a finalize method and release locks and things in there, instead of in the dispose method. Dispose may be called multiple times, at any time, and should just safely release references to other memory that might reference it back. In this case, timecodestamper would later crash in the element dispose method trying to take the freed mutex from gst_timecodestamper_release_pad(). Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/5935>
2102 lines
73 KiB
C
2102 lines
73 KiB
C
/*
|
|
* GStreamer
|
|
* Copyright (C) 2016 Vivia Nikolaidou <vivia@toolsonair.com>
|
|
* Copyright (C) 2019 Sebastian Dröge <sebastian@centricular.com>
|
|
*
|
|
* gsttimecodestamper.c
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
/**
|
|
* SECTION:element-timecodestamper
|
|
* @title: timecodestamper
|
|
* @short_description: Attach a timecode into incoming video frames
|
|
*
|
|
* This element attaches a timecode into every incoming video frame. It starts
|
|
* counting from the stream time of each segment start, which it converts into
|
|
* a timecode.
|
|
*
|
|
* ## Example launch line
|
|
* |[
|
|
* gst-launch-1.0 videotestsrc ! timecodestamper ! autovideosink
|
|
* ]|
|
|
*
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include "gsttimecodestamper.h"
|
|
|
|
#include <gst/gst.h>
|
|
#include <gst/video/video.h>
|
|
#include <gst/audio/audio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#define ABSDIFF(a,b) (((a) > (b)) ? ((a) - (b)) : ((b) - (a)))
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (timecodestamper_debug);
|
|
#define GST_CAT_DEFAULT timecodestamper_debug
|
|
|
|
/* GstTimeCodeStamper properties */
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_SOURCE,
|
|
PROP_SET,
|
|
PROP_AUTO_RESYNC,
|
|
PROP_TIMEOUT,
|
|
PROP_DROP_FRAME,
|
|
PROP_POST_MESSAGES,
|
|
PROP_SET_INTERNAL_TIMECODE,
|
|
PROP_LTC_DAILY_JAM,
|
|
PROP_LTC_AUTO_RESYNC,
|
|
PROP_LTC_EXTRA_LATENCY,
|
|
PROP_LTC_TIMEOUT,
|
|
PROP_RTC_MAX_DRIFT,
|
|
PROP_RTC_AUTO_RESYNC,
|
|
PROP_TIMECODE_OFFSET
|
|
};
|
|
|
|
#define DEFAULT_SOURCE GST_TIME_CODE_STAMPER_SOURCE_INTERNAL
|
|
#define DEFAULT_SET GST_TIME_CODE_STAMPER_SET_KEEP
|
|
#define DEFAULT_AUTO_RESYNC TRUE
|
|
#define DEFAULT_TIMEOUT GST_CLOCK_TIME_NONE
|
|
#define DEFAULT_DROP_FRAME TRUE
|
|
#define DEFAULT_POST_MESSAGES FALSE
|
|
#define DEFAULT_SET_INTERNAL_TIMECODE NULL
|
|
#define DEFAULT_LTC_DAILY_JAM NULL
|
|
#define DEFAULT_LTC_AUTO_RESYNC TRUE
|
|
#define DEFAULT_LTC_TIMEOUT GST_CLOCK_TIME_NONE
|
|
#define DEFAULT_LTC_EXTRA_LATENCY (150 * GST_MSECOND)
|
|
#define DEFAULT_RTC_MAX_DRIFT 250000000
|
|
#define DEFAULT_RTC_AUTO_RESYNC TRUE
|
|
#define DEFAULT_TIMECODE_OFFSET 0
|
|
|
|
#define DEFAULT_LTC_QUEUE 100
|
|
|
|
static GstStaticPadTemplate gst_timecodestamper_src_template =
|
|
GST_STATIC_PAD_TEMPLATE ("src",
|
|
GST_PAD_SRC,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS ("video/x-raw, framerate=[1/2147483647, 2147483647/1]; "
|
|
"closedcaption/x-cea-608, framerate=[1/2147483647, 2147483647/1]; "
|
|
"closedcaption/x-cea-708, framerate=[1/2147483647, 2147483647/1]; ")
|
|
);
|
|
|
|
static GstStaticPadTemplate gst_timecodestamper_sink_template =
|
|
GST_STATIC_PAD_TEMPLATE ("sink",
|
|
GST_PAD_SINK,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS ("video/x-raw, framerate=[1/2147483647, 2147483647/1]; "
|
|
"closedcaption/x-cea-608, framerate=[1/2147483647, 2147483647/1]; "
|
|
"closedcaption/x-cea-708, framerate=[1/2147483647, 2147483647/1]; ")
|
|
);
|
|
|
|
static GstStaticPadTemplate gst_timecodestamper_ltc_template =
|
|
GST_STATIC_PAD_TEMPLATE ("ltc_sink",
|
|
GST_PAD_SINK,
|
|
GST_PAD_REQUEST,
|
|
GST_STATIC_CAPS ("audio/x-raw,format=U8,rate=[1,max],channels=1")
|
|
);
|
|
|
|
static void gst_timecodestamper_finalize (GObject * object);
|
|
static void gst_timecodestamper_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec);
|
|
static void gst_timecodestamper_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec);
|
|
static void gst_timecodestamper_dispose (GObject * object);
|
|
static gboolean gst_timecodestamper_sink_event (GstBaseTransform * trans,
|
|
GstEvent * event);
|
|
static gboolean gst_timecodestamper_src_event (GstBaseTransform * trans,
|
|
GstEvent * event);
|
|
static GstFlowReturn gst_timecodestamper_transform_ip (GstBaseTransform *
|
|
vfilter, GstBuffer * buffer);
|
|
static gboolean gst_timecodestamper_stop (GstBaseTransform * trans);
|
|
static gboolean gst_timecodestamper_start (GstBaseTransform * trans);
|
|
static GstPad *gst_timecodestamper_request_new_pad (GstElement * element,
|
|
GstPadTemplate * temp, const gchar * unused, const GstCaps * caps);
|
|
static void gst_timecodestamper_release_pad (GstElement * element,
|
|
GstPad * pad);
|
|
|
|
#if HAVE_LTC
|
|
typedef struct
|
|
{
|
|
GstClockTime running_time;
|
|
GstVideoTimeCode timecode;
|
|
} TimestampedTimecode;
|
|
|
|
static gboolean gst_timecodestamper_query (GstBaseTransform * trans,
|
|
GstPadDirection direction, GstQuery * query);
|
|
|
|
static GstFlowReturn gst_timecodestamper_ltcpad_chain (GstPad * pad,
|
|
GstObject * parent, GstBuffer * buffer);
|
|
static gboolean gst_timecodestamper_ltcpad_event (GstPad * pad,
|
|
GstObject * parent, GstEvent * event);
|
|
static gboolean gst_timecodestamper_ltcpad_query (GstPad * pad,
|
|
GstObject * parent, GstQuery * query);
|
|
static gboolean gst_timecodestamper_ltcpad_activatemode (GstPad * pad,
|
|
GstObject * parent, GstPadMode mode, gboolean active);
|
|
|
|
static gboolean gst_timecodestamper_videopad_activatemode (GstPad * pad,
|
|
GstObject * parent, GstPadMode mode, gboolean active);
|
|
|
|
static GstIterator *gst_timecodestamper_src_iterate_internal_link (GstPad * pad,
|
|
GstObject * parent);
|
|
#endif
|
|
|
|
static void gst_timecodestamper_update_drop_frame (GstTimeCodeStamper *
|
|
timecodestamper);
|
|
|
|
G_DEFINE_TYPE (GstTimeCodeStamper, gst_timecodestamper,
|
|
GST_TYPE_BASE_TRANSFORM);
|
|
GST_ELEMENT_REGISTER_DEFINE (timecodestamper, "timecodestamper",
|
|
GST_RANK_NONE, GST_TYPE_TIME_CODE_STAMPER);
|
|
|
|
GType
|
|
gst_timecodestamper_source_get_type (void)
|
|
{
|
|
static GType type = 0;
|
|
static const GEnumValue values[] = {
|
|
{GST_TIME_CODE_STAMPER_SOURCE_INTERNAL,
|
|
"Use internal timecode counter, starting at zero or value set by property",
|
|
"internal"},
|
|
{GST_TIME_CODE_STAMPER_SOURCE_ZERO,
|
|
"Always use zero", "zero"},
|
|
{GST_TIME_CODE_STAMPER_SOURCE_LAST_KNOWN,
|
|
"Count up from the last known upstream timecode or internal if unknown",
|
|
"last-known"},
|
|
{GST_TIME_CODE_STAMPER_SOURCE_LAST_KNOWN_OR_ZERO,
|
|
"Count up from the last known upstream timecode or zero if unknown",
|
|
"last-known-or-zero"},
|
|
{GST_TIME_CODE_STAMPER_SOURCE_LTC,
|
|
"Linear timecode from an audio device", "ltc"},
|
|
{GST_TIME_CODE_STAMPER_SOURCE_RTC,
|
|
"Timecode from real time clock", "rtc"},
|
|
{0, NULL, NULL},
|
|
};
|
|
|
|
if (!type) {
|
|
type = g_enum_register_static ("GstTimeCodeStamperSource", values);
|
|
}
|
|
return type;
|
|
}
|
|
|
|
GType
|
|
gst_timecodestamper_set_get_type (void)
|
|
{
|
|
static GType type = 0;
|
|
static const GEnumValue values[] = {
|
|
{GST_TIME_CODE_STAMPER_SET_NEVER,
|
|
"Never set timecodes", "never"},
|
|
{GST_TIME_CODE_STAMPER_SET_KEEP,
|
|
"Keep upstream timecodes and only set if no upstream timecode", "keep"},
|
|
{GST_TIME_CODE_STAMPER_SET_ALWAYS,
|
|
"Always set timecode and remove upstream timecode", "always"},
|
|
{0, NULL, NULL},
|
|
};
|
|
|
|
if (!type) {
|
|
type = g_enum_register_static ("GstTimeCodeStamperSet", values);
|
|
}
|
|
return type;
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_class_init (GstTimeCodeStamperClass * klass)
|
|
{
|
|
GObjectClass *gobject_class = (GObjectClass *) klass;
|
|
GstElementClass *element_class = (GstElementClass *) klass;
|
|
GstBaseTransformClass *trans_class = (GstBaseTransformClass *) klass;
|
|
|
|
GST_DEBUG_CATEGORY_INIT (timecodestamper_debug, "timecodestamper", 0,
|
|
"timecodestamper");
|
|
gst_element_class_set_static_metadata (element_class, "Timecode stamper",
|
|
"Filter/Video", "Attaches a timecode meta into each video frame",
|
|
"Vivia Nikolaidou <vivia@toolsonair.com>");
|
|
|
|
gobject_class->set_property = gst_timecodestamper_set_property;
|
|
gobject_class->get_property = gst_timecodestamper_get_property;
|
|
gobject_class->finalize = gst_timecodestamper_finalize;
|
|
gobject_class->dispose = gst_timecodestamper_dispose;
|
|
|
|
g_object_class_install_property (gobject_class, PROP_SOURCE,
|
|
g_param_spec_enum ("source", "Timecode Source",
|
|
"Choose from what source the timecode should be taken",
|
|
GST_TYPE_TIME_CODE_STAMPER_SOURCE,
|
|
DEFAULT_SOURCE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_SET,
|
|
g_param_spec_enum ("set", "Timecode Set",
|
|
"Choose whether timecodes should be overridden or not",
|
|
GST_TYPE_TIME_CODE_STAMPER_SET,
|
|
DEFAULT_SET, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_AUTO_RESYNC,
|
|
g_param_spec_boolean ("auto-resync",
|
|
"Auto Resync",
|
|
"If true resync last known timecode from upstream, otherwise only "
|
|
"count up from the last known one",
|
|
DEFAULT_AUTO_RESYNC, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_TIMEOUT,
|
|
g_param_spec_uint64 ("timeout",
|
|
"Timeout",
|
|
"Time out upstream timecode if no new timecode was detected after this time",
|
|
0, G_MAXUINT64, DEFAULT_TIMEOUT,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_DROP_FRAME,
|
|
g_param_spec_boolean ("drop-frame", "Drop Frame",
|
|
"Use drop-frame timecodes for 29.97 and 59.94 FPS",
|
|
DEFAULT_DROP_FRAME, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_POST_MESSAGES,
|
|
g_param_spec_boolean ("post-messages", "Post element message",
|
|
"Post element message containing the current timecode",
|
|
DEFAULT_POST_MESSAGES, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_SET_INTERNAL_TIMECODE,
|
|
g_param_spec_boxed ("set-internal-timecode",
|
|
"Set Internal Timecode",
|
|
"If set, take this timecode as the internal timecode for the first "
|
|
"frame and increment from it. Only the values itself and daily jam are taken, "
|
|
"flags and frame rate are always determined by timecodestamper "
|
|
"itself. If unset, the internal timecode will start at 0 with the daily jam "
|
|
"being the current real-time clock time",
|
|
GST_TYPE_VIDEO_TIME_CODE,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_LTC_DAILY_JAM,
|
|
g_param_spec_boxed ("ltc-daily-jam",
|
|
"LTC Daily jam",
|
|
"The daily jam of the LTC timecode",
|
|
G_TYPE_DATE_TIME, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_LTC_AUTO_RESYNC,
|
|
g_param_spec_boolean ("ltc-auto-resync",
|
|
"LTC Auto Resync",
|
|
"If true the LTC timecode will be automatically resynced if it drifts, "
|
|
"otherwise it will only be counted up from the last known one",
|
|
DEFAULT_LTC_AUTO_RESYNC, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_LTC_EXTRA_LATENCY,
|
|
g_param_spec_uint64 ("ltc-extra-latency", "LTC Extra Latency",
|
|
"Extra latency to introduce for waiting for LTC timecodes",
|
|
0, G_MAXUINT64, DEFAULT_LTC_EXTRA_LATENCY,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_LTC_TIMEOUT,
|
|
g_param_spec_uint64 ("ltc-timeout", "LTC Timeout",
|
|
"Time out LTC timecode if no new timecode was detected after this time",
|
|
0, G_MAXUINT64, DEFAULT_LTC_TIMEOUT,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_RTC_MAX_DRIFT,
|
|
g_param_spec_uint64 ("rtc-max-drift",
|
|
"RTC Maximum Offset",
|
|
"Maximum number of nanoseconds the RTC clock is allowed to drift from "
|
|
"the video before it is resynced",
|
|
0, G_MAXUINT64, DEFAULT_RTC_MAX_DRIFT,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_RTC_AUTO_RESYNC,
|
|
g_param_spec_boolean ("rtc-auto-resync",
|
|
"RTC Auto Resync",
|
|
"If true the RTC timecode will be automatically resynced if it drifts, "
|
|
"otherwise it will only be counted up from the last known one",
|
|
DEFAULT_RTC_AUTO_RESYNC, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_TIMECODE_OFFSET,
|
|
g_param_spec_int ("timecode-offset",
|
|
"Timecode Offset",
|
|
"Add this offset in frames to internal, LTC or RTC timecode, "
|
|
"useful if there is an offset between the timecode source and video",
|
|
G_MININT, G_MAXINT, 0, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&gst_timecodestamper_sink_template));
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&gst_timecodestamper_src_template));
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&gst_timecodestamper_ltc_template));
|
|
|
|
element_class->request_new_pad =
|
|
GST_DEBUG_FUNCPTR (gst_timecodestamper_request_new_pad);
|
|
element_class->release_pad =
|
|
GST_DEBUG_FUNCPTR (gst_timecodestamper_release_pad);
|
|
|
|
trans_class->sink_event = GST_DEBUG_FUNCPTR (gst_timecodestamper_sink_event);
|
|
trans_class->src_event = GST_DEBUG_FUNCPTR (gst_timecodestamper_src_event);
|
|
#if HAVE_LTC
|
|
trans_class->query = GST_DEBUG_FUNCPTR (gst_timecodestamper_query);
|
|
#endif
|
|
trans_class->stop = GST_DEBUG_FUNCPTR (gst_timecodestamper_stop);
|
|
trans_class->start = GST_DEBUG_FUNCPTR (gst_timecodestamper_start);
|
|
|
|
trans_class->transform_ip =
|
|
GST_DEBUG_FUNCPTR (gst_timecodestamper_transform_ip);
|
|
|
|
gst_type_mark_as_plugin_api (GST_TYPE_TIME_CODE_STAMPER_SOURCE, 0);
|
|
gst_type_mark_as_plugin_api (GST_TYPE_TIME_CODE_STAMPER_SET, 0);
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_init (GstTimeCodeStamper * timecodestamper)
|
|
{
|
|
timecodestamper->ltcpad = NULL;
|
|
|
|
timecodestamper->tc_source = GST_TIME_CODE_STAMPER_SOURCE_INTERNAL;
|
|
timecodestamper->tc_set = GST_TIME_CODE_STAMPER_SET_KEEP;
|
|
timecodestamper->tc_auto_resync = DEFAULT_AUTO_RESYNC;
|
|
timecodestamper->tc_timeout = DEFAULT_TIMEOUT;
|
|
timecodestamper->drop_frame = DEFAULT_DROP_FRAME;
|
|
timecodestamper->post_messages = DEFAULT_POST_MESSAGES;
|
|
timecodestamper->set_internal_tc = NULL;
|
|
timecodestamper->ltc_daily_jam = DEFAULT_LTC_DAILY_JAM;
|
|
timecodestamper->ltc_auto_resync = DEFAULT_LTC_AUTO_RESYNC;
|
|
timecodestamper->ltc_extra_latency = DEFAULT_LTC_EXTRA_LATENCY;
|
|
timecodestamper->ltc_timeout = DEFAULT_LTC_TIMEOUT;
|
|
timecodestamper->rtc_max_drift = DEFAULT_RTC_MAX_DRIFT;
|
|
timecodestamper->rtc_auto_resync = DEFAULT_RTC_AUTO_RESYNC;
|
|
timecodestamper->timecode_offset = 0;
|
|
|
|
timecodestamper->internal_tc = NULL;
|
|
timecodestamper->last_tc = NULL;
|
|
timecodestamper->last_tc_running_time = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->rtc_tc = NULL;
|
|
|
|
timecodestamper->seeked_frames = -1;
|
|
|
|
#if HAVE_LTC
|
|
g_mutex_init (&timecodestamper->mutex);
|
|
g_cond_init (&timecodestamper->ltc_cond_video);
|
|
g_cond_init (&timecodestamper->ltc_cond_audio);
|
|
|
|
gst_segment_init (&timecodestamper->ltc_segment, GST_FORMAT_UNDEFINED);
|
|
timecodestamper->ltc_first_running_time = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->ltc_current_running_time = GST_CLOCK_TIME_NONE;
|
|
|
|
g_queue_init (&timecodestamper->ltc_current_tcs);
|
|
timecodestamper->ltc_internal_tc = NULL;
|
|
timecodestamper->ltc_internal_running_time = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->ltc_dec = NULL;
|
|
timecodestamper->ltc_total = 0;
|
|
|
|
timecodestamper->ltc_eos = TRUE;
|
|
timecodestamper->ltc_flushing = TRUE;
|
|
|
|
timecodestamper->audio_live = FALSE;
|
|
timecodestamper->audio_latency = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->video_live = FALSE;
|
|
timecodestamper->video_latency = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->latency = GST_CLOCK_TIME_NONE;
|
|
|
|
timecodestamper->video_activatemode_default =
|
|
GST_PAD_ACTIVATEMODEFUNC (GST_BASE_TRANSFORM_SINK_PAD (timecodestamper));
|
|
GST_PAD_ACTIVATEMODEFUNC (GST_BASE_TRANSFORM_SINK_PAD (timecodestamper)) =
|
|
gst_timecodestamper_videopad_activatemode;
|
|
gst_pad_set_iterate_internal_links_function (GST_BASE_TRANSFORM_SRC_PAD
|
|
(timecodestamper), gst_timecodestamper_src_iterate_internal_link);
|
|
#endif
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_finalize (GObject * object)
|
|
{
|
|
#if HAVE_LTC
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (object);
|
|
|
|
g_cond_clear (&timecodestamper->ltc_cond_video);
|
|
g_cond_clear (&timecodestamper->ltc_cond_audio);
|
|
g_mutex_clear (&timecodestamper->mutex);
|
|
#endif
|
|
|
|
G_OBJECT_CLASS (gst_timecodestamper_parent_class)->finalize (object);
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_dispose (GObject * object)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (object);
|
|
|
|
if (timecodestamper->ltc_daily_jam) {
|
|
g_date_time_unref (timecodestamper->ltc_daily_jam);
|
|
timecodestamper->ltc_daily_jam = NULL;
|
|
}
|
|
|
|
if (timecodestamper->internal_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->internal_tc);
|
|
timecodestamper->internal_tc = NULL;
|
|
}
|
|
|
|
if (timecodestamper->set_internal_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->set_internal_tc);
|
|
timecodestamper->set_internal_tc = NULL;
|
|
}
|
|
|
|
if (timecodestamper->last_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->last_tc);
|
|
timecodestamper->last_tc = NULL;
|
|
}
|
|
timecodestamper->last_tc_running_time = GST_CLOCK_TIME_NONE;
|
|
|
|
if (timecodestamper->rtc_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->rtc_tc);
|
|
timecodestamper->rtc_tc = NULL;
|
|
}
|
|
#if HAVE_LTC
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
{
|
|
TimestampedTimecode *tc;
|
|
while ((tc = g_queue_pop_tail (&timecodestamper->ltc_current_tcs))) {
|
|
gst_video_time_code_clear (&tc->timecode);
|
|
g_free (tc);
|
|
}
|
|
}
|
|
if (timecodestamper->ltc_internal_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->ltc_internal_tc);
|
|
timecodestamper->ltc_internal_tc = NULL;
|
|
}
|
|
timecodestamper->ltc_internal_running_time = GST_CLOCK_TIME_NONE;
|
|
|
|
if (timecodestamper->ltc_dec) {
|
|
ltc_decoder_free (timecodestamper->ltc_dec);
|
|
timecodestamper->ltc_dec = NULL;
|
|
}
|
|
|
|
if (timecodestamper->stream_align) {
|
|
gst_audio_stream_align_free (timecodestamper->stream_align);
|
|
timecodestamper->stream_align = NULL;
|
|
}
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
#endif
|
|
|
|
G_OBJECT_CLASS (gst_timecodestamper_parent_class)->dispose (object);
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (object);
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
switch (prop_id) {
|
|
case PROP_SOURCE:
|
|
timecodestamper->tc_source = (GstTimeCodeStamperSource)
|
|
g_value_get_enum (value);
|
|
break;
|
|
case PROP_SET:
|
|
timecodestamper->tc_set = (GstTimeCodeStamperSet)
|
|
g_value_get_enum (value);
|
|
break;
|
|
case PROP_AUTO_RESYNC:
|
|
timecodestamper->tc_auto_resync = g_value_get_boolean (value);
|
|
break;
|
|
case PROP_TIMEOUT:
|
|
timecodestamper->tc_timeout = g_value_get_uint64 (value);
|
|
break;
|
|
case PROP_DROP_FRAME:
|
|
timecodestamper->drop_frame = g_value_get_boolean (value);
|
|
gst_timecodestamper_update_drop_frame (timecodestamper);
|
|
break;
|
|
case PROP_LTC_DAILY_JAM:
|
|
if (timecodestamper->ltc_daily_jam)
|
|
g_date_time_unref (timecodestamper->ltc_daily_jam);
|
|
timecodestamper->ltc_daily_jam = g_value_dup_boxed (value);
|
|
|
|
#if HAVE_LTC
|
|
{
|
|
GList *l;
|
|
|
|
for (l = timecodestamper->ltc_current_tcs.head; l; l = l->next) {
|
|
TimestampedTimecode *tc = l->data;
|
|
|
|
if (tc->timecode.config.latest_daily_jam) {
|
|
g_date_time_unref (tc->timecode.config.latest_daily_jam);
|
|
}
|
|
tc->timecode.config.latest_daily_jam =
|
|
g_date_time_ref (timecodestamper->ltc_daily_jam);
|
|
}
|
|
}
|
|
|
|
if (timecodestamper->ltc_internal_tc) {
|
|
if (timecodestamper->ltc_internal_tc->config.latest_daily_jam) {
|
|
g_date_time_unref (timecodestamper->ltc_internal_tc->
|
|
config.latest_daily_jam);
|
|
}
|
|
timecodestamper->ltc_internal_tc->config.latest_daily_jam =
|
|
g_date_time_ref (timecodestamper->ltc_daily_jam);
|
|
}
|
|
#endif
|
|
break;
|
|
case PROP_POST_MESSAGES:
|
|
timecodestamper->post_messages = g_value_get_boolean (value);
|
|
break;
|
|
case PROP_SET_INTERNAL_TIMECODE:{
|
|
if (timecodestamper->set_internal_tc)
|
|
gst_video_time_code_free (timecodestamper->set_internal_tc);
|
|
timecodestamper->set_internal_tc = g_value_dup_boxed (value);
|
|
|
|
/* Reset the internal timecode on the next opportunity if a new
|
|
* timecode was set here. If none was set we just continue counting
|
|
* from the previous one */
|
|
if (timecodestamper->set_internal_tc && timecodestamper->internal_tc) {
|
|
gst_video_time_code_free (timecodestamper->internal_tc);
|
|
timecodestamper->internal_tc = NULL;
|
|
}
|
|
break;
|
|
}
|
|
case PROP_LTC_AUTO_RESYNC:
|
|
timecodestamper->ltc_auto_resync = g_value_get_boolean (value);
|
|
break;
|
|
case PROP_LTC_TIMEOUT:
|
|
timecodestamper->ltc_timeout = g_value_get_uint64 (value);
|
|
break;
|
|
case PROP_LTC_EXTRA_LATENCY:
|
|
timecodestamper->ltc_extra_latency = g_value_get_uint64 (value);
|
|
break;
|
|
case PROP_RTC_MAX_DRIFT:
|
|
timecodestamper->rtc_max_drift = g_value_get_uint64 (value);
|
|
break;
|
|
case PROP_RTC_AUTO_RESYNC:
|
|
timecodestamper->rtc_auto_resync = g_value_get_boolean (value);
|
|
break;
|
|
case PROP_TIMECODE_OFFSET:
|
|
timecodestamper->timecode_offset = g_value_get_int (value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (object);
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
switch (prop_id) {
|
|
case PROP_SOURCE:
|
|
g_value_set_enum (value, timecodestamper->tc_source);
|
|
break;
|
|
case PROP_SET:
|
|
g_value_set_enum (value, timecodestamper->tc_set);
|
|
break;
|
|
case PROP_AUTO_RESYNC:
|
|
g_value_set_boolean (value, timecodestamper->tc_auto_resync);
|
|
break;
|
|
case PROP_TIMEOUT:
|
|
g_value_set_uint64 (value, timecodestamper->tc_timeout);
|
|
break;
|
|
case PROP_DROP_FRAME:
|
|
g_value_set_boolean (value, timecodestamper->drop_frame);
|
|
break;
|
|
case PROP_LTC_DAILY_JAM:
|
|
g_value_set_boxed (value, timecodestamper->ltc_daily_jam);
|
|
break;
|
|
case PROP_POST_MESSAGES:
|
|
g_value_set_boolean (value, timecodestamper->post_messages);
|
|
break;
|
|
case PROP_SET_INTERNAL_TIMECODE:
|
|
g_value_set_boxed (value, timecodestamper->set_internal_tc);
|
|
break;
|
|
case PROP_LTC_AUTO_RESYNC:
|
|
g_value_set_boolean (value, timecodestamper->ltc_auto_resync);
|
|
break;
|
|
case PROP_LTC_TIMEOUT:
|
|
g_value_set_uint64 (value, timecodestamper->ltc_timeout);
|
|
break;
|
|
case PROP_LTC_EXTRA_LATENCY:
|
|
g_value_set_uint64 (value, timecodestamper->ltc_extra_latency);
|
|
break;
|
|
case PROP_RTC_MAX_DRIFT:
|
|
g_value_set_uint64 (value, timecodestamper->rtc_max_drift);
|
|
break;
|
|
case PROP_RTC_AUTO_RESYNC:
|
|
g_value_set_boolean (value, timecodestamper->rtc_auto_resync);
|
|
break;
|
|
case PROP_TIMECODE_OFFSET:
|
|
g_value_set_int (value, timecodestamper->timecode_offset);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_stop (GstBaseTransform * trans)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (trans);
|
|
|
|
#if HAVE_LTC
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_flushing = TRUE;
|
|
timecodestamper->video_current_running_time = GST_CLOCK_TIME_NONE;
|
|
if (timecodestamper->video_clock_id)
|
|
gst_clock_id_unschedule (timecodestamper->video_clock_id);
|
|
timecodestamper->ltc_flushing = TRUE;
|
|
g_cond_signal (&timecodestamper->ltc_cond_video);
|
|
g_cond_signal (&timecodestamper->ltc_cond_audio);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
#endif
|
|
|
|
timecodestamper->interlace_mode = GST_VIDEO_INTERLACE_MODE_PROGRESSIVE;
|
|
timecodestamper->fps_n = 0;
|
|
timecodestamper->fps_d = 1;
|
|
|
|
if (timecodestamper->internal_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->internal_tc);
|
|
timecodestamper->internal_tc = NULL;
|
|
}
|
|
|
|
if (timecodestamper->rtc_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->rtc_tc);
|
|
timecodestamper->rtc_tc = NULL;
|
|
}
|
|
|
|
if (timecodestamper->last_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->last_tc);
|
|
timecodestamper->last_tc = NULL;
|
|
}
|
|
timecodestamper->last_tc_running_time = GST_CLOCK_TIME_NONE;
|
|
#if HAVE_LTC
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
gst_audio_info_init (&timecodestamper->ainfo);
|
|
gst_segment_init (&timecodestamper->ltc_segment, GST_FORMAT_UNDEFINED);
|
|
|
|
timecodestamper->ltc_first_running_time = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->ltc_current_running_time = GST_CLOCK_TIME_NONE;
|
|
|
|
if (timecodestamper->ltc_internal_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->ltc_internal_tc);
|
|
timecodestamper->ltc_internal_tc = NULL;
|
|
}
|
|
timecodestamper->ltc_internal_running_time = GST_CLOCK_TIME_NONE;
|
|
|
|
{
|
|
TimestampedTimecode *tc;
|
|
while ((tc = g_queue_pop_tail (&timecodestamper->ltc_current_tcs))) {
|
|
gst_video_time_code_clear (&tc->timecode);
|
|
g_free (tc);
|
|
}
|
|
}
|
|
|
|
if (timecodestamper->ltc_dec) {
|
|
ltc_decoder_free (timecodestamper->ltc_dec);
|
|
timecodestamper->ltc_dec = NULL;
|
|
}
|
|
|
|
if (timecodestamper->stream_align) {
|
|
gst_audio_stream_align_free (timecodestamper->stream_align);
|
|
timecodestamper->stream_align = NULL;
|
|
}
|
|
|
|
timecodestamper->ltc_total = 0;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
#endif
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_start (GstBaseTransform * trans)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (trans);
|
|
|
|
#if HAVE_LTC
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_flushing = FALSE;
|
|
timecodestamper->video_eos = FALSE;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
#endif
|
|
|
|
timecodestamper->interlace_mode = GST_VIDEO_INTERLACE_MODE_PROGRESSIVE;
|
|
timecodestamper->fps_n = 0;
|
|
timecodestamper->fps_d = 1;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/* Must be called with object lock */
|
|
static void
|
|
gst_timecodestamper_update_drop_frame (GstTimeCodeStamper * timecodestamper)
|
|
{
|
|
if (timecodestamper->drop_frame && timecodestamper->fps_d == 1001 &&
|
|
(timecodestamper->fps_n == 30000 || timecodestamper->fps_n == 60000)) {
|
|
if (timecodestamper->internal_tc)
|
|
timecodestamper->internal_tc->config.flags |=
|
|
GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
if (timecodestamper->rtc_tc)
|
|
timecodestamper->rtc_tc->config.flags |=
|
|
GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
#if HAVE_LTC
|
|
{
|
|
GList *l;
|
|
|
|
for (l = timecodestamper->ltc_current_tcs.head; l; l = l->next) {
|
|
TimestampedTimecode *tc = l->data;
|
|
|
|
tc->timecode.config.flags |= GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
}
|
|
}
|
|
if (timecodestamper->ltc_internal_tc)
|
|
timecodestamper->ltc_internal_tc->config.flags |=
|
|
GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
#endif
|
|
} else {
|
|
if (timecodestamper->internal_tc)
|
|
timecodestamper->internal_tc->config.flags &=
|
|
~GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
if (timecodestamper->rtc_tc)
|
|
timecodestamper->rtc_tc->config.flags &=
|
|
~GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
#if HAVE_LTC
|
|
{
|
|
GList *l;
|
|
|
|
for (l = timecodestamper->ltc_current_tcs.head; l; l = l->next) {
|
|
TimestampedTimecode *tc = l->data;
|
|
|
|
tc->timecode.config.flags &= ~GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
}
|
|
}
|
|
if (timecodestamper->ltc_internal_tc)
|
|
timecodestamper->ltc_internal_tc->config.flags &=
|
|
~GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
#endif
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_update_timecode_framerate (GstTimeCodeStamper *
|
|
timecodestamper, gint fps_n, gint fps_d, GstVideoTimeCode * timecode,
|
|
gboolean is_ltc)
|
|
{
|
|
guint64 nframes;
|
|
GstClockTime time;
|
|
GDateTime *jam = NULL;
|
|
GstVideoTimeCodeFlags tc_flags = 0;
|
|
|
|
if (!timecode)
|
|
return;
|
|
|
|
if (timecodestamper->interlace_mode != GST_VIDEO_INTERLACE_MODE_PROGRESSIVE)
|
|
tc_flags |= GST_VIDEO_TIME_CODE_FLAGS_INTERLACED;
|
|
|
|
if (timecodestamper->drop_frame && timecodestamper->fps_d == 1001 &&
|
|
(timecodestamper->fps_n == 30000 || timecodestamper->fps_n == 60000))
|
|
tc_flags |= GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
|
|
/* If this is an LTC timecode and we have no framerate yet in there then
|
|
* just do nothing. We're going to set the framerate at a later time */
|
|
if (timecode->config.fps_d != 0 || !is_ltc) {
|
|
nframes = gst_video_time_code_frames_since_daily_jam (timecode);
|
|
time =
|
|
gst_util_uint64_scale (nframes,
|
|
GST_SECOND * timecodestamper->fps_d, timecodestamper->fps_n);
|
|
jam =
|
|
timecode->config.latest_daily_jam ? g_date_time_ref (timecode->config.
|
|
latest_daily_jam) : NULL;
|
|
gst_video_time_code_clear (timecode);
|
|
gst_video_time_code_init (timecode, timecodestamper->fps_n,
|
|
timecodestamper->fps_d, jam, tc_flags, 0, 0, 0, 0, 0);
|
|
if (jam)
|
|
g_date_time_unref (jam);
|
|
|
|
nframes = gst_util_uint64_scale (time, fps_n, GST_SECOND * fps_d);
|
|
gst_video_time_code_add_frames (timecode, nframes);
|
|
}
|
|
}
|
|
|
|
/* Must be called with object lock */
|
|
static gboolean
|
|
gst_timecodestamper_update_framerate (GstTimeCodeStamper * timecodestamper,
|
|
gint fps_n, gint fps_d)
|
|
{
|
|
/* Nothing changed */
|
|
if (fps_n == timecodestamper->fps_n && fps_d == timecodestamper->fps_d)
|
|
return FALSE;
|
|
|
|
gst_timecodestamper_update_timecode_framerate (timecodestamper, fps_n, fps_d,
|
|
timecodestamper->internal_tc, FALSE);
|
|
gst_timecodestamper_update_timecode_framerate (timecodestamper, fps_n, fps_d,
|
|
timecodestamper->last_tc, FALSE);
|
|
gst_timecodestamper_update_timecode_framerate (timecodestamper, fps_n, fps_d,
|
|
timecodestamper->rtc_tc, FALSE);
|
|
|
|
#if HAVE_LTC
|
|
{
|
|
GList *l;
|
|
|
|
for (l = timecodestamper->ltc_current_tcs.head; l; l = l->next) {
|
|
TimestampedTimecode *tc = l->data;
|
|
|
|
gst_timecodestamper_update_timecode_framerate (timecodestamper, fps_n,
|
|
fps_d, &tc->timecode, TRUE);
|
|
}
|
|
}
|
|
gst_timecodestamper_update_timecode_framerate (timecodestamper, fps_n, fps_d,
|
|
timecodestamper->ltc_internal_tc, FALSE);
|
|
#endif
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_sink_event (GstBaseTransform * trans, GstEvent * event)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (trans);
|
|
gboolean ret = FALSE;
|
|
|
|
GST_DEBUG_OBJECT (trans, "received event %" GST_PTR_FORMAT, event);
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_SEGMENT:
|
|
{
|
|
GstSegment segment;
|
|
|
|
gst_event_copy_segment (event, &segment);
|
|
if (segment.format != GST_FORMAT_TIME) {
|
|
GST_ERROR_OBJECT (timecodestamper, "Invalid segment format");
|
|
gst_event_unref (event);
|
|
return FALSE;
|
|
}
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
if (timecodestamper->tc_source == GST_TIME_CODE_STAMPER_SOURCE_INTERNAL
|
|
&& GST_EVENT_SEQNUM (event) == timecodestamper->prev_seek_seqnum) {
|
|
timecodestamper->reset_internal_tc_from_seek = TRUE;
|
|
timecodestamper->prev_seek_seqnum = GST_SEQNUM_INVALID;
|
|
}
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
break;
|
|
}
|
|
case GST_EVENT_CAPS:
|
|
{
|
|
GstCaps *caps;
|
|
gboolean latency_changed;
|
|
const gchar *interlace_mode;
|
|
GstStructure *s;
|
|
gint fps_n, fps_d;
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
gst_event_parse_caps (event, &caps);
|
|
|
|
s = gst_caps_get_structure (caps, 0);
|
|
|
|
if (!gst_structure_get_fraction (s, "framerate", &fps_n, &fps_d)) {
|
|
GST_ERROR_OBJECT (timecodestamper, "Expected framerate in caps");
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
gst_event_unref (event);
|
|
return FALSE;
|
|
}
|
|
|
|
if (fps_n == 0) {
|
|
GST_ERROR_OBJECT (timecodestamper,
|
|
"Non-constant frame rate found. Refusing to create a timecode");
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
gst_event_unref (event);
|
|
return FALSE;
|
|
}
|
|
|
|
if ((interlace_mode = gst_structure_get_string (s, "interlace-mode"))) {
|
|
timecodestamper->interlace_mode =
|
|
gst_video_interlace_mode_from_string (interlace_mode);
|
|
}
|
|
|
|
latency_changed =
|
|
gst_timecodestamper_update_framerate (timecodestamper, fps_n, fps_d);
|
|
|
|
timecodestamper->fps_n = fps_n;
|
|
timecodestamper->fps_d = fps_d;
|
|
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
if (latency_changed)
|
|
gst_element_post_message (GST_ELEMENT_CAST (timecodestamper),
|
|
gst_message_new_latency (GST_OBJECT_CAST (timecodestamper)));
|
|
break;
|
|
}
|
|
#if HAVE_LTC
|
|
case GST_EVENT_FLUSH_START:
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_flushing = TRUE;
|
|
timecodestamper->video_current_running_time = GST_CLOCK_TIME_NONE;
|
|
if (timecodestamper->video_clock_id)
|
|
gst_clock_id_unschedule (timecodestamper->video_clock_id);
|
|
g_cond_signal (&timecodestamper->ltc_cond_video);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
break;
|
|
case GST_EVENT_FLUSH_STOP:
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_flushing = FALSE;
|
|
timecodestamper->video_eos = FALSE;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
break;
|
|
case GST_EVENT_EOS:
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_eos = TRUE;
|
|
g_cond_signal (&timecodestamper->ltc_cond_audio);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
break;
|
|
#endif
|
|
default:
|
|
break;
|
|
}
|
|
ret =
|
|
GST_BASE_TRANSFORM_CLASS (gst_timecodestamper_parent_class)->sink_event
|
|
(trans, event);
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_src_event (GstBaseTransform * trans, GstEvent * event)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (trans);
|
|
|
|
GST_DEBUG_OBJECT (trans, "received event %" GST_PTR_FORMAT, event);
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_SEEK:
|
|
{
|
|
gdouble rate;
|
|
GstSeekType start_type;
|
|
gint64 start;
|
|
GstFormat format;
|
|
|
|
gst_event_parse_seek (event, &rate, &format, NULL, &start_type, &start,
|
|
NULL, NULL);
|
|
|
|
if (rate < 0) {
|
|
GST_ERROR_OBJECT (timecodestamper, "Reverse playback is not supported");
|
|
return FALSE;
|
|
}
|
|
|
|
if (format != GST_FORMAT_TIME) {
|
|
GST_ERROR_OBJECT (timecodestamper,
|
|
"Seeking is only supported in TIME format");
|
|
return FALSE;
|
|
}
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
if (timecodestamper->fps_d && timecodestamper->fps_n) {
|
|
timecodestamper->prev_seek_seqnum = GST_EVENT_SEQNUM (event);
|
|
timecodestamper->seeked_frames = gst_util_uint64_scale (start,
|
|
timecodestamper->fps_n, timecodestamper->fps_d * GST_SECOND);
|
|
}
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return
|
|
GST_BASE_TRANSFORM_CLASS (gst_timecodestamper_parent_class)->src_event
|
|
(trans, event);
|
|
}
|
|
|
|
#if HAVE_LTC
|
|
static gboolean
|
|
gst_timecodestamper_query (GstBaseTransform * trans,
|
|
GstPadDirection direction, GstQuery * query)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (trans);
|
|
|
|
if (direction == GST_PAD_SINK)
|
|
return
|
|
GST_BASE_TRANSFORM_CLASS (gst_timecodestamper_parent_class)->query
|
|
(trans, direction, query);
|
|
|
|
switch (GST_QUERY_TYPE (query)) {
|
|
case GST_QUERY_LATENCY:{
|
|
gboolean res;
|
|
gboolean live;
|
|
GstClockTime min_latency, max_latency;
|
|
GstClockTime latency;
|
|
|
|
res =
|
|
gst_pad_query_default (GST_BASE_TRANSFORM_SRC_PAD (trans),
|
|
GST_OBJECT_CAST (trans), query);
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
if (res && timecodestamper->fps_n && timecodestamper->fps_d) {
|
|
gst_query_parse_latency (query, &live, &min_latency, &max_latency);
|
|
if (live && timecodestamper->ltcpad) {
|
|
/* Introduce additional LTC for waiting for LTC timecodes. The
|
|
* LTC library introduces some as well as the encoding of the LTC
|
|
* signal. */
|
|
latency = timecodestamper->ltc_extra_latency;
|
|
min_latency += latency;
|
|
if (max_latency != GST_CLOCK_TIME_NONE)
|
|
max_latency += latency;
|
|
timecodestamper->latency = min_latency;
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Reporting latency min %" GST_TIME_FORMAT " max %" GST_TIME_FORMAT
|
|
" ours %" GST_TIME_FORMAT, GST_TIME_ARGS (min_latency),
|
|
GST_TIME_ARGS (max_latency), GST_TIME_ARGS (latency));
|
|
gst_query_set_latency (query, live, min_latency, max_latency);
|
|
} else {
|
|
timecodestamper->latency = 0;
|
|
}
|
|
} else if (res) {
|
|
GST_ERROR_OBJECT (timecodestamper,
|
|
"Need a known, non-variable framerate to answer LATENCY query");
|
|
res = FALSE;
|
|
timecodestamper->latency = GST_CLOCK_TIME_NONE;
|
|
}
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
|
|
return res;
|
|
}
|
|
default:
|
|
return
|
|
GST_BASE_TRANSFORM_CLASS (gst_timecodestamper_parent_class)->query
|
|
(trans, direction, query);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static gboolean
|
|
remove_timecode_meta (GstBuffer * buffer, GstMeta ** meta, gpointer user_data)
|
|
{
|
|
if (meta && *meta && (*meta)->info->api == GST_VIDEO_TIME_CODE_META_API_TYPE) {
|
|
*meta = NULL;
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
#if HAVE_LTC
|
|
static void
|
|
gst_timecodestamper_update_latency (GstTimeCodeStamper * timecodestamper,
|
|
GstPad * pad, gboolean * live, GstClockTime * latency)
|
|
{
|
|
GstQuery *query;
|
|
|
|
query = gst_query_new_latency ();
|
|
if (!gst_pad_peer_query (pad, query)) {
|
|
GST_WARNING_OBJECT (pad, "Failed to query latency");
|
|
gst_pad_mark_reconfigure (pad);
|
|
gst_query_unref (query);
|
|
return;
|
|
}
|
|
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
gst_query_parse_latency (query, live, latency, NULL);
|
|
/* If we're not live, consider a latency of 0 */
|
|
if (!*live)
|
|
*latency = 0;
|
|
GST_DEBUG_OBJECT (pad,
|
|
"Queried latency: live %d, min latency %" GST_TIME_FORMAT, *live,
|
|
GST_TIME_ARGS (*latency));
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
gst_query_unref (query);
|
|
}
|
|
#endif
|
|
|
|
static GstFlowReturn
|
|
gst_timecodestamper_transform_ip (GstBaseTransform * vfilter,
|
|
GstBuffer * buffer)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (vfilter);
|
|
GstClockTime running_time, base_time, clock_time;
|
|
GstClock *clock;
|
|
GstClockTime clock_time_now;
|
|
GDateTime *dt_now, *dt_frame;
|
|
GstVideoTimeCode *tc = NULL;
|
|
gboolean free_tc = FALSE;
|
|
GstVideoTimeCodeMeta *tc_meta;
|
|
GstFlowReturn flow_ret = GST_FLOW_OK;
|
|
GstVideoTimeCodeFlags tc_flags = 0;
|
|
|
|
if (timecodestamper->fps_n == 0 || timecodestamper->fps_d == 0
|
|
|| !GST_BUFFER_PTS_IS_VALID (buffer)) {
|
|
gst_buffer_unref (buffer);
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
}
|
|
#if HAVE_LTC
|
|
if (timecodestamper->video_latency == -1
|
|
|| gst_pad_check_reconfigure (GST_BASE_TRANSFORM_SINK_PAD (vfilter))) {
|
|
gst_timecodestamper_update_latency (timecodestamper,
|
|
GST_BASE_TRANSFORM_SINK_PAD (vfilter), &timecodestamper->video_live,
|
|
&timecodestamper->video_latency);
|
|
}
|
|
#endif
|
|
|
|
/* Collect all the current times */
|
|
base_time = gst_element_get_base_time (GST_ELEMENT (timecodestamper));
|
|
clock = gst_element_get_clock (GST_ELEMENT (timecodestamper));
|
|
if (clock) {
|
|
clock_time_now = gst_clock_get_time (clock);
|
|
gst_object_unref (clock);
|
|
} else {
|
|
clock_time_now = GST_CLOCK_TIME_NONE;
|
|
}
|
|
|
|
dt_now = g_date_time_new_now_local ();
|
|
|
|
running_time =
|
|
gst_segment_to_running_time (&vfilter->segment, GST_FORMAT_TIME,
|
|
GST_BUFFER_PTS (buffer));
|
|
|
|
if (clock_time_now != GST_CLOCK_TIME_NONE) {
|
|
gdouble seconds_diff;
|
|
|
|
clock_time = running_time + base_time;
|
|
if (clock_time_now > clock_time) {
|
|
seconds_diff = (clock_time_now - clock_time) / -1000000000.0;
|
|
} else {
|
|
seconds_diff = (clock_time - clock_time_now) / 1000000000.0;
|
|
}
|
|
dt_frame = g_date_time_add_seconds (dt_now, seconds_diff);
|
|
} else {
|
|
/* If we have no clock we can't really know the time of the frame */
|
|
dt_frame = g_date_time_ref (dt_now);
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Handling video frame with running time %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (running_time));
|
|
|
|
tc_meta = gst_buffer_get_video_time_code_meta (buffer);
|
|
|
|
/* Update all our internal timecodes as needed */
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
|
|
if (timecodestamper->interlace_mode != GST_VIDEO_INTERLACE_MODE_PROGRESSIVE)
|
|
tc_flags |= GST_VIDEO_TIME_CODE_FLAGS_INTERLACED;
|
|
|
|
if (timecodestamper->drop_frame && timecodestamper->fps_d == 1001 &&
|
|
(timecodestamper->fps_n == 30000 || timecodestamper->fps_n == 60000))
|
|
tc_flags |= GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME;
|
|
|
|
/* If we don't have an internal timecode yet then either a new one was just
|
|
* set via the property or we just started. Initialize it here, otherwise
|
|
* increment it by one */
|
|
if (!timecodestamper->internal_tc
|
|
|| timecodestamper->reset_internal_tc_from_seek) {
|
|
gchar *tc_str;
|
|
|
|
if (timecodestamper->internal_tc)
|
|
gst_video_time_code_free (timecodestamper->internal_tc);
|
|
|
|
timecodestamper->reset_internal_tc_from_seek = FALSE;
|
|
if (timecodestamper->set_internal_tc) {
|
|
timecodestamper->internal_tc =
|
|
gst_video_time_code_new (timecodestamper->fps_n,
|
|
timecodestamper->fps_d,
|
|
timecodestamper->set_internal_tc->config.latest_daily_jam, tc_flags,
|
|
timecodestamper->set_internal_tc->hours,
|
|
timecodestamper->set_internal_tc->minutes,
|
|
timecodestamper->set_internal_tc->seconds,
|
|
timecodestamper->set_internal_tc->frames,
|
|
timecodestamper->set_internal_tc->field_count);
|
|
} else {
|
|
timecodestamper->internal_tc =
|
|
gst_video_time_code_new (timecodestamper->fps_n,
|
|
timecodestamper->fps_d, dt_frame, tc_flags, 0, 0, 0, 0, 0);
|
|
if (timecodestamper->seeked_frames > 0) {
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Adding %" G_GINT64_FORMAT " frames that were seeked",
|
|
timecodestamper->seeked_frames);
|
|
gst_video_time_code_add_frames (timecodestamper->internal_tc,
|
|
timecodestamper->seeked_frames);
|
|
timecodestamper->seeked_frames = -1;
|
|
}
|
|
}
|
|
|
|
tc_str = gst_video_time_code_to_string (timecodestamper->internal_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper, "Initialized internal timecode to %s",
|
|
tc_str);
|
|
g_free (tc_str);
|
|
} else {
|
|
gchar *tc_str;
|
|
|
|
gst_video_time_code_increment_frame (timecodestamper->internal_tc);
|
|
tc_str = gst_video_time_code_to_string (timecodestamper->internal_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper, "Incremented internal timecode to %s",
|
|
tc_str);
|
|
g_free (tc_str);
|
|
}
|
|
|
|
/* If we have a new timecode on the incoming frame, update our last known
|
|
* timecode or otherwise increment it by one */
|
|
if (tc_meta && (!timecodestamper->last_tc || timecodestamper->tc_auto_resync)) {
|
|
gchar *tc_str;
|
|
|
|
if (timecodestamper->last_tc)
|
|
gst_video_time_code_free (timecodestamper->last_tc);
|
|
timecodestamper->last_tc = gst_video_time_code_copy (&tc_meta->tc);
|
|
timecodestamper->last_tc_running_time = running_time;
|
|
|
|
tc_str = gst_video_time_code_to_string (timecodestamper->last_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper, "Updated upstream timecode to %s",
|
|
tc_str);
|
|
g_free (tc_str);
|
|
} else {
|
|
if (timecodestamper->last_tc) {
|
|
if (timecodestamper->tc_auto_resync
|
|
&& timecodestamper->tc_timeout != GST_CLOCK_TIME_NONE
|
|
&& (running_time + timecodestamper->tc_timeout <
|
|
timecodestamper->last_tc_running_time
|
|
|| running_time >=
|
|
timecodestamper->last_tc_running_time +
|
|
timecodestamper->tc_timeout)) {
|
|
if (timecodestamper->last_tc)
|
|
gst_video_time_code_free (timecodestamper->last_tc);
|
|
timecodestamper->last_tc = NULL;
|
|
timecodestamper->last_tc_running_time = GST_CLOCK_TIME_NONE;
|
|
GST_DEBUG_OBJECT (timecodestamper, "Upstream timecode timed out");
|
|
} else {
|
|
gchar *tc_str;
|
|
|
|
gst_video_time_code_increment_frame (timecodestamper->last_tc);
|
|
|
|
tc_str = gst_video_time_code_to_string (timecodestamper->last_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Incremented upstream timecode to %s", tc_str);
|
|
g_free (tc_str);
|
|
}
|
|
} else {
|
|
GST_DEBUG_OBJECT (timecodestamper, "Never saw an upstream timecode");
|
|
}
|
|
}
|
|
|
|
/* Update RTC-based timecode */
|
|
{
|
|
GstVideoTimeCode rtc_timecode_now;
|
|
gchar *tc_str, *dt_str;
|
|
|
|
/* Create timecode for the current frame time */
|
|
memset (&rtc_timecode_now, 0, sizeof (rtc_timecode_now));
|
|
gst_video_time_code_init_from_date_time_full (&rtc_timecode_now,
|
|
timecodestamper->fps_n, timecodestamper->fps_d, dt_frame, tc_flags, 0);
|
|
|
|
tc_str = gst_video_time_code_to_string (&rtc_timecode_now);
|
|
dt_str = g_date_time_format (dt_frame, "%F %R %z");
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Created RTC timecode %s for %s (%06u us)", tc_str, dt_str,
|
|
g_date_time_get_microsecond (dt_frame));
|
|
g_free (dt_str);
|
|
g_free (tc_str);
|
|
|
|
/* If we don't have an RTC timecode yet, directly initialize with this one */
|
|
if (!timecodestamper->rtc_tc) {
|
|
timecodestamper->rtc_tc = gst_video_time_code_copy (&rtc_timecode_now);
|
|
tc_str = gst_video_time_code_to_string (timecodestamper->rtc_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper, "Initialized RTC timecode to %s",
|
|
tc_str);
|
|
g_free (tc_str);
|
|
} else {
|
|
GstClockTime rtc_now_time, rtc_tc_time;
|
|
GstClockTime rtc_diff;
|
|
|
|
/* Increment the old RTC timecode to this frame */
|
|
gst_video_time_code_increment_frame (timecodestamper->rtc_tc);
|
|
|
|
/* Otherwise check if we drifted too much and need to resync */
|
|
rtc_tc_time =
|
|
gst_video_time_code_nsec_since_daily_jam (timecodestamper->rtc_tc);
|
|
rtc_now_time =
|
|
gst_video_time_code_nsec_since_daily_jam (&rtc_timecode_now);
|
|
if (rtc_tc_time > rtc_now_time)
|
|
rtc_diff = rtc_tc_time - rtc_now_time;
|
|
else
|
|
rtc_diff = rtc_now_time - rtc_tc_time;
|
|
|
|
if (timecodestamper->rtc_auto_resync
|
|
&& timecodestamper->rtc_max_drift != GST_CLOCK_TIME_NONE
|
|
&& rtc_diff > timecodestamper->rtc_max_drift) {
|
|
gst_video_time_code_free (timecodestamper->rtc_tc);
|
|
timecodestamper->rtc_tc = gst_video_time_code_copy (&rtc_timecode_now);
|
|
tc_str = gst_video_time_code_to_string (timecodestamper->rtc_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Updated RTC timecode to %s (%s%" GST_TIME_FORMAT " drift)", tc_str,
|
|
(rtc_tc_time > rtc_now_time ? "-" : "+"), GST_TIME_ARGS (rtc_diff));
|
|
g_free (tc_str);
|
|
} else {
|
|
/* Else nothing to do here, we use the current one */
|
|
tc_str = gst_video_time_code_to_string (timecodestamper->rtc_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Incremented RTC timecode to %s (%s%" GST_TIME_FORMAT " drift)",
|
|
tc_str, (rtc_tc_time > rtc_now_time ? "-" : "+"),
|
|
GST_TIME_ARGS (rtc_diff));
|
|
g_free (tc_str);
|
|
}
|
|
}
|
|
|
|
gst_video_time_code_clear (&rtc_timecode_now);
|
|
}
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
/* Update LTC-based timecode as needed */
|
|
#if HAVE_LTC
|
|
if (timecodestamper->ltcpad) {
|
|
GstClockTime frame_duration;
|
|
gchar *tc_str;
|
|
TimestampedTimecode *ltc_tc;
|
|
gboolean updated_internal = FALSE;
|
|
|
|
frame_duration = gst_util_uint64_scale_int_ceil (GST_SECOND,
|
|
timecodestamper->fps_d, timecodestamper->fps_n);
|
|
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
|
|
timecodestamper->video_current_running_time = running_time;
|
|
|
|
/* Wait to compensate for the latency we introduce and to allow the LTC
|
|
* audio to provide enough audio to extract timecodes, or until the video
|
|
* pad is flushing or the LTC pad is EOS.
|
|
* In non-live mode we introduce 4 frames of latency compared to the LTC
|
|
* audio, see LATENCY query handling for details. */
|
|
if (timecodestamper->video_live) {
|
|
GstClock *clock =
|
|
gst_element_get_clock (GST_ELEMENT_CAST (timecodestamper));
|
|
|
|
if (clock) {
|
|
GstClockID clock_id;
|
|
GstClockTime base_time =
|
|
gst_element_get_base_time (GST_ELEMENT_CAST (timecodestamper));
|
|
GstClockTime wait_time;
|
|
|
|
/* If we have no latency yet then wait at least for the LTC extra
|
|
* latency. See LATENCY query handling for details. */
|
|
if (timecodestamper->latency == GST_CLOCK_TIME_NONE) {
|
|
wait_time =
|
|
base_time + running_time + timecodestamper->ltc_extra_latency;
|
|
} else {
|
|
wait_time = base_time + running_time + timecodestamper->latency;
|
|
}
|
|
|
|
GST_TRACE_OBJECT (timecodestamper,
|
|
"Waiting for clock to reach %" GST_TIME_FORMAT
|
|
" (base time %" GST_TIME_FORMAT
|
|
" + running time %" GST_TIME_FORMAT
|
|
" + latency %" GST_TIME_FORMAT
|
|
"), now %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (wait_time),
|
|
GST_TIME_ARGS (base_time),
|
|
GST_TIME_ARGS (running_time),
|
|
GST_TIME_ARGS (timecodestamper->latency ==
|
|
GST_CLOCK_TIME_NONE ? timecodestamper->ltc_extra_latency :
|
|
timecodestamper->latency),
|
|
GST_TIME_ARGS (gst_clock_get_time (clock))
|
|
);
|
|
clock_id = gst_clock_new_single_shot_id (clock, wait_time);
|
|
|
|
timecodestamper->video_clock_id = clock_id;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
gst_clock_id_wait (clock_id, NULL);
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_clock_id = NULL;
|
|
gst_clock_id_unref (clock_id);
|
|
gst_object_unref (clock);
|
|
} else {
|
|
GST_WARNING_OBJECT (timecodestamper,
|
|
"No clock in live mode, not waiting");
|
|
}
|
|
} else {
|
|
while ((timecodestamper->ltc_current_running_time == GST_CLOCK_TIME_NONE
|
|
|| timecodestamper->ltc_current_running_time <
|
|
running_time + 8 * frame_duration)
|
|
&& !timecodestamper->video_flushing && !timecodestamper->ltc_eos) {
|
|
GST_TRACE_OBJECT (timecodestamper,
|
|
"Waiting for LTC audio to advance, EOS or flushing");
|
|
g_cond_wait (&timecodestamper->ltc_cond_video, &timecodestamper->mutex);
|
|
}
|
|
}
|
|
|
|
if (timecodestamper->video_flushing) {
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
flow_ret = GST_FLOW_FLUSHING;
|
|
goto out;
|
|
}
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
/* Take timecodes out of the queue until we're at the current video
|
|
* position. */
|
|
while ((ltc_tc = g_queue_pop_head (&timecodestamper->ltc_current_tcs))) {
|
|
/* First update framerate and flags according to the video stream if not
|
|
* done yet */
|
|
if (ltc_tc->timecode.config.fps_d == 0) {
|
|
gint fps_n_div =
|
|
((gdouble) timecodestamper->fps_n) /
|
|
timecodestamper->fps_d > 30 ? 2 : 1;
|
|
|
|
ltc_tc->timecode.config.flags = tc_flags;
|
|
ltc_tc->timecode.config.fps_n = timecodestamper->fps_n / fps_n_div;
|
|
ltc_tc->timecode.config.fps_d = timecodestamper->fps_d;
|
|
}
|
|
|
|
tc_str = gst_video_time_code_to_string (<c_tc->timecode);
|
|
GST_INFO_OBJECT (timecodestamper,
|
|
"Retrieved LTC timecode %s at %" GST_TIME_FORMAT
|
|
" (%u timecodes queued)", tc_str,
|
|
GST_TIME_ARGS (ltc_tc->running_time),
|
|
g_queue_get_length (&timecodestamper->ltc_current_tcs));
|
|
g_free (tc_str);
|
|
|
|
if (!gst_video_time_code_is_valid (<c_tc->timecode)) {
|
|
tc_str = gst_video_time_code_to_string (<c_tc->timecode);
|
|
GST_INFO_OBJECT (timecodestamper, "Invalid LTC timecode %s", tc_str);
|
|
g_free (tc_str);
|
|
gst_video_time_code_clear (<c_tc->timecode);
|
|
g_free (ltc_tc);
|
|
ltc_tc = NULL;
|
|
continue;
|
|
}
|
|
|
|
/* A timecode frame that starts +/- half a frame to the
|
|
* video frame is considered belonging to that video frame.
|
|
*
|
|
* If it's further ahead than half a frame duration, break out of
|
|
* the loop here and reconsider on the next frame. */
|
|
if (ABSDIFF (running_time, ltc_tc->running_time) <= frame_duration / 2) {
|
|
/* If we're resyncing LTC in general, directly replace the current
|
|
* LTC timecode with the new one we read. Otherwise we'll continue
|
|
* counting based on the previous timecode we had
|
|
*/
|
|
if (timecodestamper->ltc_auto_resync) {
|
|
if (timecodestamper->ltc_internal_tc)
|
|
gst_video_time_code_free (timecodestamper->ltc_internal_tc);
|
|
timecodestamper->ltc_internal_tc =
|
|
gst_video_time_code_copy (<c_tc->timecode);
|
|
timecodestamper->ltc_internal_running_time = ltc_tc->running_time;
|
|
updated_internal = TRUE;
|
|
GST_INFO_OBJECT (timecodestamper, "Resynced internal LTC counter");
|
|
}
|
|
|
|
/* And store it back for the next frame in case it has more or less
|
|
* the same running time */
|
|
g_queue_push_head (&timecodestamper->ltc_current_tcs,
|
|
g_steal_pointer (<c_tc));
|
|
break;
|
|
} else if (ltc_tc->running_time > running_time
|
|
&& ltc_tc->running_time - running_time > frame_duration / 2) {
|
|
/* Store it back for the next frame */
|
|
g_queue_push_head (&timecodestamper->ltc_current_tcs,
|
|
g_steal_pointer (<c_tc));
|
|
ltc_tc = NULL;
|
|
break;
|
|
}
|
|
|
|
/* otherwise it's in the past and we need to consider the next
|
|
* timecode. Read a new one */
|
|
gst_video_time_code_clear (<c_tc->timecode);
|
|
g_free (ltc_tc);
|
|
ltc_tc = NULL;
|
|
}
|
|
|
|
/* If we didn't update from LTC above, increment our internal timecode
|
|
* for this frame */
|
|
if (!updated_internal && timecodestamper->ltc_internal_tc) {
|
|
gst_video_time_code_increment_frame (timecodestamper->ltc_internal_tc);
|
|
}
|
|
|
|
if (timecodestamper->ltc_internal_tc) {
|
|
if (timecodestamper->ltc_auto_resync
|
|
&& timecodestamper->ltc_timeout != GST_CLOCK_TIME_NONE
|
|
&& (running_time + timecodestamper->ltc_timeout <
|
|
timecodestamper->ltc_internal_running_time
|
|
|| running_time >=
|
|
timecodestamper->ltc_internal_running_time +
|
|
timecodestamper->ltc_timeout)) {
|
|
if (timecodestamper->ltc_internal_tc)
|
|
gst_video_time_code_free (timecodestamper->ltc_internal_tc);
|
|
timecodestamper->ltc_internal_tc = NULL;
|
|
GST_DEBUG_OBJECT (timecodestamper, "LTC timecode timed out");
|
|
timecodestamper->ltc_internal_running_time = GST_CLOCK_TIME_NONE;
|
|
} else {
|
|
tc_str =
|
|
gst_video_time_code_to_string (timecodestamper->ltc_internal_tc);
|
|
GST_DEBUG_OBJECT (timecodestamper, "Updated LTC timecode to %s",
|
|
tc_str);
|
|
g_free (tc_str);
|
|
}
|
|
} else {
|
|
GST_DEBUG_OBJECT (timecodestamper, "Have no LTC timecode yet");
|
|
}
|
|
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
g_cond_signal (&timecodestamper->ltc_cond_audio);
|
|
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
}
|
|
#endif
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
switch (timecodestamper->tc_source) {
|
|
case GST_TIME_CODE_STAMPER_SOURCE_INTERNAL:
|
|
tc = timecodestamper->internal_tc;
|
|
break;
|
|
case GST_TIME_CODE_STAMPER_SOURCE_ZERO:
|
|
tc = gst_video_time_code_new (timecodestamper->fps_n,
|
|
timecodestamper->fps_d, NULL, tc_flags, 0, 0, 0, 0, 0);
|
|
free_tc = TRUE;
|
|
break;
|
|
case GST_TIME_CODE_STAMPER_SOURCE_LAST_KNOWN:
|
|
tc = timecodestamper->last_tc;
|
|
if (!tc)
|
|
tc = timecodestamper->internal_tc;
|
|
break;
|
|
case GST_TIME_CODE_STAMPER_SOURCE_LAST_KNOWN_OR_ZERO:
|
|
tc = timecodestamper->last_tc;
|
|
if (!tc) {
|
|
tc = gst_video_time_code_new (timecodestamper->fps_n,
|
|
timecodestamper->fps_d, NULL, tc_flags, 0, 0, 0, 0, 0);
|
|
free_tc = TRUE;
|
|
}
|
|
break;
|
|
case GST_TIME_CODE_STAMPER_SOURCE_LTC:
|
|
#if HAVE_LTC
|
|
if (timecodestamper->ltc_internal_tc)
|
|
tc = timecodestamper->ltc_internal_tc;
|
|
#endif
|
|
if (!tc) {
|
|
tc = gst_video_time_code_new (timecodestamper->fps_n,
|
|
timecodestamper->fps_d, NULL, tc_flags, 0, 0, 0, 0, 0);
|
|
free_tc = TRUE;
|
|
}
|
|
break;
|
|
case GST_TIME_CODE_STAMPER_SOURCE_RTC:
|
|
tc = timecodestamper->rtc_tc;
|
|
break;
|
|
}
|
|
|
|
switch (timecodestamper->tc_set) {
|
|
case GST_TIME_CODE_STAMPER_SET_NEVER:
|
|
break;
|
|
case GST_TIME_CODE_STAMPER_SET_KEEP:
|
|
if (!tc_meta && tc) {
|
|
gchar *tc_str;
|
|
|
|
if (timecodestamper->timecode_offset) {
|
|
if (!free_tc) {
|
|
tc = gst_video_time_code_copy (tc);
|
|
free_tc = TRUE;
|
|
}
|
|
gst_video_time_code_add_frames (tc, timecodestamper->timecode_offset);
|
|
}
|
|
|
|
tc_str = gst_video_time_code_to_string (tc);
|
|
GST_DEBUG_OBJECT (timecodestamper, "Storing timecode %s", tc_str);
|
|
g_free (tc_str);
|
|
|
|
gst_buffer_add_video_time_code_meta (buffer, tc);
|
|
}
|
|
break;
|
|
case GST_TIME_CODE_STAMPER_SET_ALWAYS:
|
|
gst_buffer_foreach_meta (buffer, remove_timecode_meta, NULL);
|
|
if (tc) {
|
|
gchar *tc_str;
|
|
|
|
if (timecodestamper->timecode_offset) {
|
|
if (!free_tc) {
|
|
tc = gst_video_time_code_copy (tc);
|
|
free_tc = TRUE;
|
|
}
|
|
gst_video_time_code_add_frames (tc, timecodestamper->timecode_offset);
|
|
}
|
|
|
|
tc_str = gst_video_time_code_to_string (tc);
|
|
GST_DEBUG_OBJECT (timecodestamper, "Storing timecode %s", tc_str);
|
|
g_free (tc_str);
|
|
|
|
gst_buffer_add_video_time_code_meta (buffer, tc);
|
|
}
|
|
break;
|
|
}
|
|
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
if (timecodestamper->post_messages && tc) {
|
|
GstClockTime stream_time, running_time, duration;
|
|
GstStructure *s;
|
|
GstMessage *msg;
|
|
|
|
running_time =
|
|
gst_segment_to_running_time (&vfilter->segment, GST_FORMAT_TIME,
|
|
GST_BUFFER_PTS (buffer));
|
|
stream_time =
|
|
gst_segment_to_stream_time (&vfilter->segment, GST_FORMAT_TIME,
|
|
GST_BUFFER_PTS (buffer));
|
|
duration =
|
|
gst_util_uint64_scale_int (GST_SECOND, timecodestamper->fps_d,
|
|
timecodestamper->fps_n);
|
|
s = gst_structure_new ("timecodestamper", "timestamp", G_TYPE_UINT64,
|
|
GST_BUFFER_PTS (buffer), "stream-time", G_TYPE_UINT64, stream_time,
|
|
"running-time", G_TYPE_UINT64, running_time, "duration",
|
|
G_TYPE_UINT64, duration, "timecode", GST_TYPE_VIDEO_TIME_CODE, tc,
|
|
NULL);
|
|
msg = gst_message_new_element (GST_OBJECT (timecodestamper), s);
|
|
gst_element_post_message (GST_ELEMENT (timecodestamper), msg);
|
|
}
|
|
#if HAVE_LTC
|
|
out:
|
|
#endif
|
|
|
|
if (dt_now)
|
|
g_date_time_unref (dt_now);
|
|
if (dt_frame)
|
|
g_date_time_unref (dt_frame);
|
|
if (free_tc && tc)
|
|
gst_video_time_code_free (tc);
|
|
|
|
return flow_ret;
|
|
}
|
|
|
|
static GstPad *
|
|
gst_timecodestamper_request_new_pad (GstElement * element,
|
|
GstPadTemplate * templ, const gchar * name_templ, const GstCaps * caps)
|
|
{
|
|
#if HAVE_LTC
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (element);
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
if (timecodestamper->ltcpad) {
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
return NULL;
|
|
}
|
|
|
|
if (GST_STATE (timecodestamper) > GST_STATE_READY ||
|
|
GST_STATE_TARGET (timecodestamper) > GST_STATE_READY) {
|
|
GST_ERROR_OBJECT (timecodestamper,
|
|
"LTC audio pad can only be requested in NULL or READY state");
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
return NULL;
|
|
}
|
|
|
|
timecodestamper->ltcpad = gst_pad_new_from_static_template
|
|
(&gst_timecodestamper_ltc_template, "ltc_sink");
|
|
|
|
gst_pad_set_chain_function (timecodestamper->ltcpad,
|
|
GST_DEBUG_FUNCPTR (gst_timecodestamper_ltcpad_chain));
|
|
gst_pad_set_event_function (timecodestamper->ltcpad,
|
|
GST_DEBUG_FUNCPTR (gst_timecodestamper_ltcpad_event));
|
|
gst_pad_set_query_function (timecodestamper->ltcpad,
|
|
GST_DEBUG_FUNCPTR (gst_timecodestamper_ltcpad_query));
|
|
gst_pad_set_activatemode_function (timecodestamper->ltcpad,
|
|
GST_DEBUG_FUNCPTR (gst_timecodestamper_ltcpad_activatemode));
|
|
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->audio_live = FALSE;
|
|
timecodestamper->audio_latency = GST_CLOCK_TIME_NONE;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
|
|
gst_element_add_pad (element, timecodestamper->ltcpad);
|
|
|
|
gst_element_post_message (GST_ELEMENT_CAST (timecodestamper),
|
|
gst_message_new_latency (GST_OBJECT_CAST (timecodestamper)));
|
|
|
|
return timecodestamper->ltcpad;
|
|
#else
|
|
return NULL;
|
|
#endif
|
|
}
|
|
|
|
static void
|
|
gst_timecodestamper_release_pad (GstElement * element, GstPad * pad)
|
|
{
|
|
#if HAVE_LTC
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (element);
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
if (timecodestamper->ltcpad != pad) {
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
return;
|
|
}
|
|
|
|
timecodestamper->ltcpad = NULL;
|
|
|
|
if (timecodestamper->ltc_internal_tc != NULL) {
|
|
gst_video_time_code_free (timecodestamper->ltc_internal_tc);
|
|
timecodestamper->ltc_internal_tc = NULL;
|
|
}
|
|
timecodestamper->ltc_internal_running_time = GST_CLOCK_TIME_NONE;
|
|
|
|
{
|
|
TimestampedTimecode *tc;
|
|
while ((tc = g_queue_pop_tail (&timecodestamper->ltc_current_tcs))) {
|
|
gst_video_time_code_clear (&tc->timecode);
|
|
g_free (tc);
|
|
}
|
|
}
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
gst_pad_set_active (pad, FALSE);
|
|
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->ltc_flushing = TRUE;
|
|
timecodestamper->ltc_eos = TRUE;
|
|
g_cond_signal (&timecodestamper->ltc_cond_video);
|
|
g_cond_signal (&timecodestamper->ltc_cond_audio);
|
|
|
|
gst_audio_info_init (&timecodestamper->ainfo);
|
|
gst_segment_init (&timecodestamper->ltc_segment, GST_FORMAT_UNDEFINED);
|
|
|
|
timecodestamper->ltc_first_running_time = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->ltc_current_running_time = GST_CLOCK_TIME_NONE;
|
|
|
|
if (timecodestamper->ltc_dec) {
|
|
ltc_decoder_free (timecodestamper->ltc_dec);
|
|
timecodestamper->ltc_dec = NULL;
|
|
}
|
|
|
|
if (timecodestamper->stream_align) {
|
|
gst_audio_stream_align_free (timecodestamper->stream_align);
|
|
timecodestamper->stream_align = NULL;
|
|
}
|
|
|
|
timecodestamper->ltc_total = 0;
|
|
|
|
timecodestamper->audio_live = FALSE;
|
|
timecodestamper->audio_latency = GST_CLOCK_TIME_NONE;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
|
|
gst_element_post_message (GST_ELEMENT_CAST (timecodestamper),
|
|
gst_message_new_latency (GST_OBJECT_CAST (timecodestamper)));
|
|
|
|
gst_element_remove_pad (element, pad);
|
|
#endif
|
|
}
|
|
|
|
#if HAVE_LTC
|
|
static GstFlowReturn
|
|
gst_timecodestamper_ltcpad_chain (GstPad * pad,
|
|
GstObject * parent, GstBuffer * buffer)
|
|
{
|
|
GstFlowReturn fr = GST_FLOW_OK;
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (parent);
|
|
GstMapInfo map;
|
|
GstClockTime timestamp, running_time, duration;
|
|
guint nsamples;
|
|
gboolean discont;
|
|
|
|
if (timecodestamper->audio_latency == -1 || gst_pad_check_reconfigure (pad)) {
|
|
gst_timecodestamper_update_latency (timecodestamper, pad,
|
|
&timecodestamper->audio_live, &timecodestamper->audio_latency);
|
|
}
|
|
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
if (timecodestamper->ltc_flushing) {
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
gst_buffer_unref (buffer);
|
|
return GST_FLOW_FLUSHING;
|
|
}
|
|
|
|
nsamples = gst_buffer_get_size (buffer) /
|
|
GST_AUDIO_INFO_BPF (&timecodestamper->ainfo);
|
|
|
|
if (!timecodestamper->stream_align) {
|
|
timecodestamper->stream_align =
|
|
gst_audio_stream_align_new (timecodestamper->ainfo.rate,
|
|
500 * GST_MSECOND, 20 * GST_MSECOND);
|
|
}
|
|
|
|
discont =
|
|
gst_audio_stream_align_process (timecodestamper->stream_align,
|
|
GST_BUFFER_IS_DISCONT (buffer), GST_BUFFER_PTS (buffer), nsamples,
|
|
×tamp, &duration, NULL);
|
|
|
|
if (discont) {
|
|
if (timecodestamper->ltc_dec) {
|
|
GST_WARNING_OBJECT (timecodestamper, "Got discont at %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (timestamp));
|
|
ltc_decoder_queue_flush (timecodestamper->ltc_dec);
|
|
}
|
|
timecodestamper->ltc_total = 0;
|
|
}
|
|
|
|
if (!timecodestamper->ltc_dec) {
|
|
gint samples_per_frame = 1920;
|
|
|
|
GST_OBJECT_LOCK (timecodestamper);
|
|
/* This is only for initialization and needs to be somewhat close to the
|
|
* real value. It will be tracked automatically afterwards */
|
|
if (timecodestamper->fps_n) {
|
|
samples_per_frame = timecodestamper->ainfo.rate *
|
|
timecodestamper->fps_d / timecodestamper->fps_n;
|
|
}
|
|
GST_OBJECT_UNLOCK (timecodestamper);
|
|
|
|
timecodestamper->ltc_dec =
|
|
ltc_decoder_create (samples_per_frame, DEFAULT_LTC_QUEUE);
|
|
timecodestamper->ltc_total = 0;
|
|
}
|
|
|
|
running_time = gst_segment_to_running_time (&timecodestamper->ltc_segment,
|
|
GST_FORMAT_TIME, timestamp);
|
|
|
|
GST_DEBUG_OBJECT (timecodestamper,
|
|
"Handling LTC audio buffer at %" GST_TIME_FORMAT " - %" GST_TIME_FORMAT
|
|
" (offset %" G_GUINT64_FORMAT ")",
|
|
GST_TIME_ARGS (running_time),
|
|
GST_TIME_ARGS (running_time + duration),
|
|
(guint64) timecodestamper->ltc_total);
|
|
|
|
if (timecodestamper->ltc_total == 0) {
|
|
timecodestamper->ltc_first_running_time = running_time;
|
|
}
|
|
|
|
gst_buffer_map (buffer, &map, GST_MAP_READ);
|
|
ltc_decoder_write (timecodestamper->ltc_dec, map.data, map.size,
|
|
timecodestamper->ltc_total);
|
|
timecodestamper->ltc_total += map.size;
|
|
gst_buffer_unmap (buffer, &map);
|
|
|
|
/* Now read all the timecodes from the decoder that are currently available
|
|
* and store them in our own queue, which gives us more control over how
|
|
* things are working. */
|
|
{
|
|
LTCFrameExt ltc_frame;
|
|
|
|
while (ltc_decoder_read (timecodestamper->ltc_dec, <c_frame) == 1) {
|
|
SMPTETimecode stc;
|
|
TimestampedTimecode *ltc_tc;
|
|
GstClockTime ltc_running_time;
|
|
|
|
if (ltc_frame.off_start < 0) {
|
|
GstClockTime offset =
|
|
gst_util_uint64_scale (GST_SECOND, -ltc_frame.off_start,
|
|
timecodestamper->ainfo.rate);
|
|
|
|
if (offset > timecodestamper->ltc_first_running_time)
|
|
ltc_running_time = 0;
|
|
else
|
|
ltc_running_time = timecodestamper->ltc_first_running_time - offset;
|
|
} else {
|
|
ltc_running_time = timecodestamper->ltc_first_running_time +
|
|
gst_util_uint64_scale (GST_SECOND, ltc_frame.off_start,
|
|
timecodestamper->ainfo.rate);
|
|
}
|
|
|
|
ltc_frame_to_time (&stc, <c_frame.ltc, 0);
|
|
GST_INFO_OBJECT (timecodestamper,
|
|
"Got LTC timecode %02d:%02d:%02d:%02d at %" GST_TIME_FORMAT,
|
|
stc.hours, stc.mins, stc.secs, stc.frame,
|
|
GST_TIME_ARGS (ltc_running_time));
|
|
|
|
ltc_tc = g_new0 (TimestampedTimecode, 1);
|
|
ltc_tc->running_time = ltc_running_time;
|
|
/* We fill in the framerate and other metadata later */
|
|
gst_video_time_code_init (<c_tc->timecode,
|
|
0, 0, timecodestamper->ltc_daily_jam, 0,
|
|
stc.hours, stc.mins, stc.secs, stc.frame, 0);
|
|
|
|
/* If we have a discontinuity it might happen that we're getting
|
|
* timecodes that are in the past relative to timecodes we already have
|
|
* in our queue. We have to get rid of all the timecodes that are in the
|
|
* future now. */
|
|
if (discont) {
|
|
TimestampedTimecode *tmp;
|
|
|
|
while ((tmp = g_queue_peek_tail (&timecodestamper->ltc_current_tcs)) &&
|
|
tmp->running_time >= ltc_running_time) {
|
|
gst_video_time_code_clear (&tmp->timecode);
|
|
g_free (tmp);
|
|
g_queue_pop_tail (&timecodestamper->ltc_current_tcs);
|
|
}
|
|
|
|
g_queue_push_tail (&timecodestamper->ltc_current_tcs,
|
|
g_steal_pointer (<c_tc));
|
|
} else {
|
|
g_queue_push_tail (&timecodestamper->ltc_current_tcs,
|
|
g_steal_pointer (<c_tc));
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Notify the video streaming thread that new data is available */
|
|
g_cond_signal (&timecodestamper->ltc_cond_video);
|
|
|
|
/* Wait until video has caught up, if needed */
|
|
if (timecodestamper->audio_live) {
|
|
/* In live-mode, do no waiting as we're guaranteed to be more or less in
|
|
* sync (~latency) with the video */
|
|
} else {
|
|
/* If we're ahead of the video, wait until the video has caught up.
|
|
* Otherwise don't wait and drop any too old items from the ringbuffer */
|
|
while ((timecodestamper->video_current_running_time == GST_CLOCK_TIME_NONE
|
|
|| running_time + duration >=
|
|
timecodestamper->video_current_running_time)
|
|
&& timecodestamper->ltc_dec
|
|
&& g_queue_get_length (&timecodestamper->ltc_current_tcs) >
|
|
DEFAULT_LTC_QUEUE / 2 && !timecodestamper->video_eos
|
|
&& !timecodestamper->ltc_flushing) {
|
|
GST_TRACE_OBJECT (timecodestamper,
|
|
"Waiting for video to advance, EOS or flushing");
|
|
g_cond_wait (&timecodestamper->ltc_cond_audio, &timecodestamper->mutex);
|
|
}
|
|
}
|
|
|
|
if (timecodestamper->ltc_flushing)
|
|
fr = GST_FLOW_FLUSHING;
|
|
else
|
|
fr = GST_FLOW_OK;
|
|
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
|
|
gst_buffer_unref (buffer);
|
|
return fr;
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_ltcpad_event (GstPad * pad,
|
|
GstObject * parent, GstEvent * event)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (parent);
|
|
|
|
GstCaps *caps;
|
|
gboolean ret = TRUE;
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_CAPS:
|
|
gst_event_parse_caps (event, &caps);
|
|
|
|
if (!gst_audio_info_from_caps (&timecodestamper->ainfo, caps)) {
|
|
gst_event_unref (event);
|
|
return FALSE;
|
|
}
|
|
|
|
if (timecodestamper->stream_align) {
|
|
gst_audio_stream_align_set_rate (timecodestamper->stream_align,
|
|
timecodestamper->ainfo.rate);
|
|
}
|
|
|
|
break;
|
|
case GST_EVENT_SEGMENT:
|
|
gst_event_copy_segment (event, &timecodestamper->ltc_segment);
|
|
break;
|
|
|
|
case GST_EVENT_FLUSH_START:
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->ltc_flushing = TRUE;
|
|
g_cond_signal (&timecodestamper->ltc_cond_audio);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
break;
|
|
case GST_EVENT_FLUSH_STOP:
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->ltc_flushing = FALSE;
|
|
timecodestamper->ltc_eos = FALSE;
|
|
gst_segment_init (&timecodestamper->ltc_segment, GST_FORMAT_UNDEFINED);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
break;
|
|
case GST_EVENT_EOS:
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->ltc_eos = TRUE;
|
|
g_cond_signal (&timecodestamper->ltc_cond_video);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
gst_event_unref (event);
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_ltcpad_query (GstPad * pad,
|
|
GstObject * parent, GstQuery * query)
|
|
{
|
|
GstCaps *caps, *filter, *tcaps;
|
|
|
|
switch (GST_QUERY_TYPE (query)) {
|
|
case GST_QUERY_CAPS:
|
|
gst_query_parse_caps (query, &filter);
|
|
tcaps = gst_pad_get_pad_template_caps (pad);
|
|
if (filter)
|
|
caps = gst_caps_intersect_full (tcaps, filter,
|
|
GST_CAPS_INTERSECT_FIRST);
|
|
else
|
|
caps = gst_caps_ref (tcaps);
|
|
gst_query_set_caps_result (query, caps);
|
|
gst_caps_unref (tcaps);
|
|
gst_caps_unref (caps);
|
|
return TRUE;
|
|
default:
|
|
return gst_pad_query_default (pad, parent, query);
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_ltcpad_activatemode (GstPad * pad,
|
|
GstObject * parent, GstPadMode mode, gboolean active)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (parent);
|
|
|
|
if (active) {
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->ltc_flushing = FALSE;
|
|
timecodestamper->ltc_eos = FALSE;
|
|
timecodestamper->audio_live = FALSE;
|
|
timecodestamper->audio_latency = GST_CLOCK_TIME_NONE;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
} else {
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->ltc_flushing = TRUE;
|
|
timecodestamper->ltc_eos = TRUE;
|
|
g_cond_signal (&timecodestamper->ltc_cond_audio);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_timecodestamper_videopad_activatemode (GstPad * pad,
|
|
GstObject * parent, GstPadMode mode, gboolean active)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (parent);
|
|
|
|
if (active) {
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_flushing = FALSE;
|
|
timecodestamper->video_eos = FALSE;
|
|
timecodestamper->video_live = FALSE;
|
|
timecodestamper->video_latency = GST_CLOCK_TIME_NONE;
|
|
timecodestamper->video_current_running_time = GST_CLOCK_TIME_NONE;
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
} else {
|
|
g_mutex_lock (&timecodestamper->mutex);
|
|
timecodestamper->video_flushing = TRUE;
|
|
timecodestamper->video_current_running_time = GST_CLOCK_TIME_NONE;
|
|
if (timecodestamper->video_clock_id)
|
|
gst_clock_id_unschedule (timecodestamper->video_clock_id);
|
|
g_cond_signal (&timecodestamper->ltc_cond_video);
|
|
g_mutex_unlock (&timecodestamper->mutex);
|
|
}
|
|
|
|
return timecodestamper->video_activatemode_default (pad, parent, mode,
|
|
active);
|
|
}
|
|
|
|
static GstIterator *
|
|
gst_timecodestamper_src_iterate_internal_link (GstPad * pad, GstObject * parent)
|
|
{
|
|
GstTimeCodeStamper *timecodestamper = GST_TIME_CODE_STAMPER (parent);
|
|
GValue value = G_VALUE_INIT;
|
|
GstIterator *it;
|
|
|
|
g_value_init (&value, GST_TYPE_PAD);
|
|
g_value_set_object (&value, GST_BASE_TRANSFORM_SINK_PAD (timecodestamper));
|
|
it = gst_iterator_new_single (GST_TYPE_PAD, &value);
|
|
g_value_unset (&value);
|
|
|
|
return it;
|
|
}
|
|
#endif
|