mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2025-04-26 06:54:49 +00:00
decklinksrc: Keep track of a window of capture/stream time mappings and calculate relationship
Based on this we calculate the actual capture time, which should get us rid of any capturing jitter by averaging it out. Also add a output-stream-time property which forces the elements to output the stream time directly instead of doing any conversion to the pipeline clock. Use with care. https://bugzilla.gnome.org/show_bug.cgi?id=774850
This commit is contained in:
parent
881a08671e
commit
09dd64bb4f
3 changed files with 219 additions and 15 deletions
|
@ -56,7 +56,7 @@ static GstStaticPadTemplate sink_template = GST_STATIC_PAD_TEMPLATE ("src",
|
||||||
typedef struct
|
typedef struct
|
||||||
{
|
{
|
||||||
IDeckLinkAudioInputPacket *packet;
|
IDeckLinkAudioInputPacket *packet;
|
||||||
GstClockTime capture_time;
|
GstClockTime timestamp;
|
||||||
} CapturePacket;
|
} CapturePacket;
|
||||||
|
|
||||||
static void
|
static void
|
||||||
|
@ -424,11 +424,32 @@ gst_decklink_audio_src_got_packet (GstElement * element,
|
||||||
GstClockTime packet_time)
|
GstClockTime packet_time)
|
||||||
{
|
{
|
||||||
GstDecklinkAudioSrc *self = GST_DECKLINK_AUDIO_SRC_CAST (element);
|
GstDecklinkAudioSrc *self = GST_DECKLINK_AUDIO_SRC_CAST (element);
|
||||||
|
GstClockTime timestamp;
|
||||||
|
|
||||||
GST_LOG_OBJECT (self,
|
GST_LOG_OBJECT (self,
|
||||||
"Got audio packet at %" GST_TIME_FORMAT " / %" GST_TIME_FORMAT,
|
"Got audio packet at %" GST_TIME_FORMAT " / %" GST_TIME_FORMAT,
|
||||||
GST_TIME_ARGS (capture_time), GST_TIME_ARGS (packet_time));
|
GST_TIME_ARGS (capture_time), GST_TIME_ARGS (packet_time));
|
||||||
|
|
||||||
|
g_mutex_lock (&self->input->lock);
|
||||||
|
if (self->input->videosrc) {
|
||||||
|
GstDecklinkVideoSrc *videosrc =
|
||||||
|
GST_DECKLINK_VIDEO_SRC_CAST (gst_object_ref (self->input->videosrc));
|
||||||
|
|
||||||
|
if (videosrc->output_stream_time)
|
||||||
|
timestamp = packet_time;
|
||||||
|
else
|
||||||
|
timestamp = gst_clock_adjust_with_calibration (NULL, packet_time,
|
||||||
|
videosrc->current_time_mapping.xbase,
|
||||||
|
videosrc->current_time_mapping.b, videosrc->current_time_mapping.num,
|
||||||
|
videosrc->current_time_mapping.den);
|
||||||
|
} else {
|
||||||
|
timestamp = capture_time;
|
||||||
|
}
|
||||||
|
g_mutex_unlock (&self->input->lock);
|
||||||
|
|
||||||
|
GST_LOG_OBJECT (self, "Converted times to %" GST_TIME_FORMAT,
|
||||||
|
GST_TIME_ARGS (timestamp));
|
||||||
|
|
||||||
g_mutex_lock (&self->lock);
|
g_mutex_lock (&self->lock);
|
||||||
if (!self->flushing) {
|
if (!self->flushing) {
|
||||||
CapturePacket *p;
|
CapturePacket *p;
|
||||||
|
@ -436,14 +457,13 @@ gst_decklink_audio_src_got_packet (GstElement * element,
|
||||||
while (g_queue_get_length (&self->current_packets) >= self->buffer_size) {
|
while (g_queue_get_length (&self->current_packets) >= self->buffer_size) {
|
||||||
p = (CapturePacket *) g_queue_pop_head (&self->current_packets);
|
p = (CapturePacket *) g_queue_pop_head (&self->current_packets);
|
||||||
GST_WARNING_OBJECT (self, "Dropping old packet at %" GST_TIME_FORMAT,
|
GST_WARNING_OBJECT (self, "Dropping old packet at %" GST_TIME_FORMAT,
|
||||||
GST_TIME_ARGS (p->capture_time));
|
GST_TIME_ARGS (p->timestamp));
|
||||||
capture_packet_free (p);
|
capture_packet_free (p);
|
||||||
}
|
}
|
||||||
|
|
||||||
p = (CapturePacket *) g_malloc0 (sizeof (CapturePacket));
|
p = (CapturePacket *) g_malloc0 (sizeof (CapturePacket));
|
||||||
p->packet = packet;
|
p->packet = packet;
|
||||||
p->capture_time =
|
p->timestamp = timestamp;
|
||||||
capture_time != GST_CLOCK_TIME_NONE ? capture_time : packet_time;
|
|
||||||
packet->AddRef ();
|
packet->AddRef ();
|
||||||
g_queue_push_tail (&self->current_packets, p);
|
g_queue_push_tail (&self->current_packets, p);
|
||||||
g_cond_signal (&self->cond);
|
g_cond_signal (&self->cond);
|
||||||
|
@ -486,8 +506,7 @@ retry:
|
||||||
sample_count = p->packet->GetSampleFrameCount ();
|
sample_count = p->packet->GetSampleFrameCount ();
|
||||||
data_size = self->info.bpf * sample_count;
|
data_size = self->info.bpf * sample_count;
|
||||||
|
|
||||||
if (p->capture_time == GST_CLOCK_TIME_NONE
|
if (p->timestamp == GST_CLOCK_TIME_NONE && self->next_offset == (guint64) - 1) {
|
||||||
&& self->next_offset == (guint64) - 1) {
|
|
||||||
GST_DEBUG_OBJECT (self,
|
GST_DEBUG_OBJECT (self,
|
||||||
"Got packet without timestamp before initial "
|
"Got packet without timestamp before initial "
|
||||||
"timestamp after discont - dropping");
|
"timestamp after discont - dropping");
|
||||||
|
@ -507,7 +526,7 @@ retry:
|
||||||
ap->input = self->input->input;
|
ap->input = self->input->input;
|
||||||
ap->input->AddRef ();
|
ap->input->AddRef ();
|
||||||
|
|
||||||
timestamp = p->capture_time;
|
timestamp = p->timestamp;
|
||||||
|
|
||||||
// Jitter and discontinuity handling, based on audiobasesrc
|
// Jitter and discontinuity handling, based on audiobasesrc
|
||||||
start_time = timestamp;
|
start_time = timestamp;
|
||||||
|
|
|
@ -32,6 +32,7 @@ GST_DEBUG_CATEGORY_STATIC (gst_decklink_video_src_debug);
|
||||||
#define DEFAULT_MODE (GST_DECKLINK_MODE_AUTO)
|
#define DEFAULT_MODE (GST_DECKLINK_MODE_AUTO)
|
||||||
#define DEFAULT_CONNECTION (GST_DECKLINK_CONNECTION_AUTO)
|
#define DEFAULT_CONNECTION (GST_DECKLINK_CONNECTION_AUTO)
|
||||||
#define DEFAULT_BUFFER_SIZE (5)
|
#define DEFAULT_BUFFER_SIZE (5)
|
||||||
|
#define DEFAULT_OUTPUT_STREAM_TIME (FALSE)
|
||||||
|
|
||||||
enum
|
enum
|
||||||
{
|
{
|
||||||
|
@ -41,13 +42,14 @@ enum
|
||||||
PROP_DEVICE_NUMBER,
|
PROP_DEVICE_NUMBER,
|
||||||
PROP_BUFFER_SIZE,
|
PROP_BUFFER_SIZE,
|
||||||
PROP_VIDEO_FORMAT,
|
PROP_VIDEO_FORMAT,
|
||||||
PROP_TIMECODE_FORMAT
|
PROP_TIMECODE_FORMAT,
|
||||||
|
PROP_OUTPUT_STREAM_TIME
|
||||||
};
|
};
|
||||||
|
|
||||||
typedef struct
|
typedef struct
|
||||||
{
|
{
|
||||||
IDeckLinkVideoInputFrame *frame;
|
IDeckLinkVideoInputFrame *frame;
|
||||||
GstClockTime capture_time, capture_duration;
|
GstClockTime timestamp, duration;
|
||||||
GstDecklinkModeEnum mode;
|
GstDecklinkModeEnum mode;
|
||||||
BMDPixelFormat format;
|
BMDPixelFormat format;
|
||||||
GstVideoTimeCode *tc;
|
GstVideoTimeCode *tc;
|
||||||
|
@ -178,6 +180,12 @@ gst_decklink_video_src_class_init (GstDecklinkVideoSrcClass * klass)
|
||||||
(GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
|
(GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS |
|
||||||
G_PARAM_CONSTRUCT)));
|
G_PARAM_CONSTRUCT)));
|
||||||
|
|
||||||
|
g_object_class_install_property (gobject_class, PROP_OUTPUT_STREAM_TIME,
|
||||||
|
g_param_spec_boolean ("output-stream-time", "Output Stream Time",
|
||||||
|
"Output stream time directly instead of translating to pipeline clock",
|
||||||
|
DEFAULT_OUTPUT_STREAM_TIME,
|
||||||
|
(GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)));
|
||||||
|
|
||||||
templ_caps = gst_decklink_mode_get_template_caps ();
|
templ_caps = gst_decklink_mode_get_template_caps ();
|
||||||
gst_element_class_add_pad_template (element_class,
|
gst_element_class_add_pad_template (element_class,
|
||||||
gst_pad_template_new ("src", GST_PAD_SRC, GST_PAD_ALWAYS, templ_caps));
|
gst_pad_template_new ("src", GST_PAD_SRC, GST_PAD_ALWAYS, templ_caps));
|
||||||
|
@ -203,6 +211,14 @@ gst_decklink_video_src_init (GstDecklinkVideoSrc * self)
|
||||||
self->video_format = GST_DECKLINK_VIDEO_FORMAT_AUTO;
|
self->video_format = GST_DECKLINK_VIDEO_FORMAT_AUTO;
|
||||||
self->timecode_format = bmdTimecodeRP188Any;
|
self->timecode_format = bmdTimecodeRP188Any;
|
||||||
self->no_signal = FALSE;
|
self->no_signal = FALSE;
|
||||||
|
self->output_stream_time = DEFAULT_OUTPUT_STREAM_TIME;
|
||||||
|
|
||||||
|
self->window_size = 64;
|
||||||
|
self->times = g_new (GstClockTime, 4 * self->window_size);
|
||||||
|
self->times_temp = self->times + 2 * self->window_size;
|
||||||
|
self->window_fill = 0;
|
||||||
|
self->window_skip = 1;
|
||||||
|
self->window_skip_count = 0;
|
||||||
|
|
||||||
gst_base_src_set_live (GST_BASE_SRC (self), TRUE);
|
gst_base_src_set_live (GST_BASE_SRC (self), TRUE);
|
||||||
gst_base_src_set_format (GST_BASE_SRC (self), GST_FORMAT_TIME);
|
gst_base_src_set_format (GST_BASE_SRC (self), GST_FORMAT_TIME);
|
||||||
|
@ -261,6 +277,9 @@ gst_decklink_video_src_set_property (GObject * object, guint property_id,
|
||||||
gst_decklink_timecode_format_from_enum ((GstDecklinkTimecodeFormat)
|
gst_decklink_timecode_format_from_enum ((GstDecklinkTimecodeFormat)
|
||||||
g_value_get_enum (value));
|
g_value_get_enum (value));
|
||||||
break;
|
break;
|
||||||
|
case PROP_OUTPUT_STREAM_TIME:
|
||||||
|
self->output_stream_time = g_value_get_boolean (value);
|
||||||
|
break;
|
||||||
default:
|
default:
|
||||||
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec);
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec);
|
||||||
break;
|
break;
|
||||||
|
@ -293,6 +312,9 @@ gst_decklink_video_src_get_property (GObject * object, guint property_id,
|
||||||
g_value_set_enum (value,
|
g_value_set_enum (value,
|
||||||
gst_decklink_timecode_format_to_enum (self->timecode_format));
|
gst_decklink_timecode_format_to_enum (self->timecode_format));
|
||||||
break;
|
break;
|
||||||
|
case PROP_OUTPUT_STREAM_TIME:
|
||||||
|
g_value_set_boolean (value, self->output_stream_time);
|
||||||
|
break;
|
||||||
default:
|
default:
|
||||||
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec);
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec);
|
||||||
break;
|
break;
|
||||||
|
@ -304,6 +326,8 @@ gst_decklink_video_src_finalize (GObject * object)
|
||||||
{
|
{
|
||||||
GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object);
|
GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object);
|
||||||
|
|
||||||
|
g_free (self->times);
|
||||||
|
self->times = NULL;
|
||||||
g_mutex_clear (&self->lock);
|
g_mutex_clear (&self->lock);
|
||||||
g_cond_clear (&self->cond);
|
g_cond_clear (&self->cond);
|
||||||
|
|
||||||
|
@ -431,6 +455,117 @@ gst_decklink_video_src_get_caps (GstBaseSrc * bsrc, GstCaps * filter)
|
||||||
return caps;
|
return caps;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
gst_decklink_video_src_update_time_mapping (GstDecklinkVideoSrc * self,
|
||||||
|
GstClockTime capture_time, GstClockTime stream_time)
|
||||||
|
{
|
||||||
|
if (self->window_skip_count == 0) {
|
||||||
|
GstClockTime num, den, b, xbase;
|
||||||
|
gdouble r_squared;
|
||||||
|
|
||||||
|
self->times[2 * self->window_fill] = stream_time;
|
||||||
|
self->times[2 * self->window_fill + 1] = capture_time;
|
||||||
|
|
||||||
|
self->window_fill++;
|
||||||
|
self->window_skip_count++;
|
||||||
|
if (self->window_skip_count >= self->window_skip)
|
||||||
|
self->window_skip_count = 0;
|
||||||
|
|
||||||
|
if (self->window_fill >= self->window_size) {
|
||||||
|
guint fps =
|
||||||
|
((gdouble) self->info.fps_n + self->info.fps_d -
|
||||||
|
1) / ((gdouble) self->info.fps_d);
|
||||||
|
|
||||||
|
/* Start by updating first every frame, once full every second frame,
|
||||||
|
* etc. until we update once every 4 seconds */
|
||||||
|
if (self->window_skip < 4 * fps)
|
||||||
|
self->window_skip *= 2;
|
||||||
|
if (self->window_skip >= 4 * fps)
|
||||||
|
self->window_skip = 4 * fps;
|
||||||
|
|
||||||
|
self->window_fill = 0;
|
||||||
|
self->window_filled = TRUE;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* First sample ever, create some basic mapping to start */
|
||||||
|
if (!self->window_filled && self->window_fill == 1) {
|
||||||
|
self->current_time_mapping.xbase = stream_time;
|
||||||
|
self->current_time_mapping.b = capture_time;
|
||||||
|
self->current_time_mapping.num = 1;
|
||||||
|
self->current_time_mapping.den = 1;
|
||||||
|
self->next_time_mapping_pending = FALSE;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Only bother calculating anything here once we had enough measurements,
|
||||||
|
* i.e. let's take the window size as a start */
|
||||||
|
if (self->window_filled &&
|
||||||
|
gst_calculate_linear_regression (self->times, self->times_temp,
|
||||||
|
self->window_size, &num, &den, &b, &xbase, &r_squared)) {
|
||||||
|
|
||||||
|
GST_DEBUG_OBJECT (self,
|
||||||
|
"Calculated new time mapping: pipeline time = %lf * (stream time - %"
|
||||||
|
G_GUINT64_FORMAT ") + %" G_GUINT64_FORMAT " (%lf)",
|
||||||
|
((gdouble) num) / ((gdouble) den), xbase, b, r_squared);
|
||||||
|
|
||||||
|
self->next_time_mapping.xbase = xbase;
|
||||||
|
self->next_time_mapping.b = b;
|
||||||
|
self->next_time_mapping.num = num;
|
||||||
|
self->next_time_mapping.den = den;
|
||||||
|
self->next_time_mapping_pending = TRUE;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
self->window_skip_count++;
|
||||||
|
if (self->window_skip_count >= self->window_skip)
|
||||||
|
self->window_skip_count = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (self->next_time_mapping_pending) {
|
||||||
|
GstClockTime expected, new_calculated, diff, max_diff;
|
||||||
|
|
||||||
|
expected =
|
||||||
|
gst_clock_adjust_with_calibration (NULL, stream_time,
|
||||||
|
self->current_time_mapping.xbase, self->current_time_mapping.b,
|
||||||
|
self->current_time_mapping.num, self->current_time_mapping.den);
|
||||||
|
new_calculated =
|
||||||
|
gst_clock_adjust_with_calibration (NULL, stream_time,
|
||||||
|
self->next_time_mapping.xbase, self->next_time_mapping.b,
|
||||||
|
self->next_time_mapping.num, self->next_time_mapping.den);
|
||||||
|
|
||||||
|
if (new_calculated > expected)
|
||||||
|
diff = new_calculated - expected;
|
||||||
|
else
|
||||||
|
diff = expected - new_calculated;
|
||||||
|
|
||||||
|
/* At most 5% frame duration change per update */
|
||||||
|
max_diff =
|
||||||
|
gst_util_uint64_scale (GST_SECOND / 20, self->info.fps_d,
|
||||||
|
self->info.fps_n);
|
||||||
|
|
||||||
|
GST_DEBUG_OBJECT (self,
|
||||||
|
"New time mapping causes difference of %" GST_TIME_FORMAT,
|
||||||
|
GST_TIME_ARGS (diff));
|
||||||
|
GST_DEBUG_OBJECT (self, "Maximum allowed per frame %" GST_TIME_FORMAT,
|
||||||
|
GST_TIME_ARGS (max_diff));
|
||||||
|
|
||||||
|
if (diff > max_diff) {
|
||||||
|
/* adjust so that we move that much closer */
|
||||||
|
if (new_calculated > expected) {
|
||||||
|
self->current_time_mapping.b = expected + max_diff;
|
||||||
|
self->current_time_mapping.xbase = stream_time;
|
||||||
|
} else {
|
||||||
|
self->current_time_mapping.b = expected - max_diff;
|
||||||
|
self->current_time_mapping.xbase = stream_time;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
self->current_time_mapping.xbase = self->next_time_mapping.xbase;
|
||||||
|
self->current_time_mapping.b = self->next_time_mapping.b;
|
||||||
|
self->current_time_mapping.num = self->next_time_mapping.num;
|
||||||
|
self->current_time_mapping.den = self->next_time_mapping.den;
|
||||||
|
self->next_time_mapping_pending = FALSE;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
gst_decklink_video_src_got_frame (GstElement * element,
|
gst_decklink_video_src_got_frame (GstElement * element,
|
||||||
IDeckLinkVideoInputFrame * frame, GstDecklinkModeEnum mode,
|
IDeckLinkVideoInputFrame * frame, GstDecklinkModeEnum mode,
|
||||||
|
@ -439,6 +574,7 @@ gst_decklink_video_src_got_frame (GstElement * element,
|
||||||
guint frames, BMDTimecodeFlags bflags)
|
guint frames, BMDTimecodeFlags bflags)
|
||||||
{
|
{
|
||||||
GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element);
|
GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element);
|
||||||
|
GstClockTime timestamp, duration;
|
||||||
|
|
||||||
GST_LOG_OBJECT (self,
|
GST_LOG_OBJECT (self,
|
||||||
"Got video frame at %" GST_TIME_FORMAT " / %" GST_TIME_FORMAT " (%"
|
"Got video frame at %" GST_TIME_FORMAT " / %" GST_TIME_FORMAT " (%"
|
||||||
|
@ -446,6 +582,25 @@ gst_decklink_video_src_got_frame (GstElement * element,
|
||||||
GST_TIME_ARGS (stream_time), GST_TIME_ARGS (stream_duration));
|
GST_TIME_ARGS (stream_time), GST_TIME_ARGS (stream_duration));
|
||||||
|
|
||||||
g_mutex_lock (&self->lock);
|
g_mutex_lock (&self->lock);
|
||||||
|
|
||||||
|
gst_decklink_video_src_update_time_mapping (self, capture_time, stream_time);
|
||||||
|
if (self->output_stream_time) {
|
||||||
|
timestamp = stream_time;
|
||||||
|
duration = stream_duration;
|
||||||
|
} else {
|
||||||
|
timestamp =
|
||||||
|
gst_clock_adjust_with_calibration (NULL, stream_time,
|
||||||
|
self->current_time_mapping.xbase, self->current_time_mapping.b,
|
||||||
|
self->current_time_mapping.num, self->current_time_mapping.den);
|
||||||
|
duration =
|
||||||
|
gst_util_uint64_scale (stream_duration, self->current_time_mapping.num,
|
||||||
|
self->current_time_mapping.den);
|
||||||
|
}
|
||||||
|
|
||||||
|
GST_LOG_OBJECT (self,
|
||||||
|
"Converted times to %" GST_TIME_FORMAT " (%"
|
||||||
|
GST_TIME_FORMAT ")", GST_TIME_ARGS (timestamp), GST_TIME_ARGS (duration));
|
||||||
|
|
||||||
if (!self->flushing) {
|
if (!self->flushing) {
|
||||||
CaptureFrame *f;
|
CaptureFrame *f;
|
||||||
const GstDecklinkMode *bmode;
|
const GstDecklinkMode *bmode;
|
||||||
|
@ -455,15 +610,14 @@ gst_decklink_video_src_got_frame (GstElement * element,
|
||||||
while (g_queue_get_length (&self->current_frames) >= self->buffer_size) {
|
while (g_queue_get_length (&self->current_frames) >= self->buffer_size) {
|
||||||
f = (CaptureFrame *) g_queue_pop_head (&self->current_frames);
|
f = (CaptureFrame *) g_queue_pop_head (&self->current_frames);
|
||||||
GST_WARNING_OBJECT (self, "Dropping old frame at %" GST_TIME_FORMAT,
|
GST_WARNING_OBJECT (self, "Dropping old frame at %" GST_TIME_FORMAT,
|
||||||
GST_TIME_ARGS (f->capture_time));
|
GST_TIME_ARGS (f->timestamp));
|
||||||
capture_frame_free (f);
|
capture_frame_free (f);
|
||||||
}
|
}
|
||||||
|
|
||||||
f = (CaptureFrame *) g_malloc0 (sizeof (CaptureFrame));
|
f = (CaptureFrame *) g_malloc0 (sizeof (CaptureFrame));
|
||||||
f->frame = frame;
|
f->frame = frame;
|
||||||
f->capture_time =
|
f->timestamp = timestamp;
|
||||||
capture_time != GST_CLOCK_TIME_NONE ? capture_time : stream_time;
|
f->duration = duration;
|
||||||
f->capture_duration = stream_duration;
|
|
||||||
f->mode = mode;
|
f->mode = mode;
|
||||||
f->format = frame->GetPixelFormat ();
|
f->format = frame->GetPixelFormat ();
|
||||||
bmode = gst_decklink_get_mode (mode);
|
bmode = gst_decklink_get_mode (mode);
|
||||||
|
@ -594,8 +748,8 @@ gst_decklink_video_src_create (GstPushSrc * bsrc, GstBuffer ** buffer)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
GST_BUFFER_TIMESTAMP (*buffer) = f->capture_time;
|
GST_BUFFER_TIMESTAMP (*buffer) = f->timestamp;
|
||||||
GST_BUFFER_DURATION (*buffer) = f->capture_duration;
|
GST_BUFFER_DURATION (*buffer) = f->duration;
|
||||||
gst_buffer_add_video_time_code_meta (*buffer, f->tc);
|
gst_buffer_add_video_time_code_meta (*buffer, f->tc);
|
||||||
|
|
||||||
GST_DEBUG_OBJECT (self,
|
GST_DEBUG_OBJECT (self,
|
||||||
|
@ -750,6 +904,20 @@ gst_decklink_video_src_start_streams (GstElement * element)
|
||||||
|| GST_STATE_PENDING (self) == GST_STATE_PLAYING)) {
|
|| GST_STATE_PENDING (self) == GST_STATE_PLAYING)) {
|
||||||
GST_DEBUG_OBJECT (self, "Starting streams");
|
GST_DEBUG_OBJECT (self, "Starting streams");
|
||||||
|
|
||||||
|
g_mutex_lock (&self->lock);
|
||||||
|
self->window_fill = 0;
|
||||||
|
self->window_filled = FALSE;
|
||||||
|
self->window_skip = 1;
|
||||||
|
self->window_skip_count = 0;
|
||||||
|
self->current_time_mapping.xbase = 0;
|
||||||
|
self->current_time_mapping.b = 0;
|
||||||
|
self->current_time_mapping.num = 1;
|
||||||
|
self->current_time_mapping.den = 1;
|
||||||
|
self->next_time_mapping.xbase = 0;
|
||||||
|
self->next_time_mapping.b = 0;
|
||||||
|
self->next_time_mapping.num = 1;
|
||||||
|
self->next_time_mapping.den = 1;
|
||||||
|
g_mutex_unlock (&self->lock);
|
||||||
res = self->input->input->StartStreams ();
|
res = self->input->input->StartStreams ();
|
||||||
if (res != S_OK) {
|
if (res != S_OK) {
|
||||||
GST_ELEMENT_ERROR (self, STREAM, FAILED,
|
GST_ELEMENT_ERROR (self, STREAM, FAILED,
|
||||||
|
|
|
@ -55,6 +55,7 @@ struct _GstDecklinkVideoSrc
|
||||||
BMDPixelFormat caps_format;
|
BMDPixelFormat caps_format;
|
||||||
GstDecklinkConnectionEnum connection;
|
GstDecklinkConnectionEnum connection;
|
||||||
gint device_number;
|
gint device_number;
|
||||||
|
gboolean output_stream_time;
|
||||||
|
|
||||||
GstVideoInfo info;
|
GstVideoInfo info;
|
||||||
GstDecklinkVideoFormat video_format;
|
GstDecklinkVideoFormat video_format;
|
||||||
|
@ -69,6 +70,22 @@ struct _GstDecklinkVideoSrc
|
||||||
gboolean no_signal;
|
gboolean no_signal;
|
||||||
|
|
||||||
guint buffer_size;
|
guint buffer_size;
|
||||||
|
|
||||||
|
/* Protected by lock */
|
||||||
|
GstClockTime *times;
|
||||||
|
GstClockTime *times_temp;
|
||||||
|
guint window_size, window_fill;
|
||||||
|
gboolean window_filled;
|
||||||
|
guint window_skip, window_skip_count;
|
||||||
|
struct {
|
||||||
|
GstClockTime xbase, b;
|
||||||
|
GstClockTime num, den;
|
||||||
|
} current_time_mapping;
|
||||||
|
struct {
|
||||||
|
GstClockTime xbase, b;
|
||||||
|
GstClockTime num, den;
|
||||||
|
} next_time_mapping;
|
||||||
|
gboolean next_time_mapping_pending;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct _GstDecklinkVideoSrcClass
|
struct _GstDecklinkVideoSrcClass
|
||||||
|
|
Loading…
Reference in a new issue