/* GStreamer * Copyright (C) 2011 David Schleef * Copyright (C) 2014 Sebastian Dröge * Copyright (C) 2015 Florian Langlois * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin Street, Suite 500, * Boston, MA 02110-1335, USA. */ /** * SECTION:element-decklinkvideosrc * @short_description: Inputs Video from a BlackMagic DeckLink Device * * Capture Video from a BlackMagic DeckLink Device. * * ## Sample pipeline * |[ * gst-launch-1.0 \ * decklinkvideosrc device-number=0 connection=sdi mode=1080p25 ! \ * autovideosink * ]| * Capturing 1080p25 video from the SDI-In of Card 0. Devices are numbered * starting with 0. * * # Duplex-Mode: * Certain DechLink Cards like the Duo2 or the Quad2 contain two or four * independant SDI units with two connectors each. These units can operate either * in half- or in full-duplex mode. * * The Duplex-Mode of a Card can be configured using the `duplex-mode`-Property. * Cards that to not support Duplex-Modes are not influenced by the property. * * ## Half-Duplex-Mode (default): * By default decklinkvideosrc will configure them into half-duplex mode, so that * each connector acts as if it were an independant DeckLink Card which can either * be used as an Input or as an Output. In this mode the Duo2 can be used as as 4 SDI * In-/Outputs and the Quad2 as 8 SDI In-/Outputs. * * |[ * gst-launch-1.0 \ * decklinkvideosrc device-number=0 mode=1080p25 ! c. \ * decklinkvideosrc device-number=1 mode=1080p25 ! c. \ * decklinkvideosrc device-number=2 mode=1080p25 ! c. \ * decklinkvideosrc device-number=3 mode=1080p25 ! c. \ * compositor name=c \ * sink_0::xpos=0 sink_0::ypos=0 sink_0::width=960 sink_0::height=540 \ * sink_1::xpos=960 sink_1::ypos=0 sink_1::width=960 sink_1::height=540 \ * sink_2::xpos=0 sink_2::ypos=540 sink_2::width=960 sink_2::height=540 \ * sink_3::xpos=960 sink_3::ypos=540 sink_3::width=960 sink_3::height=540 ! \ * video/x-raw,width=1920,height=1080 ! \ * autovideosink * ]| * Capture 1080p25 from the first 4 units in the System (ie. the 4 Connectors of * a Duo2 Card) and compose them into a 2x2 grid. * * |[ * gst-launch-1.0 \ * videotestsrc foreground-color=0x0000ff00 ! decklinkvideosink device-number=0 mode=1080p25 \ * decklinkvideosrc device-number=1 mode=1080p25 ! autovideosink \ * decklinkvideosrc device-number=2 mode=1080p25 ! autovideosink \ * videotestsrc foreground-color=0x00ff0000 ! decklinkvideosink device-number=3 mode=1080p25 * ]| * Capture 1080p25 from the second and third unit in the System, * Playout a Test-Screen with colored Snow on the first and fourth unit * (ie. the Connectors 1-4 of a Duo2 unit). * * ## Device-Number-Mapping in Half-Duplex-Mode * The device-number to connector-mapping in half-duplex-mode is as follows for the Duo2 * - `device-number=0` SDI1 * - `device-number=1` SDI3 * - `device-number=2` SDI2 * - `device-number=3` SDI4 * * And for the Quad2 * - `device-number=0` SDI1 * - `device-number=1` SDI3 * - `device-number=2` SDI5 * - `device-number=3` SDI7 * - `device-number=4` SDI2 * - `device-number=5` SDI4 * - `device-number=6` SDI6 * - `device-number=7` SDI8 * * ## Full-Duplex-Mode: * When operating in full-duplex mode, two connectors of a unit are combined to * a single device, performing extra processing with the second connection. * * This mode is most useful for Playout. See @decklinkvideosink. * For Capturing the options are as follows: * * When capturing from a duplex-unit, the secondary port outputs the captured image * unchanged. * |[ * gst-launch-1.0 \ * decklinkvideosrc device-number=0 mode=1080p25 duplex-mode=full ! \ * autovideosink * ]| * * When simultaneously capturing and playing out onto the same device, the * secondary port outputs the played out video. Note, that this can also be * achieved using half-duplex mode. * |[ * gst-launch-1.0 \ * decklinkvideosrc device-number=0 mode=1080p25 duplex-mode=full ! \ * videoflip video-direction=vert ! \ * decklinkvideosink device-number=0 mode=1080p25 duplex-mode=full * ]| * Capturing Video on the primary port of device 0, output flipped version of the * video on secondary port of the same device. * * ## Device-Number-Mapping in Full-Duplex-Mode * The device-number to connector-mapping in full-duplex-mode is as follows for the Duo2 * - `device-number=0` SDI1 primary, SDI2 secondary * - `device-number=1` SDI3 primaty, SDI4 secondary * * And for the Quad2 * - `device-number=0` SDI1 primary, SDI2 secondary * - `device-number=1` SDI3 primaty, SDI4 secondary * - `device-number=2` SDI5 primary, SDI6 secondary * - `device-number=3` SDI7 primary, SDI8 secondary */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gstdecklinkvideosrc.h" #include GST_DEBUG_CATEGORY_STATIC (gst_decklink_video_src_debug); #define GST_CAT_DEFAULT gst_decklink_video_src_debug #define DEFAULT_MODE (GST_DECKLINK_MODE_AUTO) #define DEFAULT_CONNECTION (GST_DECKLINK_CONNECTION_AUTO) #define DEFAULT_BUFFER_SIZE (5) #define DEFAULT_OUTPUT_STREAM_TIME (FALSE) #define DEFAULT_SKIP_FIRST_TIME (0) #define DEFAULT_DROP_NO_SIGNAL_FRAMES (FALSE) #define DEFAULT_OUTPUT_CC (FALSE) #define DEFAULT_OUTPUT_AFD_BAR (FALSE) #ifndef ABSDIFF #define ABSDIFF(x, y) ( (x) > (y) ? ((x) - (y)) : ((y) - (x)) ) #endif enum { PROP_0, PROP_MODE, PROP_CONNECTION, PROP_DEVICE_NUMBER, PROP_BUFFER_SIZE, PROP_VIDEO_FORMAT, PROP_DUPLEX_MODE, PROP_TIMECODE_FORMAT, PROP_OUTPUT_STREAM_TIME, PROP_SKIP_FIRST_TIME, PROP_DROP_NO_SIGNAL_FRAMES, PROP_SIGNAL, PROP_HW_SERIAL_NUMBER, PROP_OUTPUT_CC, PROP_OUTPUT_AFD_BAR, }; typedef struct { IDeckLinkVideoInputFrame *frame; GstClockTime timestamp, duration; GstClockTime stream_timestamp; GstClockTime stream_duration; GstClockTime hardware_timestamp; GstClockTime hardware_duration; GstDecklinkModeEnum mode; BMDPixelFormat format; GstVideoTimeCode *tc; gboolean no_signal; } CaptureFrame; static void capture_frame_clear (CaptureFrame * frame) { if (frame->frame) frame->frame->Release (); if (frame->tc) gst_video_time_code_free (frame->tc); memset (frame, 0, sizeof (*frame)); } typedef struct { IDeckLinkVideoInputFrame *frame; IDeckLinkInput *input; } VideoFrame; static void video_frame_free (void *data) { VideoFrame *frame = (VideoFrame *) data; frame->frame->Release (); frame->input->Release (); g_free (frame); } static void gst_decklink_video_src_set_property (GObject * object, guint property_id, const GValue * value, GParamSpec * pspec); static void gst_decklink_video_src_get_property (GObject * object, guint property_id, GValue * value, GParamSpec * pspec); static void gst_decklink_video_src_finalize (GObject * object); static GstStateChangeReturn gst_decklink_video_src_change_state (GstElement * element, GstStateChange transition); static gboolean gst_decklink_video_src_query (GstBaseSrc * bsrc, GstQuery * query); static gboolean gst_decklink_video_src_unlock (GstBaseSrc * bsrc); static gboolean gst_decklink_video_src_unlock_stop (GstBaseSrc * bsrc); static GstFlowReturn gst_decklink_video_src_create (GstPushSrc * psrc, GstBuffer ** buffer); static gboolean gst_decklink_video_src_open (GstDecklinkVideoSrc * self); static gboolean gst_decklink_video_src_close (GstDecklinkVideoSrc * self); static gboolean gst_decklink_video_src_stop (GstDecklinkVideoSrc * self); static void gst_decklink_video_src_start_streams (GstElement * element); #define parent_class gst_decklink_video_src_parent_class G_DEFINE_TYPE (GstDecklinkVideoSrc, gst_decklink_video_src, GST_TYPE_PUSH_SRC); static void gst_decklink_video_src_class_init (GstDecklinkVideoSrcClass * klass) { GObjectClass *gobject_class = G_OBJECT_CLASS (klass); GstElementClass *element_class = GST_ELEMENT_CLASS (klass); GstBaseSrcClass *basesrc_class = GST_BASE_SRC_CLASS (klass); GstPushSrcClass *pushsrc_class = GST_PUSH_SRC_CLASS (klass); GstCaps *templ_caps; gobject_class->set_property = gst_decklink_video_src_set_property; gobject_class->get_property = gst_decklink_video_src_get_property; gobject_class->finalize = gst_decklink_video_src_finalize; element_class->change_state = GST_DEBUG_FUNCPTR (gst_decklink_video_src_change_state); basesrc_class->query = GST_DEBUG_FUNCPTR (gst_decklink_video_src_query); basesrc_class->negotiate = NULL; basesrc_class->unlock = GST_DEBUG_FUNCPTR (gst_decklink_video_src_unlock); basesrc_class->unlock_stop = GST_DEBUG_FUNCPTR (gst_decklink_video_src_unlock_stop); pushsrc_class->create = GST_DEBUG_FUNCPTR (gst_decklink_video_src_create); g_object_class_install_property (gobject_class, PROP_MODE, g_param_spec_enum ("mode", "Playback Mode", "Video Mode to use for playback", GST_TYPE_DECKLINK_MODE, DEFAULT_MODE, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT))); g_object_class_install_property (gobject_class, PROP_CONNECTION, g_param_spec_enum ("connection", "Connection", "Video input connection to use", GST_TYPE_DECKLINK_CONNECTION, DEFAULT_CONNECTION, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT))); g_object_class_install_property (gobject_class, PROP_DEVICE_NUMBER, g_param_spec_int ("device-number", "Device number", "Output device instance to use", 0, G_MAXINT, 0, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT))); g_object_class_install_property (gobject_class, PROP_BUFFER_SIZE, g_param_spec_uint ("buffer-size", "Buffer Size", "Size of internal buffer in number of video frames", 1, G_MAXINT, DEFAULT_BUFFER_SIZE, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); g_object_class_install_property (gobject_class, PROP_VIDEO_FORMAT, g_param_spec_enum ("video-format", "Video format", "Video format type to use for input (Only use auto for mode=auto)", GST_TYPE_DECKLINK_VIDEO_FORMAT, GST_DECKLINK_VIDEO_FORMAT_AUTO, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT))); g_object_class_install_property (gobject_class, PROP_DUPLEX_MODE, g_param_spec_enum ("duplex-mode", "Duplex mode", "Certain DeckLink devices such as the DeckLink Quad 2 and the " "DeckLink Duo 2 support configuration of the duplex mode of " "individual sub-devices." "A sub-device configured as full-duplex will use two connectors, " "which allows simultaneous capture and playback, internal keying, " "and fill & key scenarios." "A half-duplex sub-device will use a single connector as an " "individual capture or playback channel.", GST_TYPE_DECKLINK_DUPLEX_MODE, GST_DECKLINK_DUPLEX_MODE_HALF, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT))); g_object_class_install_property (gobject_class, PROP_TIMECODE_FORMAT, g_param_spec_enum ("timecode-format", "Timecode format", "Timecode format type to use for input", GST_TYPE_DECKLINK_TIMECODE_FORMAT, GST_DECKLINK_TIMECODE_FORMAT_RP188ANY, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT))); g_object_class_install_property (gobject_class, PROP_OUTPUT_STREAM_TIME, g_param_spec_boolean ("output-stream-time", "Output Stream Time", "Output stream time directly instead of translating to pipeline clock", DEFAULT_OUTPUT_STREAM_TIME, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); g_object_class_install_property (gobject_class, PROP_SKIP_FIRST_TIME, g_param_spec_uint64 ("skip-first-time", "Skip First Time", "Skip that much time of initial frames after starting", 0, G_MAXUINT64, DEFAULT_SKIP_FIRST_TIME, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); g_object_class_install_property (gobject_class, PROP_DROP_NO_SIGNAL_FRAMES, g_param_spec_boolean ("drop-no-signal-frames", "Drop No Signal Frames", "Drop frames that are marked as having no input signal", DEFAULT_DROP_NO_SIGNAL_FRAMES, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); g_object_class_install_property (gobject_class, PROP_SIGNAL, g_param_spec_boolean ("signal", "Input signal available", "True if there is a valid input signal available", FALSE, (GParamFlags) (G_PARAM_READABLE | G_PARAM_STATIC_STRINGS))); g_object_class_install_property (gobject_class, PROP_HW_SERIAL_NUMBER, g_param_spec_string ("hw-serial-number", "Hardware serial number", "The serial number (hardware ID) of the Decklink card", NULL, (GParamFlags) (G_PARAM_READABLE | G_PARAM_STATIC_STRINGS))); g_object_class_install_property (gobject_class, PROP_OUTPUT_CC, g_param_spec_boolean ("output-cc", "Output Closed Caption", "Extract and output CC as GstMeta (if present)", DEFAULT_OUTPUT_CC, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); g_object_class_install_property (gobject_class, PROP_OUTPUT_AFD_BAR, g_param_spec_boolean ("output-afd-bar", "Output AFD/Bar data", "Extract and output AFD/Bar as GstMeta (if present)", DEFAULT_OUTPUT_AFD_BAR, (GParamFlags) (G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS))); templ_caps = gst_decklink_mode_get_template_caps (TRUE); gst_element_class_add_pad_template (element_class, gst_pad_template_new ("src", GST_PAD_SRC, GST_PAD_ALWAYS, templ_caps)); gst_caps_unref (templ_caps); gst_element_class_set_static_metadata (element_class, "Decklink Video Source", "Video/Source/Hardware", "Decklink Source", "David Schleef , " "Sebastian Dröge "); GST_DEBUG_CATEGORY_INIT (gst_decklink_video_src_debug, "decklinkvideosrc", 0, "debug category for decklinkvideosrc element"); } static void gst_decklink_video_src_init (GstDecklinkVideoSrc * self) { self->mode = DEFAULT_MODE; self->caps_mode = GST_DECKLINK_MODE_AUTO; self->caps_format = bmdFormat8BitYUV; self->connection = DEFAULT_CONNECTION; self->device_number = 0; self->buffer_size = DEFAULT_BUFFER_SIZE; self->video_format = GST_DECKLINK_VIDEO_FORMAT_AUTO; self->duplex_mode = bmdDuplexModeHalf; self->timecode_format = bmdTimecodeRP188Any; self->signal_state = SIGNAL_STATE_UNKNOWN; self->output_stream_time = DEFAULT_OUTPUT_STREAM_TIME; self->skip_first_time = DEFAULT_SKIP_FIRST_TIME; self->drop_no_signal_frames = DEFAULT_DROP_NO_SIGNAL_FRAMES; self->output_cc = DEFAULT_OUTPUT_CC; self->output_afd_bar = DEFAULT_OUTPUT_AFD_BAR; self->window_size = 64; self->times = g_new (GstClockTime, 4 * self->window_size); self->times_temp = self->times + 2 * self->window_size; self->window_fill = 0; self->window_skip = 1; self->window_skip_count = 0; gst_base_src_set_live (GST_BASE_SRC (self), TRUE); gst_base_src_set_format (GST_BASE_SRC (self), GST_FORMAT_TIME); gst_pad_use_fixed_caps (GST_BASE_SRC_PAD (self)); g_mutex_init (&self->lock); g_cond_init (&self->cond); self->current_frames = gst_queue_array_new_for_struct (sizeof (CaptureFrame), DEFAULT_BUFFER_SIZE); } void gst_decklink_video_src_set_property (GObject * object, guint property_id, const GValue * value, GParamSpec * pspec) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object); switch (property_id) { case PROP_MODE: self->mode = (GstDecklinkModeEnum) g_value_get_enum (value); /* setting the default value for caps_mode here: if mode==auto then we * configure caps_mode from the caps, if mode!=auto we set caps_mode to * the same value as the mode. so self->caps_mode is essentially * self->mode with mode=auto filtered into whatever we got from the * negotiation */ if (self->mode != GST_DECKLINK_MODE_AUTO) self->caps_mode = self->mode; break; case PROP_CONNECTION: self->connection = (GstDecklinkConnectionEnum) g_value_get_enum (value); break; case PROP_DEVICE_NUMBER: self->device_number = g_value_get_int (value); break; case PROP_BUFFER_SIZE: self->buffer_size = g_value_get_uint (value); break; case PROP_VIDEO_FORMAT: self->video_format = (GstDecklinkVideoFormat) g_value_get_enum (value); switch (self->video_format) { case GST_DECKLINK_VIDEO_FORMAT_8BIT_YUV: case GST_DECKLINK_VIDEO_FORMAT_10BIT_YUV: case GST_DECKLINK_VIDEO_FORMAT_8BIT_ARGB: case GST_DECKLINK_VIDEO_FORMAT_8BIT_BGRA: self->caps_format = gst_decklink_pixel_format_from_type (self->video_format); case GST_DECKLINK_VIDEO_FORMAT_AUTO: break; default: GST_ELEMENT_WARNING (GST_ELEMENT (self), CORE, NOT_IMPLEMENTED, ("Format %d not supported", self->video_format), (NULL)); break; } break; case PROP_DUPLEX_MODE: self->duplex_mode = gst_decklink_duplex_mode_from_enum ((GstDecklinkDuplexMode) g_value_get_enum (value)); break; case PROP_TIMECODE_FORMAT: self->timecode_format = gst_decklink_timecode_format_from_enum ((GstDecklinkTimecodeFormat) g_value_get_enum (value)); break; case PROP_OUTPUT_STREAM_TIME: self->output_stream_time = g_value_get_boolean (value); break; case PROP_SKIP_FIRST_TIME: self->skip_first_time = g_value_get_uint64 (value); break; case PROP_DROP_NO_SIGNAL_FRAMES: self->drop_no_signal_frames = g_value_get_boolean (value); break; case PROP_OUTPUT_CC: self->output_cc = g_value_get_boolean (value); break; case PROP_OUTPUT_AFD_BAR: self->output_afd_bar = g_value_get_boolean (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); break; } } void gst_decklink_video_src_get_property (GObject * object, guint property_id, GValue * value, GParamSpec * pspec) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object); switch (property_id) { case PROP_MODE: g_value_set_enum (value, self->mode); break; case PROP_CONNECTION: g_value_set_enum (value, self->connection); break; case PROP_DEVICE_NUMBER: g_value_set_int (value, self->device_number); break; case PROP_BUFFER_SIZE: g_value_set_uint (value, self->buffer_size); break; case PROP_VIDEO_FORMAT: g_value_set_enum (value, self->video_format); break; case PROP_DUPLEX_MODE: g_value_set_enum (value, gst_decklink_duplex_mode_to_enum (self->duplex_mode)); break; case PROP_TIMECODE_FORMAT: g_value_set_enum (value, gst_decklink_timecode_format_to_enum (self->timecode_format)); break; case PROP_OUTPUT_STREAM_TIME: g_value_set_boolean (value, self->output_stream_time); break; case PROP_SKIP_FIRST_TIME: g_value_set_uint64 (value, self->skip_first_time); break; case PROP_DROP_NO_SIGNAL_FRAMES: g_value_set_boolean (value, self->drop_no_signal_frames); break; case PROP_SIGNAL: g_value_set_boolean (value, self->signal_state == SIGNAL_STATE_AVAILABLE); break; case PROP_HW_SERIAL_NUMBER: if (self->input) g_value_set_string (value, self->input->hw_serial_number); else g_value_set_string (value, NULL); break; case PROP_OUTPUT_CC: g_value_set_boolean (value, self->output_cc); break; case PROP_OUTPUT_AFD_BAR: g_value_set_boolean (value, self->output_afd_bar); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); break; } } void gst_decklink_video_src_finalize (GObject * object) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (object); g_free (self->times); self->times = NULL; g_mutex_clear (&self->lock); g_cond_clear (&self->cond); if (self->current_frames) { while (gst_queue_array_get_length (self->current_frames) > 0) { CaptureFrame *tmp = (CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); capture_frame_clear (tmp); } gst_queue_array_free (self->current_frames); self->current_frames = NULL; } G_OBJECT_CLASS (parent_class)->finalize (object); } static gboolean gst_decklink_video_src_start (GstDecklinkVideoSrc * self) { const GstDecklinkMode *mode; BMDVideoInputFlags flags; HRESULT ret; BMDPixelFormat format; g_mutex_lock (&self->input->lock); if (self->input->video_enabled) { g_mutex_unlock (&self->input->lock); return TRUE; } g_mutex_unlock (&self->input->lock); if (self->input->config && self->connection != GST_DECKLINK_CONNECTION_AUTO) { ret = self->input->config->SetInt (bmdDeckLinkConfigVideoInputConnection, gst_decklink_get_connection (self->connection)); if (ret != S_OK) { GST_ERROR_OBJECT (self, "Failed to set configuration (input source): 0x%08lx", (unsigned long) ret); return FALSE; } if (self->connection == GST_DECKLINK_CONNECTION_COMPOSITE) { ret = self->input->config->SetInt (bmdDeckLinkConfigAnalogVideoInputFlags, bmdAnalogVideoFlagCompositeSetup75); if (ret != S_OK) { GST_ERROR_OBJECT (self, "Failed to set configuration (composite setup): 0x%08lx", (unsigned long) ret); return FALSE; } } } flags = bmdVideoInputFlagDefault; if (self->mode == GST_DECKLINK_MODE_AUTO) { bool autoDetection = false; if (self->input->attributes) { ret = self->input-> attributes->GetFlag (BMDDeckLinkSupportsInputFormatDetection, &autoDetection); if (ret != S_OK) { GST_ERROR_OBJECT (self, "Failed to get attribute (autodetection): 0x%08lx", (unsigned long) ret); return FALSE; } if (autoDetection) flags |= bmdVideoInputEnableFormatDetection; } if (!autoDetection) { GST_ERROR_OBJECT (self, "Failed to activate auto-detection"); return FALSE; } } mode = gst_decklink_get_mode (self->mode); g_assert (mode != NULL); format = self->caps_format; ret = self->input->input->EnableVideoInput (mode->mode, format, flags); if (ret != S_OK) { GST_WARNING_OBJECT (self, "Failed to enable video input: 0x%08lx", (unsigned long) ret); return FALSE; } g_mutex_lock (&self->input->lock); self->input->mode = mode; self->input->video_enabled = TRUE; if (self->input->start_streams) self->input->start_streams (self->input->videosrc); g_mutex_unlock (&self->input->lock); return TRUE; } static void gst_decklink_video_src_update_time_mapping (GstDecklinkVideoSrc * self, GstClockTime capture_time, GstClockTime stream_time) { if (self->window_skip_count == 0) { GstClockTime num, den, b, xbase; gdouble r_squared; self->times[2 * self->window_fill] = stream_time; self->times[2 * self->window_fill + 1] = capture_time; self->window_fill++; self->window_skip_count++; if (self->window_skip_count >= self->window_skip) self->window_skip_count = 0; if (self->window_fill >= self->window_size) { guint fps = ((gdouble) self->info.fps_n + self->info.fps_d - 1) / ((gdouble) self->info.fps_d); /* Start by updating first every frame, once full every second frame, * etc. until we update once every 4 seconds */ if (self->window_skip < 4 * fps) self->window_skip *= 2; if (self->window_skip >= 4 * fps) self->window_skip = 4 * fps; self->window_fill = 0; self->window_filled = TRUE; } /* First sample ever, create some basic mapping to start */ if (!self->window_filled && self->window_fill == 1) { self->current_time_mapping.xbase = stream_time; self->current_time_mapping.b = capture_time; self->current_time_mapping.num = 1; self->current_time_mapping.den = 1; self->next_time_mapping_pending = FALSE; } /* Only bother calculating anything here once we had enough measurements, * i.e. let's take the window size as a start */ if (self->window_filled && gst_calculate_linear_regression (self->times, self->times_temp, self->window_size, &num, &den, &b, &xbase, &r_squared)) { GST_DEBUG_OBJECT (self, "Calculated new time mapping: pipeline time = %lf * (stream time - %" G_GUINT64_FORMAT ") + %" G_GUINT64_FORMAT " (%lf)", ((gdouble) num) / ((gdouble) den), xbase, b, r_squared); self->next_time_mapping.xbase = xbase; self->next_time_mapping.b = b; self->next_time_mapping.num = num; self->next_time_mapping.den = den; self->next_time_mapping_pending = TRUE; } } else { self->window_skip_count++; if (self->window_skip_count >= self->window_skip) self->window_skip_count = 0; } if (self->next_time_mapping_pending) { GstClockTime expected, new_calculated, diff, max_diff; expected = gst_clock_adjust_with_calibration (NULL, stream_time, self->current_time_mapping.xbase, self->current_time_mapping.b, self->current_time_mapping.num, self->current_time_mapping.den); new_calculated = gst_clock_adjust_with_calibration (NULL, stream_time, self->next_time_mapping.xbase, self->next_time_mapping.b, self->next_time_mapping.num, self->next_time_mapping.den); if (new_calculated > expected) diff = new_calculated - expected; else diff = expected - new_calculated; /* At most 5% frame duration change per update */ max_diff = gst_util_uint64_scale (GST_SECOND / 20, self->info.fps_d, self->info.fps_n); GST_DEBUG_OBJECT (self, "New time mapping causes difference of %" GST_TIME_FORMAT, GST_TIME_ARGS (diff)); GST_DEBUG_OBJECT (self, "Maximum allowed per frame %" GST_TIME_FORMAT, GST_TIME_ARGS (max_diff)); if (diff > max_diff) { /* adjust so that we move that much closer */ if (new_calculated > expected) { self->current_time_mapping.b = expected + max_diff; self->current_time_mapping.xbase = stream_time; } else { self->current_time_mapping.b = expected - max_diff; self->current_time_mapping.xbase = stream_time; } } else { self->current_time_mapping.xbase = self->next_time_mapping.xbase; self->current_time_mapping.b = self->next_time_mapping.b; self->current_time_mapping.num = self->next_time_mapping.num; self->current_time_mapping.den = self->next_time_mapping.den; self->next_time_mapping_pending = FALSE; } } } static void gst_decklink_video_src_got_frame (GstElement * element, IDeckLinkVideoInputFrame * frame, GstDecklinkModeEnum mode, GstClockTime capture_time, GstClockTime stream_time, GstClockTime stream_duration, GstClockTime hardware_time, GstClockTime hardware_duration, IDeckLinkTimecode * dtc, gboolean no_signal) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element); GstClockTime timestamp, duration; GST_LOG_OBJECT (self, "Got video frame at %" GST_TIME_FORMAT " / %" GST_TIME_FORMAT " (%" GST_TIME_FORMAT "), no signal: %d", GST_TIME_ARGS (capture_time), GST_TIME_ARGS (stream_time), GST_TIME_ARGS (stream_duration), no_signal); g_mutex_lock (&self->lock); if (self->first_time == GST_CLOCK_TIME_NONE) self->first_time = stream_time; if (self->skip_first_time > 0 && stream_time - self->first_time < self->skip_first_time) { g_mutex_unlock (&self->lock); GST_DEBUG_OBJECT (self, "Skipping frame as requested: %" GST_TIME_FORMAT " < %" GST_TIME_FORMAT, GST_TIME_ARGS (stream_time), GST_TIME_ARGS (self->skip_first_time + self->first_time)); return; } if (self->drop_no_signal_frames && no_signal) { CaptureFrame f; memset (&f, 0, sizeof (f)); /* Notify the streaming thread about the signal loss */ gst_queue_array_push_tail_struct (self->current_frames, &f); g_cond_signal (&self->cond); g_mutex_unlock (&self->lock); return; } gst_decklink_video_src_update_time_mapping (self, capture_time, stream_time); if (self->output_stream_time) { timestamp = stream_time; duration = stream_duration; } else { timestamp = gst_clock_adjust_with_calibration (NULL, stream_time, self->current_time_mapping.xbase, self->current_time_mapping.b, self->current_time_mapping.num, self->current_time_mapping.den); duration = gst_util_uint64_scale (stream_duration, self->current_time_mapping.num, self->current_time_mapping.den); } GST_LOG_OBJECT (self, "Converted times to %" GST_TIME_FORMAT " (%" GST_TIME_FORMAT ")", GST_TIME_ARGS (timestamp), GST_TIME_ARGS (duration)); if (!self->flushing) { CaptureFrame f; const GstDecklinkMode *bmode; GstVideoTimeCodeFlags flags = GST_VIDEO_TIME_CODE_FLAGS_NONE; guint field_count = 0; guint skipped_frames = 0; GstClockTime from_timestamp = GST_CLOCK_TIME_NONE; GstClockTime to_timestamp = GST_CLOCK_TIME_NONE; while (gst_queue_array_get_length (self->current_frames) >= self->buffer_size) { CaptureFrame *tmp = (CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); if (tmp->frame) { if (skipped_frames == 0) from_timestamp = tmp->timestamp; skipped_frames++; to_timestamp = tmp->timestamp; } capture_frame_clear (tmp); } if (skipped_frames > 0) GST_WARNING_OBJECT (self, "Dropped %u old frames from %" GST_TIME_FORMAT " to %" GST_TIME_FORMAT, skipped_frames, GST_TIME_ARGS (from_timestamp), GST_TIME_ARGS (to_timestamp)); memset (&f, 0, sizeof (f)); f.frame = frame; f.timestamp = timestamp; f.duration = duration; f.stream_timestamp = stream_time; f.stream_duration = stream_duration; f.hardware_timestamp = hardware_time; f.hardware_duration = hardware_duration; f.mode = mode; f.format = frame->GetPixelFormat (); f.no_signal = no_signal; if (dtc != NULL) { uint8_t hours, minutes, seconds, frames; HRESULT res; res = dtc->GetComponents (&hours, &minutes, &seconds, &frames); if (res != S_OK) { GST_ERROR ("Could not get components for timecode %p: 0x%08lx", dtc, (unsigned long) res); f.tc = NULL; } else { GST_DEBUG_OBJECT (self, "Got timecode %02d:%02d:%02d:%02d", hours, minutes, seconds, frames); bmode = gst_decklink_get_mode (mode); if (bmode->interlaced) flags = (GstVideoTimeCodeFlags) (flags | GST_VIDEO_TIME_CODE_FLAGS_INTERLACED); if (bmode->fps_d == 1001) { if (bmode->fps_n == 30000 || bmode->fps_n == 60000) { /* Some occurrences have been spotted where the driver mistakenly * fails to set the drop-frame flag for drop-frame timecodes. * Assume always drop-frame for 29.97 and 59.94 FPS */ flags = (GstVideoTimeCodeFlags) (flags | GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME); } else { /* Drop-frame isn't defined for any other framerates (e.g. 23.976) * */ flags = (GstVideoTimeCodeFlags) (flags & ~GST_VIDEO_TIME_CODE_FLAGS_DROP_FRAME); } } f.tc = gst_video_time_code_new (bmode->fps_n, bmode->fps_d, NULL, flags, hours, minutes, seconds, frames, field_count); } dtc->Release (); } else { f.tc = NULL; } frame->AddRef (); gst_queue_array_push_tail_struct (self->current_frames, &f); g_cond_signal (&self->cond); } g_mutex_unlock (&self->lock); } static void extract_vbi_line (GstDecklinkVideoSrc * self, GstBuffer ** buffer, IDeckLinkVideoFrameAncillary * vanc_frame, guint field2_offset, guint line, gboolean * found_cc_out, gboolean * found_afd_bar_out) { GstVideoAncillary gstanc; const guint8 *vancdata; gboolean found_cc = FALSE, found_afd_bar = FALSE; if (vanc_frame->GetBufferForVerticalBlankingLine (field2_offset + line, (void **) &vancdata) != S_OK) return; GST_DEBUG_OBJECT (self, "Checking for VBI data on field line %u (field %u)", field2_offset + line, field2_offset ? 2 : 1); gst_video_vbi_parser_add_line (self->vbiparser, vancdata); /* Check if CC or AFD/Bar is on this line if we didn't find any on a * previous line. Remember the line where we found them */ while (gst_video_vbi_parser_get_ancillary (self->vbiparser, &gstanc) == GST_VIDEO_VBI_PARSER_RESULT_OK) { switch (GST_VIDEO_ANCILLARY_DID16 (&gstanc)) { case GST_VIDEO_ANCILLARY_DID16_S334_EIA_708: if (*found_cc_out || !self->output_cc) continue; GST_DEBUG_OBJECT (self, "Adding CEA-708 CDP meta to buffer for line %u", field2_offset + line); GST_MEMDUMP_OBJECT (self, "CDP", gstanc.data, gstanc.data_count); gst_buffer_add_video_caption_meta (*buffer, GST_VIDEO_CAPTION_TYPE_CEA708_CDP, gstanc.data, gstanc.data_count); found_cc = TRUE; if (field2_offset) self->last_cc_vbi_line_field2 = line; else self->last_cc_vbi_line = line; break; case GST_VIDEO_ANCILLARY_DID16_S334_EIA_608: if (*found_cc_out || !self->output_cc) continue; GST_DEBUG_OBJECT (self, "Adding CEA-608 meta to buffer for line %u", field2_offset + line); GST_MEMDUMP_OBJECT (self, "CEA608", gstanc.data, gstanc.data_count); gst_buffer_add_video_caption_meta (*buffer, GST_VIDEO_CAPTION_TYPE_CEA608_S334_1A, gstanc.data, gstanc.data_count); found_cc = TRUE; if (field2_offset) self->last_cc_vbi_line_field2 = line; else self->last_cc_vbi_line = line; break; case GST_VIDEO_ANCILLARY_DID16_S2016_3_AFD_BAR:{ GstVideoAFDValue afd; gboolean is_letterbox; guint16 bar1, bar2; if (*found_afd_bar_out || !self->output_afd_bar) continue; GST_DEBUG_OBJECT (self, "Adding AFD/Bar meta to buffer for line %u", field2_offset + line); GST_MEMDUMP_OBJECT (self, "AFD/Bar", gstanc.data, gstanc.data_count); if (gstanc.data_count < 16) { GST_WARNING_OBJECT (self, "AFD/Bar data too small"); continue; } afd = (GstVideoAFDValue) ((gstanc.data[0] >> 3) & 0xf); is_letterbox = ((gstanc.data[3] >> 4) & 0x3) == 0; bar1 = GST_READ_UINT16_BE (&gstanc.data[4]); bar2 = GST_READ_UINT16_BE (&gstanc.data[6]); gst_buffer_add_video_afd_meta (*buffer, field2_offset ? 1 : 0, GST_VIDEO_AFD_SPEC_SMPTE_ST2016_1, afd); gst_buffer_add_video_bar_meta (*buffer, field2_offset ? 1 : 0, is_letterbox, bar1, bar2); found_afd_bar = TRUE; if (field2_offset) self->last_afd_bar_vbi_line_field2 = line; else self->last_afd_bar_vbi_line = line; break; } default: /* otherwise continue looking */ continue; } } if (found_cc) *found_cc_out = TRUE; if (found_afd_bar) *found_afd_bar_out = TRUE; } static void extract_vbi (GstDecklinkVideoSrc * self, GstBuffer ** buffer, VideoFrame * vf, const GstDecklinkMode * mode) { IDeckLinkVideoFrameAncillary *vanc_frame = NULL; gint line; GstVideoFormat videoformat; gboolean found_cc = FALSE, found_afd_bar = FALSE; if (vf->frame->GetAncillaryData (&vanc_frame) != S_OK) return; videoformat = gst_decklink_video_format_from_type (vanc_frame->GetPixelFormat ()); if (videoformat == GST_VIDEO_FORMAT_UNKNOWN) { GST_DEBUG_OBJECT (self, "Unknown video format for Ancillary data"); vanc_frame->Release (); return; } if (videoformat != self->anc_vformat && self->vbiparser) { gst_video_vbi_parser_free (self->vbiparser); self->vbiparser = NULL; } if (self->vbiparser == NULL) { self->vbiparser = gst_video_vbi_parser_new (videoformat, mode->width); self->anc_vformat = videoformat; } GST_DEBUG_OBJECT (self, "Checking for ancillary data in VBI"); /* First check last known lines, if any */ if (self->last_cc_vbi_line > 0) { extract_vbi_line (self, buffer, vanc_frame, 0, self->last_cc_vbi_line, &found_cc, &found_afd_bar); } if (self->last_afd_bar_vbi_line > 0 && self->last_cc_vbi_line != self->last_afd_bar_vbi_line) { extract_vbi_line (self, buffer, vanc_frame, 0, self->last_afd_bar_vbi_line, &found_cc, &found_afd_bar); } if (!found_cc) self->last_cc_vbi_line = -1; if (!found_afd_bar) self->last_afd_bar_vbi_line = -1; if ((self->output_cc && !found_cc) || (self->output_afd_bar && !found_afd_bar)) { /* Otherwise loop through the first 21 lines and hope to find the data */ /* FIXME: For the different formats the number of lines that can contain * VANC are different */ for (line = 1; line < 22; line++) { extract_vbi_line (self, buffer, vanc_frame, 0, line, &found_cc, &found_afd_bar); /* If we found everything we wanted to extract, stop here */ if ((!self->output_cc || found_cc) && (!self->output_afd_bar || found_afd_bar)) break; } } /* Do the same for field 2 in case of interlaced content */ if (GST_VIDEO_INFO_IS_INTERLACED (&self->info)) { gboolean found_cc_field2 = FALSE, found_afd_bar_field2 = FALSE; guint field2_offset = 0; /* The VANC lines for the second field are at an offset, depending on * the format in use */ switch (self->info.height) { case 486: /* NTSC: 525 / 2 + 1 */ field2_offset = 263; break; case 576: /* PAL: 625 / 2 + 1 */ field2_offset = 313; break; case 1080: /* 1080i: 1125 / 2 + 1 */ field2_offset = 563; break; default: g_assert_not_reached (); } /* First try the same lines as for field 1 if we don't know yet */ if (self->last_cc_vbi_line_field2 <= 0) self->last_cc_vbi_line_field2 = self->last_cc_vbi_line; if (self->last_afd_bar_vbi_line_field2 <= 0) self->last_afd_bar_vbi_line_field2 = self->last_afd_bar_vbi_line; if (self->last_cc_vbi_line_field2 > 0) { extract_vbi_line (self, buffer, vanc_frame, field2_offset, self->last_cc_vbi_line_field2, &found_cc_field2, &found_afd_bar_field2); } if (self->last_afd_bar_vbi_line_field2 > 0 && self->last_cc_vbi_line_field2 != self->last_afd_bar_vbi_line_field2) { extract_vbi_line (self, buffer, vanc_frame, field2_offset, self->last_afd_bar_vbi_line_field2, &found_cc_field2, &found_afd_bar_field2); } if (!found_cc_field2) self->last_cc_vbi_line_field2 = -1; if (!found_afd_bar_field2) self->last_afd_bar_vbi_line_field2 = -1; if (((self->output_cc && !found_cc_field2) || (self->output_afd_bar && !found_afd_bar_field2))) { for (line = 1; line < 22; line++) { extract_vbi_line (self, buffer, vanc_frame, field2_offset, line, &found_cc_field2, &found_afd_bar_field2); /* If we found everything we wanted to extract, stop here */ if ((!self->output_cc || found_cc_field2) && (!self->output_afd_bar || found_afd_bar_field2)) break; } } } vanc_frame->Release (); } static GstFlowReturn gst_decklink_video_src_create (GstPushSrc * bsrc, GstBuffer ** buffer) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); GstFlowReturn flow_ret = GST_FLOW_OK; const guint8 *data; gsize data_size; VideoFrame *vf; CaptureFrame f; GstCaps *caps; gboolean caps_changed = FALSE; const GstDecklinkMode *mode; static GstStaticCaps stream_reference = GST_STATIC_CAPS ("timestamp/x-decklink-stream"); static GstStaticCaps hardware_reference = GST_STATIC_CAPS ("timestamp/x-decklink-hardware"); if (!gst_decklink_video_src_start (self)) { return GST_FLOW_NOT_NEGOTIATED; } g_mutex_lock (&self->lock); retry: while (gst_queue_array_is_empty (self->current_frames) && !self->flushing) { g_cond_wait (&self->cond, &self->lock); } if (self->flushing) { GST_DEBUG_OBJECT (self, "Flushing"); g_mutex_unlock (&self->lock); return GST_FLOW_FLUSHING; } f = *(CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); // We will have no frame if frames without signal are dropped immediately // but we still have to signal that it's lost here. if (f.no_signal || !f.frame) { if (self->signal_state != SIGNAL_STATE_LOST) { self->signal_state = SIGNAL_STATE_LOST; g_object_notify (G_OBJECT (self), "signal"); GST_ELEMENT_WARNING (GST_ELEMENT (self), RESOURCE, READ, ("Signal lost"), ("No input source was detected - video frames invalid")); } // If we have no frame here, simply retry until we got one if (!f.frame) { capture_frame_clear (&f); goto retry; } } else { GstDecklinkSignalState previous_signal_state = self->signal_state; if (previous_signal_state != SIGNAL_STATE_AVAILABLE) { self->signal_state = SIGNAL_STATE_AVAILABLE; g_object_notify (G_OBJECT (self), "signal"); } if (previous_signal_state == SIGNAL_STATE_LOST) { GST_ELEMENT_INFO (GST_ELEMENT (self), RESOURCE, READ, ("Signal recovered"), ("Input source detected")); } } // If we're not flushing, we should have a valid frame from the queue g_assert (f.frame != NULL); if (!gst_pad_has_current_caps (GST_BASE_SRC_PAD (self))) { caps_changed = TRUE; } if (self->caps_mode != f.mode) { if (self->mode == GST_DECKLINK_MODE_AUTO || !gst_pad_has_current_caps (GST_BASE_SRC_PAD (self))) { GST_DEBUG_OBJECT (self, "Mode changed from %d to %d", self->caps_mode, f.mode); caps_changed = TRUE; self->caps_mode = f.mode; } else { g_mutex_unlock (&self->lock); GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, ("Invalid mode in captured frame"), ("Mode set to %d but captured %d", self->caps_mode, f.mode)); capture_frame_clear (&f); return GST_FLOW_NOT_NEGOTIATED; } } if (self->caps_format != f.format) { if (self->video_format == GST_DECKLINK_VIDEO_FORMAT_AUTO || !gst_pad_has_current_caps (GST_BASE_SRC_PAD (self))) { GST_DEBUG_OBJECT (self, "Format changed from %d to %d", self->caps_format, f.format); caps_changed = TRUE; self->caps_format = f.format; } else { g_mutex_unlock (&self->lock); GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, ("Invalid pixel format in captured frame"), ("Format set to %d but captured %d", self->caps_format, f.format)); capture_frame_clear (&f); return GST_FLOW_NOT_NEGOTIATED; } } /* 1 ns error can be just a rounding error, so that's OK. The Decklink * drivers give us a really steady stream time, so anything above 1 ns can't * be a rounding error and is therefore something to worry about */ if (self->expected_stream_time != GST_CLOCK_TIME_NONE && ABSDIFF (self->expected_stream_time, f.stream_timestamp) > 1) { GstMessage *msg; GstClockTime running_time; self->dropped += f.stream_timestamp - self->expected_stream_time; running_time = gst_segment_to_running_time (&GST_BASE_SRC (self)->segment, GST_FORMAT_TIME, f.timestamp); msg = gst_message_new_qos (GST_OBJECT (self), TRUE, running_time, f.stream_timestamp, f.timestamp, f.duration); gst_message_set_qos_stats (msg, GST_FORMAT_TIME, self->processed, self->dropped); gst_element_post_message (GST_ELEMENT (self), msg); } if (self->first_stream_time == GST_CLOCK_TIME_NONE) self->first_stream_time = f.stream_timestamp; self->processed = f.stream_timestamp - self->dropped - self->first_stream_time; self->expected_stream_time = f.stream_timestamp + f.stream_duration; g_mutex_unlock (&self->lock); if (caps_changed) { self->last_cc_vbi_line = -1; self->last_afd_bar_vbi_line = -1; self->last_cc_vbi_line_field2 = -1; self->last_afd_bar_vbi_line_field2 = -1; caps = gst_decklink_mode_get_caps (f.mode, f.format, TRUE); gst_video_info_from_caps (&self->info, caps); gst_base_src_set_caps (GST_BASE_SRC_CAST (bsrc), caps); gst_element_post_message (GST_ELEMENT_CAST (self), gst_message_new_latency (GST_OBJECT_CAST (self))); gst_caps_unref (caps); if (self->vbiparser) { gst_video_vbi_parser_free (self->vbiparser); self->vbiparser = NULL; self->anc_vformat = GST_VIDEO_FORMAT_UNKNOWN; } } f.frame->GetBytes ((gpointer *) & data); data_size = self->info.size; vf = (VideoFrame *) g_malloc0 (sizeof (VideoFrame)); *buffer = gst_buffer_new_wrapped_full ((GstMemoryFlags) GST_MEMORY_FLAG_READONLY, (gpointer) data, data_size, 0, data_size, vf, (GDestroyNotify) video_frame_free); vf->frame = f.frame; f.frame->AddRef (); vf->input = self->input->input; vf->input->AddRef (); mode = gst_decklink_get_mode (self->mode); // If we have a format that supports VANC and we are asked to extract CC, // then do it here. if ((self->output_cc || self->output_afd_bar) && self->signal_state != SIGNAL_STATE_LOST) extract_vbi (self, buffer, vf, mode); if (f.no_signal) GST_BUFFER_FLAG_SET (*buffer, GST_BUFFER_FLAG_GAP); GST_BUFFER_TIMESTAMP (*buffer) = f.timestamp; GST_BUFFER_DURATION (*buffer) = f.duration; if (f.tc != NULL) gst_buffer_add_video_time_code_meta (*buffer, f.tc); gst_buffer_add_reference_timestamp_meta (*buffer, gst_static_caps_get (&stream_reference), f.stream_timestamp, f.stream_duration); gst_buffer_add_reference_timestamp_meta (*buffer, gst_static_caps_get (&hardware_reference), f.hardware_timestamp, f.hardware_duration); if (mode->interlaced && mode->tff) GST_BUFFER_FLAG_SET (*buffer, GST_VIDEO_BUFFER_FLAG_TFF | GST_VIDEO_BUFFER_FLAG_INTERLACED); GST_DEBUG_OBJECT (self, "Outputting buffer %p with timestamp %" GST_TIME_FORMAT " and duration %" GST_TIME_FORMAT, *buffer, GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (*buffer)), GST_TIME_ARGS (GST_BUFFER_DURATION (*buffer))); capture_frame_clear (&f); return flow_ret; } static gboolean gst_decklink_video_src_query (GstBaseSrc * bsrc, GstQuery * query) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); gboolean ret = TRUE; switch (GST_QUERY_TYPE (query)) { case GST_QUERY_LATENCY:{ if (self->input) { GstClockTime min, max; const GstDecklinkMode *mode; g_mutex_lock (&self->lock); mode = gst_decklink_get_mode (self->caps_mode); g_mutex_unlock (&self->lock); min = gst_util_uint64_scale_ceil (GST_SECOND, mode->fps_d, mode->fps_n); max = self->buffer_size * min; gst_query_set_latency (query, TRUE, min, max); ret = TRUE; } else { ret = FALSE; } break; } default: ret = GST_BASE_SRC_CLASS (parent_class)->query (bsrc, query); break; } return ret; } static gboolean gst_decklink_video_src_unlock (GstBaseSrc * bsrc) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); g_mutex_lock (&self->lock); self->flushing = TRUE; g_cond_signal (&self->cond); g_mutex_unlock (&self->lock); return TRUE; } static gboolean gst_decklink_video_src_unlock_stop (GstBaseSrc * bsrc) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (bsrc); g_mutex_lock (&self->lock); self->flushing = FALSE; while (gst_queue_array_get_length (self->current_frames) > 0) { CaptureFrame *tmp = (CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); capture_frame_clear (tmp); } g_mutex_unlock (&self->lock); return TRUE; } static gboolean gst_decklink_video_src_open (GstDecklinkVideoSrc * self) { const GstDecklinkMode *mode; GST_DEBUG_OBJECT (self, "Opening"); self->input = gst_decklink_acquire_nth_input (self->device_number, GST_ELEMENT_CAST (self), FALSE); if (!self->input) { GST_ERROR_OBJECT (self, "Failed to acquire input"); return FALSE; } g_object_notify (G_OBJECT (self), "hw-serial-number"); mode = gst_decklink_get_mode (self->mode); g_assert (mode != NULL); g_mutex_lock (&self->input->lock); self->input->mode = mode; self->input->format = self->caps_format; self->input->got_video_frame = gst_decklink_video_src_got_frame; self->input->start_streams = gst_decklink_video_src_start_streams; g_mutex_unlock (&self->input->lock); return TRUE; } static gboolean gst_decklink_video_src_close (GstDecklinkVideoSrc * self) { GST_DEBUG_OBJECT (self, "Closing"); if (self->input) { g_mutex_lock (&self->input->lock); self->input->got_video_frame = NULL; self->input->mode = NULL; self->input->video_enabled = FALSE; self->input->start_streams = NULL; g_mutex_unlock (&self->input->lock); gst_decklink_release_nth_input (self->device_number, GST_ELEMENT_CAST (self), FALSE); self->input = NULL; } return TRUE; } static gboolean gst_decklink_video_src_stop (GstDecklinkVideoSrc * self) { GST_DEBUG_OBJECT (self, "Stopping"); while (gst_queue_array_get_length (self->current_frames) > 0) { CaptureFrame *tmp = (CaptureFrame *) gst_queue_array_pop_head_struct (self->current_frames); capture_frame_clear (tmp); } self->caps_mode = GST_DECKLINK_MODE_AUTO; if (self->input && self->input->video_enabled) { g_mutex_lock (&self->input->lock); self->input->video_enabled = FALSE; g_mutex_unlock (&self->input->lock); self->input->input->DisableVideoInput (); } if (self->vbiparser) { gst_video_vbi_parser_free (self->vbiparser); self->vbiparser = NULL; self->anc_vformat = GST_VIDEO_FORMAT_UNKNOWN; } return TRUE; } static void gst_decklink_video_src_start_streams (GstElement * element) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element); HRESULT res; if (self->input->video_enabled && (!self->input->audiosrc || self->input->audio_enabled) && (GST_STATE (self) == GST_STATE_PLAYING || GST_STATE_PENDING (self) == GST_STATE_PLAYING)) { GST_DEBUG_OBJECT (self, "Starting streams"); g_mutex_lock (&self->lock); self->first_time = GST_CLOCK_TIME_NONE; self->window_fill = 0; self->window_filled = FALSE; self->window_skip = 1; self->window_skip_count = 0; self->current_time_mapping.xbase = 0; self->current_time_mapping.b = 0; self->current_time_mapping.num = 1; self->current_time_mapping.den = 1; self->next_time_mapping.xbase = 0; self->next_time_mapping.b = 0; self->next_time_mapping.num = 1; self->next_time_mapping.den = 1; g_mutex_unlock (&self->lock); res = self->input->input->StartStreams (); if (res != S_OK) { GST_ELEMENT_ERROR (self, STREAM, FAILED, (NULL), ("Failed to start streams: 0x%08lx", (unsigned long) res)); return; } } else { GST_DEBUG_OBJECT (self, "Not starting streams yet"); } } static GstStateChangeReturn gst_decklink_video_src_change_state (GstElement * element, GstStateChange transition) { GstDecklinkVideoSrc *self = GST_DECKLINK_VIDEO_SRC_CAST (element); GstStateChangeReturn ret = GST_STATE_CHANGE_SUCCESS; switch (transition) { case GST_STATE_CHANGE_NULL_TO_READY: self->processed = 0; self->dropped = 0; self->expected_stream_time = GST_CLOCK_TIME_NONE; self->first_stream_time = GST_CLOCK_TIME_NONE; if (!gst_decklink_video_src_open (self)) { ret = GST_STATE_CHANGE_FAILURE; goto out; } if (self->mode == GST_DECKLINK_MODE_AUTO && self->video_format != GST_DECKLINK_VIDEO_FORMAT_AUTO) { GST_WARNING_OBJECT (self, "Warning: mode=auto and format!=auto may \ not work"); } self->vbiparser = NULL; self->anc_vformat = GST_VIDEO_FORMAT_UNKNOWN; break; case GST_STATE_CHANGE_READY_TO_PAUSED: self->flushing = FALSE; break; default: break; } if (ret == GST_STATE_CHANGE_FAILURE) return ret; ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); if (ret == GST_STATE_CHANGE_FAILURE) return ret; switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY: self->signal_state = SIGNAL_STATE_UNKNOWN; gst_decklink_video_src_stop (self); break; case GST_STATE_CHANGE_PLAYING_TO_PAUSED:{ HRESULT res; GST_DEBUG_OBJECT (self, "Stopping streams"); res = self->input->input->StopStreams (); if (res != S_OK) { GST_ELEMENT_ERROR (self, STREAM, FAILED, (NULL), ("Failed to stop streams: 0x%08lx", (unsigned long) res)); ret = GST_STATE_CHANGE_FAILURE; } break; } case GST_STATE_CHANGE_PAUSED_TO_PLAYING:{ g_mutex_lock (&self->input->lock); if (self->input->start_streams) self->input->start_streams (self->input->videosrc); g_mutex_unlock (&self->input->lock); break; } case GST_STATE_CHANGE_READY_TO_NULL: gst_decklink_video_src_close (self); break; default: break; } out: return ret; }