# Implements [`gst::BufferPoolExt`](../gst/trait.BufferPoolExt.html), [`gst::ObjectExt`](../gst/trait.ObjectExt.html), [`glib::object::ObjectExt`](../glib/object/trait.ObjectExt.html) Create a new bufferpool that can allocate video frames. This bufferpool supports all the video bufferpool options. # Returns a new `gst::BufferPool` to allocate video frames The various known types of Closed Caption (CC). Unknown type of CC CEA-608 as byte pairs. Note that this format is not recommended since is does not specify to which field the caption comes from and therefore assumes it comes from the first field (and that there is no information on the second field). Use `VideoCaptionType::Cea708Raw` if you wish to store CEA-608 from two fields and prefix each byte pair with 0xFC for the first field and 0xFD for the second field. CEA-608 as byte triplets as defined in SMPTE S334-1 Annex A. The second and third byte of the byte triplet is the raw CEA608 data, the first byte is a bitfield: The top/7th bit is 0 for the second field, 1 for the first field, bit 6 and 5 are 0 and bits 4 to 0 are a 5 bit unsigned integer that represents the line offset relative to the base-line of the original image format (line 9 for 525-line field 1, line 272 for 525-line field 2, line 5 for 625-line field 1 and line 318 for 625-line field 2). CEA-708 as cc_data byte triplets. They can also contain 608-in-708 and the first byte of each triplet has to be inspected for detecting the type. CEA-708 (and optionally CEA-608) in a CDP (Caption Distribution Packet) defined by SMPTE S-334-2. Contains the whole CDP (starting with 0x9669). Feature: `v1_16` A `VideoCodecFrame` represents a video frame both in raw and encoded form. Gets private data set on the frame by the subclass via `VideoCodecFrame::set_user_data` previously. # Returns The previously set user_data Increases the refcount of the given frame by one. # Returns `buf` Sets `user_data` on the frame and the `GDestroyNotify` that will be called when the frame is freed. Allows to attach private data by the subclass to frames. If a `user_data` was previously set, then the previous set `notify` will be called before the `user_data` is replaced. ## `user_data` private data ## `notify` a `GDestroyNotify` Decreases the refcount of the frame. If the refcount reaches 0, the frame will be freed. Structure representing the state of an incoming or outgoing video stream for encoders and decoders. Decoders and encoders will receive such a state through their respective `set_format` vmethods. Decoders and encoders can set the downstream state, by using the `VideoDecoder::set_output_state`() or `VideoEncoder::set_output_state`() methods. Increases the refcount of the given state by one. # Returns `buf` Decreases the refcount of the state. If the refcount reaches 0, the state will be freed. The color matrix is used to convert between Y'PbPr and non-linear RGB (R'G'B') unknown matrix identity matrix FCC color matrix ITU-R BT.709 color matrix ITU-R BT.601 color matrix SMPTE 240M color matrix ITU-R BT.2020 color matrix. Since: 1.6 The color primaries define the how to transform linear RGB values to and from the CIE XYZ colorspace. unknown color primaries BT709 primaries BT470M primaries BT470BG primaries SMPTE170M primaries SMPTE240M primaries Generic film BT2020 primaries. Since: 1.6 Adobe RGB primaries. Since: 1.8 SMPTE ST 428 primaries. Since: 1.16 SMPTE RP 431 primaries. Since: 1.16 SMPTE EG 432 primaries. Since: 1.16 EBU 3213 primaries. Since: 1.16 Possible color range values. These constants are defined for 8 bit color values and can be scaled for other bit depths. unknown range [0..255] for 8 bit components [16..235] for 8 bit components. Chroma has [16..240] range. Structure describing the color info. Parse the colorimetry string and update `self` with the parsed values. ## `color` a colorimetry string # Returns `true` if `color` points to valid colorimetry info. Compare the 2 colorimetry sets for equality ## `other` another `VideoColorimetry` # Returns `true` if `self` and `other` are equal. Check if the colorimetry information in `info` matches that of the string `color`. ## `color` a colorimetry string # Returns `true` if `color` conveys the same colorimetry info as the color information in `info`. Make a string representation of `self`. # Returns a string representation of `self`. This base class is for video decoders turning encoded data into raw video frames. The GstVideoDecoder base class and derived subclasses should cooperate as follows: ## Configuration * Initially, GstVideoDecoder calls `start` when the decoder element is activated, which allows the subclass to perform any global setup. * GstVideoDecoder calls `set_format` to inform the subclass of caps describing input video data that it is about to receive, including possibly configuration data. While unlikely, it might be called more than once, if changing input parameters require reconfiguration. * Incoming data buffers are processed as needed, described in Data Processing below. * GstVideoDecoder calls `stop` at end of all processing. ## Data processing * The base class gathers input data, and optionally allows subclass to parse this into subsequently manageable chunks, typically corresponding to and referred to as 'frames'. * Each input frame is provided in turn to the subclass' `handle_frame` callback. The ownership of the frame is given to the `handle_frame` callback. * If codec processing results in decoded data, the subclass should call `VideoDecoder::finish_frame` to have decoded data pushed. downstream. Otherwise, the subclass must call `VideoDecoder::drop_frame`, to allow the base class to do timestamp and offset tracking, and possibly to requeue the frame for a later attempt in the case of reverse playback. ## Shutdown phase * The GstVideoDecoder class calls `stop` to inform the subclass that data parsing will be stopped. ## Additional Notes * Seeking/Flushing * When the pipeline is seeked or otherwise flushed, the subclass is informed via a call to its `reset` callback, with the hard parameter set to true. This indicates the subclass should drop any internal data queues and timestamps and prepare for a fresh set of buffers to arrive for parsing and decoding. * End Of Stream * At end-of-stream, the subclass `parse` function may be called some final times with the at_eos parameter set to true, indicating that the element should not expect any more data to be arriving, and it should parse and remaining frames and call `VideoDecoder::have_frame` if possible. The subclass is responsible for providing pad template caps for source and sink pads. The pads need to be named "sink" and "src". It also needs to provide information about the ouptput caps, when they are known. This may be when the base class calls the subclass' `set_format` function, though it might be during decoding, before calling `VideoDecoder::finish_frame`. This is done via `VideoDecoder::set_output_state` The subclass is also responsible for providing (presentation) timestamps (likely based on corresponding input ones). If that is not applicable or possible, the base class provides limited framerate based interpolation. Similarly, the base class provides some limited (legacy) seeking support if specifically requested by the subclass, as full-fledged support should rather be left to upstream demuxer, parser or alike. This simple approach caters for seeking and duration reporting using estimated input bitrates. To enable it, a subclass should call `VideoDecoderExt::set_estimate_rate` to enable handling of incoming byte-streams. The base class provides some support for reverse playback, in particular in case incoming data is not packetized or upstream does not provide fragments on keyframe boundaries. However, the subclass should then be prepared for the parsing and frame processing stage to occur separately (in normal forward processing, the latter immediately follows the former), The subclass also needs to ensure the parsing stage properly marks keyframes, unless it knows the upstream elements will do so properly for incoming data. The bare minimum that a functional subclass needs to implement is: * Provide pad templates * Inform the base class of output caps via `VideoDecoder::set_output_state` * Parse input data, if it is not considered packetized from upstream Data will be provided to `parse` which should invoke `VideoDecoderExt::add_to_frame` and `VideoDecoder::have_frame` to separate the data belonging to each video frame. * Accept data in `handle_frame` and provide decoded results to `VideoDecoder::finish_frame`, or call `VideoDecoder::drop_frame`. # Implements [`VideoDecoderExt`](trait.VideoDecoderExt.html), [`gst::ElementExt`](../gst/trait.ElementExt.html), [`gst::ObjectExt`](../gst/trait.ObjectExt.html), [`glib::object::ObjectExt`](../glib/object/trait.ObjectExt.html) Trait containing all `VideoDecoder` methods. # Implementors [`VideoDecoder`](struct.VideoDecoder.html) Removes next `n_bytes` of input data and adds it to currently parsed frame. ## `n_bytes` the number of bytes to add Helper function that allocates a buffer to hold a video frame for `self`'s current `VideoCodecState`. You should use `VideoDecoder::allocate_output_frame` instead of this function, if possible at all. # Returns allocated buffer, or NULL if no buffer could be allocated (e.g. when downstream is flushing or shutting down) Helper function that allocates a buffer to hold a video frame for `self`'s current `VideoCodecState`. Subclass should already have configured video state and set src pad caps. The buffer allocated here is owned by the frame and you should only keep references to the frame, not the buffer. ## `frame` a `VideoCodecFrame` # Returns `gst::FlowReturn::Ok` if an output buffer could be allocated Same as `VideoDecoder::allocate_output_frame` except it allows passing `gst::BufferPoolAcquireParams` to the sub call gst_buffer_pool_acquire_buffer. Feature: `v1_12` ## `frame` a `VideoCodecFrame` ## `params` a `gst::BufferPoolAcquireParams` # Returns `gst::FlowReturn::Ok` if an output buffer could be allocated Similar to `VideoDecoder::finish_frame`, but drops `frame` in any case and posts a QoS message with the frame's details on the bus. In any case, the frame is considered finished and released. ## `frame` the `VideoCodecFrame` to drop # Returns a `gst::FlowReturn`, usually GST_FLOW_OK. `frame` should have a valid decoded data buffer, whose metadata fields are then appropriately set according to frame data and pushed downstream. If no output data is provided, `frame` is considered skipped. In any case, the frame is considered finished and released. After calling this function the output buffer of the frame is to be considered read-only. This function will also change the metadata of the buffer. ## `frame` a decoded `VideoCodecFrame` # Returns a `gst::FlowReturn` resulting from sending data downstream Lets `VideoDecoder` sub-classes to know the memory `allocator` used by the base class and its `params`. Unref the `allocator` after use it. ## `allocator` the `gst::Allocator` used ## `params` the `gst::AllocationParams` of `allocator` # Returns the instance of the `gst::BufferPool` used by the decoder; free it after use it # Returns currently configured byte to time conversion setting Get a pending unfinished `VideoCodecFrame` ## `frame_number` system_frame_number of a frame # Returns pending unfinished `VideoCodecFrame` identified by `frame_number`. Get all pending unfinished `VideoCodecFrame` # Returns pending unfinished `VideoCodecFrame`. Query the configured decoder latency. Results will be returned via `min_latency` and `max_latency`. ## `min_latency` address of variable in which to store the configured minimum latency, or `None` ## `max_latency` address of variable in which to store the configured mximum latency, or `None` Determines maximum possible decoding time for `frame` that will allow it to decode and arrive in time (as determined by QoS events). In particular, a negative result means decoding in time is no longer possible and should therefore occur as soon/skippy as possible. ## `frame` a `VideoCodecFrame` # Returns max decoding time. # Returns currently configured decoder tolerated error count. Queries decoder required format handling. # Returns `true` if required format handling is enabled. Get the oldest pending unfinished `VideoCodecFrame` # Returns oldest pending unfinished `VideoCodecFrame`. Get the `VideoCodecState` currently describing the output stream. # Returns `VideoCodecState` describing format of video data. Queries whether input data is considered packetized or not by the base class. # Returns TRUE if input data is considered packetized. Returns the number of bytes previously added to the current frame by calling `VideoDecoderExt::add_to_frame`. # Returns The number of bytes pending for the current frame # Returns The current QoS proportion. Gathers all data collected for currently parsed frame, gathers corresponding metadata and passes it along for further processing, i.e. `handle_frame`. # Returns a `gst::FlowReturn` Sets the audio decoder tags and how they should be merged with any upstream stream tags. This will override any tags previously-set with `gst_audio_decoder_merge_tags`. Note that this is provided for convenience, and the subclass is not required to use this and can still do tag handling on its own. MT safe. ## `tags` a `gst::TagList` to merge, or NULL to unset previously-set tags ## `mode` the `gst::TagMergeMode` to use, usually `gst::TagMergeMode::Replace` Negotiate with downstream elements to currently configured `VideoCodecState`. Unmark GST_PAD_FLAG_NEED_RECONFIGURE in any case. But mark it again if negotiate fails. # Returns `true` if the negotiation succeeded, else `false`. Returns caps that express `caps` (or sink template caps if `caps` == NULL) restricted to resolution/format/... combinations supported by downstream elements. ## `caps` initial caps ## `filter` filter caps # Returns a `gst::Caps` owned by caller Similar to `VideoDecoder::drop_frame`, but simply releases `frame` without any processing other than removing it from list of pending frames, after which it is considered finished and released. ## `frame` the `VideoCodecFrame` to release Allows baseclass to perform byte to time estimated conversion. ## `enabled` whether to enable byte to time conversion Same as `VideoDecoder::set_output_state`() but also allows you to also set the interlacing mode. Feature: `v1_16` ## `fmt` a `VideoFormat` ## `mode` A `VideoInterlaceMode` ## `width` The width in pixels ## `height` The height in pixels ## `reference` An optional reference `VideoCodecState` # Returns the newly configured output state. Lets `VideoDecoder` sub-classes tell the baseclass what the decoder latency is. Will also post a LATENCY message on the bus so the pipeline can reconfigure its global latency. ## `min_latency` minimum latency ## `max_latency` maximum latency Sets numbers of tolerated decoder errors, where a tolerated one is then only warned about, but more than tolerated will lead to fatal error. You can set -1 for never returning fatal errors. Default is set to GST_VIDEO_DECODER_MAX_ERRORS. The '-1' option was added in 1.4 ## `num` max tolerated errors Configures decoder format needs. If enabled, subclass needs to be negotiated with format caps before it can process any data. It will then never be handed any data before it has been configured. Otherwise, it might be handed data without having been configured and is then expected being able to do so either by default or based on the input data. ## `enabled` new state Creates a new `VideoCodecState` with the specified `fmt`, `width` and `height` as the output state for the decoder. Any previously set output state on `self` will be replaced by the newly created one. If the subclass wishes to copy over existing fields (like pixel aspec ratio, or framerate) from an existing `VideoCodecState`, it can be provided as a `reference`. If the subclass wishes to override some fields from the output state (like pixel-aspect-ratio or framerate) it can do so on the returned `VideoCodecState`. The new output state will only take effect (set on pads and buffers) starting from the next call to `VideoDecoder::finish_frame`(). ## `fmt` a `VideoFormat` ## `width` The width in pixels ## `height` The height in pixels ## `reference` An optional reference `VideoCodecState` # Returns the newly configured output state. Allows baseclass to consider input data as packetized or not. If the input is packetized, then the `parse` method will not be called. ## `packetized` whether the input data should be considered as packetized. Lets `VideoDecoder` sub-classes decide if they want the sink pad to use the default pad query handler to reply to accept-caps queries. By setting this to true it is possible to further customize the default handler with `GST_PAD_SET_ACCEPT_INTERSECT` and `GST_PAD_SET_ACCEPT_TEMPLATE` ## `use_` if the default pad accept-caps query handling should be used This base class is for video encoders turning raw video into encoded video data. GstVideoEncoder and subclass should cooperate as follows. ## Configuration * Initially, GstVideoEncoder calls `start` when the encoder element is activated, which allows subclass to perform any global setup. * GstVideoEncoder calls `set_format` to inform subclass of the format of input video data that it is about to receive. Subclass should setup for encoding and configure base class as appropriate (e.g. latency). While unlikely, it might be called more than once, if changing input parameters require reconfiguration. Baseclass will ensure that processing of current configuration is finished. * GstVideoEncoder calls `stop` at end of all processing. ## Data processing * Base class collects input data and metadata into a frame and hands this to subclass' `handle_frame`. * If codec processing results in encoded data, subclass should call `VideoEncoder::finish_frame` to have encoded data pushed downstream. * If implemented, baseclass calls subclass `pre_push` just prior to pushing to allow subclasses to modify some metadata on the buffer. If it returns GST_FLOW_OK, the buffer is pushed downstream. * GstVideoEncoderClass will handle both srcpad and sinkpad events. Sink events will be passed to subclass if `event` callback has been provided. ## Shutdown phase * GstVideoEncoder class calls `stop` to inform the subclass that data parsing will be stopped. Subclass is responsible for providing pad template caps for source and sink pads. The pads need to be named "sink" and "src". It should also be able to provide fixed src pad caps in `getcaps` by the time it calls `VideoEncoder::finish_frame`. Things that subclass need to take care of: * Provide pad templates * Provide source pad caps before pushing the first buffer * Accept data in `handle_frame` and provide encoded results to `VideoEncoder::finish_frame`. The `VideoEncoder:qos` property will enable the Quality-of-Service features of the encoder which gather statistics about the real-time performance of the downstream elements. If enabled, subclasses can use `VideoEncoderExt::get_max_encode_time` to check if input frames are already late and drop them right away to give a chance to the pipeline to catch up. # Implements [`VideoEncoderExt`](trait.VideoEncoderExt.html), [`gst::ElementExt`](../gst/trait.ElementExt.html), [`gst::ObjectExt`](../gst/trait.ObjectExt.html), [`glib::object::ObjectExt`](../glib/object/trait.ObjectExt.html) Trait containing all `VideoEncoder` methods. # Implementors [`VideoEncoder`](struct.VideoEncoder.html) Helper function that allocates a buffer to hold an encoded video frame for `self`'s current `VideoCodecState`. ## `size` size of the buffer # Returns allocated buffer Helper function that allocates a buffer to hold an encoded video frame for `self`'s current `VideoCodecState`. Subclass should already have configured video state and set src pad caps. The buffer allocated here is owned by the frame and you should only keep references to the frame, not the buffer. ## `frame` a `VideoCodecFrame` ## `size` size of the buffer # Returns `gst::FlowReturn::Ok` if an output buffer could be allocated `frame` must have a valid encoded data buffer, whose metadata fields are then appropriately set according to frame data or no buffer at all if the frame should be dropped. It is subsequently pushed downstream or provided to `pre_push`. In any case, the frame is considered finished and released. After calling this function the output buffer of the frame is to be considered read-only. This function will also change the metadata of the buffer. ## `frame` an encoded `VideoCodecFrame` # Returns a `gst::FlowReturn` resulting from sending data downstream Lets `VideoEncoder` sub-classes to know the memory `allocator` used by the base class and its `params`. Unref the `allocator` after use it. ## `allocator` the `gst::Allocator` used ## `params` the `gst::AllocationParams` of `allocator` Get a pending unfinished `VideoCodecFrame` ## `frame_number` system_frame_number of a frame # Returns pending unfinished `VideoCodecFrame` identified by `frame_number`. Get all pending unfinished `VideoCodecFrame` # Returns pending unfinished `VideoCodecFrame`. Query the configured encoding latency. Results will be returned via `min_latency` and `max_latency`. ## `min_latency` address of variable in which to store the configured minimum latency, or `None` ## `max_latency` address of variable in which to store the configured maximum latency, or `None` Determines maximum possible encoding time for `frame` that will allow it to encode and arrive in time (as determined by QoS events). In particular, a negative result means encoding in time is no longer possible and should therefore occur as soon/skippy as possible. If no QoS events have been received from downstream, or if `VideoEncoder:qos` is disabled this function returns `G_MAXINT64`. Feature: `v1_14` ## `frame` a `VideoCodecFrame` # Returns max decoding time. Get the oldest unfinished pending `VideoCodecFrame` # Returns oldest unfinished pending `VideoCodecFrame` Get the current `VideoCodecState` # Returns `VideoCodecState` describing format of video data. Checks if `self` is currently configured to handle Quality-of-Service events from downstream. Feature: `v1_14` # Returns `true` if the encoder is configured to perform Quality-of-Service. Sets the video encoder tags and how they should be merged with any upstream stream tags. This will override any tags previously-set with `VideoEncoderExt::merge_tags`. Note that this is provided for convenience, and the subclass is not required to use this and can still do tag handling on its own. MT safe. ## `tags` a `gst::TagList` to merge, or NULL to unset previously-set tags ## `mode` the `gst::TagMergeMode` to use, usually `gst::TagMergeMode::Replace` Negotiate with downstream elements to currently configured `VideoCodecState`. Unmark GST_PAD_FLAG_NEED_RECONFIGURE in any case. But mark it again if negotiate fails. # Returns `true` if the negotiation succeeded, else `false`. Returns caps that express `caps` (or sink template caps if `caps` == NULL) restricted to resolution/format/... combinations supported by downstream elements (e.g. muxers). ## `caps` initial caps ## `filter` filter caps # Returns a `gst::Caps` owned by caller Set the codec headers to be sent downstream whenever requested. ## `headers` a list of `gst::Buffer` containing the codec header Informs baseclass of encoding latency. ## `min_latency` minimum latency ## `max_latency` maximum latency Request minimal value for PTS passed to handle_frame. For streams with reordered frames this can be used to ensure that there is enough time to accomodate first DTS, which may be less than first PTS ## `min_pts` minimal PTS that will be passed to handle_frame Creates a new `VideoCodecState` with the specified caps as the output state for the encoder. Any previously set output state on `self` will be replaced by the newly created one. The specified `caps` should not contain any resolution, pixel-aspect-ratio, framerate, codec-data, .... Those should be specified instead in the returned `VideoCodecState`. If the subclass wishes to copy over existing fields (like pixel aspect ratio, or framerate) from an existing `VideoCodecState`, it can be provided as a `reference`. If the subclass wishes to override some fields from the output state (like pixel-aspect-ratio or framerate) it can do so on the returned `VideoCodecState`. The new output state will only take effect (set on pads and buffers) starting from the next call to `VideoEncoder::finish_frame`(). ## `caps` the `gst::Caps` to use for the output ## `reference` An optional reference `VideoCodecState` # Returns the newly configured output state. Configures `self` to handle Quality-of-Service events from downstream. Feature: `v1_14` ## `enabled` the new qos value. Field order of interlaced content. This is only valid for interlace-mode=interleaved and not interlace-mode=mixed. In the case of mixed or GST_VIDEO_FIELD_ORDER_UNKOWN, the field order is signalled via buffer flags. unknown field order for interlaced content. The actual field order is signalled via buffer flags. top field is first bottom field is first Feature: `v1_12` Provides useful functions and a base class for video filters. The videofilter will by default enable QoS on the parent GstBaseTransform to implement frame dropping. # Implements [`gst_base::BaseTransformExt`](../gst_base/trait.BaseTransformExt.html), [`gst::ElementExt`](../gst/trait.ElementExt.html), [`gst::ObjectExt`](../gst/trait.ObjectExt.html), [`glib::object::ObjectExt`](../glib/object/trait.ObjectExt.html) Enum value describing the most common video formats. See the [GStreamer raw video format design document](https://gstreamer.freedesktop.org/documentation/design/mediatype-video-raw.html`formats`) for details about the layout and packing of these formats in memory. Unknown or unset video format id Encoded video format. Only ever use that in caps for special video formats in combination with non-system memory GstCapsFeatures where it does not make sense to specify a real video format. planar 4:2:0 YUV planar 4:2:0 YVU (like I420 but UV planes swapped) packed 4:2:2 YUV (Y0-U0-Y1-V0 Y2-U2-Y3-V2 Y4 ...) packed 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...) packed 4:4:4 YUV with alpha channel (A0-Y0-U0-V0 ...) sparse rgb packed into 32 bit, space last sparse reverse rgb packed into 32 bit, space last sparse rgb packed into 32 bit, space first sparse reverse rgb packed into 32 bit, space first rgb with alpha channel last reverse rgb with alpha channel last rgb with alpha channel first reverse rgb with alpha channel first RGB packed into 24 bits without padding (`R-G-B-R-G-B`) reverse RGB packed into 24 bits without padding (`B-G-R-B-G-R`) planar 4:1:1 YUV planar 4:2:2 YUV packed 4:2:2 YUV (Y0-V0-Y1-U0 Y2-V2-Y3-U2 Y4 ...) planar 4:4:4 YUV packed 4:2:2 10-bit YUV, complex format packed 4:2:2 16-bit YUV, Y0-U0-Y1-V1 order planar 4:2:0 YUV with interleaved UV plane planar 4:2:0 YUV with interleaved VU plane 8-bit grayscale 16-bit grayscale, most significant byte first 16-bit grayscale, least significant byte first packed 4:4:4 YUV (Y-U-V ...) rgb 5-6-5 bits per component reverse rgb 5-6-5 bits per component rgb 5-5-5 bits per component reverse rgb 5-5-5 bits per component packed 10-bit 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...) planar 4:4:2:0 AYUV 8-bit paletted RGB planar 4:1:0 YUV planar 4:1:0 YUV (like YUV9 but UV planes swapped) packed 4:1:1 YUV (Cb-Y0-Y1-Cr-Y2-Y3 ...) rgb with alpha channel first, 16 bits per channel packed 4:4:4 YUV with alpha channel, 16 bits per channel (A0-Y0-U0-V0 ...) packed 4:4:4 RGB, 10 bits per channel planar 4:2:0 YUV, 10 bits per channel planar 4:2:0 YUV, 10 bits per channel planar 4:2:2 YUV, 10 bits per channel planar 4:2:2 YUV, 10 bits per channel planar 4:4:4 YUV, 10 bits per channel (Since: 1.2) planar 4:4:4 YUV, 10 bits per channel (Since: 1.2) planar 4:4:4 RGB, 8 bits per channel (Since: 1.2) planar 4:4:4 RGB, 10 bits per channel (Since: 1.2) planar 4:4:4 RGB, 10 bits per channel (Since: 1.2) planar 4:2:2 YUV with interleaved UV plane (Since: 1.2) planar 4:4:4 YUV with interleaved UV plane (Since: 1.2) NV12 with 64x32 tiling in zigzag pattern (Since: 1.4) planar 4:4:2:0 YUV, 10 bits per channel (Since: 1.6) planar 4:4:2:0 YUV, 10 bits per channel (Since: 1.6) planar 4:4:2:2 YUV, 10 bits per channel (Since: 1.6) planar 4:4:2:2 YUV, 10 bits per channel (Since: 1.6) planar 4:4:4:4 YUV, 10 bits per channel (Since: 1.6) planar 4:4:4:4 YUV, 10 bits per channel (Since: 1.6) planar 4:2:2 YUV with interleaved VU plane (Since: 1.6) planar 4:2:0 YUV with interleaved UV plane, 10 bits per channel (Since: 1.10) planar 4:2:0 YUV with interleaved UV plane, 10 bits per channel (Since: 1.10) packed 4:4:4 YUV (U-Y-V ...) (Since: 1.10) packed 4:2:2 YUV (V0-Y0-U0-Y1 V2-Y2-U2-Y3 V4 ...) planar 4:4:4:4 ARGB, 8 bits per channel (Since: 1.12) planar 4:4:4:4 ARGB, 10 bits per channel (Since: 1.12) planar 4:4:4:4 ARGB, 10 bits per channel (Since: 1.12) planar 4:4:4 RGB, 12 bits per channel (Since: 1.12) planar 4:4:4 RGB, 12 bits per channel (Since: 1.12) planar 4:4:4:4 ARGB, 12 bits per channel (Since: 1.12) planar 4:4:4:4 ARGB, 12 bits per channel (Since: 1.12) planar 4:2:0 YUV, 12 bits per channel (Since: 1.12) planar 4:2:0 YUV, 12 bits per channel (Since: 1.12) planar 4:2:2 YUV, 12 bits per channel (Since: 1.12) planar 4:2:2 YUV, 12 bits per channel (Since: 1.12) planar 4:4:4 YUV, 12 bits per channel (Since: 1.12) planar 4:4:4 YUV, 12 bits per channel (Since: 1.12) 10-bit grayscale, packed into 32bit words (2 bits padding) (Since: 1.14) 10-bit variant of `VideoFormat::Nv12`, packed into 32bit words (MSB 2 bits padding) (Since: 1.14) 10-bit variant of `VideoFormat::Nv16`, packed into 32bit words (MSB 2 bits padding) (Since: 1.14) Fully packed variant of NV12_10LE32 (Since: 1.16) packed 4:2:2 YUV, 10 bits per channel (Since: 1.16) packed 4:4:4 YUV, 10 bits per channel(A-V-Y-U...) (Since: 1.16) packed 4:4:4 YUV with alpha channel (V0-U0-Y0-A0...) (Since: 1.16) packed 4:4:4 RGB with alpha channel(B-G-R-A), 10 bits for R/G/B channel and MSB 2 bits for alpha channel (Since: 1.16) Information for a video format. A video frame obtained from `VideoFrame::map` Copy the contents from `src` to `self`. ## `src` a `VideoFrame` # Returns TRUE if the contents could be copied. Copy the plane with index `plane` from `src` to `self`. ## `src` a `VideoFrame` ## `plane` a plane # Returns TRUE if the contents could be copied. Use `info` and `buffer` to fill in the values of `self`. `self` is usually allocated on the stack, and you will pass the address to the `VideoFrame` structure allocated on the stack; `VideoFrame::map` will then fill in the structures with the various video-specific information you need to access the pixels of the video buffer. You can then use accessor macros such as GST_VIDEO_FRAME_COMP_DATA(), GST_VIDEO_FRAME_PLANE_DATA(), GST_VIDEO_FRAME_COMP_STRIDE(), GST_VIDEO_FRAME_PLANE_STRIDE() etc. to get to the pixels. ```C GstVideoFrame vframe; ... // set RGB pixels to black one at a time if (gst_video_frame_map (&vframe, video_info, video_buffer, GST_MAP_WRITE)) { guint8 *pixels = GST_VIDEO_FRAME_PLANE_DATA (vframe, 0); guint stride = GST_VIDEO_FRAME_PLANE_STRIDE (vframe, 0); guint pixel_stride = GST_VIDEO_FRAME_COMP_PSTRIDE (vframe, 0); for (h = 0; h < height; ++h) { for (w = 0; w < width; ++w) { guint8 *pixel = pixels + h * stride + w * pixel_stride; memset (pixel, 0, pixel_stride); } } gst_video_frame_unmap (&vframe); } ... ``` All video planes of `buffer` will be mapped and the pointers will be set in `self`->data. The purpose of this function is to make it easy for you to get to the video pixels in a generic way, without you having to worry too much about details such as whether the video data is allocated in one contiguous memory chunk or multiple memory chunks (e.g. one for each plane); or if custom strides and custom plane offsets are used or not (as signalled by GstVideoMeta on each buffer). This function will just fill the `VideoFrame` structure with the right values and if you use the accessor macros everything will just work and you can access the data easily. It also maps the underlying memory chunks for you. ## `info` a `VideoInfo` ## `buffer` the buffer to map ## `flags` `gst::MapFlags` # Returns `true` on success. Use `info` and `buffer` to fill in the values of `self` with the video frame information of frame `id`. When `id` is -1, the default frame is mapped. When `id` != -1, this function will return `false` when there is no GstVideoMeta with that id. All video planes of `buffer` will be mapped and the pointers will be set in `self`->data. ## `info` a `VideoInfo` ## `buffer` the buffer to map ## `id` the frame id to map ## `flags` `gst::MapFlags` # Returns `true` on success. Unmap the memory previously mapped with gst_video_frame_map. Information describing image properties. This information can be filled in from GstCaps with `VideoInfo::from_caps`. The information is also used to store the specific video info when mapping a video frame with `VideoFrame::map`. Use the provided macros to access the info in this structure. Allocate a new `VideoInfo` that is also initialized with `VideoInfo::init`. # Returns a new `VideoInfo`. free with `VideoInfo::free`. Adjust the offset and stride fields in `self` so that the padding and stride alignment in `align` is respected. Extra padding will be added to the right side when stride alignment padding is required and `align` will be updated with the new padding values. ## `align` alignment parameters # Returns `false` if alignment could not be applied, e.g. because the size of a frame can't be represented as a 32 bit integer (Since: 1.12) Converts among various `gst::Format` types. This function handles GST_FORMAT_BYTES, GST_FORMAT_TIME, and GST_FORMAT_DEFAULT. For raw video, GST_FORMAT_DEFAULT corresponds to video frames. This function can be used to handle pad queries of the type GST_QUERY_CONVERT. ## `src_format` `gst::Format` of the `src_value` ## `src_value` value to convert ## `dest_format` `gst::Format` of the `dest_value` ## `dest_value` pointer to destination value # Returns TRUE if the conversion was successful. Copy a GstVideoInfo structure. # Returns a new `VideoInfo`. free with gst_video_info_free. Free a GstVideoInfo structure previously allocated with `VideoInfo::new` or `VideoInfo::copy`. Parse `caps` and update `self`. ## `caps` a `gst::Caps` # Returns TRUE if `caps` could be parsed Initialize `self` with default values. Compares two `VideoInfo` and returns whether they are equal or not ## `other` a `VideoInfo` # Returns `true` if `self` and `other` are equal, else `false`. Set the default info for a video frame of `format` and `width` and `height`. Note: This initializes `self` first, no values are preserved. This function does not set the offsets correctly for interlaced vertically subsampled formats. ## `format` the format ## `width` a width ## `height` a height # Returns `false` if the returned video info is invalid, e.g. because the size of a frame can't be represented as a 32 bit integer (Since: 1.12) Same as `VideoInfo::set_format` but also allowing to set the interlaced mode. Feature: `v1_16` ## `format` the format ## `mode` a `VideoInterlaceMode` ## `width` a width ## `height` a height # Returns `false` if the returned video info is invalid, e.g. because the size of a frame can't be represented as a 32 bit integer. Convert the values of `self` into a `gst::Caps`. # Returns a new `gst::Caps` containing the info of `self`. The possible values of the `VideoInterlaceMode` describing the interlace mode of the stream. all frames are progressive 2 fields are interleaved in one video frame. Extra buffer flags describe the field order. frames contains both interlaced and progressive video, the buffer flags describe the frame and fields. 2 fields are stored in one buffer, use the frame ID to get access to the required field. For multiview (the 'views' property > 1) the fields of view N can be found at frame ID (N * 2) and (N * 2) + 1. Each field has only half the amount of lines as noted in the height property. This mode requires multiple GstVideoMeta metadata to describe the fields. 1 field is stored in one buffer, `GST_VIDEO_BUFFER_FLAG_TF` or `GST_VIDEO_BUFFER_FLAG_BF` indicates if the buffer is carrying the top or bottom field, respectively. The top and bottom buffers are expected to alternate in the pipeline, with this mode (Since: 1.16). `VideoMultiviewFramePacking` represents the subset of `VideoMultiviewMode` values that can be applied to any video frame without needing extra metadata. It can be used by elements that provide a property to override the multiview interpretation of a video stream when the video doesn't contain any markers. This enum is used (for example) on playbin, to re-interpret a played video stream as a stereoscopic video. The individual enum values are equivalent to and have the same value as the matching `VideoMultiviewMode`. A special value indicating no frame packing info. All frames are monoscopic. All frames represent a left-eye view. All frames represent a right-eye view. Left and right eye views are provided in the left and right half of the frame respectively. Left and right eye views are provided in the left and right half of the frame, but have been sampled using quincunx method, with half-pixel offset between the 2 views. Alternating vertical columns of pixels represent the left and right eye view respectively. Alternating horizontal rows of pixels represent the left and right eye view respectively. The top half of the frame contains the left eye, and the bottom half the right eye. Pixels are arranged with alternating pixels representing left and right eye views in a checkerboard fashion. All possible stereoscopic 3D and multiview representations. In conjunction with `VideoMultiviewFlags`, describes how multiview content is being transported in the stream. A special value indicating no multiview information. Used in GstVideoInfo and other places to indicate that no specific multiview handling has been requested or provided. This value is never carried on caps. All frames are monoscopic. All frames represent a left-eye view. All frames represent a right-eye view. Left and right eye views are provided in the left and right half of the frame respectively. Left and right eye views are provided in the left and right half of the frame, but have been sampled using quincunx method, with half-pixel offset between the 2 views. Alternating vertical columns of pixels represent the left and right eye view respectively. Alternating horizontal rows of pixels represent the left and right eye view respectively. The top half of the frame contains the left eye, and the bottom half the right eye. Pixels are arranged with alternating pixels representing left and right eye views in a checkerboard fashion. Left and right eye views are provided in separate frames alternately. Multiple independent views are provided in separate frames in sequence. This method only applies to raw video buffers at the moment. Specific view identification is via the `GstVideoMultiviewMeta` and `VideoMeta`(s) on raw video buffers. Multiple views are provided as separate `gst::Memory` framebuffers attached to each `gst::Buffer`, described by the `GstVideoMultiviewMeta` and `VideoMeta`(s) The `VideoOverlay` interface is used for 2 main purposes : * To get a grab on the Window where the video sink element is going to render. This is achieved by either being informed about the Window identifier that the video sink element generated, or by forcing the video sink element to use a specific Window identifier for rendering. * To force a redrawing of the latest video frame the video sink element displayed on the Window. Indeed if the `gst::Pipeline` is in `gst::State::Paused` state, moving the Window around will damage its content. Application developers will want to handle the Expose events themselves and force the video sink element to refresh the Window's content. Using the Window created by the video sink is probably the simplest scenario, in some cases, though, it might not be flexible enough for application developers if they need to catch events such as mouse moves and button clicks. Setting a specific Window identifier on the video sink element is the most flexible solution but it has some issues. Indeed the application needs to set its Window identifier at the right time to avoid internal Window creation from the video sink element. To solve this issue a `gst::Message` is posted on the bus to inform the application that it should set the Window identifier immediately. Here is an example on how to do that correctly: ```text static GstBusSyncReply create_window (GstBus * bus, GstMessage * message, GstPipeline * pipeline) { // ignore anything but 'prepare-window-handle' element messages if (!gst_is_video_overlay_prepare_window_handle_message (message)) return GST_BUS_PASS; win = XCreateSimpleWindow (disp, root, 0, 0, 320, 240, 0, 0, 0); XSetWindowBackgroundPixmap (disp, win, None); XMapRaised (disp, win); XSync (disp, FALSE); gst_video_overlay_set_window_handle (GST_VIDEO_OVERLAY (GST_MESSAGE_SRC (message)), win); gst_message_unref (message); return GST_BUS_DROP; } ... int main (int argc, char **argv) { ... bus = gst_pipeline_get_bus (GST_PIPELINE (pipeline)); gst_bus_set_sync_handler (bus, (GstBusSyncHandler) create_window, pipeline, NULL); ... } ``` ## Two basic usage scenarios There are two basic usage scenarios: in the simplest case, the application uses `playbin` or `plasink` or knows exactly what particular element is used for video output, which is usually the case when the application creates the videosink to use (e.g. `xvimagesink`, `ximagesink`, etc.) itself; in this case, the application can just create the videosink element, create and realize the window to render the video on and then call `VideoOverlay::set_window_handle` directly with the XID or native window handle, before starting up the pipeline. As `playbin` and `playsink` implement the video overlay interface and proxy it transparently to the actual video sink even if it is created later, this case also applies when using these elements. In the other and more common case, the application does not know in advance what GStreamer video sink element will be used for video output. This is usually the case when an element such as `autovideosink` is used. In this case, the video sink element itself is created asynchronously from a GStreamer streaming thread some time after the pipeline has been started up. When that happens, however, the video sink will need to know right then whether to render onto an already existing application window or whether to create its own window. This is when it posts a prepare-window-handle message, and that is also why this message needs to be handled in a sync bus handler which will be called from the streaming thread directly (because the video sink will need an answer right then). As response to the prepare-window-handle element message in the bus sync handler, the application may use `VideoOverlay::set_window_handle` to tell the video sink to render onto an existing window surface. At this point the application should already have obtained the window handle / XID, so it just needs to set it. It is generally not advisable to call any GUI toolkit functions or window system functions from the streaming thread in which the prepare-window-handle message is handled, because most GUI toolkits and windowing systems are not thread-safe at all and a lot of care would be required to co-ordinate the toolkit and window system calls of the different threads (Gtk+ users please note: prior to Gtk+ 2.18 GDK_WINDOW_XID() was just a simple structure access, so generally fine to do within the bus sync handler; this macro was changed to a function call in Gtk+ 2.18 and later, which is likely to cause problems when called from a sync handler; see below for a better approach without GDK_WINDOW_XID() used in the callback). ## GstVideoOverlay and Gtk+ ```text #include <gst/video/videooverlay.h> #include <gtk/gtk.h> #ifdef GDK_WINDOWING_X11 #include <gdk/gdkx.h> // for GDK_WINDOW_XID #endif #ifdef GDK_WINDOWING_WIN32 #include <gdk/gdkwin32.h> // for GDK_WINDOW_HWND #endif ... static guintptr video_window_handle = 0; ... static GstBusSyncReply bus_sync_handler (GstBus * bus, GstMessage * message, gpointer user_data) { // ignore anything but 'prepare-window-handle' element messages if (!gst_is_video_overlay_prepare_window_handle_message (message)) return GST_BUS_PASS; if (video_window_handle != 0) { GstVideoOverlay *overlay; // GST_MESSAGE_SRC (message) will be the video sink element overlay = GST_VIDEO_OVERLAY (GST_MESSAGE_SRC (message)); gst_video_overlay_set_window_handle (overlay, video_window_handle); } else { g_warning ("Should have obtained video_window_handle by now!"); } gst_message_unref (message); return GST_BUS_DROP; } ... static void video_widget_realize_cb (GtkWidget * widget, gpointer data) { #if GTK_CHECK_VERSION(2,18,0) // Tell Gtk+/Gdk to create a native window for this widget instead of // drawing onto the parent widget. // This is here just for pedagogical purposes, GDK_WINDOW_XID will call // it as well in newer Gtk versions if (!gdk_window_ensure_native (widget->window)) g_error ("Couldn't create native window needed for GstVideoOverlay!"); #endif #ifdef GDK_WINDOWING_X11 { gulong xid = GDK_WINDOW_XID (gtk_widget_get_window (video_window)); video_window_handle = xid; } #endif #ifdef GDK_WINDOWING_WIN32 { HWND wnd = GDK_WINDOW_HWND (gtk_widget_get_window (video_window)); video_window_handle = (guintptr) wnd; } #endif } ... int main (int argc, char **argv) { GtkWidget *video_window; GtkWidget *app_window; ... app_window = gtk_window_new (GTK_WINDOW_TOPLEVEL); ... video_window = gtk_drawing_area_new (); g_signal_connect (video_window, "realize", G_CALLBACK (video_widget_realize_cb), NULL); gtk_widget_set_double_buffered (video_window, FALSE); ... // usually the video_window will not be directly embedded into the // application window like this, but there will be many other widgets // and the video window will be embedded in one of them instead gtk_container_add (GTK_CONTAINER (ap_window), video_window); ... // show the GUI gtk_widget_show_all (app_window); // realize window now so that the video window gets created and we can // obtain its XID/HWND before the pipeline is started up and the videosink // asks for the XID/HWND of the window to render onto gtk_widget_realize (video_window); // we should have the XID/HWND now g_assert (video_window_handle != 0); ... // set up sync handler for setting the xid once the pipeline is started bus = gst_pipeline_get_bus (GST_PIPELINE (pipeline)); gst_bus_set_sync_handler (bus, (GstBusSyncHandler) bus_sync_handler, NULL, NULL); gst_object_unref (bus); ... gst_element_set_state (pipeline, GST_STATE_PLAYING); ... } ``` ## GstVideoOverlay and Qt ```text #include <glib.h> #include <gst/gst.h> #include <gst/video/videooverlay.h> #include <QApplication> #include <QTimer> #include <QWidget> int main(int argc, char *argv[]) { if (!g_thread_supported ()) g_thread_init (NULL); gst_init (&argc, &argv); QApplication app(argc, argv); app.connect(&app, SIGNAL(lastWindowClosed()), &app, SLOT(quit ())); // prepare the pipeline GstElement *pipeline = gst_pipeline_new ("xvoverlay"); GstElement *src = gst_element_factory_make ("videotestsrc", NULL); GstElement *sink = gst_element_factory_make ("xvimagesink", NULL); gst_bin_add_many (GST_BIN (pipeline), src, sink, NULL); gst_element_link (src, sink); // prepare the ui QWidget window; window.resize(320, 240); window.show(); WId xwinid = window.winId(); gst_video_overlay_set_window_handle (GST_VIDEO_OVERLAY (sink), xwinid); // run the pipeline GstStateChangeReturn sret = gst_element_set_state (pipeline, GST_STATE_PLAYING); if (sret == GST_STATE_CHANGE_FAILURE) { gst_element_set_state (pipeline, GST_STATE_NULL); gst_object_unref (pipeline); // Exit application QTimer::singleShot(0, QApplication::activeWindow(), SLOT(quit())); } int ret = app.exec(); window.hide(); gst_element_set_state (pipeline, GST_STATE_NULL); gst_object_unref (pipeline); return ret; } ``` # Implements [`VideoOverlayExt`](trait.VideoOverlayExt.html) Trait containing all `VideoOverlay` methods. # Implementors [`VideoOverlay`](struct.VideoOverlay.html) This helper shall be used by classes implementing the `VideoOverlay` interface that want the render rectangle to be controllable using properties. This helper will install "render-rectangle" property into the class. Feature: `v1_14` ## `oclass` The class on which the properties will be installed ## `last_prop_id` The first free property ID to use This helper shall be used by classes implementing the `VideoOverlay` interface that want the render rectangle to be controllable using properties. This helper will parse and set the render rectangle calling `VideoOverlay::set_render_rectangle`. Feature: `v1_14` ## `object` The instance on which the property is set ## `last_prop_id` The highest property ID. ## `property_id` The property ID ## `value` The `gobject::Value` to be set # Returns `true` if the `property_id` matches the GstVideoOverlay property Tell an overlay that it has been exposed. This will redraw the current frame in the drawable even if the pipeline is PAUSED. This will post a "have-window-handle" element message on the bus. This function should only be used by video overlay plugin developers. ## `handle` a platform-specific handle referencing the window Tell an overlay that it should handle events from the window system. These events are forwarded upstream as navigation events. In some window system, events are not propagated in the window hierarchy if a client is listening for them. This method allows you to disable events handling completely from the `VideoOverlay`. ## `handle_events` a `gboolean` indicating if events should be handled or not. This will post a "prepare-window-handle" element message on the bus to give applications an opportunity to call `VideoOverlay::set_window_handle` before a plugin creates its own window. This function should only be used by video overlay plugin developers. Configure a subregion as a video target within the window set by `VideoOverlay::set_window_handle`. If this is not used or not supported the video will fill the area of the window set as the overlay to 100%. By specifying the rectangle, the video can be overlayed to a specific region of that window only. After setting the new rectangle one should call `VideoOverlay::expose` to force a redraw. To unset the region pass -1 for the `width` and `height` parameters. This method is needed for non fullscreen video overlay in UI toolkits that do not support subwindows. ## `x` the horizontal offset of the render area inside the window ## `y` the vertical offset of the render area inside the window ## `width` the width of the render area inside the window ## `height` the height of the render area inside the window # Returns `false` if not supported by the sink. This will call the video overlay's set_window_handle method. You should use this method to tell to an overlay to display video output to a specific window (e.g. an XWindow on X11). Passing 0 as the `handle` will tell the overlay to stop using that window and create an internal one. ## `handle` a handle referencing the window. Enum value describing the available tiling modes. Unknown or unset tile mode Every four adjacent blocks - two horizontally and two vertically are grouped together and are located in memory in Z or flipped Z order. In case of odd rows, the last row of blocks is arranged in linear order. `field_count` must be 0 for progressive video and 1 or 2 for interlaced. A representation of a SMPTE time code. `hours` must be positive and less than 24. Will wrap around otherwise. `minutes` and `seconds` must be positive and less than 60. `frames` must be less than or equal to `config.fps_n` / `config.fps_d` These values are *NOT* automatically normalized. Feature: `v1_10` `field_count` is 0 for progressive, 1 or 2 for interlaced. `latest_daiy_jam` reference is stolen from caller. Feature: `v1_10` ## `fps_n` Numerator of the frame rate ## `fps_d` Denominator of the frame rate ## `latest_daily_jam` The latest daily jam of the `VideoTimeCode` ## `flags` `VideoTimeCodeFlags` ## `hours` the hours field of `VideoTimeCode` ## `minutes` the minutes field of `VideoTimeCode` ## `seconds` the seconds field of `VideoTimeCode` ## `frames` the frames field of `VideoTimeCode` ## `field_count` Interlaced video field count # Returns a new `VideoTimeCode` with the given values. The values are not checked for being in a valid range. To see if your timecode actually has valid content, use `VideoTimeCode::is_valid`. Feature: `v1_10` # Returns a new empty, invalid `VideoTimeCode` The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time. This might return a completely invalid timecode, use `VideoTimeCode::new_from_date_time_full` to ensure that you would get `None` instead in that case. Feature: `v1_12` ## `fps_n` Numerator of the frame rate ## `fps_d` Denominator of the frame rate ## `dt` `glib::DateTime` to convert ## `flags` `VideoTimeCodeFlags` ## `field_count` Interlaced video field count # Returns the `VideoTimeCode` representation of `dt`. The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time. Feature: `v1_16` ## `fps_n` Numerator of the frame rate ## `fps_d` Denominator of the frame rate ## `dt` `glib::DateTime` to convert ## `flags` `VideoTimeCodeFlags` ## `field_count` Interlaced video field count # Returns the `VideoTimeCode` representation of `dt`, or `None` if no valid timecode could be created. Feature: `v1_12` ## `tc_str` The string that represents the `VideoTimeCode` # Returns a new `VideoTimeCode` from the given string or `None` if the string could not be passed. Adds or subtracts `frames` amount of frames to `self`. tc needs to contain valid data, as verified by `VideoTimeCode::is_valid`. Feature: `v1_10` ## `frames` How many frames to add or subtract This makes a component-wise addition of `tc_inter` to `self`. For example, adding ("01:02:03:04", "00:01:00:00") will return "01:03:03:04". When it comes to drop-frame timecodes, adding ("00:00:00;00", "00:01:00:00") will return "00:01:00;02" because of drop-frame oddities. However, adding ("00:09:00;02", "00:01:00:00") will return "00:10:00;00" because this time we can have an exact minute. Feature: `v1_12` ## `tc_inter` The `VideoTimeCodeInterval` to add to `self`. The interval must contain valid values, except that for drop-frame timecode, it may also contain timecodes which would normally be dropped. These are then corrected to the next reasonable timecode. # Returns A new `VideoTimeCode` with `tc_inter` added or `None` if the interval can't be added. Initializes `self` with empty/zero/NULL values and frees any memory it might currently use. Feature: `v1_10` Compares `self` and `tc2`. If both have latest daily jam information, it is taken into account. Otherwise, it is assumed that the daily jam of both `self` and `tc2` was at the same time. Both time codes must be valid. Feature: `v1_10` ## `tc2` another valid `VideoTimeCode` # Returns 1 if `self` is after `tc2`, -1 if `self` is before `tc2`, 0 otherwise. Feature: `v1_10` # Returns a new `VideoTimeCode` with the same values as `self`. Feature: `v1_10` # Returns how many frames have passed since the daily jam of `self`. Frees `self`. Feature: `v1_10` Adds one frame to `self`. Feature: `v1_10` `field_count` is 0 for progressive, 1 or 2 for interlaced. `latest_daiy_jam` reference is stolen from caller. Initializes `self` with the given values. The values are not checked for being in a valid range. To see if your timecode actually has valid content, use `VideoTimeCode::is_valid`. Feature: `v1_10` ## `fps_n` Numerator of the frame rate ## `fps_d` Denominator of the frame rate ## `latest_daily_jam` The latest daily jam of the `VideoTimeCode` ## `flags` `VideoTimeCodeFlags` ## `hours` the hours field of `VideoTimeCode` ## `minutes` the minutes field of `VideoTimeCode` ## `seconds` the seconds field of `VideoTimeCode` ## `frames` the frames field of `VideoTimeCode` ## `field_count` Interlaced video field count The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time. Will assert on invalid parameters, use `VideoTimeCode::init_from_date_time_full` for being able to handle invalid parameters. Feature: `v1_12` ## `fps_n` Numerator of the frame rate ## `fps_d` Denominator of the frame rate ## `dt` `glib::DateTime` to convert ## `flags` `VideoTimeCodeFlags` ## `field_count` Interlaced video field count The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time. Feature: `v1_16` ## `fps_n` Numerator of the frame rate ## `fps_d` Denominator of the frame rate ## `dt` `glib::DateTime` to convert ## `flags` `VideoTimeCodeFlags` ## `field_count` Interlaced video field count # Returns `true` if `self` could be correctly initialized to a valid timecode Feature: `v1_10` # Returns whether `self` is a valid timecode (supported frame rate, hours/minutes/seconds/frames not overflowing) Feature: `v1_10` # Returns how many nsec have passed since the daily jam of `self`. The `self.config`->latest_daily_jam is required to be non-NULL. Feature: `v1_10` # Returns the `glib::DateTime` representation of `self` or `None` if `self` has no daily jam. Feature: `v1_10` # Returns the SMPTE ST 2059-1:2015 string representation of `self`. That will take the form hh:mm:ss:ff. The last separator (between seconds and frames) may vary: ';' for drop-frame, non-interlaced content and for drop-frame interlaced field 2 ',' for drop-frame interlaced field 1 ':' for non-drop-frame, non-interlaced content and for non-drop-frame interlaced field 2 '.' for non-drop-frame interlaced field 1 A representation of a difference between two `VideoTimeCode` instances. Will not necessarily correspond to a real timecode (e.g. 00:00:10;00) Feature: `v1_12` Feature: `v1_12` ## `hours` the hours field of `VideoTimeCodeInterval` ## `minutes` the minutes field of `VideoTimeCodeInterval` ## `seconds` the seconds field of `VideoTimeCodeInterval` ## `frames` the frames field of `VideoTimeCodeInterval` # Returns a new `VideoTimeCodeInterval` with the given values. `tc_inter_str` must only have ":" as separators. Feature: `v1_12` ## `tc_inter_str` The string that represents the `VideoTimeCodeInterval` # Returns a new `VideoTimeCodeInterval` from the given string or `None` if the string could not be passed. Initializes `self` with empty/zero/NULL values. Feature: `v1_12` Feature: `v1_12` # Returns a new `VideoTimeCodeInterval` with the same values as `self`. Frees `self`. Feature: `v1_12` Initializes `self` with the given values. Feature: `v1_12` ## `hours` the hours field of `VideoTimeCodeInterval` ## `minutes` the minutes field of `VideoTimeCodeInterval` ## `seconds` the seconds field of `VideoTimeCodeInterval` ## `frames` the frames field of `VideoTimeCodeInterval` The video transfer function defines the formula for converting between non-linear RGB (R'G'B') and linear RGB unknown transfer function linear RGB, gamma 1.0 curve Gamma 1.8 curve Gamma 2.0 curve Gamma 2.2 curve Gamma 2.2 curve with a linear segment in the lower range Gamma 2.2 curve with a linear segment in the lower range Gamma 2.4 curve with a linear segment in the lower range Gamma 2.8 curve Logarithmic transfer characteristic 100:1 range Logarithmic transfer characteristic 316.22777:1 range Gamma 2.2 curve with a linear segment in the lower range. Used for BT.2020 with 12 bits per component. Since: 1.6 Gamma 2.19921875. Since: 1.8