73 KiB
Implements
gst::BufferPoolExt
, gst::ObjectExt
, glib::object::ObjectExt
Create a new bufferpool that can allocate video frames. This bufferpool supports all the video bufferpool options.
Returns
a new gst::BufferPool
to allocate video frames
The various known types of Closed Caption (CC).
Unknown type of CC
CEA-608 as byte pairs. Note that
this format is not recommended since is does not specify to
which field the caption comes from and therefore assumes
it comes from the first field (and that there is no information
on the second field). Use VideoCaptionType::Cea708Raw
if you wish to store CEA-608 from two fields and prefix each byte pair
with 0xFC for the first field and 0xFD for the second field.
CEA-608 as byte triplets as defined in SMPTE S334-1 Annex A. The second and third byte of the byte triplet is the raw CEA608 data, the first byte is a bitfield: The top/7th bit is 0 for the second field, 1 for the first field, bit 6 and 5 are 0 and bits 4 to 0 are a 5 bit unsigned integer that represents the line offset relative to the base-line of the original image format (line 9 for 525-line field 1, line 272 for 525-line field 2, line 5 for 625-line field 1 and line 318 for 625-line field 2).
CEA-708 as cc_data byte triplets. They can also contain 608-in-708 and the first byte of each triplet has to be inspected for detecting the type.
CEA-708 (and optionally CEA-608) in a CDP (Caption Distribution Packet) defined by SMPTE S-334-2. Contains the whole CDP (starting with 0x9669).
Feature: v1_16
A VideoCodecFrame
represents a video frame both in raw and
encoded form.
Gets private data set on the frame by the subclass via
VideoCodecFrame::set_user_data
previously.
Returns
The previously set user_data
Increases the refcount of the given frame by one.
Returns
buf
Sets user_data
on the frame and the GDestroyNotify
that will be called when
the frame is freed. Allows to attach private data by the subclass to frames.
If a user_data
was previously set, then the previous set notify
will be called
before the user_data
is replaced.
user_data
private data
notify
a GDestroyNotify
Decreases the refcount of the frame. If the refcount reaches 0, the frame will be freed.
Structure representing the state of an incoming or outgoing video stream for encoders and decoders.
Decoders and encoders will receive such a state through their
respective set_format
vmethods.
Decoders and encoders can set the downstream state, by using the
VideoDecoder::set_output_state
() or
VideoEncoder::set_output_state
() methods.
Increases the refcount of the given state by one.
Returns
buf
Decreases the refcount of the state. If the refcount reaches 0, the state will be freed.
The color matrix is used to convert between Y'PbPr and non-linear RGB (R'G'B')
unknown matrix
identity matrix
FCC color matrix
ITU-R BT.709 color matrix
ITU-R BT.601 color matrix
SMPTE 240M color matrix
ITU-R BT.2020 color matrix. Since: 1.6
The color primaries define the how to transform linear RGB values to and from the CIE XYZ colorspace.
unknown color primaries
BT709 primaries
BT470M primaries
BT470BG primaries
SMPTE170M primaries
SMPTE240M primaries
Generic film
BT2020 primaries. Since: 1.6
Adobe RGB primaries. Since: 1.8
SMPTE ST 428 primaries. Since: 1.16
SMPTE RP 431 primaries. Since: 1.16
SMPTE EG 432 primaries. Since: 1.16
EBU 3213 primaries. Since: 1.16
Possible color range values. These constants are defined for 8 bit color values and can be scaled for other bit depths.
unknown range
[0..255] for 8 bit components
[16..235] for 8 bit components. Chroma has [16..240] range.
Structure describing the color info.
Parse the colorimetry string and update self
with the parsed
values.
color
a colorimetry string
Returns
true
if color
points to valid colorimetry info.
Compare the 2 colorimetry sets for equality
other
another VideoColorimetry
Returns
true
if self
and other
are equal.
Check if the colorimetry information in info
matches that of the
string color
.
color
a colorimetry string
Returns
true
if color
conveys the same colorimetry info as the color
information in info
.
Make a string representation of self
.
Returns
a string representation of self
.
This base class is for video decoders turning encoded data into raw video frames.
The GstVideoDecoder base class and derived subclasses should cooperate as follows:
Configuration
-
Initially, GstVideoDecoder calls
start
when the decoder element is activated, which allows the subclass to perform any global setup. -
GstVideoDecoder calls
set_format
to inform the subclass of caps describing input video data that it is about to receive, including possibly configuration data. While unlikely, it might be called more than once, if changing input parameters require reconfiguration. -
Incoming data buffers are processed as needed, described in Data Processing below.
-
GstVideoDecoder calls
stop
at end of all processing.
Data processing
-
The base class gathers input data, and optionally allows subclass to parse this into subsequently manageable chunks, typically corresponding to and referred to as 'frames'.
-
Each input frame is provided in turn to the subclass'
handle_frame
callback. The ownership of the frame is given to thehandle_frame
callback. -
If codec processing results in decoded data, the subclass should call
VideoDecoder::finish_frame
to have decoded data pushed. downstream. Otherwise, the subclass must callVideoDecoder::drop_frame
, to allow the base class to do timestamp and offset tracking, and possibly to requeue the frame for a later attempt in the case of reverse playback.
Shutdown phase
- The GstVideoDecoder class calls
stop
to inform the subclass that data parsing will be stopped.
Additional Notes
-
Seeking/Flushing
-
When the pipeline is seeked or otherwise flushed, the subclass is informed via a call to its
reset
callback, with the hard parameter set to true. This indicates the subclass should drop any internal data queues and timestamps and prepare for a fresh set of buffers to arrive for parsing and decoding. -
End Of Stream
-
At end-of-stream, the subclass
parse
function may be called some final times with the at_eos parameter set to true, indicating that the element should not expect any more data to be arriving, and it should parse and remaining frames and callVideoDecoder::have_frame
if possible.
The subclass is responsible for providing pad template caps for
source and sink pads. The pads need to be named "sink" and "src". It also
needs to provide information about the ouptput caps, when they are known.
This may be when the base class calls the subclass' set_format
function,
though it might be during decoding, before calling
VideoDecoder::finish_frame
. This is done via
VideoDecoder::set_output_state
The subclass is also responsible for providing (presentation) timestamps (likely based on corresponding input ones). If that is not applicable or possible, the base class provides limited framerate based interpolation.
Similarly, the base class provides some limited (legacy) seeking support
if specifically requested by the subclass, as full-fledged support
should rather be left to upstream demuxer, parser or alike. This simple
approach caters for seeking and duration reporting using estimated input
bitrates. To enable it, a subclass should call
VideoDecoderExt::set_estimate_rate
to enable handling of incoming
byte-streams.
The base class provides some support for reverse playback, in particular in case incoming data is not packetized or upstream does not provide fragments on keyframe boundaries. However, the subclass should then be prepared for the parsing and frame processing stage to occur separately (in normal forward processing, the latter immediately follows the former), The subclass also needs to ensure the parsing stage properly marks keyframes, unless it knows the upstream elements will do so properly for incoming data.
The bare minimum that a functional subclass needs to implement is:
-
Provide pad templates
-
Inform the base class of output caps via
VideoDecoder::set_output_state
-
Parse input data, if it is not considered packetized from upstream Data will be provided to
parse
which should invokeVideoDecoderExt::add_to_frame
andVideoDecoder::have_frame
to separate the data belonging to each video frame. -
Accept data in
handle_frame
and provide decoded results toVideoDecoder::finish_frame
, or callVideoDecoder::drop_frame
.
Implements
VideoDecoderExt
, gst::ElementExt
, gst::ObjectExt
, glib::object::ObjectExt
Trait containing all VideoDecoder
methods.
Implementors
Removes next n_bytes
of input data and adds it to currently parsed frame.
n_bytes
the number of bytes to add
Helper function that allocates a buffer to hold a video frame for self
's
current VideoCodecState
.
You should use VideoDecoder::allocate_output_frame
instead of this
function, if possible at all.
Returns
allocated buffer, or NULL if no buffer could be allocated (e.g. when downstream is flushing or shutting down)
Helper function that allocates a buffer to hold a video frame for self
's
current VideoCodecState
. Subclass should already have configured video
state and set src pad caps.
The buffer allocated here is owned by the frame and you should only keep references to the frame, not the buffer.
frame
a VideoCodecFrame
Returns
gst::FlowReturn::Ok
if an output buffer could be allocated
Same as VideoDecoder::allocate_output_frame
except it allows passing
gst::BufferPoolAcquireParams
to the sub call gst_buffer_pool_acquire_buffer.
Feature: v1_12
frame
a VideoCodecFrame
params
a gst::BufferPoolAcquireParams
Returns
gst::FlowReturn::Ok
if an output buffer could be allocated
Similar to VideoDecoder::finish_frame
, but drops frame
in any
case and posts a QoS message with the frame's details on the bus.
In any case, the frame is considered finished and released.
frame
the VideoCodecFrame
to drop
Returns
a gst::FlowReturn
, usually GST_FLOW_OK.
frame
should have a valid decoded data buffer, whose metadata fields
are then appropriately set according to frame data and pushed downstream.
If no output data is provided, frame
is considered skipped.
In any case, the frame is considered finished and released.
After calling this function the output buffer of the frame is to be considered read-only. This function will also change the metadata of the buffer.
frame
a decoded VideoCodecFrame
Returns
a gst::FlowReturn
resulting from sending data downstream
Lets VideoDecoder
sub-classes to know the memory allocator
used by the base class and its params
.
Unref the allocator
after use it.
allocator
the gst::Allocator
used
params
the
gst::AllocationParams
of allocator
Returns
the instance of the gst::BufferPool
used
by the decoder; free it after use it
Returns
currently configured byte to time conversion setting
Get a pending unfinished VideoCodecFrame
frame_number
system_frame_number of a frame
Returns
pending unfinished VideoCodecFrame
identified by frame_number
.
Get all pending unfinished VideoCodecFrame
Returns
pending unfinished VideoCodecFrame
.
Query the configured decoder latency. Results will be returned via
min_latency
and max_latency
.
min_latency
address of variable in which to store the
configured minimum latency, or None
max_latency
address of variable in which to store the
configured mximum latency, or None
Determines maximum possible decoding time for frame
that will
allow it to decode and arrive in time (as determined by QoS events).
In particular, a negative result means decoding in time is no longer possible
and should therefore occur as soon/skippy as possible.
frame
a VideoCodecFrame
Returns
max decoding time.
Returns
currently configured decoder tolerated error count.
Queries decoder required format handling.
Returns
true
if required format handling is enabled.
Get the oldest pending unfinished VideoCodecFrame
Returns
oldest pending unfinished VideoCodecFrame
.
Get the VideoCodecState
currently describing the output stream.
Returns
VideoCodecState
describing format of video data.
Queries whether input data is considered packetized or not by the base class.
Returns
TRUE if input data is considered packetized.
Returns the number of bytes previously added to the current frame
by calling VideoDecoderExt::add_to_frame
.
Returns
The number of bytes pending for the current frame
Returns
The current QoS proportion.
Gathers all data collected for currently parsed frame, gathers corresponding
metadata and passes it along for further processing, i.e. handle_frame
.
Returns
a gst::FlowReturn
Sets the audio decoder tags and how they should be merged with any
upstream stream tags. This will override any tags previously-set
with gst_audio_decoder_merge_tags
.
Note that this is provided for convenience, and the subclass is not required to use this and can still do tag handling on its own.
MT safe.
tags
a gst::TagList
to merge, or NULL to unset
previously-set tags
mode
the gst::TagMergeMode
to use, usually gst::TagMergeMode::Replace
Negotiate with downstream elements to currently configured VideoCodecState
.
Unmark GST_PAD_FLAG_NEED_RECONFIGURE in any case. But mark it again if
negotiate fails.
Returns
true
if the negotiation succeeded, else false
.
Returns caps that express caps
(or sink template caps if caps
== NULL)
restricted to resolution/format/... combinations supported by downstream
elements.
caps
initial caps
filter
filter caps
Returns
a gst::Caps
owned by caller
Similar to VideoDecoder::drop_frame
, but simply releases frame
without any processing other than removing it from list of pending frames,
after which it is considered finished and released.
frame
the VideoCodecFrame
to release
Allows baseclass to perform byte to time estimated conversion.
enabled
whether to enable byte to time conversion
Same as VideoDecoder::set_output_state
() but also allows you to also set
the interlacing mode.
Feature: v1_16
fmt
a VideoFormat
mode
A VideoInterlaceMode
width
The width in pixels
height
The height in pixels
reference
An optional reference VideoCodecState
Returns
the newly configured output state.
Lets VideoDecoder
sub-classes tell the baseclass what the decoder
latency is. Will also post a LATENCY message on the bus so the pipeline
can reconfigure its global latency.
min_latency
minimum latency
max_latency
maximum latency
Sets numbers of tolerated decoder errors, where a tolerated one is then only warned about, but more than tolerated will lead to fatal error. You can set -1 for never returning fatal errors. Default is set to GST_VIDEO_DECODER_MAX_ERRORS.
The '-1' option was added in 1.4
num
max tolerated errors
Configures decoder format needs. If enabled, subclass needs to be negotiated with format caps before it can process any data. It will then never be handed any data before it has been configured. Otherwise, it might be handed data without having been configured and is then expected being able to do so either by default or based on the input data.
enabled
new state
Creates a new VideoCodecState
with the specified fmt
, width
and height
as the output state for the decoder.
Any previously set output state on self
will be replaced by the newly
created one.
If the subclass wishes to copy over existing fields (like pixel aspec ratio,
or framerate) from an existing VideoCodecState
, it can be provided as a
reference
.
If the subclass wishes to override some fields from the output state (like
pixel-aspect-ratio or framerate) it can do so on the returned VideoCodecState
.
The new output state will only take effect (set on pads and buffers) starting
from the next call to VideoDecoder::finish_frame
().
fmt
a VideoFormat
width
The width in pixels
height
The height in pixels
reference
An optional reference VideoCodecState
Returns
the newly configured output state.
Allows baseclass to consider input data as packetized or not. If the
input is packetized, then the parse
method will not be called.
packetized
whether the input data should be considered as packetized.
Lets VideoDecoder
sub-classes decide if they want the sink pad
to use the default pad query handler to reply to accept-caps queries.
By setting this to true it is possible to further customize the default
handler with GST_PAD_SET_ACCEPT_INTERSECT
and
GST_PAD_SET_ACCEPT_TEMPLATE
use_
if the default pad accept-caps query handling should be used
This base class is for video encoders turning raw video into encoded video data.
GstVideoEncoder and subclass should cooperate as follows.
Configuration
- Initially, GstVideoEncoder calls
start
when the encoder element is activated, which allows subclass to perform any global setup. - GstVideoEncoder calls
set_format
to inform subclass of the format of input video data that it is about to receive. Subclass should setup for encoding and configure base class as appropriate (e.g. latency). While unlikely, it might be called more than once, if changing input parameters require reconfiguration. Baseclass will ensure that processing of current configuration is finished. - GstVideoEncoder calls
stop
at end of all processing.
Data processing
-
Base class collects input data and metadata into a frame and hands this to subclass'
handle_frame
. -
If codec processing results in encoded data, subclass should call
VideoEncoder::finish_frame
to have encoded data pushed downstream. -
If implemented, baseclass calls subclass
pre_push
just prior to pushing to allow subclasses to modify some metadata on the buffer. If it returns GST_FLOW_OK, the buffer is pushed downstream. -
GstVideoEncoderClass will handle both srcpad and sinkpad events. Sink events will be passed to subclass if
event
callback has been provided.
Shutdown phase
- GstVideoEncoder class calls
stop
to inform the subclass that data parsing will be stopped.
Subclass is responsible for providing pad template caps for
source and sink pads. The pads need to be named "sink" and "src". It should
also be able to provide fixed src pad caps in getcaps
by the time it calls
VideoEncoder::finish_frame
.
Things that subclass need to take care of:
- Provide pad templates
- Provide source pad caps before pushing the first buffer
- Accept data in
handle_frame
and provide encoded results toVideoEncoder::finish_frame
.
The VideoEncoder:qos
property will enable the Quality-of-Service
features of the encoder which gather statistics about the real-time
performance of the downstream elements. If enabled, subclasses can
use VideoEncoderExt::get_max_encode_time
to check if input frames
are already late and drop them right away to give a chance to the
pipeline to catch up.
Implements
VideoEncoderExt
, gst::ElementExt
, gst::ObjectExt
, glib::object::ObjectExt
Trait containing all VideoEncoder
methods.
Implementors
Helper function that allocates a buffer to hold an encoded video frame
for self
's current VideoCodecState
.
size
size of the buffer
Returns
allocated buffer
Helper function that allocates a buffer to hold an encoded video frame for self
's
current VideoCodecState
. Subclass should already have configured video
state and set src pad caps.
The buffer allocated here is owned by the frame and you should only keep references to the frame, not the buffer.
frame
a VideoCodecFrame
size
size of the buffer
Returns
gst::FlowReturn::Ok
if an output buffer could be allocated
frame
must have a valid encoded data buffer, whose metadata fields
are then appropriately set according to frame data or no buffer at
all if the frame should be dropped.
It is subsequently pushed downstream or provided to pre_push
.
In any case, the frame is considered finished and released.
After calling this function the output buffer of the frame is to be considered read-only. This function will also change the metadata of the buffer.
frame
an encoded VideoCodecFrame
Returns
a gst::FlowReturn
resulting from sending data downstream
Lets VideoEncoder
sub-classes to know the memory allocator
used by the base class and its params
.
Unref the allocator
after use it.
allocator
the gst::Allocator
used
params
the
gst::AllocationParams
of allocator
Get a pending unfinished VideoCodecFrame
frame_number
system_frame_number of a frame
Returns
pending unfinished VideoCodecFrame
identified by frame_number
.
Get all pending unfinished VideoCodecFrame
Returns
pending unfinished VideoCodecFrame
.
Query the configured encoding latency. Results will be returned via
min_latency
and max_latency
.
min_latency
address of variable in which to store the
configured minimum latency, or None
max_latency
address of variable in which to store the
configured maximum latency, or None
Determines maximum possible encoding time for frame
that will
allow it to encode and arrive in time (as determined by QoS events).
In particular, a negative result means encoding in time is no longer possible
and should therefore occur as soon/skippy as possible.
If no QoS events have been received from downstream, or if
VideoEncoder:qos
is disabled this function returns G_MAXINT64
.
Feature: v1_14
frame
a VideoCodecFrame
Returns
max decoding time.
Get the oldest unfinished pending VideoCodecFrame
Returns
oldest unfinished pending VideoCodecFrame
Get the current VideoCodecState
Returns
VideoCodecState
describing format of video data.
Checks if self
is currently configured to handle Quality-of-Service
events from downstream.
Feature: v1_14
Returns
true
if the encoder is configured to perform Quality-of-Service.
Sets the video encoder tags and how they should be merged with any
upstream stream tags. This will override any tags previously-set
with VideoEncoderExt::merge_tags
.
Note that this is provided for convenience, and the subclass is not required to use this and can still do tag handling on its own.
MT safe.
tags
a gst::TagList
to merge, or NULL to unset
previously-set tags
mode
the gst::TagMergeMode
to use, usually gst::TagMergeMode::Replace
Negotiate with downstream elements to currently configured VideoCodecState
.
Unmark GST_PAD_FLAG_NEED_RECONFIGURE in any case. But mark it again if
negotiate fails.
Returns
true
if the negotiation succeeded, else false
.
Returns caps that express caps
(or sink template caps if caps
== NULL)
restricted to resolution/format/... combinations supported by downstream
elements (e.g. muxers).
caps
initial caps
filter
filter caps
Returns
a gst::Caps
owned by caller
Set the codec headers to be sent downstream whenever requested.
headers
a list of gst::Buffer
containing the codec header
Informs baseclass of encoding latency.
min_latency
minimum latency
max_latency
maximum latency
Request minimal value for PTS passed to handle_frame.
For streams with reordered frames this can be used to ensure that there is enough time to accomodate first DTS, which may be less than first PTS
min_pts
minimal PTS that will be passed to handle_frame
Creates a new VideoCodecState
with the specified caps as the output state
for the encoder.
Any previously set output state on self
will be replaced by the newly
created one.
The specified caps
should not contain any resolution, pixel-aspect-ratio,
framerate, codec-data, .... Those should be specified instead in the returned
VideoCodecState
.
If the subclass wishes to copy over existing fields (like pixel aspect ratio,
or framerate) from an existing VideoCodecState
, it can be provided as a
reference
.
If the subclass wishes to override some fields from the output state (like
pixel-aspect-ratio or framerate) it can do so on the returned VideoCodecState
.
The new output state will only take effect (set on pads and buffers) starting
from the next call to VideoEncoder::finish_frame
().
caps
the gst::Caps
to use for the output
reference
An optional reference VideoCodecState
Returns
the newly configured output state.
Configures self
to handle Quality-of-Service events from downstream.
Feature: v1_14
enabled
the new qos value.
Field order of interlaced content. This is only valid for interlace-mode=interleaved and not interlace-mode=mixed. In the case of mixed or GST_VIDEO_FIELD_ORDER_UNKOWN, the field order is signalled via buffer flags.
unknown field order for interlaced content. The actual field order is signalled via buffer flags.
top field is first
bottom field is first
Feature: v1_12
Provides useful functions and a base class for video filters.
The videofilter will by default enable QoS on the parent GstBaseTransform to implement frame dropping.
Implements
gst_base::BaseTransformExt
, gst::ElementExt
, gst::ObjectExt
, glib::object::ObjectExt
Enum value describing the most common video formats.
See the GStreamer raw video format design document for details about the layout and packing of these formats in memory.
Unknown or unset video format id
Encoded video format. Only ever use that in caps for special video formats in combination with non-system memory GstCapsFeatures where it does not make sense to specify a real video format.
planar 4:2:0 YUV
planar 4:2:0 YVU (like I420 but UV planes swapped)
packed 4:2:2 YUV (Y0-U0-Y1-V0 Y2-U2-Y3-V2 Y4 ...)
packed 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...)
packed 4:4:4 YUV with alpha channel (A0-Y0-U0-V0 ...)
sparse rgb packed into 32 bit, space last
sparse reverse rgb packed into 32 bit, space last
sparse rgb packed into 32 bit, space first
sparse reverse rgb packed into 32 bit, space first
rgb with alpha channel last
reverse rgb with alpha channel last
rgb with alpha channel first
reverse rgb with alpha channel first
RGB packed into 24 bits without padding (R-G-B-R-G-B
)
reverse RGB packed into 24 bits without padding (B-G-R-B-G-R
)
planar 4:1:1 YUV
planar 4:2:2 YUV
packed 4:2:2 YUV (Y0-V0-Y1-U0 Y2-V2-Y3-U2 Y4 ...)
planar 4:4:4 YUV
packed 4:2:2 10-bit YUV, complex format
packed 4:2:2 16-bit YUV, Y0-U0-Y1-V1 order
planar 4:2:0 YUV with interleaved UV plane
planar 4:2:0 YUV with interleaved VU plane
8-bit grayscale
16-bit grayscale, most significant byte first
16-bit grayscale, least significant byte first
packed 4:4:4 YUV (Y-U-V ...)
rgb 5-6-5 bits per component
reverse rgb 5-6-5 bits per component
rgb 5-5-5 bits per component
reverse rgb 5-5-5 bits per component
packed 10-bit 4:2:2 YUV (U0-Y0-V0-Y1 U2-Y2-V2-Y3 U4 ...)
planar 4:4:2:0 AYUV
8-bit paletted RGB
planar 4:1:0 YUV
planar 4:1:0 YUV (like YUV9 but UV planes swapped)
packed 4:1:1 YUV (Cb-Y0-Y1-Cr-Y2-Y3 ...)
rgb with alpha channel first, 16 bits per channel
packed 4:4:4 YUV with alpha channel, 16 bits per channel (A0-Y0-U0-V0 ...)
packed 4:4:4 RGB, 10 bits per channel
planar 4:2:0 YUV, 10 bits per channel
planar 4:2:0 YUV, 10 bits per channel
planar 4:2:2 YUV, 10 bits per channel
planar 4:2:2 YUV, 10 bits per channel
planar 4:4:4 YUV, 10 bits per channel (Since: 1.2)
planar 4:4:4 YUV, 10 bits per channel (Since: 1.2)
planar 4:4:4 RGB, 8 bits per channel (Since: 1.2)
planar 4:4:4 RGB, 10 bits per channel (Since: 1.2)
planar 4:4:4 RGB, 10 bits per channel (Since: 1.2)
planar 4:2:2 YUV with interleaved UV plane (Since: 1.2)
planar 4:4:4 YUV with interleaved UV plane (Since: 1.2)
NV12 with 64x32 tiling in zigzag pattern (Since: 1.4)
planar 4:4:2:0 YUV, 10 bits per channel (Since: 1.6)
planar 4:4:2:0 YUV, 10 bits per channel (Since: 1.6)
planar 4:4:2:2 YUV, 10 bits per channel (Since: 1.6)
planar 4:4:2:2 YUV, 10 bits per channel (Since: 1.6)
planar 4:4:4:4 YUV, 10 bits per channel (Since: 1.6)
planar 4:4:4:4 YUV, 10 bits per channel (Since: 1.6)
planar 4:2:2 YUV with interleaved VU plane (Since: 1.6)
planar 4:2:0 YUV with interleaved UV plane, 10 bits per channel (Since: 1.10)
planar 4:2:0 YUV with interleaved UV plane, 10 bits per channel (Since: 1.10)
packed 4:4:4 YUV (U-Y-V ...) (Since: 1.10)
packed 4:2:2 YUV (V0-Y0-U0-Y1 V2-Y2-U2-Y3 V4 ...)
planar 4:4:4:4 ARGB, 8 bits per channel (Since: 1.12)
planar 4:4:4:4 ARGB, 10 bits per channel (Since: 1.12)
planar 4:4:4:4 ARGB, 10 bits per channel (Since: 1.12)
planar 4:4:4 RGB, 12 bits per channel (Since: 1.12)
planar 4:4:4 RGB, 12 bits per channel (Since: 1.12)
planar 4:4:4:4 ARGB, 12 bits per channel (Since: 1.12)
planar 4:4:4:4 ARGB, 12 bits per channel (Since: 1.12)
planar 4:2:0 YUV, 12 bits per channel (Since: 1.12)
planar 4:2:0 YUV, 12 bits per channel (Since: 1.12)
planar 4:2:2 YUV, 12 bits per channel (Since: 1.12)
planar 4:2:2 YUV, 12 bits per channel (Since: 1.12)
planar 4:4:4 YUV, 12 bits per channel (Since: 1.12)
planar 4:4:4 YUV, 12 bits per channel (Since: 1.12)
10-bit grayscale, packed into 32bit words (2 bits padding) (Since: 1.14)
10-bit variant of VideoFormat::Nv12
, packed into 32bit words (MSB 2 bits padding) (Since: 1.14)
10-bit variant of VideoFormat::Nv16
, packed into 32bit words (MSB 2 bits padding) (Since: 1.14)
Fully packed variant of NV12_10LE32 (Since: 1.16)
packed 4:2:2 YUV, 10 bits per channel (Since: 1.16)
packed 4:4:4 YUV, 10 bits per channel(A-V-Y-U...) (Since: 1.16)
packed 4:4:4 YUV with alpha channel (V0-U0-Y0-A0...) (Since: 1.16)
packed 4:4:4 RGB with alpha channel(B-G-R-A), 10 bits for R/G/B channel and MSB 2 bits for alpha channel (Since: 1.16)
Information for a video format.
A video frame obtained from VideoFrame::map
Copy the contents from src
to self
.
src
a VideoFrame
Returns
TRUE if the contents could be copied.
Copy the plane with index plane
from src
to self
.
src
a VideoFrame
plane
a plane
Returns
TRUE if the contents could be copied.
Use info
and buffer
to fill in the values of self
. self
is usually
allocated on the stack, and you will pass the address to the VideoFrame
structure allocated on the stack; VideoFrame::map
will then fill in
the structures with the various video-specific information you need to access
the pixels of the video buffer. You can then use accessor macros such as
GST_VIDEO_FRAME_COMP_DATA(), GST_VIDEO_FRAME_PLANE_DATA(),
GST_VIDEO_FRAME_COMP_STRIDE(), GST_VIDEO_FRAME_PLANE_STRIDE() etc.
to get to the pixels.
GstVideoFrame vframe;
...
// set RGB pixels to black one at a time
if (gst_video_frame_map (&vframe, video_info, video_buffer, GST_MAP_WRITE)) {
guint8 *pixels = GST_VIDEO_FRAME_PLANE_DATA (vframe, 0);
guint stride = GST_VIDEO_FRAME_PLANE_STRIDE (vframe, 0);
guint pixel_stride = GST_VIDEO_FRAME_COMP_PSTRIDE (vframe, 0);
for (h = 0; h < height; ++h) {
for (w = 0; w < width; ++w) {
guint8 *pixel = pixels + h * stride + w * pixel_stride;
memset (pixel, 0, pixel_stride);
}
}
gst_video_frame_unmap (&vframe);
}
...
All video planes of buffer
will be mapped and the pointers will be set in
self
->data.
The purpose of this function is to make it easy for you to get to the video
pixels in a generic way, without you having to worry too much about details
such as whether the video data is allocated in one contiguous memory chunk
or multiple memory chunks (e.g. one for each plane); or if custom strides
and custom plane offsets are used or not (as signalled by GstVideoMeta on
each buffer). This function will just fill the VideoFrame
structure
with the right values and if you use the accessor macros everything will
just work and you can access the data easily. It also maps the underlying
memory chunks for you.
info
a VideoInfo
buffer
the buffer to map
flags
gst::MapFlags
Returns
true
on success.
Use info
and buffer
to fill in the values of self
with the video frame
information of frame id
.
When id
is -1, the default frame is mapped. When id
!= -1, this function
will return false
when there is no GstVideoMeta with that id.
All video planes of buffer
will be mapped and the pointers will be set in
self
->data.
info
a VideoInfo
buffer
the buffer to map
id
the frame id to map
flags
gst::MapFlags
Returns
true
on success.
Unmap the memory previously mapped with gst_video_frame_map.
Information describing image properties. This information can be filled
in from GstCaps with VideoInfo::from_caps
. The information is also used
to store the specific video info when mapping a video frame with
VideoFrame::map
.
Use the provided macros to access the info in this structure.
Allocate a new VideoInfo
that is also initialized with
VideoInfo::init
.
Returns
a new VideoInfo
. free with VideoInfo::free
.
Adjust the offset and stride fields in self
so that the padding and
stride alignment in align
is respected.
Extra padding will be added to the right side when stride alignment padding
is required and align
will be updated with the new padding values.
align
alignment parameters
Returns
false
if alignment could not be applied, e.g. because the
size of a frame can't be represented as a 32 bit integer (Since: 1.12)
Converts among various gst::Format
types. This function handles
GST_FORMAT_BYTES, GST_FORMAT_TIME, and GST_FORMAT_DEFAULT. For
raw video, GST_FORMAT_DEFAULT corresponds to video frames. This
function can be used to handle pad queries of the type GST_QUERY_CONVERT.
src_format
gst::Format
of the src_value
src_value
value to convert
dest_format
gst::Format
of the dest_value
dest_value
pointer to destination value
Returns
TRUE if the conversion was successful.
Copy a GstVideoInfo structure.
Returns
a new VideoInfo
. free with gst_video_info_free.
Free a GstVideoInfo structure previously allocated with VideoInfo::new
or VideoInfo::copy
.
Parse caps
and update self
.
caps
a gst::Caps
Returns
TRUE if caps
could be parsed
Initialize self
with default values.
Compares two VideoInfo
and returns whether they are equal or not
other
a VideoInfo
Returns
true
if self
and other
are equal, else false
.
Set the default info for a video frame of format
and width
and height
.
Note: This initializes self
first, no values are preserved. This function
does not set the offsets correctly for interlaced vertically
subsampled formats.
format
the format
width
a width
height
a height
Returns
false
if the returned video info is invalid, e.g. because the
size of a frame can't be represented as a 32 bit integer (Since: 1.12)
Same as VideoInfo::set_format
but also allowing to set the interlaced
mode.
Feature: v1_16
format
the format
mode
a VideoInterlaceMode
width
a width
height
a height
Returns
false
if the returned video info is invalid, e.g. because the
size of a frame can't be represented as a 32 bit integer.
Convert the values of self
into a gst::Caps
.
Returns
a new gst::Caps
containing the info of self
.
The possible values of the VideoInterlaceMode
describing the interlace
mode of the stream.
all frames are progressive
2 fields are interleaved in one video frame. Extra buffer flags describe the field order.
frames contains both interlaced and progressive video, the buffer flags describe the frame and fields.
2 fields are stored in one buffer, use the frame ID to get access to the required field. For multiview (the 'views' property > 1) the fields of view N can be found at frame ID (N * 2) and (N * 2) + 1. Each field has only half the amount of lines as noted in the height property. This mode requires multiple GstVideoMeta metadata to describe the fields.
1 field is stored in one buffer,
GST_VIDEO_BUFFER_FLAG_TF
or GST_VIDEO_BUFFER_FLAG_BF
indicates if
the buffer is carrying the top or bottom field, respectively. The top and
bottom buffers are expected to alternate in the pipeline, with this mode
(Since: 1.16).
VideoMultiviewFramePacking
represents the subset of VideoMultiviewMode
values that can be applied to any video frame without needing extra metadata.
It can be used by elements that provide a property to override the
multiview interpretation of a video stream when the video doesn't contain
any markers.
This enum is used (for example) on playbin, to re-interpret a played
video stream as a stereoscopic video. The individual enum values are
equivalent to and have the same value as the matching VideoMultiviewMode
.
A special value indicating no frame packing info.
All frames are monoscopic.
All frames represent a left-eye view.
All frames represent a right-eye view.
Left and right eye views are provided in the left and right half of the frame respectively.
Left and right eye views are provided in the left and right half of the frame, but have been sampled using quincunx method, with half-pixel offset between the 2 views.
Alternating vertical columns of pixels represent the left and right eye view respectively.
Alternating horizontal rows of pixels represent the left and right eye view respectively.
The top half of the frame contains the left eye, and the bottom half the right eye.
Pixels are arranged with alternating pixels representing left and right eye views in a checkerboard fashion.
All possible stereoscopic 3D and multiview representations.
In conjunction with VideoMultiviewFlags
, describes how
multiview content is being transported in the stream.
A special value indicating no multiview information. Used in GstVideoInfo and other places to indicate that no specific multiview handling has been requested or provided. This value is never carried on caps.
All frames are monoscopic.
All frames represent a left-eye view.
All frames represent a right-eye view.
Left and right eye views are provided in the left and right half of the frame respectively.
Left and right eye views are provided in the left and right half of the frame, but have been sampled using quincunx method, with half-pixel offset between the 2 views.
Alternating vertical columns of pixels represent the left and right eye view respectively.
Alternating horizontal rows of pixels represent the left and right eye view respectively.
The top half of the frame contains the left eye, and the bottom half the right eye.
Pixels are arranged with alternating pixels representing left and right eye views in a checkerboard fashion.
Left and right eye views are provided in separate frames alternately.
Multiple
independent views are provided in separate frames in sequence.
This method only applies to raw video buffers at the moment.
Specific view identification is via the GstVideoMultiviewMeta
and VideoMeta
(s) on raw video buffers.
Multiple views are
provided as separate gst::Memory
framebuffers attached to each
gst::Buffer
, described by the GstVideoMultiviewMeta
and VideoMeta
(s)
The VideoOverlay
interface is used for 2 main purposes :
- To get a grab on the Window where the video sink element is going to render. This is achieved by either being informed about the Window identifier that the video sink element generated, or by forcing the video sink element to use a specific Window identifier for rendering.
- To force a redrawing of the latest video frame the video sink element
displayed on the Window. Indeed if the
gst::Pipeline
is ingst::State::Paused
state, moving the Window around will damage its content. Application developers will want to handle the Expose events themselves and force the video sink element to refresh the Window's content.
Using the Window created by the video sink is probably the simplest scenario, in some cases, though, it might not be flexible enough for application developers if they need to catch events such as mouse moves and button clicks.
Setting a specific Window identifier on the video sink element is the most
flexible solution but it has some issues. Indeed the application needs to set
its Window identifier at the right time to avoid internal Window creation
from the video sink element. To solve this issue a gst::Message
is posted on
the bus to inform the application that it should set the Window identifier
immediately. Here is an example on how to do that correctly:
static GstBusSyncReply
create_window (GstBus * bus, GstMessage * message, GstPipeline * pipeline)
{
// ignore anything but 'prepare-window-handle' element messages
if (!gst_is_video_overlay_prepare_window_handle_message (message))
return GST_BUS_PASS;
win = XCreateSimpleWindow (disp, root, 0, 0, 320, 240, 0, 0, 0);
XSetWindowBackgroundPixmap (disp, win, None);
XMapRaised (disp, win);
XSync (disp, FALSE);
gst_video_overlay_set_window_handle (GST_VIDEO_OVERLAY (GST_MESSAGE_SRC (message)),
win);
gst_message_unref (message);
return GST_BUS_DROP;
}
...
int
main (int argc, char **argv)
{
...
bus = gst_pipeline_get_bus (GST_PIPELINE (pipeline));
gst_bus_set_sync_handler (bus, (GstBusSyncHandler) create_window, pipeline,
NULL);
...
}
Two basic usage scenarios
There are two basic usage scenarios: in the simplest case, the application
uses playbin
or plasink
or knows exactly what particular element is used
for video output, which is usually the case when the application creates
the videosink to use (e.g. xvimagesink
, ximagesink
, etc.) itself; in this
case, the application can just create the videosink element, create and
realize the window to render the video on and then
call VideoOverlay::set_window_handle
directly with the XID or native
window handle, before starting up the pipeline.
As playbin
and playsink
implement the video overlay interface and proxy
it transparently to the actual video sink even if it is created later, this
case also applies when using these elements.
In the other and more common case, the application does not know in advance
what GStreamer video sink element will be used for video output. This is
usually the case when an element such as autovideosink
is used.
In this case, the video sink element itself is created
asynchronously from a GStreamer streaming thread some time after the
pipeline has been started up. When that happens, however, the video sink
will need to know right then whether to render onto an already existing
application window or whether to create its own window. This is when it
posts a prepare-window-handle message, and that is also why this message needs
to be handled in a sync bus handler which will be called from the streaming
thread directly (because the video sink will need an answer right then).
As response to the prepare-window-handle element message in the bus sync
handler, the application may use VideoOverlay::set_window_handle
to tell
the video sink to render onto an existing window surface. At this point the
application should already have obtained the window handle / XID, so it
just needs to set it. It is generally not advisable to call any GUI toolkit
functions or window system functions from the streaming thread in which the
prepare-window-handle message is handled, because most GUI toolkits and
windowing systems are not thread-safe at all and a lot of care would be
required to co-ordinate the toolkit and window system calls of the
different threads (Gtk+ users please note: prior to Gtk+ 2.18
GDK_WINDOW_XID() was just a simple structure access, so generally fine to do
within the bus sync handler; this macro was changed to a function call in
Gtk+ 2.18 and later, which is likely to cause problems when called from a
sync handler; see below for a better approach without GDK_WINDOW_XID()
used in the callback).
GstVideoOverlay and Gtk+
#include <gst/video/videooverlay.h>
#include <gtk/gtk.h>
#ifdef GDK_WINDOWING_X11
#include <gdk/gdkx.h> // for GDK_WINDOW_XID
#endif
#ifdef GDK_WINDOWING_WIN32
#include <gdk/gdkwin32.h> // for GDK_WINDOW_HWND
#endif
...
static guintptr video_window_handle = 0;
...
static GstBusSyncReply
bus_sync_handler (GstBus * bus, GstMessage * message, gpointer user_data)
{
// ignore anything but 'prepare-window-handle' element messages
if (!gst_is_video_overlay_prepare_window_handle_message (message))
return GST_BUS_PASS;
if (video_window_handle != 0) {
GstVideoOverlay *overlay;
// GST_MESSAGE_SRC (message) will be the video sink element
overlay = GST_VIDEO_OVERLAY (GST_MESSAGE_SRC (message));
gst_video_overlay_set_window_handle (overlay, video_window_handle);
} else {
g_warning ("Should have obtained video_window_handle by now!");
}
gst_message_unref (message);
return GST_BUS_DROP;
}
...
static void
video_widget_realize_cb (GtkWidget * widget, gpointer data)
{
#if GTK_CHECK_VERSION(2,18,0)
// Tell Gtk+/Gdk to create a native window for this widget instead of
// drawing onto the parent widget.
// This is here just for pedagogical purposes, GDK_WINDOW_XID will call
// it as well in newer Gtk versions
if (!gdk_window_ensure_native (widget->window))
g_error ("Couldn't create native window needed for GstVideoOverlay!");
#endif
#ifdef GDK_WINDOWING_X11
{
gulong xid = GDK_WINDOW_XID (gtk_widget_get_window (video_window));
video_window_handle = xid;
}
#endif
#ifdef GDK_WINDOWING_WIN32
{
HWND wnd = GDK_WINDOW_HWND (gtk_widget_get_window (video_window));
video_window_handle = (guintptr) wnd;
}
#endif
}
...
int
main (int argc, char **argv)
{
GtkWidget *video_window;
GtkWidget *app_window;
...
app_window = gtk_window_new (GTK_WINDOW_TOPLEVEL);
...
video_window = gtk_drawing_area_new ();
g_signal_connect (video_window, "realize",
G_CALLBACK (video_widget_realize_cb), NULL);
gtk_widget_set_double_buffered (video_window, FALSE);
...
// usually the video_window will not be directly embedded into the
// application window like this, but there will be many other widgets
// and the video window will be embedded in one of them instead
gtk_container_add (GTK_CONTAINER (ap_window), video_window);
...
// show the GUI
gtk_widget_show_all (app_window);
// realize window now so that the video window gets created and we can
// obtain its XID/HWND before the pipeline is started up and the videosink
// asks for the XID/HWND of the window to render onto
gtk_widget_realize (video_window);
// we should have the XID/HWND now
g_assert (video_window_handle != 0);
...
// set up sync handler for setting the xid once the pipeline is started
bus = gst_pipeline_get_bus (GST_PIPELINE (pipeline));
gst_bus_set_sync_handler (bus, (GstBusSyncHandler) bus_sync_handler, NULL,
NULL);
gst_object_unref (bus);
...
gst_element_set_state (pipeline, GST_STATE_PLAYING);
...
}
GstVideoOverlay and Qt
#include <glib.h>
#include <gst/gst.h>
#include <gst/video/videooverlay.h>
#include <QApplication>
#include <QTimer>
#include <QWidget>
int main(int argc, char *argv[])
{
if (!g_thread_supported ())
g_thread_init (NULL);
gst_init (&argc, &argv);
QApplication app(argc, argv);
app.connect(&app, SIGNAL(lastWindowClosed()), &app, SLOT(quit ()));
// prepare the pipeline
GstElement *pipeline = gst_pipeline_new ("xvoverlay");
GstElement *src = gst_element_factory_make ("videotestsrc", NULL);
GstElement *sink = gst_element_factory_make ("xvimagesink", NULL);
gst_bin_add_many (GST_BIN (pipeline), src, sink, NULL);
gst_element_link (src, sink);
// prepare the ui
QWidget window;
window.resize(320, 240);
window.show();
WId xwinid = window.winId();
gst_video_overlay_set_window_handle (GST_VIDEO_OVERLAY (sink), xwinid);
// run the pipeline
GstStateChangeReturn sret = gst_element_set_state (pipeline,
GST_STATE_PLAYING);
if (sret == GST_STATE_CHANGE_FAILURE) {
gst_element_set_state (pipeline, GST_STATE_NULL);
gst_object_unref (pipeline);
// Exit application
QTimer::singleShot(0, QApplication::activeWindow(), SLOT(quit()));
}
int ret = app.exec();
window.hide();
gst_element_set_state (pipeline, GST_STATE_NULL);
gst_object_unref (pipeline);
return ret;
}
Implements
Trait containing all VideoOverlay
methods.
Implementors
This helper shall be used by classes implementing the VideoOverlay
interface that want the render rectangle to be controllable using
properties. This helper will install "render-rectangle" property into the
class.
Feature: v1_14
oclass
The class on which the properties will be installed
last_prop_id
The first free property ID to use
This helper shall be used by classes implementing the VideoOverlay
interface that want the render rectangle to be controllable using
properties. This helper will parse and set the render rectangle calling
VideoOverlay::set_render_rectangle
.
Feature: v1_14
object
The instance on which the property is set
last_prop_id
The highest property ID.
property_id
The property ID
value
The gobject::Value
to be set
Returns
true
if the property_id
matches the GstVideoOverlay property
Tell an overlay that it has been exposed. This will redraw the current frame in the drawable even if the pipeline is PAUSED.
This will post a "have-window-handle" element message on the bus.
This function should only be used by video overlay plugin developers.
handle
a platform-specific handle referencing the window
Tell an overlay that it should handle events from the window system. These
events are forwarded upstream as navigation events. In some window system,
events are not propagated in the window hierarchy if a client is listening
for them. This method allows you to disable events handling completely
from the VideoOverlay
.
handle_events
a gboolean
indicating if events should be handled or not.
This will post a "prepare-window-handle" element message on the bus
to give applications an opportunity to call
VideoOverlay::set_window_handle
before a plugin creates its own
window.
This function should only be used by video overlay plugin developers.
Configure a subregion as a video target within the window set by
VideoOverlay::set_window_handle
. If this is not used or not supported
the video will fill the area of the window set as the overlay to 100%.
By specifying the rectangle, the video can be overlayed to a specific region
of that window only. After setting the new rectangle one should call
VideoOverlay::expose
to force a redraw. To unset the region pass -1 for
the width
and height
parameters.
This method is needed for non fullscreen video overlay in UI toolkits that do not support subwindows.
x
the horizontal offset of the render area inside the window
y
the vertical offset of the render area inside the window
width
the width of the render area inside the window
height
the height of the render area inside the window
Returns
false
if not supported by the sink.
This will call the video overlay's set_window_handle method. You
should use this method to tell to an overlay to display video output to a
specific window (e.g. an XWindow on X11). Passing 0 as the handle
will
tell the overlay to stop using that window and create an internal one.
handle
a handle referencing the window.
Enum value describing the available tiling modes.
Unknown or unset tile mode
Every four adjacent blocks - two horizontally and two vertically are grouped together and are located in memory in Z or flipped Z order. In case of odd rows, the last row of blocks is arranged in linear order.
field_count
must be 0 for progressive video and 1 or 2 for interlaced.
A representation of a SMPTE time code.
hours
must be positive and less than 24. Will wrap around otherwise.
minutes
and seconds
must be positive and less than 60.
frames
must be less than or equal to config.fps_n
/ config.fps_d
These values are NOT automatically normalized.
Feature: v1_10
field_count
is 0 for progressive, 1 or 2 for interlaced.
latest_daiy_jam
reference is stolen from caller.
Feature: v1_10
fps_n
Numerator of the frame rate
fps_d
Denominator of the frame rate
latest_daily_jam
The latest daily jam of the VideoTimeCode
flags
VideoTimeCodeFlags
hours
the hours field of VideoTimeCode
minutes
the minutes field of VideoTimeCode
seconds
the seconds field of VideoTimeCode
frames
the frames field of VideoTimeCode
field_count
Interlaced video field count
Returns
a new VideoTimeCode
with the given values.
The values are not checked for being in a valid range. To see if your
timecode actually has valid content, use VideoTimeCode::is_valid
.
Feature: v1_10
Returns
a new empty, invalid VideoTimeCode
The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time.
This might return a completely invalid timecode, use
VideoTimeCode::new_from_date_time_full
to ensure
that you would get None
instead in that case.
Feature: v1_12
fps_n
Numerator of the frame rate
fps_d
Denominator of the frame rate
dt
glib::DateTime
to convert
flags
VideoTimeCodeFlags
field_count
Interlaced video field count
Returns
the VideoTimeCode
representation of dt
.
The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time.
Feature: v1_16
fps_n
Numerator of the frame rate
fps_d
Denominator of the frame rate
dt
glib::DateTime
to convert
flags
VideoTimeCodeFlags
field_count
Interlaced video field count
Returns
the VideoTimeCode
representation of dt
, or None
if
no valid timecode could be created.
Feature: v1_12
tc_str
The string that represents the VideoTimeCode
Returns
a new VideoTimeCode
from the given string or None
if the string could not be passed.
Adds or subtracts frames
amount of frames to self
. tc needs to
contain valid data, as verified by VideoTimeCode::is_valid
.
Feature: v1_10
frames
How many frames to add or subtract
This makes a component-wise addition of tc_inter
to self
. For example,
adding ("01:02:03:04", "00:01:00:00") will return "01:03:03:04".
When it comes to drop-frame timecodes,
adding ("00:00:00;00", "00:01:00:00") will return "00:01:00;02"
because of drop-frame oddities. However,
adding ("00:09:00;02", "00:01:00:00") will return "00:10:00;00"
because this time we can have an exact minute.
Feature: v1_12
tc_inter
The VideoTimeCodeInterval
to add to self
.
The interval must contain valid values, except that for drop-frame
timecode, it may also contain timecodes which would normally
be dropped. These are then corrected to the next reasonable timecode.
Returns
A new VideoTimeCode
with tc_inter
added or None
if the interval can't be added.
Initializes self
with empty/zero/NULL values and frees any memory
it might currently use.
Feature: v1_10
Compares self
and tc2
. If both have latest daily jam information, it is
taken into account. Otherwise, it is assumed that the daily jam of both
self
and tc2
was at the same time. Both time codes must be valid.
Feature: v1_10
tc2
another valid VideoTimeCode
Returns
1 if self
is after tc2
, -1 if self
is before tc2
, 0 otherwise.
Feature: v1_10
Returns
a new VideoTimeCode
with the same values as self
.
Feature: v1_10
Returns
how many frames have passed since the daily jam of self
.
Frees self
.
Feature: v1_10
Adds one frame to self
.
Feature: v1_10
field_count
is 0 for progressive, 1 or 2 for interlaced.
latest_daiy_jam
reference is stolen from caller.
Initializes self
with the given values.
The values are not checked for being in a valid range. To see if your
timecode actually has valid content, use VideoTimeCode::is_valid
.
Feature: v1_10
fps_n
Numerator of the frame rate
fps_d
Denominator of the frame rate
latest_daily_jam
The latest daily jam of the VideoTimeCode
flags
VideoTimeCodeFlags
hours
the hours field of VideoTimeCode
minutes
the minutes field of VideoTimeCode
seconds
the seconds field of VideoTimeCode
frames
the frames field of VideoTimeCode
field_count
Interlaced video field count
The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time.
Will assert on invalid parameters, use VideoTimeCode::init_from_date_time_full
for being able to handle invalid parameters.
Feature: v1_12
fps_n
Numerator of the frame rate
fps_d
Denominator of the frame rate
dt
glib::DateTime
to convert
flags
VideoTimeCodeFlags
field_count
Interlaced video field count
The resulting config->latest_daily_jam is set to midnight, and timecode is set to the given time.
Feature: v1_16
fps_n
Numerator of the frame rate
fps_d
Denominator of the frame rate
dt
glib::DateTime
to convert
flags
VideoTimeCodeFlags
field_count
Interlaced video field count
Returns
true
if self
could be correctly initialized to a valid timecode
Feature: v1_10
Returns
whether self
is a valid timecode (supported frame rate,
hours/minutes/seconds/frames not overflowing)
Feature: v1_10
Returns
how many nsec have passed since the daily jam of self
.
The self.config
->latest_daily_jam is required to be non-NULL.
Feature: v1_10
Returns
the glib::DateTime
representation of self
or None
if self
has no daily jam.
Feature: v1_10
Returns
the SMPTE ST 2059-1:2015 string representation of self
. That will
take the form hh:mm:ss:ff. The last separator (between seconds and frames)
may vary:
';' for drop-frame, non-interlaced content and for drop-frame interlaced field 2 ',' for drop-frame interlaced field 1 ':' for non-drop-frame, non-interlaced content and for non-drop-frame interlaced field 2 '.' for non-drop-frame interlaced field 1
A representation of a difference between two VideoTimeCode
instances.
Will not necessarily correspond to a real timecode (e.g. 00:00:10;00)
Feature: v1_12
Feature: v1_12
hours
the hours field of VideoTimeCodeInterval
minutes
the minutes field of VideoTimeCodeInterval
seconds
the seconds field of VideoTimeCodeInterval
frames
the frames field of VideoTimeCodeInterval
Returns
a new VideoTimeCodeInterval
with the given values.
tc_inter_str
must only have ":" as separators.
Feature: v1_12
tc_inter_str
The string that represents the VideoTimeCodeInterval
Returns
a new VideoTimeCodeInterval
from the given string
or None
if the string could not be passed.
Initializes self
with empty/zero/NULL values.
Feature: v1_12
Feature: v1_12
Returns
a new VideoTimeCodeInterval
with the same values as self
.
Frees self
.
Feature: v1_12
Initializes self
with the given values.
Feature: v1_12
hours
the hours field of VideoTimeCodeInterval
minutes
the minutes field of VideoTimeCodeInterval
seconds
the seconds field of VideoTimeCodeInterval
frames
the frames field of VideoTimeCodeInterval
The video transfer function defines the formula for converting between non-linear RGB (R'G'B') and linear RGB
unknown transfer function
linear RGB, gamma 1.0 curve
Gamma 1.8 curve
Gamma 2.0 curve
Gamma 2.2 curve
Gamma 2.2 curve with a linear segment in the lower range
Gamma 2.2 curve with a linear segment in the lower range
Gamma 2.4 curve with a linear segment in the lower range
Gamma 2.8 curve
Logarithmic transfer characteristic 100:1 range
Logarithmic transfer characteristic 316.22777:1 range
Gamma 2.2 curve with a linear segment in the lower range. Used for BT.2020 with 12 bits per component. Since: 1.6
Gamma 2.19921875. Since: 1.8