/* Gstreamer * Copyright (C) <2011> Intel Corporation * Copyright (C) <2011> Collabora Ltd. * Copyright (C) <2011> Thibault Saunier * * Some bits C-c,C-v'ed and s/4/3 from h264parse and videoparsers/h264parse.c: * Copyright (C) <2010> Mark Nauwelaerts * Copyright (C) <2010> Collabora Multimedia * Copyright (C) <2010> Nokia Corporation * * (C) 2005 Michal Benes * (C) 2008 Wim Taymans * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ #ifndef __GST_H264_PARSER_H__ #define __GST_H264_PARSER_H__ #ifndef GST_USE_UNSTABLE_API #warning "The H.264 parsing library is unstable API and may change in future." #warning "You can define GST_USE_UNSTABLE_API to avoid this warning." #endif #include #include G_BEGIN_DECLS #define GST_H264_MAX_SPS_COUNT 32 #define GST_H264_MAX_PPS_COUNT 256 #define GST_H264_MAX_VIEW_COUNT 1024 #define GST_H264_MAX_VIEW_ID (GST_H264_MAX_VIEW_COUNT - 1) #define GST_H264_IS_P_SLICE(slice) (((slice)->type % 5) == GST_H264_P_SLICE) #define GST_H264_IS_B_SLICE(slice) (((slice)->type % 5) == GST_H264_B_SLICE) #define GST_H264_IS_I_SLICE(slice) (((slice)->type % 5) == GST_H264_I_SLICE) #define GST_H264_IS_SP_SLICE(slice) (((slice)->type % 5) == GST_H264_SP_SLICE) #define GST_H264_IS_SI_SLICE(slice) (((slice)->type % 5) == GST_H264_SI_SLICE) /** * GST_H264_IS_SVC_NALU: * @nalu: a #GstH264NalUnit * * Check if @nalu is a scalable extension NAL unit. * * Since: 1.6 */ #define GST_H264_IS_SVC_NALU(nalu) \ ((nalu)->extension_type == GST_H264_NAL_EXTENSION_SVC) /** * GST_H264_IS_MVC_NALU: * @nalu: a #GstH264NalUnit * * Check if @nalu is a multiview extension NAL unit. * * Since: 1.6 */ #define GST_H264_IS_MVC_NALU(nalu) \ ((nalu)->extension_type == GST_H264_NAL_EXTENSION_MVC) /** * GstH264Profile: * @GST_H264_PROFILE_BASELINE: Baseline profile (A.2.1) * @GST_H264_PROFILE_MAIN: Main profile (A.2.2) * @GST_H264_PROFILE_EXTENDED: Extended profile (A.2.3) * @GST_H264_PROFILE_HIGH: High profile (A.2.4), * or Progressive High profile (A.2.4.1), or Constrained High profile (A.2.4.2) * depending on constraint_set4_flag and constraint_set5_flag * @GST_H264_PROFILE_HIGH10: High 10 profile (A.2.5) or High 10 Intra * profile (A.2.8), or Progressive High 10 profile (A.2.5.1) depending on * constraint_set3_flag and constraint_set4_flag * @GST_H264_PROFILE_HIGH_422: High 4:2:2 profile (A.2.6) or High * 4:2:2 Intra profile (A.2.9), depending on constraint_set3_flag * @GST_H264_PROFILE_HIGH_444: High 4:4:4 Predictive profile (A.2.7) * or High 4:4:4 Intra profile (A.2.10), depending on the value of * constraint_set3_flag * @GST_H264_PROFILE_MULTIVIEW_HIGH: Multiview High profile (H.10.1.1) * @GST_H264_PROFILE_STEREO_HIGH: Stereo High profile (H.10.1.2) * @GST_H264_PROFILE_SCALABLE_BASELINE: Scalable Baseline profile (G.10.1.1) * @GST_H264_PROFILE_SCALABLE_HIGH: Scalable High profile (G.10.1.2) * or Scalable High Intra profile (G.10.1.3), depending on the value * of constraint_set3_flag * * H.264 Profiles. * * Since: 1.2 */ typedef enum { GST_H264_PROFILE_BASELINE = 66, GST_H264_PROFILE_MAIN = 77, GST_H264_PROFILE_EXTENDED = 88, GST_H264_PROFILE_HIGH = 100, GST_H264_PROFILE_HIGH10 = 110, GST_H264_PROFILE_HIGH_422 = 122, GST_H264_PROFILE_HIGH_444 = 244, GST_H264_PROFILE_MULTIVIEW_HIGH = 118, GST_H264_PROFILE_STEREO_HIGH = 128, GST_H264_PROFILE_SCALABLE_BASELINE = 83, GST_H264_PROFILE_SCALABLE_HIGH = 86 } GstH264Profile; /** * GstH264NalUnitType: * @GST_H264_NAL_UNKNOWN: Unknown nal type * @GST_H264_NAL_SLICE: Slice nal * @GST_H264_NAL_SLICE_DPA: DPA slice nal * @GST_H264_NAL_SLICE_DPB: DPB slice nal * @GST_H264_NAL_SLICE_DPC: DPC slice nal * @GST_H264_NAL_SLICE_IDR: DPR slice nal * @GST_H264_NAL_SEI: Supplemental enhancement information (SEI) nal unit * @GST_H264_NAL_SPS: Sequence parameter set (SPS) nal unit * @GST_H264_NAL_PPS: Picture parameter set (PPS) nal unit * @GST_H264_NAL_AU_DELIMITER: Access unit (AU) delimiter nal unit * @GST_H264_NAL_SEQ_END: End of sequence nal unit * @GST_H264_NAL_STREAM_END: End of stream nal unit * @GST_H264_NAL_FILLER_DATA: Filler data nal lunit * @GST_H264_NAL_SPS_EXT: Sequence parameter set (SPS) extension NAL unit * @GST_H264_NAL_PREFIX_UNIT: Prefix NAL unit * @GST_H264_NAL_SUBSET_SPS: Subset sequence parameter set (SSPS) NAL unit * @GST_H264_NAL_DEPTH_SPS: Depth parameter set (DPS) NAL unit * @GST_H264_NAL_SLICE_AUX: Auxiliary coded picture without partitioning NAL unit * @GST_H264_NAL_SLICE_EXT: Coded slice extension NAL unit * @GST_H264_NAL_SLICE_DEPTH: Coded slice extension for depth or 3D-AVC texture view * * Indicates the type of H264 Nal Units */ typedef enum { GST_H264_NAL_UNKNOWN = 0, GST_H264_NAL_SLICE = 1, GST_H264_NAL_SLICE_DPA = 2, GST_H264_NAL_SLICE_DPB = 3, GST_H264_NAL_SLICE_DPC = 4, GST_H264_NAL_SLICE_IDR = 5, GST_H264_NAL_SEI = 6, GST_H264_NAL_SPS = 7, GST_H264_NAL_PPS = 8, GST_H264_NAL_AU_DELIMITER = 9, GST_H264_NAL_SEQ_END = 10, GST_H264_NAL_STREAM_END = 11, GST_H264_NAL_FILLER_DATA = 12, GST_H264_NAL_SPS_EXT = 13, GST_H264_NAL_PREFIX_UNIT = 14, GST_H264_NAL_SUBSET_SPS = 15, GST_H264_NAL_DEPTH_SPS = 16, GST_H264_NAL_SLICE_AUX = 19, GST_H264_NAL_SLICE_EXT = 20, GST_H264_NAL_SLICE_DEPTH = 21 } GstH264NalUnitType; /** * GstH264NalUnitExtensionType: * @GST_H264_NAL_EXTENSION_NONE: No NAL unit header extension is available * @GST_H264_NAL_EXTENSION_SVC: NAL unit header extension for SVC (Annex G) * @GST_H264_NAL_EXTENSION_MVC: NAL unit header extension for MVC (Annex H) * * Indicates the type of H.264 NAL unit extension. * * Since: 1.6 */ typedef enum { GST_H264_NAL_EXTENSION_NONE = 0, GST_H264_NAL_EXTENSION_SVC, GST_H264_NAL_EXTENSION_MVC, } GstH264NalUnitExtensionType; /** * GstH264ParserResult: * @GST_H264_PARSER_OK: The parsing succeeded * @GST_H264_PARSER_BROKEN_DATA: The data to parse is broken * @GST_H264_PARSER_BROKEN_LINK: The link to structure needed for the parsing couldn't be found * @GST_H264_PARSER_ERROR: An error occurred when parsing * @GST_H264_PARSER_NO_NAL: No NAL unit found during the parsing * @GST_H264_PARSER_NO_NAL_END: Start of the NAL unit found, but not the end. * This will be returned if no start/sync marker for the next NAL unit was * found. In this case the parser will assume that the end of the data is * also the end of the NAL unit. Whether this assumption is correct or not * depends on the context, which only the caller can know, which is why a * special result value is returned in this case. If the data is NAL-aligned * then #GST_H264_PARSER_NO_NAL_END can be treated just like * #GST_H264_PARSER_OK. If the data is not guaranteed to be NAL-aligned, * then the caller probably wants to collect more data until there's another * sync marker or the end of the stream has been reached. * * The result of parsing H264 data. */ typedef enum { GST_H264_PARSER_OK, GST_H264_PARSER_BROKEN_DATA, GST_H264_PARSER_BROKEN_LINK, GST_H264_PARSER_ERROR, GST_H264_PARSER_NO_NAL, GST_H264_PARSER_NO_NAL_END } GstH264ParserResult; /** * GstH264FramePackingType: * @GST_H264_FRAME_PACKING_NONE: A complete 2D frame without any frame packing * @GST_H264_FRAME_PACKING_CHECKERBOARD_INTERLEAVING: Checkerboard * based interleaving * @GST_H264_FRAME_PACKING_COLUMN_INTERLEAVING: Column based interleaving * @GST_H264_FRAME_PACKING_ROW_INTERLEAVING: Row based interleaving * @GST_H264_FRAME_PACKING_SIDE_BY_SIDE: Side-by-side packing * @GST_H264_FRMAE_PACKING_TOP_BOTTOM: Top-Bottom packing * @GST_H264_FRAME_PACKING_TEMPORAL_INTERLEAVING: Temporal interleaving * * Frame packing arrangement types. * * Since: 1.6 */ typedef enum { GST_H264_FRAME_PACKING_NONE = 6, GST_H264_FRAME_PACKING_CHECKERBOARD_INTERLEAVING = 0, GST_H264_FRAME_PACKING_COLUMN_INTERLEAVING = 1, GST_H264_FRAME_PACKING_ROW_INTERLEAVING = 2, GST_H264_FRAME_PACKING_SIDE_BY_SIDE = 3, GST_H264_FRMAE_PACKING_TOP_BOTTOM = 4, GST_H264_FRAME_PACKING_TEMPORAL_INTERLEAVING = 5 } GstH264FramePackingType; /** * GstH264SEIPayloadType: * @GST_H264_SEI_BUF_PERIOD: Buffering Period SEI Message * @GST_H264_SEI_PIC_TIMING: Picture Timing SEI Message * @GST_H264_SEI_REGISTERED_USER_DATA: Registered user data (D.2.5) * @GST_H264_SEI_RECOVERY_POINT: Recovery Point SEI Message (D.2.7) * @GST_H264_SEI_STEREO_VIDEO_INFO: stereo video info SEI message (Since: 1.6) * @GST_H264_SEI_FRAME_PACKING: Frame Packing Arrangement (FPA) message that * contains the 3D arrangement for stereoscopic 3D video (Since: 1.6) * ... * * The type of SEI message. */ typedef enum { GST_H264_SEI_BUF_PERIOD = 0, GST_H264_SEI_PIC_TIMING = 1, GST_H264_SEI_REGISTERED_USER_DATA = 4, GST_H264_SEI_RECOVERY_POINT = 6, GST_H264_SEI_STEREO_VIDEO_INFO = 21, GST_H264_SEI_FRAME_PACKING = 45 /* and more... */ } GstH264SEIPayloadType; /** * GstH264SEIPicStructType: * @GST_H264_SEI_PIC_STRUCT_FRAME: Picture is a frame * @GST_H264_SEI_PIC_STRUCT_TOP_FIELD: Top field of frame * @GST_H264_SEI_PIC_STRUCT_BOTTOM_FIELD: Bottom field of frame * @GST_H264_SEI_PIC_STRUCT_TOP_BOTTOM: Top bottom field of frame * @GST_H264_SEI_PIC_STRUCT_BOTTOM_TOP: bottom top field of frame * @GST_H264_SEI_PIC_STRUCT_TOP_BOTTOM_TOP: top bottom top field of frame * @GST_H264_SEI_PIC_STRUCT_BOTTOM_TOP_BOTTOM: bottom top bottom field of frame * @GST_H264_SEI_PIC_STRUCT_FRAME_DOUBLING: indicates that the frame should * be displayed two times consecutively * @GST_H264_SEI_PIC_STRUCT_FRAME_TRIPLING: indicates that the frame should be * displayed three times consecutively * * SEI pic_struct type */ typedef enum { GST_H264_SEI_PIC_STRUCT_FRAME = 0, GST_H264_SEI_PIC_STRUCT_TOP_FIELD = 1, GST_H264_SEI_PIC_STRUCT_BOTTOM_FIELD = 2, GST_H264_SEI_PIC_STRUCT_TOP_BOTTOM = 3, GST_H264_SEI_PIC_STRUCT_BOTTOM_TOP = 4, GST_H264_SEI_PIC_STRUCT_TOP_BOTTOM_TOP = 5, GST_H264_SEI_PIC_STRUCT_BOTTOM_TOP_BOTTOM = 6, GST_H264_SEI_PIC_STRUCT_FRAME_DOUBLING = 7, GST_H264_SEI_PIC_STRUCT_FRAME_TRIPLING = 8 } GstH264SEIPicStructType; /** * GstH264SliceType: * * Type of Picture slice */ typedef enum { GST_H264_P_SLICE = 0, GST_H264_B_SLICE = 1, GST_H264_I_SLICE = 2, GST_H264_SP_SLICE = 3, GST_H264_SI_SLICE = 4, GST_H264_S_P_SLICE = 5, GST_H264_S_B_SLICE = 6, GST_H264_S_I_SLICE = 7, GST_H264_S_SP_SLICE = 8, GST_H264_S_SI_SLICE = 9 } GstH264SliceType; /** * GstH264CtType * * Mapping of ct_type to source picture scan * * Since: 1.16 */ typedef enum { GST_H264_CT_TYPE_PROGRESSIVE = 0, GST_H264_CT_TYPE_INTERLACED = 1, GST_H264_CT_TYPE_UNKNOWN = 2, } GstCtType; typedef struct _GstH264NalParser GstH264NalParser; typedef struct _GstH264NalUnit GstH264NalUnit; typedef struct _GstH264NalUnitExtensionMVC GstH264NalUnitExtensionMVC; typedef struct _GstH264SPSExtMVCView GstH264SPSExtMVCView; typedef struct _GstH264SPSExtMVCLevelValue GstH264SPSExtMVCLevelValue; typedef struct _GstH264SPSExtMVCLevelValueOp GstH264SPSExtMVCLevelValueOp; typedef struct _GstH264SPSExtMVC GstH264SPSExtMVC; typedef struct _GstH264SPS GstH264SPS; typedef struct _GstH264PPS GstH264PPS; typedef struct _GstH264HRDParams GstH264HRDParams; typedef struct _GstH264VUIParams GstH264VUIParams; typedef struct _GstH264RefPicListModification GstH264RefPicListModification; typedef struct _GstH264DecRefPicMarking GstH264DecRefPicMarking; typedef struct _GstH264RefPicMarking GstH264RefPicMarking; typedef struct _GstH264PredWeightTable GstH264PredWeightTable; typedef struct _GstH264SliceHdr GstH264SliceHdr; typedef struct _GstH264ClockTimestamp GstH264ClockTimestamp; typedef struct _GstH264PicTiming GstH264PicTiming; typedef struct _GstH264RegisteredUserData GstH264RegisteredUserData; typedef struct _GstH264BufferingPeriod GstH264BufferingPeriod; typedef struct _GstH264RecoveryPoint GstH264RecoveryPoint; typedef struct _GstH264StereoVideoInfo GstH264StereoVideoInfo; typedef struct _GstH264FramePacking GstH264FramePacking; typedef struct _GstH264SEIMessage GstH264SEIMessage; /** * GstH264NalUnitExtensionMVC: * @non_idr_flag: If equal to 0, it specifies that the current access * unit is an IDR access unit * @priority_id: The priority identifier for the NAL unit * @view_id: The view identifier for the NAL unit * @temporal_id: The temporal identifier for the NAL unit * @anchor_pic_flag: If equal to 1, it specifies that the current * access unit is an anchor access unit * @inter_view_flag: If equal to 0, it specifies that the current view * component is not used for inter-view prediction by any other view * component in the current access unit * * Since: 1.6 */ struct _GstH264NalUnitExtensionMVC { guint8 non_idr_flag; guint8 priority_id; guint16 view_id; guint8 temporal_id; guint8 anchor_pic_flag; guint8 inter_view_flag; }; /** * GstH264NalUnit: * @ref_idc: not equal to 0 specifies that the content of the NAL unit * contains a sequence parameter set, a sequence parameter set * extension, a subset sequence parameter set, a picture parameter * set, a slice of a reference picture, a slice data partition of a * reference picture, or a prefix NAL unit preceding a slice of a * reference picture. * @type: A #GstH264NalUnitType * @idr_pic_flag: calculated idr_pic_flag * @size: The size of the NAL unit starting from @offset, thus * including the header bytes. e.g. @type (nal_unit_type), * but not the start code. * @offset: The offset of the first byte of the NAL unit header, * just after the start code. * @sc_offset: The offset of the first byte of the start code of * the NAL unit. * @valid: If the NAL unit is valid, which means it has * already been parsed * @data: The data array from which the NAL unit has been parsed, * into which the offset and sc_offset apply. * @header_bytes: The size of the NALU header in bytes. The NALU * header is the 1-byte type code, and for extension / prefix NALs * includes the extension header bytes. @offset + @header_bytes is * therefore the first byte of the actual packet payload. * (Since: 1.6) * @extension_type: the extension type for prefix NAL/MVC/SVC * (Since: 1.6) * * Structure defining the NAL unit headers */ struct _GstH264NalUnit { guint16 ref_idc; guint16 type; /* calculated values */ guint8 idr_pic_flag; guint size; guint offset; guint sc_offset; gboolean valid; guint8 *data; guint8 header_bytes; guint8 extension_type; union { GstH264NalUnitExtensionMVC mvc; } extension; }; /** * GstH264HRDParams: * @cpb_cnt_minus1: plus 1 specifies the number of alternative * CPB specifications in the bitstream * @bit_rate_scale: specifies the maximum input bit rate of the * SchedSelIdx-th CPB * @cpb_size_scale: specifies the CPB size of the SchedSelIdx-th CPB * @guint32 bit_rate_value_minus1: specifies the maximum input bit rate for the * SchedSelIdx-th CPB * @cpb_size_value_minus1: is used together with cpb_size_scale to specify the * SchedSelIdx-th CPB size * @cbr_flag: Specifies if running in constant or intermittent bit rate mode * @initial_cpb_removal_delay_length_minus1: specifies the length in bits of * the cpb_removal_delay syntax element * @cpb_removal_delay_length_minus1: specifies the length in bits of the * dpb_output_delay syntax element * @dpb_output_delay_length_minus1: >0 specifies the length in bits of the time_offset syntax element. * =0 specifies that the time_offset syntax element is not present * @time_offset_length: Length of the time offset * * Defines the HRD parameters */ struct _GstH264HRDParams { guint8 cpb_cnt_minus1; guint8 bit_rate_scale; guint8 cpb_size_scale; guint32 bit_rate_value_minus1[32]; guint32 cpb_size_value_minus1[32]; guint8 cbr_flag[32]; guint8 initial_cpb_removal_delay_length_minus1; guint8 cpb_removal_delay_length_minus1; guint8 dpb_output_delay_length_minus1; guint8 time_offset_length; }; /** * GstH264VUIParams: * @aspect_ratio_info_present_flag: %TRUE specifies that aspect_ratio_idc is present. * %FALSE specifies that aspect_ratio_idc is not present * @aspect_ratio_idc specifies the value of the sample aspect ratio of the luma samples * @sar_width indicates the horizontal size of the sample aspect ratio * @sar_height indicates the vertical size of the sample aspect ratio * @overscan_info_present_flag: %TRUE overscan_appropriate_flag is present %FALSE otherwise * @overscan_appropriate_flag: %TRUE indicates that the cropped decoded pictures * output are suitable for display using overscan. %FALSE the cropped decoded pictures * output contain visually important information * @video_signal_type_present_flag: %TRUE specifies that video_format, video_full_range_flag and * colour_description_present_flag are present. * @video_format: indicates the representation of the picture * @video_full_range_flag: indicates the black level and range of the luma and chroma signals * @colour_description_present_flag: %TRUE specifies that colour_primaries, * transfer_characteristics and matrix_coefficients are present * @colour_primaries: indicates the chromaticity coordinates of the source primaries * @transfer_characteristics: indicates the opto-electronic transfer characteristic * @matrix_coefficients: describes the matrix coefficients used in deriving luma and chroma signals * @chroma_loc_info_present_flag: %TRUE specifies that chroma_sample_loc_type_top_field and * chroma_sample_loc_type_bottom_field are present, %FALSE otherwise * @chroma_sample_loc_type_top_field: specify the location of chroma for top field * @chroma_sample_loc_type_bottom_field specify the location of chroma for bottom field * @timing_info_present_flag: %TRUE specifies that num_units_in_tick, * time_scale and fixed_frame_rate_flag are present in the bitstream * @num_units_in_tick: is the number of time units of a clock operating at the frequency time_scale Hz * time_scale: is the number of time units that pass in one second * @fixed_frame_rate_flag: %TRUE indicates that the temporal distance between the HRD output times * of any two consecutive pictures in output order is constrained as specified in the spec, %FALSE * otherwise. * @nal_hrd_parameters_present_flag: %TRUE if NAL HRD parameters exist in the bitstream * @vcl_hrd_parameters_present_flag: %TRUE if VCL HRD parameters exist in the bitstream * @low_delay_hrd_flag: specifies the HRD operational mode * @pic_struct_present_flag: %TRUE specifies that picture timing SEI messages are present or not * @bitstream_restriction_flag: %TRUE specifies that the following coded video sequence bitstream restriction * parameters are present * @motion_vectors_over_pic_boundaries_flag: %FALSE indicates that no sample outside the * picture boundaries and no sample at a fractional sample position, %TRUE indicates that one or more * samples outside picture boundaries may be used in inter prediction * @max_bytes_per_pic_denom: indicates a number of bytes not exceeded by the sum of the sizes of * the VCL NAL units associated with any coded picture in the coded video sequence. * @max_bits_per_mb_denom: indicates the maximum number of coded bits of macroblock_layer * @log2_max_mv_length_horizontal: indicate the maximum absolute value of a decoded horizontal * motion vector component * @log2_max_mv_length_vertical: indicate the maximum absolute value of a decoded vertical * motion vector component * @num_reorder_frames: indicates the maximum number of frames, complementary field pairs, * or non-paired fields that precede any frame, * @max_dec_frame_buffering: specifies the required size of the HRD decoded picture buffer in * units of frame buffers. * * The structure representing the VUI parameters. */ struct _GstH264VUIParams { guint8 aspect_ratio_info_present_flag; guint8 aspect_ratio_idc; /* if aspect_ratio_idc == 255 */ guint16 sar_width; guint16 sar_height; guint8 overscan_info_present_flag; /* if overscan_info_present_flag */ guint8 overscan_appropriate_flag; guint8 video_signal_type_present_flag; guint8 video_format; guint8 video_full_range_flag; guint8 colour_description_present_flag; guint8 colour_primaries; guint8 transfer_characteristics; guint8 matrix_coefficients; guint8 chroma_loc_info_present_flag; guint8 chroma_sample_loc_type_top_field; guint8 chroma_sample_loc_type_bottom_field; guint8 timing_info_present_flag; /* if timing_info_present_flag */ guint32 num_units_in_tick; guint32 time_scale; guint8 fixed_frame_rate_flag; guint8 nal_hrd_parameters_present_flag; /* if nal_hrd_parameters_present_flag */ GstH264HRDParams nal_hrd_parameters; guint8 vcl_hrd_parameters_present_flag; /* if vcl_hrd_parameters_present_flag */ GstH264HRDParams vcl_hrd_parameters; guint8 low_delay_hrd_flag; guint8 pic_struct_present_flag; guint8 bitstream_restriction_flag; /* if bitstream_restriction_flag */ guint8 motion_vectors_over_pic_boundaries_flag; guint32 max_bytes_per_pic_denom; guint32 max_bits_per_mb_denom; guint32 log2_max_mv_length_horizontal; guint32 log2_max_mv_length_vertical; guint32 num_reorder_frames; guint32 max_dec_frame_buffering; /* calculated values */ guint par_n; guint par_d; }; /** * GstH264SPSExtMVCView: * @num_anchor_refs_l0: specifies the number of view components for * inter-view prediction in the initialized RefPicList0 in decoding * anchor view components. * @anchor_ref_l0: specifies the view_id for inter-view prediction in * the initialized RefPicList0 in decoding anchor view components. * @num_anchor_refs_l1: specifies the number of view components for * inter-view prediction in the initialized RefPicList1 in decoding * anchor view components. * @anchor_ref_l1: specifies the view_id for inter-view prediction in * the initialized RefPicList1 in decoding anchor view components. * @num_non_anchor_refs_l0: specifies the number of view components * for inter-view prediction in the initialized RefPicList0 in * decoding non-anchor view components. * @non_anchor_ref_l0: specifies the view_id for inter-view prediction * in the initialized RefPicList0 in decoding non-anchor view * components. * @num_non_anchor_refs_l1: specifies the number of view components * for inter-view prediction in the initialized RefPicList1 in * decoding non-anchor view components. * @non_anchor_ref_l1: specifies the view_id for inter-view prediction * in the initialized RefPicList1 in decoding non-anchor view * components. * * Represents inter-view dependency relationships for the coded video * sequence. * * Since: 1.6 */ struct _GstH264SPSExtMVCView { guint16 view_id; guint8 num_anchor_refs_l0; guint16 anchor_ref_l0[15]; guint8 num_anchor_refs_l1; guint16 anchor_ref_l1[15]; guint8 num_non_anchor_refs_l0; guint16 non_anchor_ref_l0[15]; guint8 num_non_anchor_refs_l1; guint16 non_anchor_ref_l1[15]; }; /** * GstH264SPSExtMVCLevelValueOp: * * Represents an operation point for the coded video sequence. * * Since: 1.6 */ struct _GstH264SPSExtMVCLevelValueOp { guint8 temporal_id; guint16 num_target_views_minus1; guint16 *target_view_id; guint16 num_views_minus1; }; /** * GstH264SPSExtMVCLevelValue: * @level_idc: specifies the level value signalled for the coded video * sequence * @num_applicable_ops_minus1: plus 1 specifies the number of * operation points to which the level indicated by level_idc applies * @applicable_op: specifies the applicable operation point * * Represents level values for a subset of the operation points for * the coded video sequence. * * Since: 1.6 */ struct _GstH264SPSExtMVCLevelValue { guint8 level_idc; guint16 num_applicable_ops_minus1; GstH264SPSExtMVCLevelValueOp *applicable_op; }; /** * GstH264SPSExtMVC: * @num_views_minus1: plus 1 specifies the maximum number of coded * views in the coded video sequence * @view: array of #GstH264SPSExtMVCView * @num_level_values_signalled_minus1: plus 1 specifies the number of * level values signalled for the coded video sequence. * @level_value: array of #GstH264SPSExtMVCLevelValue * * Represents the parsed `seq_parameter_set_mvc_extension()`. * * Since: 1.6 */ struct _GstH264SPSExtMVC { guint16 num_views_minus1; GstH264SPSExtMVCView *view; guint8 num_level_values_signalled_minus1; GstH264SPSExtMVCLevelValue *level_value; }; /** * GstH264SPS: * @id: The ID of the sequence parameter set * @profile_idc: indicate the profile to which the coded video sequence conforms * * H264 Sequence Parameter Set (SPS) */ struct _GstH264SPS { gint id; guint8 profile_idc; guint8 constraint_set0_flag; guint8 constraint_set1_flag; guint8 constraint_set2_flag; guint8 constraint_set3_flag; guint8 constraint_set4_flag; guint8 constraint_set5_flag; guint8 level_idc; guint8 chroma_format_idc; guint8 separate_colour_plane_flag; guint8 bit_depth_luma_minus8; guint8 bit_depth_chroma_minus8; guint8 qpprime_y_zero_transform_bypass_flag; guint8 scaling_matrix_present_flag; guint8 scaling_lists_4x4[6][16]; guint8 scaling_lists_8x8[6][64]; guint8 log2_max_frame_num_minus4; guint8 pic_order_cnt_type; /* if pic_order_cnt_type == 0 */ guint8 log2_max_pic_order_cnt_lsb_minus4; /* else if pic_order_cnt_type == 1 */ guint8 delta_pic_order_always_zero_flag; gint32 offset_for_non_ref_pic; gint32 offset_for_top_to_bottom_field; guint8 num_ref_frames_in_pic_order_cnt_cycle; gint32 offset_for_ref_frame[255]; guint32 num_ref_frames; guint8 gaps_in_frame_num_value_allowed_flag; guint32 pic_width_in_mbs_minus1; guint32 pic_height_in_map_units_minus1; guint8 frame_mbs_only_flag; guint8 mb_adaptive_frame_field_flag; guint8 direct_8x8_inference_flag; guint8 frame_cropping_flag; /* if frame_cropping_flag */ guint32 frame_crop_left_offset; guint32 frame_crop_right_offset; guint32 frame_crop_top_offset; guint32 frame_crop_bottom_offset; guint8 vui_parameters_present_flag; /* if vui_parameters_present_flag */ GstH264VUIParams vui_parameters; /* calculated values */ guint8 chroma_array_type; guint32 max_frame_num; gint width, height; gint crop_rect_width, crop_rect_height; gint crop_rect_x, crop_rect_y; gint fps_num_removed, fps_den_removed; /* FIXME: remove */ gboolean valid; /* Subset SPS extensions */ guint8 extension_type; union { GstH264SPSExtMVC mvc; } extension; }; /** * GstH264PPS: * * H264 Picture Parameter Set */ struct _GstH264PPS { gint id; GstH264SPS *sequence; guint8 entropy_coding_mode_flag; guint8 pic_order_present_flag; guint32 num_slice_groups_minus1; /* if num_slice_groups_minus1 > 0 */ guint8 slice_group_map_type; /* and if slice_group_map_type == 0 */ guint32 run_length_minus1[8]; /* or if slice_group_map_type == 2 */ guint32 top_left[8]; guint32 bottom_right[8]; /* or if slice_group_map_type == (3, 4, 5) */ guint8 slice_group_change_direction_flag; guint32 slice_group_change_rate_minus1; /* or if slice_group_map_type == 6 */ guint32 pic_size_in_map_units_minus1; guint8 *slice_group_id; guint8 num_ref_idx_l0_active_minus1; guint8 num_ref_idx_l1_active_minus1; guint8 weighted_pred_flag; guint8 weighted_bipred_idc; gint8 pic_init_qp_minus26; gint8 pic_init_qs_minus26; gint8 chroma_qp_index_offset; guint8 deblocking_filter_control_present_flag; guint8 constrained_intra_pred_flag; guint8 redundant_pic_cnt_present_flag; guint8 transform_8x8_mode_flag; guint8 scaling_lists_4x4[6][16]; guint8 scaling_lists_8x8[6][64]; guint8 second_chroma_qp_index_offset; gboolean valid; }; struct _GstH264RefPicListModification { guint8 modification_of_pic_nums_idc; union { /* if modification_of_pic_nums_idc == 0 || 1 */ guint32 abs_diff_pic_num_minus1; /* if modification_of_pic_nums_idc == 2 */ guint32 long_term_pic_num; /* if modification_of_pic_nums_idc == 4 || 5 */ guint32 abs_diff_view_idx_minus1; } value; }; struct _GstH264PredWeightTable { guint8 luma_log2_weight_denom; guint8 chroma_log2_weight_denom; gint16 luma_weight_l0[32]; gint8 luma_offset_l0[32]; /* if seq->ChromaArrayType != 0 */ gint16 chroma_weight_l0[32][2]; gint8 chroma_offset_l0[32][2]; /* if slice->slice_type % 5 == 1 */ gint16 luma_weight_l1[32]; gint8 luma_offset_l1[32]; /* and if seq->ChromaArrayType != 0 */ gint16 chroma_weight_l1[32][2]; gint8 chroma_offset_l1[32][2]; }; struct _GstH264RefPicMarking { guint8 memory_management_control_operation; guint32 difference_of_pic_nums_minus1; guint32 long_term_pic_num; guint32 long_term_frame_idx; guint32 max_long_term_frame_idx_plus1; }; struct _GstH264DecRefPicMarking { /* if slice->nal_unit.IdrPicFlag */ guint8 no_output_of_prior_pics_flag; guint8 long_term_reference_flag; guint8 adaptive_ref_pic_marking_mode_flag; GstH264RefPicMarking ref_pic_marking[10]; guint8 n_ref_pic_marking; }; struct _GstH264SliceHdr { guint32 first_mb_in_slice; guint32 type; GstH264PPS *pps; /* if seq->separate_colour_plane_flag */ guint8 colour_plane_id; guint16 frame_num; guint8 field_pic_flag; guint8 bottom_field_flag; /* if nal_unit.type == 5 */ guint16 idr_pic_id; /* if seq->pic_order_cnt_type == 0 */ guint16 pic_order_cnt_lsb; /* if seq->pic_order_present_flag && !field_pic_flag */ gint32 delta_pic_order_cnt_bottom; gint32 delta_pic_order_cnt[2]; guint8 redundant_pic_cnt; /* if slice_type == B_SLICE */ guint8 direct_spatial_mv_pred_flag; guint8 num_ref_idx_l0_active_minus1; guint8 num_ref_idx_l1_active_minus1; guint8 ref_pic_list_modification_flag_l0; guint8 n_ref_pic_list_modification_l0; GstH264RefPicListModification ref_pic_list_modification_l0[32]; guint8 ref_pic_list_modification_flag_l1; guint8 n_ref_pic_list_modification_l1; GstH264RefPicListModification ref_pic_list_modification_l1[32]; GstH264PredWeightTable pred_weight_table; /* if nal_unit.ref_idc != 0 */ GstH264DecRefPicMarking dec_ref_pic_marking; guint8 cabac_init_idc; gint8 slice_qp_delta; gint8 slice_qs_delta; guint8 disable_deblocking_filter_idc; gint8 slice_alpha_c0_offset_div2; gint8 slice_beta_offset_div2; guint16 slice_group_change_cycle; /* calculated values */ guint32 max_pic_num; gboolean valid; /* Size of the slice_header() in bits */ guint header_size; /* Number of emulation prevention bytes (EPB) in this slice_header() */ guint n_emulation_prevention_bytes; /* Since: 1.18 */ guint8 num_ref_idx_active_override_flag; guint8 sp_for_switch_flag; }; struct _GstH264ClockTimestamp { guint8 ct_type; guint8 nuit_field_based_flag; guint8 counting_type; guint8 discontinuity_flag; guint8 cnt_dropped_flag; guint8 n_frames; guint8 seconds_flag; guint8 seconds_value; guint8 minutes_flag; guint8 minutes_value; guint8 hours_flag; guint8 hours_value; guint32 time_offset; }; /** * GstH264FramePacking: * * Since: 1.6 */ struct _GstH264FramePacking { guint32 frame_packing_id; guint8 frame_packing_cancel_flag; guint8 frame_packing_type; /* GstH264FramePackingType */ guint8 quincunx_sampling_flag; guint8 content_interpretation_type; guint8 spatial_flipping_flag; guint8 frame0_flipped_flag; guint8 field_views_flag; guint8 current_frame_is_frame0_flag; guint8 frame0_self_contained_flag; guint8 frame1_self_contained_flag; guint8 frame0_grid_position_x; guint8 frame0_grid_position_y; guint8 frame1_grid_position_x; guint8 frame1_grid_position_y; guint16 frame_packing_repetition_period; }; /** * GstH264StereoVideoInfo: * * Since: 1.6 */ struct _GstH264StereoVideoInfo { guint8 field_views_flag; guint8 top_field_is_left_view_flag; guint8 current_frame_is_left_view_flag; guint8 next_frame_is_second_view_flag; guint8 left_view_self_contained_flag; guint8 right_view_self_contained_flag; }; struct _GstH264PicTiming { guint32 cpb_removal_delay; guint32 dpb_output_delay; guint8 pic_struct_present_flag; /* if pic_struct_present_flag */ guint8 pic_struct; guint8 clock_timestamp_flag[3]; GstH264ClockTimestamp clock_timestamp[3]; }; struct _GstH264RegisteredUserData { guint8 country_code; guint8 country_code_extension; const guint8 *data; guint size; }; struct _GstH264BufferingPeriod { GstH264SPS *sps; /* seq->vui_parameters->nal_hrd_parameters_present_flag */ guint32 nal_initial_cpb_removal_delay[32]; guint32 nal_initial_cpb_removal_delay_offset[32]; /* seq->vui_parameters->vcl_hrd_parameters_present_flag */ guint32 vcl_initial_cpb_removal_delay[32]; guint32 vcl_initial_cpb_removal_delay_offset[32]; }; struct _GstH264RecoveryPoint { guint32 recovery_frame_cnt; guint8 exact_match_flag; guint8 broken_link_flag; guint8 changing_slice_group_idc; }; struct _GstH264SEIMessage { GstH264SEIPayloadType payloadType; union { GstH264BufferingPeriod buffering_period; GstH264PicTiming pic_timing; GstH264RegisteredUserData registered_user_data; GstH264RecoveryPoint recovery_point; GstH264StereoVideoInfo stereo_video_info; GstH264FramePacking frame_packing; /* ... could implement more */ } payload; }; /** * GstH264NalParser: * * H264 NAL Parser (opaque structure). */ struct _GstH264NalParser { /*< private >*/ GstH264SPS sps[GST_H264_MAX_SPS_COUNT]; GstH264PPS pps[GST_H264_MAX_PPS_COUNT]; GstH264SPS *last_sps; GstH264PPS *last_pps; }; GST_CODEC_PARSERS_API GstH264NalParser *gst_h264_nal_parser_new (void); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_identify_nalu (GstH264NalParser *nalparser, const guint8 *data, guint offset, gsize size, GstH264NalUnit *nalu); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_identify_nalu_unchecked (GstH264NalParser *nalparser, const guint8 *data, guint offset, gsize size, GstH264NalUnit *nalu); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_identify_nalu_avc (GstH264NalParser *nalparser, const guint8 *data, guint offset, gsize size, guint8 nal_length_size, GstH264NalUnit *nalu); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_parse_nal (GstH264NalParser *nalparser, GstH264NalUnit *nalu); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_parse_slice_hdr (GstH264NalParser *nalparser, GstH264NalUnit *nalu, GstH264SliceHdr *slice, gboolean parse_pred_weight_table, gboolean parse_dec_ref_pic_marking); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_parse_subset_sps (GstH264NalParser *nalparser, GstH264NalUnit *nalu, GstH264SPS *sps); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_parse_sps (GstH264NalParser *nalparser, GstH264NalUnit *nalu, GstH264SPS *sps); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_parse_pps (GstH264NalParser *nalparser, GstH264NalUnit *nalu, GstH264PPS *pps); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parser_parse_sei (GstH264NalParser *nalparser, GstH264NalUnit *nalu, GArray ** messages); GST_CODEC_PARSERS_API void gst_h264_nal_parser_free (GstH264NalParser *nalparser); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parse_subset_sps (GstH264NalUnit *nalu, GstH264SPS *sps); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parse_sps (GstH264NalUnit *nalu, GstH264SPS *sps); GST_CODEC_PARSERS_API GstH264ParserResult gst_h264_parse_pps (GstH264NalParser *nalparser, GstH264NalUnit *nalu, GstH264PPS *pps); GST_CODEC_PARSERS_API void gst_h264_sps_clear (GstH264SPS *sps); GST_CODEC_PARSERS_API void gst_h264_pps_clear (GstH264PPS *pps); GST_CODEC_PARSERS_API void gst_h264_quant_matrix_8x8_get_zigzag_from_raster (guint8 out_quant[64], const guint8 quant[64]); GST_CODEC_PARSERS_API void gst_h264_quant_matrix_8x8_get_raster_from_zigzag (guint8 out_quant[64], const guint8 quant[64]); GST_CODEC_PARSERS_API void gst_h264_quant_matrix_4x4_get_zigzag_from_raster (guint8 out_quant[16], const guint8 quant[16]); GST_CODEC_PARSERS_API void gst_h264_quant_matrix_4x4_get_raster_from_zigzag (guint8 out_quant[16], const guint8 quant[16]); GST_CODEC_PARSERS_API void gst_h264_video_calculate_framerate (const GstH264SPS * sps, guint field_pic_flag, guint pic_struct, gint * fps_num, gint * fps_den); G_END_DECLS #endif