/* Gstreamer * Copyright (C) <2011> Intel Corporation * Copyright (C) <2011> Collabora Ltd. * Copyright (C) <2011> Thibault Saunier * * Some bits C-c,C-v'ed and s/4/3 from h264parse and videoparsers/h264parse.c: * Copyright (C) <2010> Mark Nauwelaerts * Copyright (C) <2010> Collabora Multimedia * Copyright (C) <2010> Nokia Corporation * * (C) 2005 Michal Benes * (C) 2008 Wim Taymans * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:gsth264parser * @title: GstH264Parser * @short_description: Convenience library for h264 video * bitstream parsing. * * It offers bitstream parsing in both AVC (length-prefixed) and Annex B * (0x000001 start code prefix) format. To identify a NAL unit in a bitstream * and parse its headers, first call: * * * #gst_h264_parser_identify_nalu to identify a NAL unit in an Annex B type bitstream * * * #gst_h264_parser_identify_nalu_avc to identify a NAL unit in an AVC type bitstream * * The following functions are then available for parsing the structure of the * #GstH264NalUnit, depending on the #GstH264NalUnitType: * * * From #GST_H264_NAL_SLICE to #GST_H264_NAL_SLICE_IDR: #gst_h264_parser_parse_slice_hdr * * * #GST_H264_NAL_SEI: #gst_h264_parser_parse_sei * * * #GST_H264_NAL_SPS: #gst_h264_parser_parse_sps * * * #GST_H264_NAL_PPS: #gst_h264_parser_parse_pps * * * Any other: #gst_h264_parser_parse_nal * * One of these functions *must* be called on every NAL unit in the bitstream, * in order to keep the internal structures of the #GstH264NalParser up to * date. It is legal to call #gst_h264_parser_parse_nal on NAL units of any * type, if no special parsing of the current NAL unit is required by the * application. * * For more details about the structures, look at the ITU-T H.264 and ISO/IEC 14496-10 – MPEG-4 * Part 10 specifications, available at: * * * ITU-T H.264: http://www.itu.int/rec/T-REC-H.264 * * * ISO/IEC 14496-10: http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=56538 * */ #ifdef HAVE_CONFIG_H # include "config.h" #endif #include "nalutils.h" #include "gsth264parser.h" #include #include #include #ifndef GST_DISABLE_GST_DEBUG #define GST_CAT_DEFAULT gst_h264_debug_category_get() static GstDebugCategory * gst_h264_debug_category_get (void) { static gsize cat_gonce = 0; if (g_once_init_enter (&cat_gonce)) { GstDebugCategory *cat = NULL; GST_DEBUG_CATEGORY_INIT (cat, "codecparsers_h264", 0, "h264 parse library"); g_once_init_leave (&cat_gonce, (gsize) cat); } return (GstDebugCategory *) cat_gonce; } #endif /* GST_DISABLE_GST_DEBUG */ /**** Default scaling_lists according to Table 7-2 *****/ static const guint8 default_4x4_intra[16] = { 6, 13, 13, 20, 20, 20, 28, 28, 28, 28, 32, 32, 32, 37, 37, 42 }; static const guint8 default_4x4_inter[16] = { 10, 14, 14, 20, 20, 20, 24, 24, 24, 24, 27, 27, 27, 30, 30, 34 }; static const guint8 default_8x8_intra[64] = { 6, 10, 10, 13, 11, 13, 16, 16, 16, 16, 18, 18, 18, 18, 18, 23, 23, 23, 23, 23, 23, 25, 25, 25, 25, 25, 25, 25, 27, 27, 27, 27, 27, 27, 27, 27, 29, 29, 29, 29, 29, 29, 29, 31, 31, 31, 31, 31, 31, 33, 33, 33, 33, 33, 36, 36, 36, 36, 38, 38, 38, 40, 40, 42 }; static const guint8 default_8x8_inter[64] = { 9, 13, 13, 15, 13, 15, 17, 17, 17, 17, 19, 19, 19, 19, 19, 21, 21, 21, 21, 21, 21, 22, 22, 22, 22, 22, 22, 22, 24, 24, 24, 24, 24, 24, 24, 24, 25, 25, 25, 25, 25, 25, 25, 27, 27, 27, 27, 27, 27, 28, 28, 28, 28, 28, 30, 30, 30, 30, 32, 32, 32, 33, 33, 35 }; static const guint8 zigzag_8x8[64] = { 0, 1, 8, 16, 9, 2, 3, 10, 17, 24, 32, 25, 18, 11, 4, 5, 12, 19, 26, 33, 40, 48, 41, 34, 27, 20, 13, 6, 7, 14, 21, 28, 35, 42, 49, 56, 57, 50, 43, 36, 29, 22, 15, 23, 30, 37, 44, 51, 58, 59, 52, 45, 38, 31, 39, 46, 53, 60, 61, 54, 47, 55, 62, 63 }; static const guint8 zigzag_4x4[16] = { 0, 1, 4, 8, 5, 2, 3, 6, 9, 12, 13, 10, 7, 11, 14, 15, }; typedef struct { guint par_n, par_d; } PAR; /* Table E-1 - Meaning of sample aspect ratio indicator (1..16) */ static const PAR aspect_ratios[17] = { {0, 0}, {1, 1}, {12, 11}, {10, 11}, {16, 11}, {40, 33}, {24, 11}, {20, 11}, {32, 11}, {80, 33}, {18, 11}, {15, 11}, {64, 33}, {160, 99}, {4, 3}, {3, 2}, {2, 1} }; /***** Utils ****/ #define EXTENDED_SAR 255 static GstH264SPS * gst_h264_parser_get_sps (GstH264NalParser * nalparser, guint8 sps_id) { GstH264SPS *sps; sps = &nalparser->sps[sps_id]; if (sps->valid) return sps; return NULL; } static GstH264PPS * gst_h264_parser_get_pps (GstH264NalParser * nalparser, guint8 pps_id) { GstH264PPS *pps; pps = &nalparser->pps[pps_id]; if (pps->valid) return pps; return NULL; } static gboolean gst_h264_parse_nalu_header (GstH264NalUnit * nalu) { guint8 *data = nalu->data + nalu->offset; guint8 svc_extension_flag; GstBitReader br; if (nalu->size < 1) return FALSE; nalu->type = (data[0] & 0x1f); nalu->ref_idc = (data[0] & 0x60) >> 5; nalu->idr_pic_flag = (nalu->type == 5 ? 1 : 0); nalu->header_bytes = 1; nalu->extension_type = GST_H264_NAL_EXTENSION_NONE; switch (nalu->type) { case GST_H264_NAL_PREFIX_UNIT: case GST_H264_NAL_SLICE_EXT: if (nalu->size < 4) return FALSE; gst_bit_reader_init (&br, nalu->data + nalu->offset + nalu->header_bytes, nalu->size - nalu->header_bytes); svc_extension_flag = gst_bit_reader_get_bits_uint8_unchecked (&br, 1); if (svc_extension_flag) { /* SVC */ nalu->extension_type = GST_H264_NAL_EXTENSION_SVC; } else { /* MVC */ GstH264NalUnitExtensionMVC *const mvc = &nalu->extension.mvc; nalu->extension_type = GST_H264_NAL_EXTENSION_MVC; mvc->non_idr_flag = gst_bit_reader_get_bits_uint8_unchecked (&br, 1); mvc->priority_id = gst_bit_reader_get_bits_uint8_unchecked (&br, 6); mvc->view_id = gst_bit_reader_get_bits_uint16_unchecked (&br, 10); mvc->temporal_id = gst_bit_reader_get_bits_uint8_unchecked (&br, 3); mvc->anchor_pic_flag = gst_bit_reader_get_bits_uint8_unchecked (&br, 1); mvc->inter_view_flag = gst_bit_reader_get_bits_uint8_unchecked (&br, 1); /* Update IdrPicFlag (H.7.4.1.1) */ nalu->idr_pic_flag = !mvc->non_idr_flag; } nalu->header_bytes += 3; break; default: break; } GST_DEBUG ("Nal type %u, ref_idc %u", nalu->type, nalu->ref_idc); return TRUE; } /* * gst_h264_pps_copy: * @dst_pps: The destination #GstH264PPS to copy into * @src_pps: The source #GstH264PPS to copy from * * Copies @src_pps into @dst_pps. * * Returns: %TRUE if everything went fine, %FALSE otherwise */ static gboolean gst_h264_pps_copy (GstH264PPS * dst_pps, const GstH264PPS * src_pps) { g_return_val_if_fail (dst_pps != NULL, FALSE); g_return_val_if_fail (src_pps != NULL, FALSE); gst_h264_pps_clear (dst_pps); *dst_pps = *src_pps; if (src_pps->slice_group_id) dst_pps->slice_group_id = g_memdup2 (src_pps->slice_group_id, src_pps->pic_size_in_map_units_minus1 + 1); return TRUE; } /* Copy MVC-specific data for subset SPS header */ static gboolean gst_h264_sps_mvc_copy (GstH264SPS * dst_sps, const GstH264SPS * src_sps) { GstH264SPSExtMVC *const dst_mvc = &dst_sps->extension.mvc; const GstH264SPSExtMVC *const src_mvc = &src_sps->extension.mvc; guint i, j, k; g_assert (dst_sps->extension_type == GST_H264_NAL_EXTENSION_MVC); dst_mvc->num_views_minus1 = src_mvc->num_views_minus1; dst_mvc->view = g_new0 (GstH264SPSExtMVCView, dst_mvc->num_views_minus1 + 1); if (!dst_mvc->view) return FALSE; dst_mvc->view[0].view_id = src_mvc->view[0].view_id; for (i = 1; i <= dst_mvc->num_views_minus1; i++) { GstH264SPSExtMVCView *const dst_view = &dst_mvc->view[i]; const GstH264SPSExtMVCView *const src_view = &src_mvc->view[i]; dst_view->view_id = src_view->view_id; dst_view->num_anchor_refs_l0 = src_view->num_anchor_refs_l0; for (j = 0; j < dst_view->num_anchor_refs_l0; j++) dst_view->anchor_ref_l0[j] = src_view->anchor_ref_l0[j]; dst_view->num_anchor_refs_l1 = src_view->num_anchor_refs_l1; for (j = 0; j < dst_view->num_anchor_refs_l1; j++) dst_view->anchor_ref_l1[j] = src_view->anchor_ref_l1[j]; dst_view->num_non_anchor_refs_l0 = src_view->num_non_anchor_refs_l0; for (j = 0; j < dst_view->num_non_anchor_refs_l0; j++) dst_view->non_anchor_ref_l0[j] = src_view->non_anchor_ref_l0[j]; dst_view->num_non_anchor_refs_l1 = src_view->num_non_anchor_refs_l1; for (j = 0; j < dst_view->num_non_anchor_refs_l1; j++) dst_view->non_anchor_ref_l1[j] = src_view->non_anchor_ref_l1[j]; } dst_mvc->num_level_values_signalled_minus1 = src_mvc->num_level_values_signalled_minus1; dst_mvc->level_value = g_new0 (GstH264SPSExtMVCLevelValue, dst_mvc->num_level_values_signalled_minus1 + 1); if (!dst_mvc->level_value) return FALSE; for (i = 0; i <= dst_mvc->num_level_values_signalled_minus1; i++) { GstH264SPSExtMVCLevelValue *const dst_value = &dst_mvc->level_value[i]; const GstH264SPSExtMVCLevelValue *const src_value = &src_mvc->level_value[i]; dst_value->level_idc = src_value->level_idc; dst_value->num_applicable_ops_minus1 = src_value->num_applicable_ops_minus1; dst_value->applicable_op = g_new0 (GstH264SPSExtMVCLevelValueOp, dst_value->num_applicable_ops_minus1 + 1); if (!dst_value->applicable_op) return FALSE; for (j = 0; j <= dst_value->num_applicable_ops_minus1; j++) { GstH264SPSExtMVCLevelValueOp *const dst_op = &dst_value->applicable_op[j]; const GstH264SPSExtMVCLevelValueOp *const src_op = &src_value->applicable_op[j]; dst_op->temporal_id = src_op->temporal_id; dst_op->num_target_views_minus1 = src_op->num_target_views_minus1; dst_op->target_view_id = g_new (guint16, dst_op->num_target_views_minus1 + 1); if (!dst_op->target_view_id) return FALSE; for (k = 0; k <= dst_op->num_target_views_minus1; k++) dst_op->target_view_id[k] = src_op->target_view_id[k]; dst_op->num_views_minus1 = src_op->num_views_minus1; } } return TRUE; } /* * gst_h264_sps_copy: * @dst_sps: The destination #GstH264SPS to copy into * @src_sps: The source #GstH264SPS to copy from * * Copies @src_sps into @dst_sps. * * Returns: %TRUE if everything went fine, %FALSE otherwise */ static gboolean gst_h264_sps_copy (GstH264SPS * dst_sps, const GstH264SPS * src_sps) { g_return_val_if_fail (dst_sps != NULL, FALSE); g_return_val_if_fail (src_sps != NULL, FALSE); gst_h264_sps_clear (dst_sps); *dst_sps = *src_sps; switch (dst_sps->extension_type) { case GST_H264_NAL_EXTENSION_MVC: if (!gst_h264_sps_mvc_copy (dst_sps, src_sps)) return FALSE; break; } return TRUE; } /****** Parsing functions *****/ static gboolean gst_h264_parse_hrd_parameters (GstH264HRDParams * hrd, NalReader * nr) { guint sched_sel_idx; GST_DEBUG ("parsing \"HRD Parameters\""); READ_UE_MAX (nr, hrd->cpb_cnt_minus1, 31); READ_UINT8 (nr, hrd->bit_rate_scale, 4); READ_UINT8 (nr, hrd->cpb_size_scale, 4); for (sched_sel_idx = 0; sched_sel_idx <= hrd->cpb_cnt_minus1; sched_sel_idx++) { READ_UE (nr, hrd->bit_rate_value_minus1[sched_sel_idx]); READ_UE (nr, hrd->cpb_size_value_minus1[sched_sel_idx]); READ_UINT8 (nr, hrd->cbr_flag[sched_sel_idx], 1); } READ_UINT8 (nr, hrd->initial_cpb_removal_delay_length_minus1, 5); READ_UINT8 (nr, hrd->cpb_removal_delay_length_minus1, 5); READ_UINT8 (nr, hrd->dpb_output_delay_length_minus1, 5); READ_UINT8 (nr, hrd->time_offset_length, 5); return TRUE; error: GST_WARNING ("error parsing \"HRD Parameters\""); return FALSE; } static gboolean gst_h264_parse_vui_parameters (GstH264SPS * sps, NalReader * nr) { GstH264VUIParams *vui = &sps->vui_parameters; GST_DEBUG ("parsing \"VUI Parameters\""); /* set default values for fields that might not be present in the bitstream and have valid defaults */ vui->video_format = 5; vui->colour_primaries = 2; vui->transfer_characteristics = 2; vui->matrix_coefficients = 2; READ_UINT8 (nr, vui->aspect_ratio_info_present_flag, 1); if (vui->aspect_ratio_info_present_flag) { READ_UINT8 (nr, vui->aspect_ratio_idc, 8); if (vui->aspect_ratio_idc == EXTENDED_SAR) { READ_UINT16 (nr, vui->sar_width, 16); READ_UINT16 (nr, vui->sar_height, 16); vui->par_n = vui->sar_width; vui->par_d = vui->sar_height; } else if (vui->aspect_ratio_idc <= 16) { vui->par_n = aspect_ratios[vui->aspect_ratio_idc].par_n; vui->par_d = aspect_ratios[vui->aspect_ratio_idc].par_d; } } READ_UINT8 (nr, vui->overscan_info_present_flag, 1); if (vui->overscan_info_present_flag) READ_UINT8 (nr, vui->overscan_appropriate_flag, 1); READ_UINT8 (nr, vui->video_signal_type_present_flag, 1); if (vui->video_signal_type_present_flag) { READ_UINT8 (nr, vui->video_format, 3); READ_UINT8 (nr, vui->video_full_range_flag, 1); READ_UINT8 (nr, vui->colour_description_present_flag, 1); if (vui->colour_description_present_flag) { READ_UINT8 (nr, vui->colour_primaries, 8); READ_UINT8 (nr, vui->transfer_characteristics, 8); READ_UINT8 (nr, vui->matrix_coefficients, 8); } } READ_UINT8 (nr, vui->chroma_loc_info_present_flag, 1); if (vui->chroma_loc_info_present_flag) { READ_UE_MAX (nr, vui->chroma_sample_loc_type_top_field, 5); READ_UE_MAX (nr, vui->chroma_sample_loc_type_bottom_field, 5); } READ_UINT8 (nr, vui->timing_info_present_flag, 1); if (vui->timing_info_present_flag) { READ_UINT32 (nr, vui->num_units_in_tick, 32); if (vui->num_units_in_tick == 0) GST_WARNING ("num_units_in_tick = 0 detected in stream " "(incompliant to H.264 E.2.1)."); READ_UINT32 (nr, vui->time_scale, 32); if (vui->time_scale == 0) GST_WARNING ("time_scale = 0 detected in stream " "(incompliant to H.264 E.2.1)."); READ_UINT8 (nr, vui->fixed_frame_rate_flag, 1); } READ_UINT8 (nr, vui->nal_hrd_parameters_present_flag, 1); if (vui->nal_hrd_parameters_present_flag) { if (!gst_h264_parse_hrd_parameters (&vui->nal_hrd_parameters, nr)) goto error; } READ_UINT8 (nr, vui->vcl_hrd_parameters_present_flag, 1); if (vui->vcl_hrd_parameters_present_flag) { if (!gst_h264_parse_hrd_parameters (&vui->vcl_hrd_parameters, nr)) goto error; } if (vui->nal_hrd_parameters_present_flag || vui->vcl_hrd_parameters_present_flag) READ_UINT8 (nr, vui->low_delay_hrd_flag, 1); READ_UINT8 (nr, vui->pic_struct_present_flag, 1); READ_UINT8 (nr, vui->bitstream_restriction_flag, 1); if (vui->bitstream_restriction_flag) { READ_UINT8 (nr, vui->motion_vectors_over_pic_boundaries_flag, 1); READ_UE (nr, vui->max_bytes_per_pic_denom); READ_UE_MAX (nr, vui->max_bits_per_mb_denom, 16); READ_UE_MAX (nr, vui->log2_max_mv_length_horizontal, 16); READ_UE_MAX (nr, vui->log2_max_mv_length_vertical, 16); READ_UE (nr, vui->num_reorder_frames); READ_UE (nr, vui->max_dec_frame_buffering); } return TRUE; error: GST_WARNING ("error parsing \"VUI Parameters\""); return FALSE; } static gboolean gst_h264_parser_parse_scaling_list (NalReader * nr, guint8 scaling_lists_4x4[6][16], guint8 scaling_lists_8x8[6][64], const guint8 fallback_4x4_inter[16], const guint8 fallback_4x4_intra[16], const guint8 fallback_8x8_inter[64], const guint8 fallback_8x8_intra[64], guint8 n_lists) { guint i; static const guint8 *default_lists[12] = { default_4x4_intra, default_4x4_intra, default_4x4_intra, default_4x4_inter, default_4x4_inter, default_4x4_inter, default_8x8_intra, default_8x8_inter, default_8x8_intra, default_8x8_inter, default_8x8_intra, default_8x8_inter }; GST_DEBUG ("parsing scaling lists"); for (i = 0; i < 12; i++) { gboolean use_default = FALSE; if (i < n_lists) { guint8 scaling_list_present_flag; READ_UINT8 (nr, scaling_list_present_flag, 1); if (scaling_list_present_flag) { guint8 *scaling_list; guint size; guint j; guint8 last_scale, next_scale; if (i < 6) { scaling_list = scaling_lists_4x4[i]; size = 16; } else { scaling_list = scaling_lists_8x8[i - 6]; size = 64; } last_scale = 8; next_scale = 8; for (j = 0; j < size; j++) { if (next_scale != 0) { gint32 delta_scale; READ_SE (nr, delta_scale); next_scale = (last_scale + delta_scale) & 0xff; } if (j == 0 && next_scale == 0) { /* Use default scaling lists (7.4.2.1.1.1) */ memcpy (scaling_list, default_lists[i], size); break; } last_scale = scaling_list[j] = (next_scale == 0) ? last_scale : next_scale; } } else use_default = TRUE; } else use_default = TRUE; if (use_default) { switch (i) { case 0: memcpy (scaling_lists_4x4[0], fallback_4x4_intra, 16); break; case 1: memcpy (scaling_lists_4x4[1], scaling_lists_4x4[0], 16); break; case 2: memcpy (scaling_lists_4x4[2], scaling_lists_4x4[1], 16); break; case 3: memcpy (scaling_lists_4x4[3], fallback_4x4_inter, 16); break; case 4: memcpy (scaling_lists_4x4[4], scaling_lists_4x4[3], 16); break; case 5: memcpy (scaling_lists_4x4[5], scaling_lists_4x4[4], 16); break; case 6: memcpy (scaling_lists_8x8[0], fallback_8x8_intra, 64); break; case 7: memcpy (scaling_lists_8x8[1], fallback_8x8_inter, 64); break; case 8: memcpy (scaling_lists_8x8[2], scaling_lists_8x8[0], 64); break; case 9: memcpy (scaling_lists_8x8[3], scaling_lists_8x8[1], 64); break; case 10: memcpy (scaling_lists_8x8[4], scaling_lists_8x8[2], 64); break; case 11: memcpy (scaling_lists_8x8[5], scaling_lists_8x8[3], 64); break; default: break; } } } return TRUE; error: GST_WARNING ("error parsing scaling lists"); return FALSE; } static gboolean slice_parse_ref_pic_list_modification_1 (GstH264SliceHdr * slice, NalReader * nr, guint list, gboolean is_mvc) { GstH264RefPicListModification *entries; guint8 *ref_pic_list_modification_flag, *n_ref_pic_list_modification; guint32 modification_of_pic_nums_idc; gsize max_entries; guint i = 0; if (list == 0) { entries = slice->ref_pic_list_modification_l0; max_entries = G_N_ELEMENTS (slice->ref_pic_list_modification_l0); ref_pic_list_modification_flag = &slice->ref_pic_list_modification_flag_l0; n_ref_pic_list_modification = &slice->n_ref_pic_list_modification_l0; } else { entries = slice->ref_pic_list_modification_l1; max_entries = G_N_ELEMENTS (slice->ref_pic_list_modification_l1); ref_pic_list_modification_flag = &slice->ref_pic_list_modification_flag_l1; n_ref_pic_list_modification = &slice->n_ref_pic_list_modification_l1; } READ_UINT8 (nr, *ref_pic_list_modification_flag, 1); if (*ref_pic_list_modification_flag) { while (1) { READ_UE (nr, modification_of_pic_nums_idc); if (modification_of_pic_nums_idc == 0 || modification_of_pic_nums_idc == 1) { READ_UE_MAX (nr, entries[i].value.abs_diff_pic_num_minus1, slice->max_pic_num - 1); } else if (modification_of_pic_nums_idc == 2) { READ_UE (nr, entries[i].value.long_term_pic_num); } else if (is_mvc && (modification_of_pic_nums_idc == 4 || modification_of_pic_nums_idc == 5)) { READ_UE (nr, entries[i].value.abs_diff_view_idx_minus1); } entries[i++].modification_of_pic_nums_idc = modification_of_pic_nums_idc; if (modification_of_pic_nums_idc == 3) break; if (i >= max_entries) goto error; } } *n_ref_pic_list_modification = i; return TRUE; error: GST_WARNING ("error parsing \"Reference picture list %u modification\"", list); return FALSE; } static gboolean slice_parse_ref_pic_list_modification (GstH264SliceHdr * slice, NalReader * nr, gboolean is_mvc) { if (!GST_H264_IS_I_SLICE (slice) && !GST_H264_IS_SI_SLICE (slice)) { if (!slice_parse_ref_pic_list_modification_1 (slice, nr, 0, is_mvc)) return FALSE; } if (GST_H264_IS_B_SLICE (slice)) { if (!slice_parse_ref_pic_list_modification_1 (slice, nr, 1, is_mvc)) return FALSE; } return TRUE; } static gboolean gst_h264_slice_parse_dec_ref_pic_marking (GstH264SliceHdr * slice, GstH264NalUnit * nalu, NalReader * nr) { GstH264DecRefPicMarking *dec_ref_pic_m; guint start_pos, start_epb; GST_DEBUG ("parsing \"Decoded reference picture marking\""); start_pos = nal_reader_get_pos (nr); start_epb = nal_reader_get_epb_count (nr); dec_ref_pic_m = &slice->dec_ref_pic_marking; if (nalu->idr_pic_flag) { READ_UINT8 (nr, dec_ref_pic_m->no_output_of_prior_pics_flag, 1); READ_UINT8 (nr, dec_ref_pic_m->long_term_reference_flag, 1); } else { READ_UINT8 (nr, dec_ref_pic_m->adaptive_ref_pic_marking_mode_flag, 1); if (dec_ref_pic_m->adaptive_ref_pic_marking_mode_flag) { guint32 mem_mgmt_ctrl_op; GstH264RefPicMarking *refpicmarking; dec_ref_pic_m->n_ref_pic_marking = 0; while (1) { READ_UE_MAX (nr, mem_mgmt_ctrl_op, 6); if (mem_mgmt_ctrl_op == 0) break; if (dec_ref_pic_m->n_ref_pic_marking >= G_N_ELEMENTS (dec_ref_pic_m->ref_pic_marking)) goto error; refpicmarking = &dec_ref_pic_m->ref_pic_marking[dec_ref_pic_m->n_ref_pic_marking]; refpicmarking->memory_management_control_operation = mem_mgmt_ctrl_op; if (mem_mgmt_ctrl_op == 1 || mem_mgmt_ctrl_op == 3) READ_UE (nr, refpicmarking->difference_of_pic_nums_minus1); if (mem_mgmt_ctrl_op == 2) READ_UE (nr, refpicmarking->long_term_pic_num); if (mem_mgmt_ctrl_op == 3 || mem_mgmt_ctrl_op == 6) READ_UE (nr, refpicmarking->long_term_frame_idx); if (mem_mgmt_ctrl_op == 4) READ_UE (nr, refpicmarking->max_long_term_frame_idx_plus1); dec_ref_pic_m->n_ref_pic_marking++; } } } dec_ref_pic_m->bit_size = (nal_reader_get_pos (nr) - start_pos) - (8 * (nal_reader_get_epb_count (nr) - start_epb)); return TRUE; error: GST_WARNING ("error parsing \"Decoded reference picture marking\""); return FALSE; } static gboolean gst_h264_slice_parse_pred_weight_table (GstH264SliceHdr * slice, NalReader * nr, guint8 chroma_array_type) { GstH264PredWeightTable *p; gint16 default_luma_weight, default_chroma_weight; gint i; GST_DEBUG ("parsing \"Prediction weight table\""); p = &slice->pred_weight_table; READ_UE_MAX (nr, p->luma_log2_weight_denom, 7); /* set default values */ default_luma_weight = 1 << p->luma_log2_weight_denom; for (i = 0; i <= slice->num_ref_idx_l0_active_minus1; i++) p->luma_weight_l0[i] = default_luma_weight; if (GST_H264_IS_B_SLICE (slice)) { for (i = 0; i <= slice->num_ref_idx_l1_active_minus1; i++) p->luma_weight_l1[i] = default_luma_weight; } if (chroma_array_type != 0) { READ_UE_MAX (nr, p->chroma_log2_weight_denom, 7); /* set default values */ default_chroma_weight = 1 << p->chroma_log2_weight_denom; for (i = 0; i <= slice->num_ref_idx_l0_active_minus1; i++) { p->chroma_weight_l0[i][0] = default_chroma_weight; p->chroma_weight_l0[i][1] = default_chroma_weight; } if (GST_H264_IS_B_SLICE (slice)) { for (i = 0; i <= slice->num_ref_idx_l0_active_minus1; i++) { p->chroma_weight_l1[i][0] = default_chroma_weight; p->chroma_weight_l1[i][1] = default_chroma_weight; } } } for (i = 0; i <= slice->num_ref_idx_l0_active_minus1; i++) { guint8 luma_weight_l0_flag; READ_UINT8 (nr, luma_weight_l0_flag, 1); if (luma_weight_l0_flag) { READ_SE_ALLOWED (nr, p->luma_weight_l0[i], -128, 127); READ_SE_ALLOWED (nr, p->luma_offset_l0[i], -128, 127); } if (chroma_array_type != 0) { guint8 chroma_weight_l0_flag; gint j; READ_UINT8 (nr, chroma_weight_l0_flag, 1); if (chroma_weight_l0_flag) { for (j = 0; j < 2; j++) { READ_SE_ALLOWED (nr, p->chroma_weight_l0[i][j], -128, 127); READ_SE_ALLOWED (nr, p->chroma_offset_l0[i][j], -128, 127); } } } } if (GST_H264_IS_B_SLICE (slice)) { for (i = 0; i <= slice->num_ref_idx_l1_active_minus1; i++) { guint8 luma_weight_l1_flag; READ_UINT8 (nr, luma_weight_l1_flag, 1); if (luma_weight_l1_flag) { READ_SE_ALLOWED (nr, p->luma_weight_l1[i], -128, 127); READ_SE_ALLOWED (nr, p->luma_offset_l1[i], -128, 127); } if (chroma_array_type != 0) { guint8 chroma_weight_l1_flag; gint j; READ_UINT8 (nr, chroma_weight_l1_flag, 1); if (chroma_weight_l1_flag) { for (j = 0; j < 2; j++) { READ_SE_ALLOWED (nr, p->chroma_weight_l1[i][j], -128, 127); READ_SE_ALLOWED (nr, p->chroma_offset_l1[i][j], -128, 127); } } } } } return TRUE; error: GST_WARNING ("error parsing \"Prediction weight table\""); return FALSE; } static GstH264ParserResult gst_h264_parser_parse_buffering_period (GstH264NalParser * nalparser, GstH264BufferingPeriod * per, NalReader * nr) { GstH264SPS *sps; guint8 sps_id; GST_DEBUG ("parsing \"Buffering period\""); READ_UE_MAX (nr, sps_id, GST_H264_MAX_SPS_COUNT - 1); sps = gst_h264_parser_get_sps (nalparser, sps_id); if (!sps) { GST_WARNING ("couldn't find associated sequence parameter set with id: %d", sps_id); return GST_H264_PARSER_BROKEN_LINK; } per->sps = sps; if (sps->vui_parameters_present_flag) { GstH264VUIParams *vui = &sps->vui_parameters; if (vui->nal_hrd_parameters_present_flag) { GstH264HRDParams *hrd = &vui->nal_hrd_parameters; const guint8 nbits = hrd->initial_cpb_removal_delay_length_minus1 + 1; guint8 sched_sel_idx; for (sched_sel_idx = 0; sched_sel_idx <= hrd->cpb_cnt_minus1; sched_sel_idx++) { READ_UINT32 (nr, per->nal_initial_cpb_removal_delay[sched_sel_idx], nbits); READ_UINT32 (nr, per->nal_initial_cpb_removal_delay_offset[sched_sel_idx], nbits); } } if (vui->vcl_hrd_parameters_present_flag) { GstH264HRDParams *hrd = &vui->vcl_hrd_parameters; const guint8 nbits = hrd->initial_cpb_removal_delay_length_minus1 + 1; guint8 sched_sel_idx; for (sched_sel_idx = 0; sched_sel_idx <= hrd->cpb_cnt_minus1; sched_sel_idx++) { READ_UINT32 (nr, per->vcl_initial_cpb_removal_delay[sched_sel_idx], nbits); READ_UINT32 (nr, per->vcl_initial_cpb_removal_delay_offset[sched_sel_idx], nbits); } } } return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Buffering period\""); return GST_H264_PARSER_ERROR; } static gboolean gst_h264_parse_clock_timestamp (GstH264ClockTimestamp * tim, guint8 time_offset_length, NalReader * nr) { GST_DEBUG ("parsing \"Clock timestamp\""); /* default values */ tim->time_offset = 0; READ_UINT8 (nr, tim->ct_type, 2); READ_UINT8 (nr, tim->nuit_field_based_flag, 1); READ_UINT8 (nr, tim->counting_type, 5); READ_UINT8 (nr, tim->full_timestamp_flag, 1); READ_UINT8 (nr, tim->discontinuity_flag, 1); READ_UINT8 (nr, tim->cnt_dropped_flag, 1); READ_UINT8 (nr, tim->n_frames, 8); if (tim->full_timestamp_flag) { tim->seconds_flag = TRUE; READ_UINT8 (nr, tim->seconds_value, 6); tim->minutes_flag = TRUE; READ_UINT8 (nr, tim->minutes_value, 6); tim->hours_flag = TRUE; READ_UINT8 (nr, tim->hours_value, 5); } else { READ_UINT8 (nr, tim->seconds_flag, 1); if (tim->seconds_flag) { READ_UINT8 (nr, tim->seconds_value, 6); READ_UINT8 (nr, tim->minutes_flag, 1); if (tim->minutes_flag) { READ_UINT8 (nr, tim->minutes_value, 6); READ_UINT8 (nr, tim->hours_flag, 1); if (tim->hours_flag) READ_UINT8 (nr, tim->hours_value, 5); } } } if (time_offset_length > 0) READ_UINT32 (nr, tim->time_offset, time_offset_length); return TRUE; error: GST_WARNING ("error parsing \"Clock timestamp\""); return FALSE; } static GstH264ParserResult gst_h264_parser_parse_pic_timing (GstH264NalParser * nalparser, GstH264PicTiming * tim, NalReader * nr) { GstH264ParserResult error = GST_H264_PARSER_ERROR; GST_DEBUG ("parsing \"Picture timing\""); if (!nalparser->last_sps || !nalparser->last_sps->valid) { GST_WARNING ("didn't get the associated sequence parameter set for the " "current access unit"); error = GST_H264_PARSER_BROKEN_LINK; goto error; } if (nalparser->last_sps->vui_parameters_present_flag) { GstH264VUIParams *vui = &nalparser->last_sps->vui_parameters; GstH264HRDParams *hrd = NULL; if (vui->nal_hrd_parameters_present_flag) { hrd = &vui->nal_hrd_parameters; } else if (vui->vcl_hrd_parameters_present_flag) { hrd = &vui->vcl_hrd_parameters; } tim->CpbDpbDelaysPresentFlag = ! !hrd; tim->pic_struct_present_flag = vui->pic_struct_present_flag; if (tim->CpbDpbDelaysPresentFlag) { tim->cpb_removal_delay_length_minus1 = hrd->cpb_removal_delay_length_minus1; tim->dpb_output_delay_length_minus1 = hrd->dpb_output_delay_length_minus1; READ_UINT32 (nr, tim->cpb_removal_delay, tim->cpb_removal_delay_length_minus1 + 1); READ_UINT32 (nr, tim->dpb_output_delay, tim->dpb_output_delay_length_minus1 + 1); } if (tim->pic_struct_present_flag) { const guint8 num_clock_ts_table[9] = { 1, 1, 1, 2, 2, 3, 3, 2, 3 }; guint8 num_clock_num_ts; guint i; READ_UINT8 (nr, tim->pic_struct, 4); CHECK_ALLOWED ((gint8) tim->pic_struct, 0, 8); tim->time_offset_length = 24; if (hrd) tim->time_offset_length = hrd->time_offset_length; num_clock_num_ts = num_clock_ts_table[tim->pic_struct]; for (i = 0; i < num_clock_num_ts; i++) { READ_UINT8 (nr, tim->clock_timestamp_flag[i], 1); if (tim->clock_timestamp_flag[i]) { if (!gst_h264_parse_clock_timestamp (&tim->clock_timestamp[i], tim->time_offset_length, nr)) goto error; } } } } if (!tim->CpbDpbDelaysPresentFlag && !tim->pic_struct_present_flag) { GST_WARNING ("Invalid pic_timing SEI NAL with neither CpbDpbDelays nor pic_struct"); return GST_H264_PARSER_BROKEN_DATA; } return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Picture timing\""); return error; } static GstH264ParserResult gst_h264_parser_parse_registered_user_data (GstH264NalParser * nalparser, GstH264RegisteredUserData * rud, NalReader * nr, guint payload_size) { guint8 *data = NULL; guint i; rud->data = NULL; rud->size = 0; if (payload_size < 2) { GST_WARNING ("Too small payload size %d", payload_size); return GST_H264_PARSER_BROKEN_DATA; } READ_UINT8 (nr, rud->country_code, 8); --payload_size; if (rud->country_code == 0xFF) { READ_UINT8 (nr, rud->country_code_extension, 8); --payload_size; } else { rud->country_code_extension = 0; } if (payload_size < 1) { GST_WARNING ("No more remaining payload data to store"); return GST_H264_PARSER_BROKEN_DATA; } data = g_malloc (payload_size); for (i = 0; i < payload_size; ++i) { READ_UINT8 (nr, data[i], 8); } GST_MEMDUMP ("SEI user data", data, payload_size); rud->data = data; rud->size = payload_size; return GST_H264_PARSER_OK; error: { GST_WARNING ("error parsing \"Registered User Data\""); g_free (data); return GST_H264_PARSER_ERROR; } } static GstH264ParserResult gst_h264_parser_parse_recovery_point (GstH264NalParser * nalparser, GstH264RecoveryPoint * rp, NalReader * nr) { GstH264SPS *const sps = nalparser->last_sps; GST_DEBUG ("parsing \"Recovery point\""); if (!sps || !sps->valid) { GST_WARNING ("didn't get the associated sequence parameter set for the " "current access unit"); goto error; } READ_UE_MAX (nr, rp->recovery_frame_cnt, sps->max_frame_num - 1); READ_UINT8 (nr, rp->exact_match_flag, 1); READ_UINT8 (nr, rp->broken_link_flag, 1); READ_UINT8 (nr, rp->changing_slice_group_idc, 2); return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Recovery point\""); return GST_H264_PARSER_ERROR; } /* Parse SEI stereo_video_info() message */ static GstH264ParserResult gst_h264_parser_parse_stereo_video_info (GstH264NalParser * nalparser, GstH264StereoVideoInfo * info, NalReader * nr) { GST_DEBUG ("parsing \"Stereo Video info\""); READ_UINT8 (nr, info->field_views_flag, 1); if (info->field_views_flag) { READ_UINT8 (nr, info->top_field_is_left_view_flag, 1); } else { READ_UINT8 (nr, info->current_frame_is_left_view_flag, 1); READ_UINT8 (nr, info->next_frame_is_second_view_flag, 1); } READ_UINT8 (nr, info->left_view_self_contained_flag, 1); READ_UINT8 (nr, info->right_view_self_contained_flag, 1); return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Stereo Video info\""); return GST_H264_PARSER_ERROR; } /* Parse SEI frame_packing_arrangement() message */ static GstH264ParserResult gst_h264_parser_parse_frame_packing (GstH264NalParser * nalparser, GstH264FramePacking * frame_packing, NalReader * nr, guint payload_size) { guint8 frame_packing_extension_flag; guint start_pos; GST_DEBUG ("parsing \"Frame Packing Arrangement\""); start_pos = nal_reader_get_pos (nr); READ_UE (nr, frame_packing->frame_packing_id); READ_UINT8 (nr, frame_packing->frame_packing_cancel_flag, 1); if (!frame_packing->frame_packing_cancel_flag) { READ_UINT8 (nr, frame_packing->frame_packing_type, 7); READ_UINT8 (nr, frame_packing->quincunx_sampling_flag, 1); READ_UINT8 (nr, frame_packing->content_interpretation_type, 6); READ_UINT8 (nr, frame_packing->spatial_flipping_flag, 1); READ_UINT8 (nr, frame_packing->frame0_flipped_flag, 1); READ_UINT8 (nr, frame_packing->field_views_flag, 1); READ_UINT8 (nr, frame_packing->current_frame_is_frame0_flag, 1); READ_UINT8 (nr, frame_packing->frame0_self_contained_flag, 1); READ_UINT8 (nr, frame_packing->frame1_self_contained_flag, 1); if (!frame_packing->quincunx_sampling_flag && frame_packing->frame_packing_type != GST_H264_FRAME_PACKING_TEMPORAL_INTERLEAVING) { READ_UINT8 (nr, frame_packing->frame0_grid_position_x, 4); READ_UINT8 (nr, frame_packing->frame0_grid_position_y, 4); READ_UINT8 (nr, frame_packing->frame1_grid_position_x, 4); READ_UINT8 (nr, frame_packing->frame1_grid_position_y, 4); } /* Skip frame_packing_arrangement_reserved_byte */ if (!nal_reader_skip (nr, 8)) goto error; READ_UE_MAX (nr, frame_packing->frame_packing_repetition_period, 16384); } READ_UINT8 (nr, frame_packing_extension_flag, 1); /* All data that follows within a frame packing arrangement SEI message after the value 1 for frame_packing_arrangement_extension_flag shall be ignored (D.2.25) */ if (frame_packing_extension_flag) { nal_reader_skip_long (nr, payload_size - (nal_reader_get_pos (nr) - start_pos)); } return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Frame Packing Arrangement\""); return GST_H264_PARSER_ERROR; } static GstH264ParserResult gst_h264_parser_parse_mastering_display_colour_volume (GstH264NalParser * parser, GstH264MasteringDisplayColourVolume * mdcv, NalReader * nr) { guint i; GST_DEBUG ("parsing \"Mastering display colour volume\""); for (i = 0; i < 3; i++) { READ_UINT16 (nr, mdcv->display_primaries_x[i], 16); READ_UINT16 (nr, mdcv->display_primaries_y[i], 16); } READ_UINT16 (nr, mdcv->white_point_x, 16); READ_UINT16 (nr, mdcv->white_point_y, 16); READ_UINT32 (nr, mdcv->max_display_mastering_luminance, 32); READ_UINT32 (nr, mdcv->min_display_mastering_luminance, 32); return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Mastering display colour volume\""); return GST_H264_PARSER_ERROR; } static GstH264ParserResult gst_h264_parser_parse_content_light_level_info (GstH264NalParser * parser, GstH264ContentLightLevel * cll, NalReader * nr) { GST_DEBUG ("parsing \"Content light level\""); READ_UINT16 (nr, cll->max_content_light_level, 16); READ_UINT16 (nr, cll->max_pic_average_light_level, 16); return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Content light level\""); return GST_H264_PARSER_ERROR; } static GstH264ParserResult gst_h264_parser_parse_sei_unhandled_payload (GstH264NalParser * parser, GstH264SEIUnhandledPayload * payload, NalReader * nr, guint payload_type, guint payload_size) { guint8 *data = NULL; gint i; payload->payloadType = payload_type; data = g_malloc0 (payload_size); for (i = 0; i < payload_size; ++i) { READ_UINT8 (nr, data[i], 8); } payload->size = payload_size; payload->data = data; return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Unhandled payload\""); g_free (data); return GST_H264_PARSER_ERROR; } static GstH264ParserResult gst_h264_parser_parse_sei_message (GstH264NalParser * nalparser, NalReader * nr, GstH264SEIMessage * sei) { guint32 payloadSize; guint8 payload_type_byte, payload_size_byte; guint remaining, payload_size, next; GstH264ParserResult res; GST_DEBUG ("parsing \"SEI message\""); memset (sei, 0, sizeof (*sei)); do { READ_UINT8 (nr, payload_type_byte, 8); sei->payloadType += payload_type_byte; } while (payload_type_byte == 0xff); payloadSize = 0; do { READ_UINT8 (nr, payload_size_byte, 8); payloadSize += payload_size_byte; } while (payload_size_byte == 0xff); remaining = nal_reader_get_remaining (nr); payload_size = payloadSize * 8 < remaining ? payloadSize * 8 : remaining; next = nal_reader_get_pos (nr) + payload_size; GST_DEBUG ("SEI message received: payloadType %u, payloadSize = %u bits", sei->payloadType, payload_size); switch (sei->payloadType) { case GST_H264_SEI_BUF_PERIOD: /* size not set; might depend on emulation_prevention_three_byte */ res = gst_h264_parser_parse_buffering_period (nalparser, &sei->payload.buffering_period, nr); break; case GST_H264_SEI_PIC_TIMING: /* size not set; might depend on emulation_prevention_three_byte */ res = gst_h264_parser_parse_pic_timing (nalparser, &sei->payload.pic_timing, nr); break; case GST_H264_SEI_REGISTERED_USER_DATA: res = gst_h264_parser_parse_registered_user_data (nalparser, &sei->payload.registered_user_data, nr, payload_size >> 3); break; case GST_H264_SEI_RECOVERY_POINT: res = gst_h264_parser_parse_recovery_point (nalparser, &sei->payload.recovery_point, nr); break; case GST_H264_SEI_STEREO_VIDEO_INFO: res = gst_h264_parser_parse_stereo_video_info (nalparser, &sei->payload.stereo_video_info, nr); break; case GST_H264_SEI_FRAME_PACKING: res = gst_h264_parser_parse_frame_packing (nalparser, &sei->payload.frame_packing, nr, payload_size); break; case GST_H264_SEI_MASTERING_DISPLAY_COLOUR_VOLUME: res = gst_h264_parser_parse_mastering_display_colour_volume (nalparser, &sei->payload.mastering_display_colour_volume, nr); break; case GST_H264_SEI_CONTENT_LIGHT_LEVEL: res = gst_h264_parser_parse_content_light_level_info (nalparser, &sei->payload.content_light_level, nr); break; default: res = gst_h264_parser_parse_sei_unhandled_payload (nalparser, &sei->payload.unhandled_payload, nr, sei->payloadType, payload_size >> 3); sei->payloadType = GST_H264_SEI_UNHANDLED_PAYLOAD; break; } /* When SEI message doesn't end at byte boundary, * check remaining bits fit the specification. */ if (!nal_reader_is_byte_aligned (nr)) { guint8 bit_equal_to_one; READ_UINT8 (nr, bit_equal_to_one, 1); if (!bit_equal_to_one) GST_WARNING ("Bit non equal to one."); while (!nal_reader_is_byte_aligned (nr)) { guint8 bit_equal_to_zero; READ_UINT8 (nr, bit_equal_to_zero, 1); if (bit_equal_to_zero) GST_WARNING ("Bit non equal to zero."); } } /* Always make sure all the advertised SEI bits * were consumed during parsing */ if (next > nal_reader_get_pos (nr)) { GST_LOG ("Skipping %u unused SEI bits", next - nal_reader_get_pos (nr)); if (!nal_reader_skip_long (nr, next - nal_reader_get_pos (nr))) goto error; } return res; error: GST_WARNING ("error parsing \"Sei message\""); return GST_H264_PARSER_ERROR; } /******** API *************/ /** * gst_h264_nal_parser_new: * * Creates a new #GstH264NalParser. It should be freed with * gst_h264_nal_parser_free after use. * * Returns: a new #GstH264NalParser */ GstH264NalParser * gst_h264_nal_parser_new (void) { GstH264NalParser *nalparser; nalparser = g_slice_new0 (GstH264NalParser); return nalparser; } /** * gst_h264_nal_parser_free: * @nalparser: the #GstH264NalParser to free * * Frees @nalparser and sets it to %NULL */ void gst_h264_nal_parser_free (GstH264NalParser * nalparser) { guint i; for (i = 0; i < GST_H264_MAX_SPS_COUNT; i++) gst_h264_sps_clear (&nalparser->sps[i]); for (i = 0; i < GST_H264_MAX_PPS_COUNT; i++) gst_h264_pps_clear (&nalparser->pps[i]); g_slice_free (GstH264NalParser, nalparser); nalparser = NULL; } /** * gst_h264_parser_identify_nalu_unchecked: * @nalparser: a #GstH264NalParser * @data: The data to parse * @offset: the offset from which to parse @data * @size: the size of @data * @nalu: The #GstH264NalUnit where to store parsed nal headers * * Parses @data and fills @nalu from the next nalu data from @data. * * This differs from @gst_h264_parser_identify_nalu in that it doesn't * check whether the packet is complete or not. * * Note: Only use this function if you already know the provided @data * is a complete NALU, else use @gst_h264_parser_identify_nalu. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_identify_nalu_unchecked (GstH264NalParser * nalparser, const guint8 * data, guint offset, gsize size, GstH264NalUnit * nalu) { gint off1; memset (nalu, 0, sizeof (*nalu)); if (size < offset + 4) { GST_DEBUG ("Can't parse, buffer has too small size %" G_GSIZE_FORMAT ", offset %u", size, offset); return GST_H264_PARSER_ERROR; } off1 = scan_for_start_codes (data + offset, size - offset); if (off1 < 0) { GST_DEBUG ("No start code prefix in this buffer"); return GST_H264_PARSER_NO_NAL; } nalu->sc_offset = offset + off1; /* sc might have 2 or 3 0-bytes */ if (nalu->sc_offset > 0 && data[nalu->sc_offset - 1] == 00) nalu->sc_offset--; nalu->offset = offset + off1 + 3; nalu->data = (guint8 *) data; nalu->size = size - nalu->offset; if (!gst_h264_parse_nalu_header (nalu)) { GST_WARNING ("error parsing \"NAL unit header\""); nalu->size = 0; return GST_H264_PARSER_BROKEN_DATA; } nalu->valid = TRUE; if (nalu->type == GST_H264_NAL_SEQ_END || nalu->type == GST_H264_NAL_STREAM_END) { GST_DEBUG ("end-of-seq or end-of-stream nal found"); nalu->size = 1; return GST_H264_PARSER_OK; } return GST_H264_PARSER_OK; } /** * gst_h264_parser_identify_nalu: * @nalparser: a #GstH264NalParser * @data: The data to parse, containing an Annex B coded NAL unit * @offset: the offset in @data from which to parse the NAL unit * @size: the size of @data * @nalu: The #GstH264NalUnit to store the identified NAL unit in * * Parses the headers of an Annex B coded NAL unit from @data and puts the * result into @nalu. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_identify_nalu (GstH264NalParser * nalparser, const guint8 * data, guint offset, gsize size, GstH264NalUnit * nalu) { GstH264ParserResult res; gint off2; res = gst_h264_parser_identify_nalu_unchecked (nalparser, data, offset, size, nalu); if (res != GST_H264_PARSER_OK) goto beach; /* The two NALs are exactly 1 byte size and are placed at the end of an AU, * there is no need to wait for the following */ if (nalu->type == GST_H264_NAL_SEQ_END || nalu->type == GST_H264_NAL_STREAM_END) goto beach; off2 = scan_for_start_codes (data + nalu->offset, size - nalu->offset); if (off2 < 0) { GST_DEBUG ("Nal start %d, No end found", nalu->offset); return GST_H264_PARSER_NO_NAL_END; } /* Mini performance improvement: * We could have a way to store how many 0s were skipped to avoid * parsing them again on the next NAL */ while (off2 > 0 && data[nalu->offset + off2 - 1] == 00) off2--; nalu->size = off2; if (nalu->size < 2) return GST_H264_PARSER_BROKEN_DATA; GST_DEBUG ("Complete nal found. Off: %d, Size: %d", nalu->offset, nalu->size); beach: return res; } /** * gst_h264_parser_identify_nalu_avc: * @nalparser: a #GstH264NalParser * @data: The data to parse, containing an AVC coded NAL unit * @offset: the offset in @data from which to parse the NAL unit * @size: the size of @data * @nal_length_size: the size in bytes of the AVC nal length prefix. * @nalu: The #GstH264NalUnit to store the identified NAL unit in * * Parses the headers of an AVC coded NAL unit from @data and puts the result * into @nalu. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_identify_nalu_avc (GstH264NalParser * nalparser, const guint8 * data, guint offset, gsize size, guint8 nal_length_size, GstH264NalUnit * nalu) { GstBitReader br; memset (nalu, 0, sizeof (*nalu)); /* Would overflow guint below otherwise: the callers needs to ensure that * this never happens */ if (offset > G_MAXUINT32 - nal_length_size) { GST_WARNING ("offset + nal_length_size overflow"); nalu->size = 0; return GST_H264_PARSER_BROKEN_DATA; } if (size < offset + nal_length_size) { GST_DEBUG ("Can't parse, buffer has too small size %" G_GSIZE_FORMAT ", offset %u", size, offset); return GST_H264_PARSER_ERROR; } size = size - offset; gst_bit_reader_init (&br, data + offset, size); nalu->size = gst_bit_reader_get_bits_uint32_unchecked (&br, nal_length_size * 8); nalu->sc_offset = offset; nalu->offset = offset + nal_length_size; if (nalu->size > G_MAXUINT32 - nal_length_size) { GST_WARNING ("NALU size + nal_length_size overflow"); nalu->size = 0; return GST_H264_PARSER_BROKEN_DATA; } if (size < (gsize) nalu->size + nal_length_size) { nalu->size = 0; return GST_H264_PARSER_NO_NAL_END; } nalu->data = (guint8 *) data; if (!gst_h264_parse_nalu_header (nalu)) { GST_WARNING ("error parsing \"NAL unit header\""); nalu->size = 0; return GST_H264_PARSER_BROKEN_DATA; } nalu->valid = TRUE; return GST_H264_PARSER_OK; } /** * gst_h264_parser_parse_nal: * @nalparser: a #GstH264NalParser * @nalu: The #GstH264NalUnit to parse * * This function should be called in the case one doesn't need to * parse a specific structure. It is necessary to do so to make * sure @nalparser is up to date. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_parse_nal (GstH264NalParser * nalparser, GstH264NalUnit * nalu) { GstH264SPS sps; GstH264PPS pps; switch (nalu->type) { case GST_H264_NAL_SPS: return gst_h264_parser_parse_sps (nalparser, nalu, &sps); break; case GST_H264_NAL_PPS: return gst_h264_parser_parse_pps (nalparser, nalu, &pps); } return GST_H264_PARSER_OK; } /** * gst_h264_parser_parse_sps: * @nalparser: a #GstH264NalParser * @nalu: The #GST_H264_NAL_SPS #GstH264NalUnit to parse * @sps: The #GstH264SPS to fill. * * Parses @nalu containing a Sequence Parameter Set, and fills @sps. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_parse_sps (GstH264NalParser * nalparser, GstH264NalUnit * nalu, GstH264SPS * sps) { GstH264ParserResult res = gst_h264_parse_sps (nalu, sps); if (res == GST_H264_PARSER_OK) { GST_DEBUG ("adding sequence parameter set with id: %d to array", sps->id); if (!gst_h264_sps_copy (&nalparser->sps[sps->id], sps)) return GST_H264_PARSER_ERROR; nalparser->last_sps = &nalparser->sps[sps->id]; } return res; } /* Parse seq_parameter_set_data() */ static gboolean gst_h264_parse_sps_data (NalReader * nr, GstH264SPS * sps) { gint width, height; guint subwc[] = { 1, 2, 2, 1 }; guint subhc[] = { 1, 2, 1, 1 }; memset (sps, 0, sizeof (*sps)); /* set default values for fields that might not be present in the bitstream and have valid defaults */ sps->extension_type = GST_H264_NAL_EXTENSION_NONE; sps->chroma_format_idc = 1; memset (sps->scaling_lists_4x4, 16, 96); memset (sps->scaling_lists_8x8, 16, 384); READ_UINT8 (nr, sps->profile_idc, 8); READ_UINT8 (nr, sps->constraint_set0_flag, 1); READ_UINT8 (nr, sps->constraint_set1_flag, 1); READ_UINT8 (nr, sps->constraint_set2_flag, 1); READ_UINT8 (nr, sps->constraint_set3_flag, 1); READ_UINT8 (nr, sps->constraint_set4_flag, 1); READ_UINT8 (nr, sps->constraint_set5_flag, 1); /* skip reserved_zero_2bits */ if (!nal_reader_skip (nr, 2)) goto error; READ_UINT8 (nr, sps->level_idc, 8); READ_UE_MAX (nr, sps->id, GST_H264_MAX_SPS_COUNT - 1); if (sps->profile_idc == 100 || sps->profile_idc == 110 || sps->profile_idc == 122 || sps->profile_idc == 244 || sps->profile_idc == 44 || sps->profile_idc == 83 || sps->profile_idc == 86 || sps->profile_idc == 118 || sps->profile_idc == 128 || sps->profile_idc == 138 || sps->profile_idc == 139 || sps->profile_idc == 134 || sps->profile_idc == 135) { READ_UE_MAX (nr, sps->chroma_format_idc, 3); if (sps->chroma_format_idc == 3) READ_UINT8 (nr, sps->separate_colour_plane_flag, 1); READ_UE_MAX (nr, sps->bit_depth_luma_minus8, 6); READ_UE_MAX (nr, sps->bit_depth_chroma_minus8, 6); READ_UINT8 (nr, sps->qpprime_y_zero_transform_bypass_flag, 1); READ_UINT8 (nr, sps->scaling_matrix_present_flag, 1); if (sps->scaling_matrix_present_flag) { guint8 n_lists; n_lists = (sps->chroma_format_idc != 3) ? 8 : 12; if (!gst_h264_parser_parse_scaling_list (nr, sps->scaling_lists_4x4, sps->scaling_lists_8x8, default_4x4_inter, default_4x4_intra, default_8x8_inter, default_8x8_intra, n_lists)) goto error; } } READ_UE_MAX (nr, sps->log2_max_frame_num_minus4, 12); sps->max_frame_num = 1 << (sps->log2_max_frame_num_minus4 + 4); READ_UE_MAX (nr, sps->pic_order_cnt_type, 2); if (sps->pic_order_cnt_type == 0) { READ_UE_MAX (nr, sps->log2_max_pic_order_cnt_lsb_minus4, 12); } else if (sps->pic_order_cnt_type == 1) { guint i; READ_UINT8 (nr, sps->delta_pic_order_always_zero_flag, 1); READ_SE (nr, sps->offset_for_non_ref_pic); READ_SE (nr, sps->offset_for_top_to_bottom_field); READ_UE_MAX (nr, sps->num_ref_frames_in_pic_order_cnt_cycle, 255); for (i = 0; i < sps->num_ref_frames_in_pic_order_cnt_cycle; i++) READ_SE (nr, sps->offset_for_ref_frame[i]); } READ_UE (nr, sps->num_ref_frames); READ_UINT8 (nr, sps->gaps_in_frame_num_value_allowed_flag, 1); READ_UE (nr, sps->pic_width_in_mbs_minus1); READ_UE (nr, sps->pic_height_in_map_units_minus1); READ_UINT8 (nr, sps->frame_mbs_only_flag, 1); if (!sps->frame_mbs_only_flag) READ_UINT8 (nr, sps->mb_adaptive_frame_field_flag, 1); READ_UINT8 (nr, sps->direct_8x8_inference_flag, 1); READ_UINT8 (nr, sps->frame_cropping_flag, 1); if (sps->frame_cropping_flag) { READ_UE (nr, sps->frame_crop_left_offset); READ_UE (nr, sps->frame_crop_right_offset); READ_UE (nr, sps->frame_crop_top_offset); READ_UE (nr, sps->frame_crop_bottom_offset); } READ_UINT8 (nr, sps->vui_parameters_present_flag, 1); if (sps->vui_parameters_present_flag) if (!gst_h264_parse_vui_parameters (sps, nr)) goto error; /* calculate ChromaArrayType */ if (!sps->separate_colour_plane_flag) sps->chroma_array_type = sps->chroma_format_idc; /* Calculate width and height */ width = (sps->pic_width_in_mbs_minus1 + 1); width *= 16; height = (sps->pic_height_in_map_units_minus1 + 1); height *= 16 * (2 - sps->frame_mbs_only_flag); GST_LOG ("initial width=%d, height=%d", width, height); if (width < 0 || height < 0) { GST_WARNING ("invalid width/height in SPS"); goto error; } sps->width = width; sps->height = height; if (sps->frame_cropping_flag) { const guint crop_unit_x = subwc[sps->chroma_format_idc]; const guint crop_unit_y = subhc[sps->chroma_format_idc] * (2 - sps->frame_mbs_only_flag); width -= (sps->frame_crop_left_offset + sps->frame_crop_right_offset) * crop_unit_x; height -= (sps->frame_crop_top_offset + sps->frame_crop_bottom_offset) * crop_unit_y; sps->crop_rect_width = width; sps->crop_rect_height = height; sps->crop_rect_x = sps->frame_crop_left_offset * crop_unit_x; sps->crop_rect_y = sps->frame_crop_top_offset * crop_unit_y; GST_LOG ("crop_rectangle x=%u y=%u width=%u, height=%u", sps->crop_rect_x, sps->crop_rect_y, width, height); } sps->fps_num_removed = 0; sps->fps_den_removed = 1; return TRUE; error: return FALSE; } /* Parse subset_seq_parameter_set() data for MVC */ static gboolean gst_h264_parse_sps_mvc_data (NalReader * nr, GstH264SPS * sps) { GstH264SPSExtMVC *const mvc = &sps->extension.mvc; guint8 bit_equal_to_one; guint i, j, k; READ_UINT8 (nr, bit_equal_to_one, 1); if (!bit_equal_to_one) return FALSE; sps->extension_type = GST_H264_NAL_EXTENSION_MVC; READ_UE_MAX (nr, mvc->num_views_minus1, GST_H264_MAX_VIEW_COUNT - 1); mvc->view = g_new0 (GstH264SPSExtMVCView, mvc->num_views_minus1 + 1); if (!mvc->view) goto error_allocation_failed; for (i = 0; i <= mvc->num_views_minus1; i++) READ_UE_MAX (nr, mvc->view[i].view_id, GST_H264_MAX_VIEW_ID); for (i = 1; i <= mvc->num_views_minus1; i++) { /* for RefPicList0 */ READ_UE_MAX (nr, mvc->view[i].num_anchor_refs_l0, 15); for (j = 0; j < mvc->view[i].num_anchor_refs_l0; j++) { READ_UE_MAX (nr, mvc->view[i].anchor_ref_l0[j], GST_H264_MAX_VIEW_ID); } /* for RefPicList1 */ READ_UE_MAX (nr, mvc->view[i].num_anchor_refs_l1, 15); for (j = 0; j < mvc->view[i].num_anchor_refs_l1; j++) { READ_UE_MAX (nr, mvc->view[i].anchor_ref_l1[j], GST_H264_MAX_VIEW_ID); } } for (i = 1; i <= mvc->num_views_minus1; i++) { /* for RefPicList0 */ READ_UE_MAX (nr, mvc->view[i].num_non_anchor_refs_l0, 15); for (j = 0; j < mvc->view[i].num_non_anchor_refs_l0; j++) { READ_UE_MAX (nr, mvc->view[i].non_anchor_ref_l0[j], GST_H264_MAX_VIEW_ID); } /* for RefPicList1 */ READ_UE_MAX (nr, mvc->view[i].num_non_anchor_refs_l1, 15); for (j = 0; j < mvc->view[i].num_non_anchor_refs_l1; j++) { READ_UE_MAX (nr, mvc->view[i].non_anchor_ref_l1[j], GST_H264_MAX_VIEW_ID); } } READ_UE_MAX (nr, mvc->num_level_values_signalled_minus1, 63); mvc->level_value = g_new0 (GstH264SPSExtMVCLevelValue, mvc->num_level_values_signalled_minus1 + 1); if (!mvc->level_value) goto error_allocation_failed; for (i = 0; i <= mvc->num_level_values_signalled_minus1; i++) { GstH264SPSExtMVCLevelValue *const level_value = &mvc->level_value[i]; READ_UINT8 (nr, level_value->level_idc, 8); READ_UE_MAX (nr, level_value->num_applicable_ops_minus1, 1023); level_value->applicable_op = g_new0 (GstH264SPSExtMVCLevelValueOp, level_value->num_applicable_ops_minus1 + 1); if (!level_value->applicable_op) goto error_allocation_failed; for (j = 0; j <= level_value->num_applicable_ops_minus1; j++) { GstH264SPSExtMVCLevelValueOp *const op = &level_value->applicable_op[j]; READ_UINT8 (nr, op->temporal_id, 3); READ_UE_MAX (nr, op->num_target_views_minus1, 1023); op->target_view_id = g_new (guint16, op->num_target_views_minus1 + 1); if (!op->target_view_id) goto error_allocation_failed; for (k = 0; k <= op->num_target_views_minus1; k++) READ_UE_MAX (nr, op->target_view_id[k], GST_H264_MAX_VIEW_ID); READ_UE_MAX (nr, op->num_views_minus1, 1023); } } return TRUE; error_allocation_failed: GST_WARNING ("failed to allocate memory"); gst_h264_sps_clear (sps); return FALSE; error: gst_h264_sps_clear (sps); return FALSE; } /** * gst_h264_parse_sps: * @nalu: The #GST_H264_NAL_SPS #GstH264NalUnit to parse * @sps: The #GstH264SPS to fill. * * Parses @data, and fills the @sps structure. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parse_sps (GstH264NalUnit * nalu, GstH264SPS * sps) { NalReader nr; GST_DEBUG ("parsing SPS"); nal_reader_init (&nr, nalu->data + nalu->offset + nalu->header_bytes, nalu->size - nalu->header_bytes); if (!gst_h264_parse_sps_data (&nr, sps)) goto error; sps->valid = TRUE; return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Sequence parameter set\""); sps->valid = FALSE; return GST_H264_PARSER_ERROR; } /** * gst_h264_parser_parse_subset_sps: * @nalparser: a #GstH264NalParser * @nalu: The #GST_H264_NAL_SUBSET_SPS #GstH264NalUnit to parse * @sps: The #GstH264SPS to fill. * * Parses @data, and fills in the @sps structure. * * This function fully parses @data and allocates all the necessary * data structures needed for MVC extensions. The resulting @sps * structure shall be deallocated with gst_h264_sps_clear() when it is * no longer needed. * * Note: if the caller doesn't need any of the MVC-specific data, then * gst_h264_parser_parse_sps() is more efficient because those extra * syntax elements are not parsed and no extra memory is allocated. * * Returns: a #GstH264ParserResult * * Since: 1.6 */ GstH264ParserResult gst_h264_parser_parse_subset_sps (GstH264NalParser * nalparser, GstH264NalUnit * nalu, GstH264SPS * sps) { GstH264ParserResult res; res = gst_h264_parse_subset_sps (nalu, sps); if (res == GST_H264_PARSER_OK) { GST_DEBUG ("adding sequence parameter set with id: %d to array", sps->id); if (!gst_h264_sps_copy (&nalparser->sps[sps->id], sps)) { gst_h264_sps_clear (sps); return GST_H264_PARSER_ERROR; } nalparser->last_sps = &nalparser->sps[sps->id]; } return res; } /** * gst_h264_parse_subset_sps: * @nalu: The #GST_H264_NAL_SUBSET_SPS #GstH264NalUnit to parse * @sps: The #GstH264SPS to fill. * * Parses @data, and fills in the @sps structure. * * This function fully parses @data and allocates all the necessary * data structures needed for MVC extensions. The resulting @sps * structure shall be deallocated with gst_h264_sps_clear() when it is * no longer needed. * * Note: if the caller doesn't need any of the MVC-specific data, then * gst_h264_parser_parse_sps() is more efficient because those extra * syntax elements are not parsed and no extra memory is allocated. * * Returns: a #GstH264ParserResult * * Since: 1.6 */ GstH264ParserResult gst_h264_parse_subset_sps (GstH264NalUnit * nalu, GstH264SPS * sps) { NalReader nr; GST_DEBUG ("parsing Subset SPS"); nal_reader_init (&nr, nalu->data + nalu->offset + nalu->header_bytes, nalu->size - nalu->header_bytes); if (!gst_h264_parse_sps_data (&nr, sps)) goto error; if (sps->profile_idc == GST_H264_PROFILE_MULTIVIEW_HIGH || sps->profile_idc == GST_H264_PROFILE_STEREO_HIGH) { if (!gst_h264_parse_sps_mvc_data (&nr, sps)) goto error; } sps->valid = TRUE; return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Subset sequence parameter set\""); gst_h264_sps_clear (sps); sps->valid = FALSE; return GST_H264_PARSER_ERROR; } /** * gst_h264_parse_pps: * @nalparser: a #GstH264NalParser * @nalu: The #GST_H264_NAL_PPS #GstH264NalUnit to parse * @pps: The #GstH264PPS to fill. * * Parses @data, and fills the @pps structure. * * The resulting @pps data structure shall be deallocated with the * gst_h264_pps_clear() function when it is no longer needed, or prior * to parsing a new PPS NAL unit. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parse_pps (GstH264NalParser * nalparser, GstH264NalUnit * nalu, GstH264PPS * pps) { NalReader nr; GstH264SPS *sps; gint sps_id; gint qp_bd_offset; GST_DEBUG ("parsing PPS"); nal_reader_init (&nr, nalu->data + nalu->offset + nalu->header_bytes, nalu->size - nalu->header_bytes); memset (pps, 0, sizeof (*pps)); READ_UE_MAX (&nr, pps->id, GST_H264_MAX_PPS_COUNT - 1); READ_UE_MAX (&nr, sps_id, GST_H264_MAX_SPS_COUNT - 1); sps = gst_h264_parser_get_sps (nalparser, sps_id); if (!sps) { GST_WARNING ("couldn't find associated sequence parameter set with id: %d", sps_id); return GST_H264_PARSER_BROKEN_LINK; } pps->sequence = sps; qp_bd_offset = 6 * (sps->bit_depth_luma_minus8 + sps->separate_colour_plane_flag); /* set default values for fields that might not be present in the bitstream and have valid defaults */ memcpy (&pps->scaling_lists_4x4, &sps->scaling_lists_4x4, 96); memcpy (&pps->scaling_lists_8x8, &sps->scaling_lists_8x8, 384); READ_UINT8 (&nr, pps->entropy_coding_mode_flag, 1); READ_UINT8 (&nr, pps->pic_order_present_flag, 1); READ_UE_MAX (&nr, pps->num_slice_groups_minus1, 7); if (pps->num_slice_groups_minus1 > 0) { READ_UE_MAX (&nr, pps->slice_group_map_type, 6); if (pps->slice_group_map_type == 0) { gint i; for (i = 0; i <= pps->num_slice_groups_minus1; i++) READ_UE (&nr, pps->run_length_minus1[i]); } else if (pps->slice_group_map_type == 2) { gint i; for (i = 0; i < pps->num_slice_groups_minus1; i++) { READ_UE (&nr, pps->top_left[i]); READ_UE (&nr, pps->bottom_right[i]); } } else if (pps->slice_group_map_type >= 3 && pps->slice_group_map_type <= 5) { READ_UINT8 (&nr, pps->slice_group_change_direction_flag, 1); READ_UE (&nr, pps->slice_group_change_rate_minus1); } else if (pps->slice_group_map_type == 6) { gint bits; gint i; READ_UE (&nr, pps->pic_size_in_map_units_minus1); bits = g_bit_storage (pps->num_slice_groups_minus1); pps->slice_group_id = g_new (guint8, pps->pic_size_in_map_units_minus1 + 1); for (i = 0; i <= pps->pic_size_in_map_units_minus1; i++) READ_UINT8 (&nr, pps->slice_group_id[i], bits); } } READ_UE_MAX (&nr, pps->num_ref_idx_l0_active_minus1, 31); READ_UE_MAX (&nr, pps->num_ref_idx_l1_active_minus1, 31); READ_UINT8 (&nr, pps->weighted_pred_flag, 1); READ_UINT8 (&nr, pps->weighted_bipred_idc, 2); READ_SE_ALLOWED (&nr, pps->pic_init_qp_minus26, -(26 + qp_bd_offset), 25); READ_SE_ALLOWED (&nr, pps->pic_init_qs_minus26, -26, 25); READ_SE_ALLOWED (&nr, pps->chroma_qp_index_offset, -12, 12); pps->second_chroma_qp_index_offset = pps->chroma_qp_index_offset; READ_UINT8 (&nr, pps->deblocking_filter_control_present_flag, 1); READ_UINT8 (&nr, pps->constrained_intra_pred_flag, 1); READ_UINT8 (&nr, pps->redundant_pic_cnt_present_flag, 1); if (!nal_reader_has_more_data (&nr)) goto done; READ_UINT8 (&nr, pps->transform_8x8_mode_flag, 1); READ_UINT8 (&nr, pps->pic_scaling_matrix_present_flag, 1); if (pps->pic_scaling_matrix_present_flag) { guint8 n_lists; n_lists = 6 + ((sps->chroma_format_idc != 3) ? 2 : 6) * pps->transform_8x8_mode_flag; if (sps->scaling_matrix_present_flag) { if (!gst_h264_parser_parse_scaling_list (&nr, pps->scaling_lists_4x4, pps->scaling_lists_8x8, sps->scaling_lists_4x4[3], sps->scaling_lists_4x4[0], sps->scaling_lists_8x8[3], sps->scaling_lists_8x8[0], n_lists)) goto error; } else { if (!gst_h264_parser_parse_scaling_list (&nr, pps->scaling_lists_4x4, pps->scaling_lists_8x8, default_4x4_inter, default_4x4_intra, default_8x8_inter, default_8x8_intra, n_lists)) goto error; } } READ_SE_ALLOWED (&nr, pps->second_chroma_qp_index_offset, -12, 12); done: pps->valid = TRUE; return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Picture parameter set\""); pps->valid = FALSE; gst_h264_pps_clear (pps); return GST_H264_PARSER_ERROR; } /** * gst_h264_parser_parse_pps: * @nalparser: a #GstH264NalParser * @nalu: The #GST_H264_NAL_PPS #GstH264NalUnit to parse * @pps: The #GstH264PPS to fill. * * Parses @nalu containing a Picture Parameter Set, and fills @pps. * * The resulting @pps data structure must be deallocated by the caller using * gst_h264_pps_clear(). * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_parse_pps (GstH264NalParser * nalparser, GstH264NalUnit * nalu, GstH264PPS * pps) { GstH264ParserResult res = gst_h264_parse_pps (nalparser, nalu, pps); if (res == GST_H264_PARSER_OK) { GST_DEBUG ("adding picture parameter set with id: %d to array", pps->id); if (!gst_h264_pps_copy (&nalparser->pps[pps->id], pps)) return GST_H264_PARSER_ERROR; nalparser->last_pps = &nalparser->pps[pps->id]; } return res; } /** * gst_h264_pps_clear: * @pps: The #GstH264PPS to free * * Clears all @pps internal resources. * * Since: 1.4 */ void gst_h264_pps_clear (GstH264PPS * pps) { g_return_if_fail (pps != NULL); g_free (pps->slice_group_id); pps->slice_group_id = NULL; } /** * gst_h264_parser_parse_slice_hdr: * @nalparser: a #GstH264NalParser * @nalu: The #GST_H264_NAL_SLICE to #GST_H264_NAL_SLICE_IDR #GstH264NalUnit to parse * @slice: The #GstH264SliceHdr to fill. * @parse_pred_weight_table: Whether to parse the pred_weight_table or not * @parse_dec_ref_pic_marking: Whether to parse the dec_ref_pic_marking or not * * Parses @nalu containing a coded slice, and fills @slice. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_parse_slice_hdr (GstH264NalParser * nalparser, GstH264NalUnit * nalu, GstH264SliceHdr * slice, gboolean parse_pred_weight_table, gboolean parse_dec_ref_pic_marking) { NalReader nr; gint pps_id; GstH264PPS *pps; GstH264SPS *sps; guint start_pos, start_epb; memset (slice, 0, sizeof (*slice)); if (!nalu->size) { GST_DEBUG ("Invalid Nal Unit"); return GST_H264_PARSER_ERROR; } nal_reader_init (&nr, nalu->data + nalu->offset + nalu->header_bytes, nalu->size - nalu->header_bytes); READ_UE (&nr, slice->first_mb_in_slice); READ_UE (&nr, slice->type); GST_DEBUG ("parsing \"Slice header\", slice type %u", slice->type); READ_UE_MAX (&nr, pps_id, GST_H264_MAX_PPS_COUNT - 1); pps = gst_h264_parser_get_pps (nalparser, pps_id); if (!pps) { GST_WARNING ("couldn't find associated picture parameter set with id: %d", pps_id); return GST_H264_PARSER_BROKEN_LINK; } slice->pps = pps; sps = pps->sequence; if (!sps) { GST_WARNING ("couldn't find associated sequence parameter set with id: %d", pps->id); return GST_H264_PARSER_BROKEN_LINK; } /* Check we can actually parse this slice (AVC, MVC headers only) */ if (sps->extension_type && sps->extension_type != GST_H264_NAL_EXTENSION_MVC) { GST_WARNING ("failed to parse unsupported slice header"); return GST_H264_PARSER_BROKEN_DATA; } /* set default values for fields that might not be present in the bitstream and have valid defaults */ if (GST_H264_IS_I_SLICE (slice)) { slice->num_ref_idx_l0_active_minus1 = 0; slice->num_ref_idx_l1_active_minus1 = 0; } else { slice->num_ref_idx_l0_active_minus1 = pps->num_ref_idx_l0_active_minus1; if (GST_H264_IS_B_SLICE (slice)) slice->num_ref_idx_l1_active_minus1 = pps->num_ref_idx_l1_active_minus1; else slice->num_ref_idx_l1_active_minus1 = 0; } if (sps->separate_colour_plane_flag) READ_UINT8 (&nr, slice->colour_plane_id, 2); READ_UINT16 (&nr, slice->frame_num, sps->log2_max_frame_num_minus4 + 4); if (!sps->frame_mbs_only_flag) { READ_UINT8 (&nr, slice->field_pic_flag, 1); if (slice->field_pic_flag) READ_UINT8 (&nr, slice->bottom_field_flag, 1); } /* calculate MaxPicNum */ if (slice->field_pic_flag) slice->max_pic_num = 2 * sps->max_frame_num; else slice->max_pic_num = sps->max_frame_num; if (nalu->idr_pic_flag) READ_UE_MAX (&nr, slice->idr_pic_id, G_MAXUINT16); start_pos = nal_reader_get_pos (&nr); start_epb = nal_reader_get_epb_count (&nr); if (sps->pic_order_cnt_type == 0) { READ_UINT16 (&nr, slice->pic_order_cnt_lsb, sps->log2_max_pic_order_cnt_lsb_minus4 + 4); if (pps->pic_order_present_flag && !slice->field_pic_flag) READ_SE (&nr, slice->delta_pic_order_cnt_bottom); } if (sps->pic_order_cnt_type == 1 && !sps->delta_pic_order_always_zero_flag) { READ_SE (&nr, slice->delta_pic_order_cnt[0]); if (pps->pic_order_present_flag && !slice->field_pic_flag) READ_SE (&nr, slice->delta_pic_order_cnt[1]); } slice->pic_order_cnt_bit_size = (nal_reader_get_pos (&nr) - start_pos) - (8 * (nal_reader_get_epb_count (&nr) - start_epb)); if (pps->redundant_pic_cnt_present_flag) READ_UE_MAX (&nr, slice->redundant_pic_cnt, G_MAXINT8); if (GST_H264_IS_B_SLICE (slice)) READ_UINT8 (&nr, slice->direct_spatial_mv_pred_flag, 1); if (GST_H264_IS_P_SLICE (slice) || GST_H264_IS_SP_SLICE (slice) || GST_H264_IS_B_SLICE (slice)) { READ_UINT8 (&nr, slice->num_ref_idx_active_override_flag, 1); if (slice->num_ref_idx_active_override_flag) { READ_UE_MAX (&nr, slice->num_ref_idx_l0_active_minus1, 31); if (GST_H264_IS_B_SLICE (slice)) READ_UE_MAX (&nr, slice->num_ref_idx_l1_active_minus1, 31); } } if (!slice_parse_ref_pic_list_modification (slice, &nr, GST_H264_IS_MVC_NALU (nalu))) goto error; if ((pps->weighted_pred_flag && (GST_H264_IS_P_SLICE (slice) || GST_H264_IS_SP_SLICE (slice))) || (pps->weighted_bipred_idc == 1 && GST_H264_IS_B_SLICE (slice))) { if (!gst_h264_slice_parse_pred_weight_table (slice, &nr, sps->chroma_array_type)) goto error; } if (nalu->ref_idc != 0) { if (!gst_h264_slice_parse_dec_ref_pic_marking (slice, nalu, &nr)) goto error; } if (pps->entropy_coding_mode_flag && !GST_H264_IS_I_SLICE (slice) && !GST_H264_IS_SI_SLICE (slice)) READ_UE_MAX (&nr, slice->cabac_init_idc, 2); READ_SE_ALLOWED (&nr, slice->slice_qp_delta, -87, 77); if (GST_H264_IS_SP_SLICE (slice) || GST_H264_IS_SI_SLICE (slice)) { if (GST_H264_IS_SP_SLICE (slice)) READ_UINT8 (&nr, slice->sp_for_switch_flag, 1); READ_SE_ALLOWED (&nr, slice->slice_qs_delta, -51, 51); } if (pps->deblocking_filter_control_present_flag) { READ_UE_MAX (&nr, slice->disable_deblocking_filter_idc, 2); if (slice->disable_deblocking_filter_idc != 1) { READ_SE_ALLOWED (&nr, slice->slice_alpha_c0_offset_div2, -6, 6); READ_SE_ALLOWED (&nr, slice->slice_beta_offset_div2, -6, 6); } } if (pps->num_slice_groups_minus1 > 0 && pps->slice_group_map_type >= 3 && pps->slice_group_map_type <= 5) { /* Ceil(Log2(PicSizeInMapUnits / SliceGroupChangeRate + 1)) [7-33] */ guint32 PicWidthInMbs = sps->pic_width_in_mbs_minus1 + 1; guint32 PicHeightInMapUnits = sps->pic_height_in_map_units_minus1 + 1; guint32 PicSizeInMapUnits = PicWidthInMbs * PicHeightInMapUnits; guint32 SliceGroupChangeRate = pps->slice_group_change_rate_minus1 + 1; const guint n = ceil_log2 (PicSizeInMapUnits / SliceGroupChangeRate + 1); READ_UINT16 (&nr, slice->slice_group_change_cycle, n); } slice->header_size = nal_reader_get_pos (&nr); slice->n_emulation_prevention_bytes = nal_reader_get_epb_count (&nr); return GST_H264_PARSER_OK; error: GST_WARNING ("error parsing \"Slice header\""); return GST_H264_PARSER_ERROR; } /* Free MVC-specific data from subset SPS header */ static void gst_h264_sps_mvc_clear (GstH264SPS * sps) { GstH264SPSExtMVC *const mvc = &sps->extension.mvc; guint i, j; g_assert (sps->extension_type == GST_H264_NAL_EXTENSION_MVC); g_free (mvc->view); mvc->view = NULL; for (i = 0; i <= mvc->num_level_values_signalled_minus1; i++) { GstH264SPSExtMVCLevelValue *const level_value = &mvc->level_value[i]; for (j = 0; j <= level_value->num_applicable_ops_minus1; j++) { g_free (level_value->applicable_op[j].target_view_id); level_value->applicable_op[j].target_view_id = NULL; } g_free (level_value->applicable_op); level_value->applicable_op = NULL; } g_free (mvc->level_value); mvc->level_value = NULL; /* All meaningful MVC info are now gone, just pretend to be a * standard AVC struct now */ sps->extension_type = GST_H264_NAL_EXTENSION_NONE; } /** * gst_h264_sps_clear: * @sps: The #GstH264SPS to free * * Clears all @sps internal resources. * * Since: 1.6 */ void gst_h264_sps_clear (GstH264SPS * sps) { g_return_if_fail (sps != NULL); switch (sps->extension_type) { case GST_H264_NAL_EXTENSION_MVC: gst_h264_sps_mvc_clear (sps); break; } } /** * gst_h264_sei_clear: * sei: The #GstH264SEIMessage to clear * * Frees allocated data in @sei if any. * * Since: 1.18 */ void gst_h264_sei_clear (GstH264SEIMessage * sei) { switch (sei->payloadType) { case GST_H264_SEI_REGISTERED_USER_DATA:{ GstH264RegisteredUserData *rud = &sei->payload.registered_user_data; g_free ((guint8 *) rud->data); rud->data = NULL; break; } case GST_H264_SEI_UNHANDLED_PAYLOAD:{ GstH264SEIUnhandledPayload *payload = &sei->payload.unhandled_payload; g_free (payload->data); payload->data = NULL; payload->size = 0; break; } default: break; } } /** * gst_h264_parser_parse_sei: * @nalparser: a #GstH264NalParser * @nalu: The #GST_H264_NAL_SEI #GstH264NalUnit to parse * @messages: The GArray of #GstH264SEIMessage to fill. The caller must free it when done. * * Parses @nalu containing one or more Supplementary Enhancement Information messages, * and allocates and fills the @messages array. * * Returns: a #GstH264ParserResult */ GstH264ParserResult gst_h264_parser_parse_sei (GstH264NalParser * nalparser, GstH264NalUnit * nalu, GArray ** messages) { NalReader nr; GstH264SEIMessage sei; GstH264ParserResult res; GST_DEBUG ("parsing SEI nal"); nal_reader_init (&nr, nalu->data + nalu->offset + nalu->header_bytes, nalu->size - nalu->header_bytes); *messages = g_array_new (FALSE, FALSE, sizeof (GstH264SEIMessage)); g_array_set_clear_func (*messages, (GDestroyNotify) gst_h264_sei_clear); do { res = gst_h264_parser_parse_sei_message (nalparser, &nr, &sei); if (res == GST_H264_PARSER_OK) g_array_append_val (*messages, sei); else break; } while (nal_reader_has_more_data (&nr)); return res; } /** * gst_h264_parser_update_sps: * @nalparser: a #GstH264NalParser * @sps: (transfer none): a #GstH264SPS. * * Replace internal Sequence Parameter Set struct corresponding to id of @sps * with @sps. @nalparser will mark @sps as last parsed sps. * * Returns: a #GstH264ParserResult * * Since: 1.18 */ GstH264ParserResult gst_h264_parser_update_sps (GstH264NalParser * nalparser, GstH264SPS * sps) { g_return_val_if_fail (nalparser != NULL, GST_H264_PARSER_ERROR); g_return_val_if_fail (sps != NULL, GST_H264_PARSER_ERROR); g_return_val_if_fail (sps->id >= 0 && sps->id < GST_H264_MAX_SPS_COUNT, GST_H264_PARSER_ERROR); if (!sps->valid) { GST_WARNING ("Cannot update with invalid SPS"); return GST_H264_PARSER_ERROR; } GST_DEBUG ("Updating sequence parameter set with id: %d", sps->id); if (!gst_h264_sps_copy (&nalparser->sps[sps->id], sps)) return GST_H264_PARSER_ERROR; nalparser->last_sps = &nalparser->sps[sps->id]; return GST_H264_PARSER_OK; } /** * gst_h264_parser_update_pps: * @nalparser: a #GstH264NalParser * @pps: (transfer none): a #GstH264PPS. * * Replace internal Picture Parameter Set struct corresponding to id of @pps * with @pps. @nalparser will mark @pps as last parsed pps. * * Returns: a #GstH264ParserResult * * Since: 1.18 */ GstH264ParserResult gst_h264_parser_update_pps (GstH264NalParser * nalparser, GstH264PPS * pps) { GstH264SPS *sps; g_return_val_if_fail (nalparser != NULL, GST_H264_PARSER_ERROR); g_return_val_if_fail (pps != NULL, GST_H264_PARSER_ERROR); g_return_val_if_fail (pps->id >= 0 && pps->id < GST_H264_MAX_PPS_COUNT, GST_H264_PARSER_ERROR); if (!pps->valid) { GST_WARNING ("Cannot update with invalid PPS"); return GST_H264_PARSER_ERROR; } if (!pps->sequence) { GST_WARNING ("No linked SPS struct"); return GST_H264_PARSER_BROKEN_LINK; } sps = gst_h264_parser_get_sps (nalparser, pps->sequence->id); if (!sps || sps != pps->sequence) { GST_WARNING ("Linked SPS is not identical to internal SPS"); return GST_H264_PARSER_BROKEN_LINK; } GST_DEBUG ("Updating picture parameter set with id: %d", pps->id); if (!gst_h264_pps_copy (&nalparser->pps[pps->id], pps)) return GST_H264_PARSER_ERROR; nalparser->last_pps = &nalparser->pps[pps->id]; return GST_H264_PARSER_OK; } /** * gst_h264_quant_matrix_8x8_get_zigzag_from_raster: * @out_quant: (out): The resulting quantization matrix * @quant: The source quantization matrix * * Converts quantization matrix @quant from raster scan order to * zigzag scan order and store the resulting factors into @out_quant. * * Note: it is an error to pass the same table in both @quant and * @out_quant arguments. * * Since: 1.4 */ void gst_h264_quant_matrix_8x8_get_zigzag_from_raster (guint8 out_quant[64], const guint8 quant[64]) { guint i; g_return_if_fail (out_quant != quant); for (i = 0; i < 64; i++) out_quant[i] = quant[zigzag_8x8[i]]; } /** * gst_h264_quant_matrix_8x8_get_raster_from_zigzag: * @out_quant: (out): The resulting quantization matrix * @quant: The source quantization matrix * * Converts quantization matrix @quant from zigzag scan order to * raster scan order and store the resulting factors into @out_quant. * * Note: it is an error to pass the same table in both @quant and * @out_quant arguments. * * Since: 1.4 */ void gst_h264_quant_matrix_8x8_get_raster_from_zigzag (guint8 out_quant[64], const guint8 quant[64]) { guint i; g_return_if_fail (out_quant != quant); for (i = 0; i < 64; i++) out_quant[zigzag_8x8[i]] = quant[i]; } /** * gst_h264_quant_matrix_4x4_get_zigzag_from_raster: * @out_quant: (out): The resulting quantization matrix * @quant: The source quantization matrix * * Converts quantization matrix @quant from raster scan order to * zigzag scan order and store the resulting factors into @out_quant. * * Note: it is an error to pass the same table in both @quant and * @out_quant arguments. * * Since: 1.4 */ void gst_h264_quant_matrix_4x4_get_zigzag_from_raster (guint8 out_quant[16], const guint8 quant[16]) { guint i; g_return_if_fail (out_quant != quant); for (i = 0; i < 16; i++) out_quant[i] = quant[zigzag_4x4[i]]; } /** * gst_h264_quant_matrix_4x4_get_raster_from_zigzag: * @out_quant: (out): The resulting quantization matrix * @quant: The source quantization matrix * * Converts quantization matrix @quant from zigzag scan order to * raster scan order and store the resulting factors into @out_quant. * * Note: it is an error to pass the same table in both @quant and * @out_quant arguments. * * Since: 1.4 */ void gst_h264_quant_matrix_4x4_get_raster_from_zigzag (guint8 out_quant[16], const guint8 quant[16]) { guint i; g_return_if_fail (out_quant != quant); for (i = 0; i < 16; i++) out_quant[zigzag_4x4[i]] = quant[i]; } /** * gst_h264_video_calculate_framerate: * @sps: Current Sequence Parameter Set * @field_pic_flag: Current @field_pic_flag, obtained from latest slice header * @pic_struct: @pic_struct value if available, 0 otherwise * @fps_num: (out): The resulting fps numerator * @fps_den: (out): The resulting fps denominator * * Calculate framerate of a video sequence using @sps VUI information, * @field_pic_flag from a slice header and @pic_struct from #GstH264PicTiming SEI * message. * * If framerate is variable or can't be determined, @fps_num will be set to 0 * and @fps_den to 1. */ void gst_h264_video_calculate_framerate (const GstH264SPS * sps, guint field_pic_flag, guint pic_struct, gint * fps_num, gint * fps_den) { gint num = 0; gint den = 1; /* To calculate framerate, we use this formula: * time_scale 1 1 * fps = ----------------- x --------------- x ------------------------ * num_units_in_tick DeltaTfiDivisor (field_pic_flag ? 2 : 1) * * See H264 specification E2.1 for more details. */ if (sps) { if (sps->vui_parameters_present_flag) { const GstH264VUIParams *vui = &sps->vui_parameters; if (vui->timing_info_present_flag) { int delta_tfi_divisor = 1; num = vui->time_scale; den = vui->num_units_in_tick; if (vui->pic_struct_present_flag) { switch (pic_struct) { case 1: case 2: delta_tfi_divisor = 1; break; case 0: case 3: case 4: delta_tfi_divisor = 2; break; case 5: case 6: delta_tfi_divisor = 3; break; case 7: delta_tfi_divisor = 4; break; case 8: delta_tfi_divisor = 6; break; } } else { delta_tfi_divisor = field_pic_flag ? 1 : 2; } den *= delta_tfi_divisor; /* Picture is two fields ? */ den *= (field_pic_flag ? 2 : 1); } } } *fps_num = num; *fps_den = den; } static gboolean gst_h264_write_sei_registered_user_data (NalWriter * nw, GstH264RegisteredUserData * rud) { WRITE_UINT8 (nw, rud->country_code, 8); if (rud->country_code == 0xff) WRITE_UINT8 (nw, rud->country_code_extension, 8); WRITE_BYTES (nw, rud->data, rud->size); return TRUE; error: return FALSE; } static gboolean gst_h264_write_sei_frame_packing (NalWriter * nw, GstH264FramePacking * frame_packing) { WRITE_UE (nw, frame_packing->frame_packing_id); WRITE_UINT8 (nw, frame_packing->frame_packing_cancel_flag, 1); if (!frame_packing->frame_packing_cancel_flag) { WRITE_UINT8 (nw, frame_packing->frame_packing_type, 7); WRITE_UINT8 (nw, frame_packing->quincunx_sampling_flag, 1); WRITE_UINT8 (nw, frame_packing->content_interpretation_type, 6); WRITE_UINT8 (nw, frame_packing->spatial_flipping_flag, 1); WRITE_UINT8 (nw, frame_packing->frame0_flipped_flag, 1); WRITE_UINT8 (nw, frame_packing->field_views_flag, 1); WRITE_UINT8 (nw, frame_packing->current_frame_is_frame0_flag, 1); WRITE_UINT8 (nw, frame_packing->frame0_self_contained_flag, 1); WRITE_UINT8 (nw, frame_packing->frame1_self_contained_flag, 1); if (!frame_packing->quincunx_sampling_flag && frame_packing->frame_packing_type != GST_H264_FRAME_PACKING_TEMPORAL_INTERLEAVING) { WRITE_UINT8 (nw, frame_packing->frame0_grid_position_x, 4); WRITE_UINT8 (nw, frame_packing->frame0_grid_position_y, 4); WRITE_UINT8 (nw, frame_packing->frame1_grid_position_x, 4); WRITE_UINT8 (nw, frame_packing->frame1_grid_position_y, 4); } /* frame_packing_arrangement_reserved_byte */ WRITE_UINT8 (nw, 0, 8); WRITE_UE (nw, frame_packing->frame_packing_repetition_period); } /* frame_packing_arrangement_extension_flag */ WRITE_UINT8 (nw, 0, 1); return TRUE; error: return FALSE; } static gboolean gst_h264_write_sei_mastering_display_colour_volume (NalWriter * nw, GstH264MasteringDisplayColourVolume * mdcv) { gint i; for (i = 0; i < 3; i++) { WRITE_UINT16 (nw, mdcv->display_primaries_x[i], 16); WRITE_UINT16 (nw, mdcv->display_primaries_y[i], 16); } WRITE_UINT16 (nw, mdcv->white_point_x, 16); WRITE_UINT16 (nw, mdcv->white_point_y, 16); WRITE_UINT32 (nw, mdcv->max_display_mastering_luminance, 32); WRITE_UINT32 (nw, mdcv->min_display_mastering_luminance, 32); return TRUE; error: return FALSE; } static gboolean gst_h264_write_sei_content_light_level_info (NalWriter * nw, GstH264ContentLightLevel * cll) { WRITE_UINT16 (nw, cll->max_content_light_level, 16); WRITE_UINT16 (nw, cll->max_pic_average_light_level, 16); return TRUE; error: return FALSE; } static gboolean gst_h264_write_sei_pic_timing (NalWriter * nw, GstH264PicTiming * tim) { if (tim->CpbDpbDelaysPresentFlag) { WRITE_UINT32 (nw, tim->cpb_removal_delay, tim->cpb_removal_delay_length_minus1 + 1); WRITE_UINT32 (nw, tim->dpb_output_delay, tim->dpb_output_delay_length_minus1 + 1); } if (tim->pic_struct_present_flag) { const guint8 num_clock_ts_table[9] = { 1, 1, 1, 2, 2, 3, 3, 2, 3 }; guint8 num_clock_num_ts; guint i; WRITE_UINT8 (nw, tim->pic_struct, 4); num_clock_num_ts = num_clock_ts_table[tim->pic_struct]; for (i = 0; i < num_clock_num_ts; i++) { WRITE_UINT8 (nw, tim->clock_timestamp_flag[i], 1); if (tim->clock_timestamp_flag[i]) { GstH264ClockTimestamp *timestamp = &tim->clock_timestamp[i]; WRITE_UINT8 (nw, timestamp->ct_type, 2); WRITE_UINT8 (nw, timestamp->nuit_field_based_flag, 1); WRITE_UINT8 (nw, timestamp->counting_type, 5); WRITE_UINT8 (nw, timestamp->full_timestamp_flag, 1); WRITE_UINT8 (nw, timestamp->discontinuity_flag, 1); WRITE_UINT8 (nw, timestamp->cnt_dropped_flag, 1); WRITE_UINT8 (nw, timestamp->n_frames, 8); if (timestamp->full_timestamp_flag) { WRITE_UINT8 (nw, timestamp->seconds_value, 6); WRITE_UINT8 (nw, timestamp->minutes_value, 6); WRITE_UINT8 (nw, timestamp->hours_value, 5); } else { WRITE_UINT8 (nw, timestamp->seconds_flag, 1); if (timestamp->seconds_flag) { WRITE_UINT8 (nw, timestamp->seconds_value, 6); WRITE_UINT8 (nw, timestamp->minutes_flag, 1); if (timestamp->minutes_flag) { WRITE_UINT8 (nw, timestamp->minutes_value, 6); WRITE_UINT8 (nw, timestamp->hours_flag, 1); if (timestamp->hours_flag) WRITE_UINT8 (nw, timestamp->hours_value, 5); } } } if (tim->time_offset_length > 0) { WRITE_UINT32 (nw, timestamp->time_offset, tim->time_offset_length); } } } } return TRUE; error: return FALSE; } static GstMemory * gst_h264_create_sei_memory_internal (guint8 nal_prefix_size, gboolean packetized, GArray * messages) { NalWriter nw; gint i; gboolean have_written_data = FALSE; nal_writer_init (&nw, nal_prefix_size, packetized); if (messages->len == 0) goto error; GST_DEBUG ("Create SEI nal from array, len: %d", messages->len); /* nal header */ /* forbidden_zero_bit */ WRITE_UINT8 (&nw, 0, 1); /* nal_ref_idc, zero for sei nalu */ WRITE_UINT8 (&nw, 0, 2); /* nal_unit_type */ WRITE_UINT8 (&nw, GST_H264_NAL_SEI, 5); for (i = 0; i < messages->len; i++) { GstH264SEIMessage *msg = &g_array_index (messages, GstH264SEIMessage, i); guint32 payload_size_data = 0; guint32 payload_size_in_bits = 0; guint32 payload_type_data = msg->payloadType; gboolean need_align = FALSE; switch (payload_type_data) { case GST_H264_SEI_REGISTERED_USER_DATA:{ GstH264RegisteredUserData *rud = &msg->payload.registered_user_data; /* itu_t_t35_country_code: 8 bits */ payload_size_data = 1; if (rud->country_code == 0xff) { /* itu_t_t35_country_code_extension_byte */ payload_size_data++; } payload_size_data += rud->size; break; } case GST_H264_SEI_FRAME_PACKING:{ GstH264FramePacking *frame_packing = &msg->payload.frame_packing; guint leading_zeros, rest; /* frame_packing_arrangement_id: exp-golomb bits */ count_exp_golomb_bits (frame_packing->frame_packing_id, &leading_zeros, &rest); payload_size_in_bits = leading_zeros + rest; /* frame_packing_arrangement_cancel_flag: 1 bit */ payload_size_in_bits++; if (!frame_packing->frame_packing_cancel_flag) { /* frame_packing_arrangement_type: 7 bits * quincunx_sampling_flag: 1 bit * content_interpretation_type: 6 bit * spatial_flipping_flag: 1 bit * frame0_flipped_flag: 1 bit * field_views_flag: 1 bit * current_frame_is_frame0_flag: 1 bit * frame0_self_contained_flag: 1 bit * frame1_self_contained_flag: 1 bit */ payload_size_in_bits += 20; if (!frame_packing->quincunx_sampling_flag && frame_packing->frame_packing_type != GST_H264_FRAME_PACKING_TEMPORAL_INTERLEAVING) { /* frame0_grid_position_x: 4bits * frame0_grid_position_y: 4bits * frame1_grid_position_x: 4bits * frame1_grid_position_y: 4bits */ payload_size_in_bits += 16; } /* frame_packing_arrangement_reserved_byte: 8 bits */ payload_size_in_bits += 8; /* frame_packing_arrangement_repetition_period: exp-golomb bits */ count_exp_golomb_bits (frame_packing->frame_packing_repetition_period, &leading_zeros, &rest); payload_size_in_bits += (leading_zeros + rest); } /* frame_packing_arrangement_extension_flag: 1 bit */ payload_size_in_bits++; payload_size_data = payload_size_in_bits >> 3; if ((payload_size_in_bits & 0x7) != 0) { GST_INFO ("Bits for Frame Packing SEI is not byte aligned"); payload_size_data++; need_align = TRUE; } break; } case GST_H264_SEI_MASTERING_DISPLAY_COLOUR_VOLUME: /* x, y 16 bits per RGB channel * x, y 16 bits white point * max, min luminance 32 bits * * (2 * 2 * 3) + (2 * 2) + (4 * 2) = 24 bytes */ payload_size_data = 24; break; case GST_H264_SEI_CONTENT_LIGHT_LEVEL: /* maxCLL and maxFALL per 16 bits * * 2 * 2 = 4 bytes */ payload_size_data = 4; break; case GST_H264_SEI_PIC_TIMING:{ GstH264PicTiming *tim = &msg->payload.pic_timing; const guint8 num_clock_ts_table[9] = { 1, 1, 1, 2, 2, 3, 3, 2, 3 }; guint8 num_clock_num_ts; guint i; if (!tim->CpbDpbDelaysPresentFlag && !tim->pic_struct_present_flag) { GST_WARNING ("Both CpbDpbDelaysPresentFlag and pic_struct_present_flag are zero"); break; } if (tim->CpbDpbDelaysPresentFlag) { payload_size_in_bits = tim->cpb_removal_delay_length_minus1 + 1; payload_size_in_bits += tim->dpb_output_delay_length_minus1 + 1; } if (tim->pic_struct_present_flag) { /* pic_struct: 4bits */ payload_size_in_bits += 4; num_clock_num_ts = num_clock_ts_table[tim->pic_struct]; for (i = 0; i < num_clock_num_ts; i++) { /* clock_timestamp_flag: 1bit */ payload_size_in_bits++; if (tim->clock_timestamp_flag[i]) { GstH264ClockTimestamp *timestamp = &tim->clock_timestamp[i]; /* ct_type: 2bits * nuit_field_based_flag: 1bit * counting_type: 5bits * full_timestamp_flag: 1bit * discontinuity_flag: 1bit * cnt_dropped_flag: 1bit * n_frames: 8bits */ payload_size_in_bits += 19; if (timestamp->full_timestamp_flag) { /* seconds_value: 6bits * minutes_value: 6bits * hours_value: 5bits */ payload_size_in_bits += 17; } else { /* seconds_flag: 1bit */ payload_size_in_bits++; if (timestamp->seconds_flag) { /* seconds_value: 6bits * minutes_flag: 1bit */ payload_size_in_bits += 7; if (timestamp->minutes_flag) { /* minutes_value: 6bits * hours_flag: 1bits */ payload_size_in_bits += 7; if (timestamp->hours_flag) { /* hours_value: 5bits */ payload_size_in_bits += 5; } } } } /* time_offset_length bits */ payload_size_in_bits += tim->time_offset_length; } } } payload_size_data = payload_size_in_bits >> 3; if ((payload_size_in_bits & 0x7) != 0) { GST_INFO ("Bits for Picture Timing SEI is not byte aligned"); payload_size_data++; need_align = TRUE; } break; } default: break; } if (payload_size_data == 0) { GST_FIXME ("Unsupported SEI type %d", msg->payloadType); continue; } /* write payload type bytes */ while (payload_type_data >= 0xff) { WRITE_UINT8 (&nw, 0xff, 8); payload_type_data -= -0xff; } WRITE_UINT8 (&nw, payload_type_data, 8); /* write payload size bytes */ while (payload_size_data >= 0xff) { WRITE_UINT8 (&nw, 0xff, 8); payload_size_data -= -0xff; } WRITE_UINT8 (&nw, payload_size_data, 8); switch (msg->payloadType) { case GST_H264_SEI_REGISTERED_USER_DATA: GST_DEBUG ("Writing \"Registered user data\""); if (!gst_h264_write_sei_registered_user_data (&nw, &msg->payload.registered_user_data)) { GST_WARNING ("Failed to write \"Registered user data\""); goto error; } have_written_data = TRUE; break; case GST_H264_SEI_FRAME_PACKING: GST_DEBUG ("Writing \"Frame packing\""); if (!gst_h264_write_sei_frame_packing (&nw, &msg->payload.frame_packing)) { GST_WARNING ("Failed to write \"Frame packing\""); goto error; } have_written_data = TRUE; break; case GST_H264_SEI_MASTERING_DISPLAY_COLOUR_VOLUME: GST_DEBUG ("Writing \"Mastering display colour volume\""); if (!gst_h264_write_sei_mastering_display_colour_volume (&nw, &msg->payload.mastering_display_colour_volume)) { GST_WARNING ("Failed to write \"Mastering display colour volume\""); goto error; } have_written_data = TRUE; break; case GST_H264_SEI_CONTENT_LIGHT_LEVEL: GST_DEBUG ("Writing \"Content light level\""); if (!gst_h264_write_sei_content_light_level_info (&nw, &msg->payload.content_light_level)) { GST_WARNING ("Failed to write \"Content light level\""); goto error; } have_written_data = TRUE; break; case GST_H264_SEI_PIC_TIMING: GST_DEBUG ("Writing \"Picture timing\""); if (!gst_h264_write_sei_pic_timing (&nw, &msg->payload.pic_timing)) { GST_WARNING ("Failed to write \"Picture timing\""); goto error; } have_written_data = TRUE; break; default: break; } if (need_align && !nal_writer_do_rbsp_trailing_bits (&nw)) { GST_WARNING ("Cannot insert traling bits"); goto error; } } if (!have_written_data) { GST_WARNING ("No written sei data"); goto error; } if (!nal_writer_do_rbsp_trailing_bits (&nw)) { GST_WARNING ("Failed to insert rbsp trailing bits"); goto error; } return nal_writer_reset_and_get_memory (&nw); error: nal_writer_reset (&nw); return NULL; } /** * gst_h264_create_sei_memory: * @start_code_prefix_length: a length of start code prefix, must be 3 or 4 * @messages: (transfer none): a GArray of #GstH264SEIMessage * * Creates raw byte-stream format (a.k.a Annex B type) SEI nal unit data * from @messages * * Returns: a #GstMemory containing a SEI nal unit * * Since: 1.18 */ GstMemory * gst_h264_create_sei_memory (guint8 start_code_prefix_length, GArray * messages) { g_return_val_if_fail (start_code_prefix_length == 3 || start_code_prefix_length == 4, NULL); g_return_val_if_fail (messages != NULL, NULL); g_return_val_if_fail (messages->len > 0, NULL); return gst_h264_create_sei_memory_internal (start_code_prefix_length, FALSE, messages); } /** * gst_h264_create_sei_memory_avc: * @nal_length_size: a size of nal length field, allowed range is [1, 4] * @messages: (transfer none): a GArray of #GstH264SEIMessage * * Creates raw packetized format SEI nal unit data from @messages * * Returns: a #GstMemory containing a SEI nal unit * * Since: 1.18 */ GstMemory * gst_h264_create_sei_memory_avc (guint8 nal_length_size, GArray * messages) { g_return_val_if_fail (nal_length_size > 0 && nal_length_size < 5, NULL); g_return_val_if_fail (messages != NULL, NULL); g_return_val_if_fail (messages->len > 0, NULL); return gst_h264_create_sei_memory_internal (nal_length_size, TRUE, messages); } static GstBuffer * gst_h264_parser_insert_sei_internal (GstH264NalParser * nalparser, guint8 nal_prefix_size, gboolean packetized, GstBuffer * au, GstMemory * sei) { GstH264NalUnit nalu; GstMapInfo info; GstH264ParserResult pres; guint offset = 0; GstBuffer *new_buffer = NULL; if (!gst_buffer_map (au, &info, GST_MAP_READ)) { GST_ERROR ("Cannot map au buffer"); return NULL; } /* Find the offset of the first slice */ do { if (packetized) { pres = gst_h264_parser_identify_nalu_avc (nalparser, info.data, offset, info.size, nal_prefix_size, &nalu); } else { pres = gst_h264_parser_identify_nalu (nalparser, info.data, offset, info.size, &nalu); } if (pres != GST_H264_PARSER_OK && pres != GST_H264_PARSER_NO_NAL_END) { GST_DEBUG ("Failed to identify nal unit, ret: %d", pres); gst_buffer_unmap (au, &info); return NULL; } if ((nalu.type >= GST_H264_NAL_SLICE && nalu.type <= GST_H264_NAL_SLICE_IDR) || (nalu.type >= GST_H264_NAL_SLICE_AUX && nalu.type <= GST_H264_NAL_SLICE_DEPTH)) { GST_DEBUG ("Found slice nal type %d at offset %d", nalu.type, nalu.sc_offset); break; } offset = nalu.offset + nalu.size; } while (pres == GST_H264_PARSER_OK); gst_buffer_unmap (au, &info); /* found the best position now, create new buffer */ new_buffer = gst_buffer_new (); /* copy all metadata */ if (!gst_buffer_copy_into (new_buffer, au, GST_BUFFER_COPY_METADATA, 0, -1)) { GST_ERROR ("Failed to copy metadata into new buffer"); gst_clear_buffer (&new_buffer); goto out; } /* copy non-slice nal */ if (nalu.sc_offset > 0) { if (!gst_buffer_copy_into (new_buffer, au, GST_BUFFER_COPY_MEMORY, 0, nalu.sc_offset)) { GST_ERROR ("Failed to copy buffer"); gst_clear_buffer (&new_buffer); goto out; } } /* insert sei */ gst_buffer_append_memory (new_buffer, gst_memory_ref (sei)); /* copy the rest */ if (!gst_buffer_copy_into (new_buffer, au, GST_BUFFER_COPY_MEMORY, nalu.sc_offset, -1)) { GST_ERROR ("Failed to copy buffer"); gst_clear_buffer (&new_buffer); goto out; } out: return new_buffer; } /** * gst_h264_parser_insert_sei: * @nalparser: a #GstH264NalParser * @au: (transfer none): a #GstBuffer containing AU data * @sei: (transfer none): a #GstMemory containing a SEI nal * * Copy @au into new #GstBuffer and insert @sei into the #GstBuffer. * The validation for completeness of @au and @sei is caller's responsibility. * Both @au and @sei must be byte-stream formatted * * Returns: (nullable): a SEI inserted #GstBuffer or %NULL * if cannot figure out proper position to insert a @sei * * Since: 1.18 */ GstBuffer * gst_h264_parser_insert_sei (GstH264NalParser * nalparser, GstBuffer * au, GstMemory * sei) { g_return_val_if_fail (nalparser != NULL, NULL); g_return_val_if_fail (GST_IS_BUFFER (au), NULL); g_return_val_if_fail (sei != NULL, NULL); /* the size of start code prefix (3 or 4) is not matter since it will be * scanned */ return gst_h264_parser_insert_sei_internal (nalparser, 4, FALSE, au, sei); } /** * gst_h264_parser_insert_sei_avc: * @nalparser: a #GstH264NalParser * @nal_length_size: a size of nal length field, allowed range is [1, 4] * @au: (transfer none): a #GstBuffer containing AU data * @sei: (transfer none): a #GstMemory containing a SEI nal * * Copy @au into new #GstBuffer and insert @sei into the #GstBuffer. * The validation for completeness of @au and @sei is caller's responsibility. * Nal prefix type of both @au and @sei must be packetized, and * also the size of nal length field must be identical to @nal_length_size * * Returns: (nullable): a SEI inserted #GstBuffer or %NULL * if cannot figure out proper position to insert a @sei * * Since: 1.18 */ GstBuffer * gst_h264_parser_insert_sei_avc (GstH264NalParser * nalparser, guint8 nal_length_size, GstBuffer * au, GstMemory * sei) { g_return_val_if_fail (nalparser != NULL, NULL); g_return_val_if_fail (nal_length_size > 0 && nal_length_size < 5, NULL); g_return_val_if_fail (GST_IS_BUFFER (au), NULL); g_return_val_if_fail (sei != NULL, NULL); /* the size of start code prefix (3 or 4) is not matter since it will be * scanned */ return gst_h264_parser_insert_sei_internal (nalparser, nal_length_size, TRUE, au, sei); }